TY - GEN
T1 - Optimal self-driven sampling for estimation based on value of information
AU - Soleymani, Touraj
AU - Hirche, Sandra
AU - Baras, John S.
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2016/6/22
Y1 - 2016/6/22
N2 - Consider an observer (reporter) who desires to inform optimally a distant agent regarding a physical stochastic process in the environment while the directed communication of the observer to the agent has a price. We define a metric, from a task oriented perspective, for the information transferred from the observer to the agent. We develop a framework for optimizing an augmented cost function which is a convex combination of the transferred information and the paid price over a finite horizon. We suppose that the decision making takes place inside a source encoder, and that the sampling schedule is the decision variable. Moreover, we assume that no measurement at the current time is available to the observer for the decision making. We derive the optimal self-driven sampling policy using dynamic programming, and we show that this policy corresponds to a self-driven sampling policy based on a quantity that is in fact the value of information at each time instant. In addition, we use a semi-definite programming relaxation to provide a suboptimal sampling policy. Numerical and simulation results are presented for a simple unstable system.
AB - Consider an observer (reporter) who desires to inform optimally a distant agent regarding a physical stochastic process in the environment while the directed communication of the observer to the agent has a price. We define a metric, from a task oriented perspective, for the information transferred from the observer to the agent. We develop a framework for optimizing an augmented cost function which is a convex combination of the transferred information and the paid price over a finite horizon. We suppose that the decision making takes place inside a source encoder, and that the sampling schedule is the decision variable. Moreover, we assume that no measurement at the current time is available to the observer for the decision making. We derive the optimal self-driven sampling policy using dynamic programming, and we show that this policy corresponds to a self-driven sampling policy based on a quantity that is in fact the value of information at each time instant. In addition, we use a semi-definite programming relaxation to provide a suboptimal sampling policy. Numerical and simulation results are presented for a simple unstable system.
KW - Approximate Dynamic Programming
KW - Estimation
KW - Self-Driven Sampling
KW - Value of Information
UR - http://www.scopus.com/inward/record.url?scp=84981326616&partnerID=8YFLogxK
U2 - 10.1109/WODES.2016.7497846
DO - 10.1109/WODES.2016.7497846
M3 - Conference contribution
AN - SCOPUS:84981326616
T3 - 2016 13th International Workshop on Discrete Event Systems, WODES 2016
SP - 183
EP - 188
BT - 2016 13th International Workshop on Discrete Event Systems, WODES 2016
A2 - Cassandras, Christos G.
A2 - Giua, Alessandro
A2 - Giua, Alessandro
A2 - Li, Zhiwu
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 13th International Workshop on Discrete Event Systems, WODES 2016
Y2 - 30 May 2016 through 1 June 2016
ER -