Directed information

Directed information, , is a measure of information theory and it measures the amount of information that flows from the process to , where denotes the vector and denotes . The term "directed information" was coined by James Massey and is defined as

,

where is the conditional mutual information.

Note that if , directed information becomes mutual information . Directed information has many applications in problems where causality plays an important role such as capacity of channel with feedback,[1][2] capacity of discrete memoryless networks with feedback,[3] gambling with causal side information,[4] and compression with causal side information.[5]

References

  1. Massey, James (1990). "Causality, Feedback And Directed Information" (ISITA).
  2. Permuter, Haim Henry; Weissman, Tsachy; Goldsmith, Andrea J. (February 2009). "Finite State Channels With Time-Invariant Deterministic Feedback". IEEE Transactions on Information Theory. 55 (2): 644–662. doi:10.1109/TIT.2008.2009849.
  3. Kramer, G. (January 2003). "Capacity results for the discrete memoryless network". IEEE Transactions on Information Theory. 49 (1): 4–21. doi:10.1109/TIT.2002.806135.
  4. Permuter, Haim H.; Kim, Young-Han; Weissman, Tsachy (June 2011). "Interpretations of Directed Information in Portfolio Theory, Data Compression, and Hypothesis Testing". IEEE Transactions on Information Theory. 57 (6): 3248–3259. doi:10.1109/TIT.2011.2136270.
  5. Simeone, Osvaldo; Permuter, Haim Henri (June 2013). "Source Coding When the Side Information May Be Delayed". IEEE Transactions on Information Theory. 59 (6): 3607–3618. doi:10.1109/TIT.2013.2248192.
This article is issued from Wikipedia - version of the 11/20/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.