Timezone: »
For planning with high uncertainty, or with too many possible end positions as in games like Go or even chess, one can almost never solve the optimal control problem and must use some receding horizon heuristics. One such heuristics is based on the idea of maximizing empowerment, namely, keep the number of possible options maximal. This has been formulated using information theoretic ideas as maximizing the information capacity between the sequence of actions and the possible state of the system at some finite horizon, but no efficient algorithm for calculating this capacity was suggested. In this work we propose a concrete and efficient way for calculating the capacity between a sequence of actions and future states, based on local linearization of the dynamics and Gaussian channel capacity calculation. I will describe the new algorithm and some of its interesting implications.
Author Information
Naftali Tishby (The Hebrew University Jerusalem)
Naftali Tishby, is a professor of computer science and the director of the Interdisciplinary Center for Neural Computation (ICNC) at the Hebrew university of Jerusalem. He received his Ph.D. in theoretical physics from the Hebrew University and was a research staff member at MIT and Bell Labs from 1985 to 1991. He was also a visiting professor at Princeton NECI, the University of Pennsylvania and the University of California at Santa Barbara. Dr. Tishby is a leader of machine learning research and computational neuroscience. He was among the first to introduce methods from statistical physics into learning theory, and dynamical systems techniques in speech processing. His current research is at the interface between computer science, statistical physics and computational neuroscience and concerns the foundations of biological information processing and the connections between dynamics and information.
More from the Same Authors
-
2017 : How do the Deep Learning layers converge to the Information Bottleneck limit by Stochastic Gradient Descent? »
Naftali Tishby -
2014 Workshop: Novel Trends and Applications in Reinforcement Learning »
Csaba Szepesvari · Marc Deisenroth · Sergey Levine · Pedro Ortega · Brian Ziebart · Emma Brunskill · Naftali Tishby · Gerhard Neumann · Daniel Lee · Sridhar Mahadevan · Pieter Abbeel · David Silver · Vicenç Gómez -
2013 Workshop: Planning with Information Constraints for Control, Reinforcement Learning, Computational Neuroscience, Robotics and Games. »
Hilbert J Kappen · Naftali Tishby · Jan Peters · Evangelos Theodorou · David H Wolpert · Pedro Ortega -
2012 Workshop: Information in Perception and Action »
Naftali Tishby · Daniel Polani · Tobias Jung -
2011 Tutorial: Information Theory in Learning and Control »
Naftali Tishby -
2010 Poster: Tight Sample Complexity of Large-Margin Learning »
Sivan Sabato · Nati Srebro · Naftali Tishby -
2008 Workshop: Principled Theoretical Frameworks for the Perception-Action Cycle »
Daniel Polani · Naftali Tishby -
2008 Mini Symposium: Principled Theoretical Frameworks for the Perception-Action Cycle »
Daniel Polani · Naftali Tishby -
2008 Poster: On the Reliability of Clustering Stability in the Large Sample Regime »
Ohad Shamir · Naftali Tishby -
2008 Spotlight: On the Reliability of Clustering Stability in the Large Sample Regime »
Ohad Shamir · Naftali Tishby -
2007 Oral: Cluster Stability for Finite Samples »
Ohad Shamir · Naftali Tishby -
2007 Poster: Cluster Stability for Finite Samples »
Ohad Shamir · Naftali Tishby -
2006 Workshop: Revealing Hidden Elements of Dynamical Systems »
Naftali Tishby -
2006 Poster: Information Bottleneck for Non Co-Occurrence Data »
Yevgeny Seldin · Noam Slonim · Naftali Tishby