previous up next
Next: About this document ... Up: Reinforcement Learning of Active Previous: Conclusion

References

1
Aloimonos, Y., ed., Active Perception, Lawrence Erlbaum Associates, 1993.

2
Bajcsy, R., Active Perception, Proceedings of the IEEE, 76(8):996-1005, August 1988

3
Ballard, D., Reference Frames for Animate Vision, in Proc. 11th IJCAI, pp. 1635-1641, August 1989.

4
Bandera, C., Vico, J., Bravo, J., Harmon, M., Baird, L., Residual Q-Learning Applied to Visual Attention, Proc. 13th Intl. Conf. on Machine Learning, Bari, Italy, July 1996.

5
Barto, A. G., Bradtke, S. J., and Singh, S. P.,
Real-time learning and control using asynchronous dynamic programming.
Computer Science Technical Report 91-57, University of Massachusetts, August 1991.

6
Callari, F.G., Ferrie, F.P., Autonomous Recognition: Driven by Ambiguity, Proc. CVPR-96, pp. 701-707. 1996.

7
Cassandra, A., Kaelbling, L. P., and Littman, M.,
Acting optimally in partially observable stochastic domains.
In Proc. AAAI-94, pages 1023-1028. Morgan Kaufmann, 1994.

8
Chrisman, L., Reinforcement learning with perceptual aliasing: The perceptual distinctions approach. In Proc. AAAI-92, pages 183-188. Morgan Kaufmann, Los Altos, California, 1992.

9
Darrell, T., and Pentland, A., Attention-driven Expression and Gesture Analysis in an Interactive Environment, in Proc. Intl. Workshop on Automatic Face and Gesture Recognition (IWAFGR '95), Zurich, Switzerland, 1995.

10
Darrell, T., Moghaddam, B., and Pentland, A., Active Face Tracking and Pose Estimation in an Interactive Room, Proc. CVPR-96. 1996.

11
Darrell, T., and Pentland, A., Active Gesture Recognition using Learned Visual Attention, in D. S. Touretzky, M. Mozer, and M. Hasselmo, eds., Advances in Neural Information Processing Systems (NIPS) 8 , MIT Press, Cambridge MA, 1996.

12
Darrell, T., and Pentland, A., Active Gesture Recognition using Partially Observable Markov Decision Processes, in Proc. IEEE International Conference on Pattern Recognition, IEEE Computer Society Press, Vienna, 1996.

13
Dickinson, S.J., Christensen, H.I., Tsotsos, J., and Olofsson, G., Active Object Recognition Integrating Attention and Viewpoint Control, Proc. ECCV-94, 1994.

14
Jaakkola, T., Singh, S., and Jordan, M.,
Reinforcement Learning Algorithm for Partially Observable Markov Decision Problems.
In Advances In Neural Information Processing Systems 7, MIT Press, 1995.

15
Kosecka, J., Christensen, H.I., Bajcsy, R., Discrete-Event Modeling of Visually Guided Behaviors, IJCV(14), No. 2, pp. 179-191, March 1995.

16
Lin, L., and Michell, T.,
Reinforcement learning with hidden states.
In Proc. AAAI-92. Morgan Kaufmann, 1992.

17
Loeve, M.M., Probability Theory, Van Nostrand, Princeton, 1955.

18
Lovejoy, W.,
A survey of algorithmic methods of partially observed markov decision processes.
Annals of Operation Reserach, 28:47-66, 1991.

19
McCallum., R. A., Overcoming incompleat perception with utile distinction memory. In Proceedings Tenth Machine Learning Conference. Morgan Kaufmann, 1993.

20
McCallum, R. A., Instance-based State Identification for Reinforcement Learning. In Advances In Neural Information Processing Systems 7, MIT Press, 1995.

21
Nakamura, N., and Asada, M.,, Motion Sketch: Acquisition of Visual Motion Guided Behaviors, in Proc. of Int. Joint Conference on Aritificial Intelligence, pp.126-132, 1995.

22
Peng, J., Bhanu, B., Delayed Reinforcement Learning for Closed-Loop Object Recognition, Proc. Intl. Conf. Pattern Recognition '96, Vienna, Austria. 1996.

23
Rong, S., Bhanu, B., Reinforcement Learning for Integrating Context with Clutter Models for Target Detection, Proc. ARPA IU Workshop '96, pp. 1389-1394, 1996.

24
Sondik, E. J.
The optimal control of partially observable markov processes over the infinite horizon: Discounted costs.
Operations Reserach, 26(2):282-304, 1978.

25
Sutton, R. S.,
Learning to predict by the method of temporal differences.
Machine Learning, 3:9-44, 1988.

26
Watkins, C., and Dayan, P.,
Q-learning.
Machine Learning, 8:279-292, 1992.

27
Whitehead, S., Active perception and reinforcement learning. In Proc. 7th Intl. Conf. ML, June 1990.

28
Wren, C., Darrell, T., Starner, T., Johnston, M., Russell, K., Azarbayejani, A., and Pentland, A. pfinder: A Real-Time System for Tracking People, SPIE Conference on Real-Time Vision, M. Bove, Ed., Philadelpia, PA, July 1995.

previous up next
Next: About this document ... Up: Reinforcement Learning of Active Previous: Conclusion
Trevor Darrell
9/14/1998