PLAYBOT A visually-guided robot for physically disabled children

Abstract This paper overviews the PLAYBOT project, a long-term, large-scale research program whose goal is to provide a directable robot which may enable physically disabled children to access and manipulate toys. This domain is the first test domain, but there is nothing inherent in the design of PLAYBOT that prohibits its extension to other tasks. The research is guided by several important goals: vision is the primary sensor; vision is task directed; the robot must be able to visually search its environment; object and event recognition are basic capabilities; environments must be natural and dynamic; users and environments are assumed to be unpredictable; task direction and reactivity must be smoothly integrated; and safety is of high importance. The emphasis of the research has been on vision for the robot this is the most challenging research aspect and the major bottleneck to the development of intelligent robots. Since the control framework is behavior-based, the visual capabilities of PLAYBOT are described in terms of visual behaviors. Many of the components of PLAYBOT are briefly described and several examples of implemented sub-systems are shown. The paper concludes with a description of the current overall system implementation, and a complete example of PLAYBOT performing a simple task.

[1]  Allan D. Jepson,et al.  Priors, preferences and categorical percepts , 1996 .

[2]  Sven J. Dickinson,et al.  Integrating task-directed planning with reactive object recognition , 1993, Other Conferences.

[3]  John K. Tsotsos,et al.  A binocular robotic head system with torsional eye movements , 1993, [1993] Proceedings IEEE International Conference on Robotics and Automation.

[4]  John K. Tsotsos,et al.  The role of feature visibility constraints in perspective alignment , 1995, Proceedings., International Conference on Image Processing.

[5]  John K. Tsotsos,et al.  Techniques for disparity measurement , 1991, CVGIP Image Underst..

[6]  Azriel Rosenfeld,et al.  From volumes to views: An approach to 3-D object recognition , 1992, CVGIP Image Underst..

[7]  Elizabeth S. Helfman,et al.  Blissymbolics, speaking without speech , 1980 .

[8]  John K. Tsotsos,et al.  Modeling Visual Attention via Selective Tuning , 1995, Artif. Intell..

[9]  Michael J. Black,et al.  Mixture models for optical flow computation , 1993, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[10]  John K. Tsotsos,et al.  Integration of camera motion behaviours for active object recognition , 1994 .

[11]  John K. Tsotsos Behaviorist Intelligence and the Scaling Problem , 1995, Artif. Intell..

[12]  Sven J. Dickinson,et al.  Active Object Recognition Integrating Attention and Viewpoint Control , 1994, Comput. Vis. Image Underst..

[13]  John K. Tsotsos,et al.  Active stereo vision and cyclotorsion , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[14]  John K. Tsotsos Intelligent control for perceptually attentive agents: The S* proposal , 1997, Robotics Auton. Syst..

[15]  Allan D. Jepson,et al.  The Computational Perception of Scene Dynamics , 1997, Comput. Vis. Image Underst..

[16]  Allan D. Jepson,et al.  Computational Perception of Scene Dynamics , 1996, ECCV.

[17]  John K. Tsotsos,et al.  Design and Performance of Trish, a Binocular Robot Head with Torsional Eye Movements , 1993, Int. J. Pattern Recognit. Artif. Intell..

[18]  Yiming Ye,et al.  Where to look next in 3D object search , 1995, Proceedings of International Symposium on Computer Vision - ISCV.

[19]  John K. Tsotsos,et al.  Real-time Model-based Tracking Using Perspective Alignment: Parallel Implementation and Stability Analysis , 1995, Research in Computer and Robot Vision.

[20]  Yiming Ye,et al.  Sensor planning in 3d object search: its formu-lation and complexity , 1995 .