Fast saliency map extraction from video: A hardware approach

Saliency map is a central part of many visual attention systems, particularly during learning and control of bottom-up attention. In this research we developed a hardware tool to extract saliency map from a video sequence. Saliency map is obtained by aggregating primary features of each frame, such as intensity, color, and lines orientation, along with temporal difference. The system is designed to provide both high speed and acceptable accuracy for real-time applications, such as machine vision and robotics. A versatile Verilog model for realization of the video processing system is developed, which can easily be mapped and synthesized on various FPGA or ASIC platforms. The proposed parallel hardware can process over 50 million pixels in a second, which is about 2x faster than the state-of-the-art designs. Experimental results on sample images justify the applicability and efficiency of the developed system in real-time applications.

[1]  Berin Martini,et al.  Hardware accelerated visual attention algorithm , 2011, 2011 45th Annual Conference on Information Sciences and Systems.

[2]  Nanning Zheng,et al.  A Visual Brain Chip Based on Selective Attention for Robot Vision Application , 2009, 2009 Third IEEE International Conference on Space Mission Challenges for Information Technology.

[3]  Christof Koch,et al.  Feature combination strategies for saliency-based visual attention systems , 2001, J. Electronic Imaging.

[4]  Christof Koch,et al.  A Model of Saliency-Based Visual Attention for Rapid Scene Analysis , 2009 .

[5]  Narayanan Vijaykrishnan,et al.  Emulating Mammalian Vision on Reconfigurable Hardware , 2012, 2012 IEEE 20th International Symposium on Field-Programmable Custom Computing Machines.

[6]  Narayanan Vijaykrishnan,et al.  SHARC: A streaming model for FPGA accelerators and its application to Saliency , 2011, 2011 Design, Automation & Test in Europe.

[7]  Fei Zheng,et al.  An Improved Model of Producing Saliency Map for Visual Attention System , 2008, ICIC.

[8]  Eduardo Ros,et al.  Bottom-up visual attention model based on FPGA , 2012, 2012 19th IEEE International Conference on Electronics, Circuits, and Systems (ICECS 2012).