کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
413352 680437 2015 14 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
A developmental where–what network for concurrent and interactive visual attention and recognition
ترجمه فارسی عنوان
کجا رشد می کند؟ چه شبکه ای برای توجه و تشخیص همزمان و تعاملی تصویری
کلمات کلیدی
یادگیری پیشرفته مسیرهای حسگر حرکتی کجاست؟ توجه به رسمیت شناختن، مغز الهام گرفته از شبکه عصبی
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر هوش مصنوعی
چکیده انگلیسی


• Computational modeling of attention and recognition as ubiquitous internal actions.
• Development of sensory invariance through motor-specific entropy reduction.
• Spatiotemporal optimization of cortical self-organization.
• Non-iterative bidirectional sparse coding model.
• Divide-and-conquer solution to learn deep neural network with bidirectional flows.

This paper presents a brain-inspired developmental architecture called Where–What Network (WWN). In this second version of WWN, WWN-2 is learned for concurrent and interactive visual attention and recognition, via complementary pathways guided by “type” motor and “location” motor. The motor-driven top-down signals, together with bottom-up excitatory activities from the visual input, shape three possible information flows through a Y-shaped network. Using ℓ0ℓ0 constrained sparse coding scheme, the top-down and bottom-up co-firing leads to a non-iterative cell-centered synaptic update model, entailing the strict entropy reduction from early to later layers, as well as a dual optimization of update directions and step sizes that dynamically depend on the firing ages of the neurons. Three operational modes for cluttered scenes emerge from the learning process, depending on what is available in the motor area: context-free mode for detection and recognition from a cluttered scene for a learned object, location-context mode for doing object recognition, and type-context mode for doing object search, all by a single network. To demonstrate the attention capabilities along with their interaction of visual processing, the proposed network is in the presence of complex backgrounds, learns on the fly, and produces engineering graded performance regarding attended pixel errors and recognition accuracy. As the proposed architecture is developmental, meaning that the internal representations are learned from pairs of input and motor signal, and thereby not manipulated internally for a specific task, we argue that the same learning principles and computational architecture can be potentially applicable to other sensory modalities, such as audition and touch.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: Robotics and Autonomous Systems - Volume 71, September 2015, Pages 35–48
نویسندگان
, ,