Integration of Computer Vision Components into a Multi-modal Cognitive System
Proceedings of the Fourteenth Computer Vision Winter Workshop (CVWW), 2009
We present a general method for integrating visual components into a multi-modal cognitive system. The integration is very generic and can work with an arbitrary set of other modalities. We illustrate our integration approach with a specific instantiation of the architecture schema that focuses on integration of vision and language: a cognitive system able to collaborate with a human, learn and display some understanding of its surroundings. As examples of cross-modal interaction we describe mechanisms for clarification and visual learning.