Information fusion for object & situation assessment in sensor networks

Abhishek Srivastav, Yicheng Wen, Evan Hendrick, Ishanu Chattopadhyay, Asok Ray, Shashi Phoha

Research output: Chapter in Book/Report/Conference proceedingChapterpeer-review

Abstract

A semantic framework for information fusion in sensor networks for object and situation assessment is proposed. The overall vision is to construct machine representations that would enable human-like perceptual understanding of observed scenes via fusion of heterogeneous sensor data. In this regard, a hierarchical framework is proposed that is based on the Data Fusion Information Group (DFIG) model. Unlike a simple set-theoretic information fusion methodology that leads to loss of information, relational dependencies are modeled as cross-machines called relational Probabilistic Finite State Automata using the xD-Markov machine construction. This leads to a tractable approach for modeling composite patterns as structured sets for both object and scene representation. An illustrative example demonstrates the superior capability of the proposed methodology for pattern classification in urban scenarios. © 2011 AACC American Automatic Control Council.
Original languageAmerican English
Title of host publicationProceedings of the American Control Conference
Pages1274-1279
Number of pages6
DOIs
StatePublished - 2011

Publication series

NameProceedings of the American Control Conference

Fingerprint

Dive into the research topics of 'Information fusion for object & situation assessment in sensor networks'. Together they form a unique fingerprint.

Cite this