A joint sparsity model for video anomaly detection

Xuan Mo, Vishal Monga, Raja Bala, Zhigang Fan

Research output: Chapter in Book/Report/Conference proceedingConference contribution

2 Scopus citations

Abstract

Video anomaly detection can be used in the transportation domain to identify unusual patterns such as traffic violations, accidents, unsafe driver behavior, street crime, and other suspicious activities. A common class of approaches relies upon object tracking and trajectory analysis. A key challenge is the ability to effectively handle occlusions among objects and their trajectories. Another challenge is the detection of joint anomalies between multiple moving objects. Recently sparse reconstruction techniques have been used for image classification, and shown to provide excellent robustness to occlusion. This paper proposes a new joint sparsity model for anomaly detection that effectively addresses both the robustness to occlusion and the detection of joint anomalies involving multiple objects. Experimental results on real and synthetic data demonstrate the effectiveness of our approach for both single-object and multi-object anomalies.

Original languageEnglish (US)
Title of host publicationConference Record of the 46th Asilomar Conference on Signals, Systems and Computers, ASILOMAR 2012
Pages1969-1973
Number of pages5
DOIs
StatePublished - 2012
Event46th Asilomar Conference on Signals, Systems and Computers, ASILOMAR 2012 - Pacific Grove, CA, United States
Duration: Nov 4 2012Nov 7 2012

Publication series

NameConference Record - Asilomar Conference on Signals, Systems and Computers
ISSN (Print)1058-6393

Other

Other46th Asilomar Conference on Signals, Systems and Computers, ASILOMAR 2012
Country/TerritoryUnited States
CityPacific Grove, CA
Period11/4/1211/7/12

All Science Journal Classification (ASJC) codes

  • Signal Processing
  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'A joint sparsity model for video anomaly detection'. Together they form a unique fingerprint.

Cite this