Globality–Locality-Based Based Consistent Discriminant Feature Ensemble for Multicamera Tracking
Abstract: Spatiotemporal data association and fusion is a well well-known NP--hard problem even in a small number of cameras and frames. Although it is difficult to be tractable, solving them is pivotal for tracking in a multicamera network. Most approaches model associat association ion maladaptively toward properties and contents of video, and hence they produce suboptimal associations and association errors propagate over time to adversely affect fusion. In this paper, we present an online multicamera multitarget tracking framework that performs adaptive tracklet correspondence by analyzing and understanding contents and properties of video. Unlike other methods that work only on synchronous videos, our approach uses dynamic time warping to establish correspondence even if videos have hav linear or nonlinear time asynchronous relationship. Association is a two two--stage process based on geometric and appearance descriptor space ranked by their interinter and intra-camera camera consistency and discriminancy. Fusion is reinforced by weighting the associated ted tracklets with a confidence score calculated using reliability of individual camera tracklets. Our robust ranking and election learning algorithm dynamically selects appropriate features for any given video. Our method establishes that, given the right ensemble of features, even computationally efficient optimization yields better accuracy in tracking over time and provides faster convergence that is suitable for real real-time time application. For evaluation on