Abstract
Recognizing persons over a system of disjunct cameras is a hard task for human operators and even harder for automated systems. In particular, realistic setups show difficulties such as different camera angles or different camera properties. Additionally, also the appearance of exactly the same person can change dramatically due to different views (e.g., frontal/back) of carried objects. In this paper, we mainly address the first problem by learning the transition from one camera to the other. This is realized by learning a Mahalanobis metric using pairs of labeled samples from different cameras. Building on the ideas of Large Margin Nearest Neighbor classification, we obtain a more efficient solution which additionally provides much better generalization properties. To demonstrate these benefits, we run experiments on three different publicly available datasets, showing state-of-the-art or even better results, however, on much lower computational efforts. This is in particular interesting since we use quite simple color and texture features, whereas other approaches build on rather complex image descriptions!
Original language | English |
---|---|
Title of host publication | Proceedings of the IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS) |
Publisher | . |
Pages | 203-208 |
DOIs | |
Publication status | Published - 2012 |
Event | 9th IEEE International Conference on Advanced Video and Signal-Based Surveillance: AVSS 2012 - Beijing, China Duration: 18 Sept 2012 → 21 Sept 2012 |
Conference
Conference | 9th IEEE International Conference on Advanced Video and Signal-Based Surveillance |
---|---|
Country/Territory | China |
City | Beijing |
Period | 18/09/12 → 21/09/12 |
Fields of Expertise
- Information, Communication & Computing