Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/87288
Citations
Scopus Web of Science® Altmetric
?
?
Type: Conference paper
Title: Robust real-time visual tracking using pixel-wise posteriors
Author: Bibby, C.
Reid, I.
Citation: Lecture Notes in Artificial Intelligence, 2008 / Forsyth, D., Torr, P., Zisserman, A. (ed./s), vol.5303 LNCS, iss.PART 2, pp.831-844
Publisher: Springer Berlin Heidelberg
Publisher Place: Germany
Issue Date: 2008
Series/Report no.: Lecture Notes in Computer Science, 2008, vol. 5303
ISBN: 3540886850
9783540886884
ISSN: 0302-9743
1611-3349
Conference Name: 10th European Conference on Computer Vision (ECCV) (12 Oct 2008 - 18 Oct 2008 : Marseille, France)
Editor: Forsyth, D.
Torr, P.
Zisserman, A.
Statement of
Responsibility: 
Charles Bibby and Ian Reid
Abstract: We derive a probabilistic framework for robust, real-time, visual tracking of previously unseen objects from a moving camera. The tracking problem is handled using a bag-of-pixels representation and comprises a rigid registration between frames, a segmentation and online appearance learning. The registration compensates for rigid motion, segmentation models any residual shape deformation and the online appearance learning provides continual refinement of both the object and background appearance models. The key to the success of our method is the use of pixel-wise posteriors, as opposed to likelihoods. We demonstrate the superior performance of our tracker by comparing cost function statistics against those commonly used in the visual tracking literature. Our comparison method provides a way of summarising tracking performance using lots of data from a variety of different sequences.
Rights: © Springer-Verlag Berlin Heidelberg 2008
DOI: 10.1007/978-3-540-88688-4_61
Published version: http://dx.doi.org/10.1007/978-3-540-88688-4_61
Appears in Collections:Aurora harvest 7
Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.