Please use this identifier to cite or link to this item:
https://hdl.handle.net/2440/90334
Type: | Thesis |
Title: | Efficient and robust image ranking for object retrieval. |
Author: | Chen, Yanzhi |
Issue Date: | 2013 |
School/Discipline: | School of Computer Science |
Abstract: | This thesis focuses on efficient and effective object retrieval from an unlabelled collection of images. The goal of object retrieval is to, given a query image depicting an object, return the dataset images containing that same object, quickly and accurately. Due to its simplicity and efficiency, it is common to use a “Bag-of-Words” (BoW) model in which each image is represented as a weighted vector of quantised features, known as visual words. Although the BoW retrieval system is efficient, the extraction and quantisation of local image features introduces errors into the retrieval results. We build our retrieval system on the BoW model, proposing three kinds of method to improve the retrieval accuracy: i) refinement of BoW image representation; ii) refinement of image similarity; iii) retrieval result re-ranking. Firstly, a visual thesaurus structure is proposed to discover the spatial relatedness of visual words. Based on these, a spatial expansion method is able to enrich the original query with those spatially related visual words (enriched by a general thesaurus) and spatially related foreground words (enriched by an object-based thesaurus). Therefore, the BoW image representation is improved. The second contribution improves the standard image similarity used in the BoW retrieval system such that the similarity between query/dataset images is better described. We do this by a cross-word image matching scheme, such that matching features mapped to different visual words are able to contribute to the similarity score. Thirdly, we also aim at efficient result re-ranking methods to improve the initial retrieval results. We present two re-ranking methods in this thesis. A context based re-ranking method is based on the analysis of correlated subsets of image dataset, called “contexts”. Images that share contexts are weakly correlated to each other, and should therefore mutually influence each other’s ranking. The initial ranking scores are refined by this contextual information. We also present a ranking verification method that is able to extract a set of reliable query relevant images from the retrieved results and thus can be applied in a number of object retrieval applications. Note that neither method needs to recover low level feature information or prior knowledge from the dataset. Instead, they utilize ranking information during run time. We also revisit the definition of the object retrieval problem and propose a group-query method, in which the query is a collection of images depicting the same object instead of a single query image used in the traditional “query-by-example” methods. |
Advisor: | Dick, Anthony Robert Li, Xi |
Dissertation Note: | Thesis (Ph.D.) -- University of Adelaide, School of Computer Science, 2013 |
Keywords: | object retrieval; bag-of-words; visual thesaurus; ranking verification; group-query |
Provenance: | This electronic version is made publicly available by the University of Adelaide in accordance with its open access policy for student theses. Copyright in this thesis remains with the author. This thesis may incorporate third party material which has been used by the author pursuant to Fair Dealing exceptions. If you are the owner of any included third party copyright material you wish to be removed from this electronic version, please complete the take down form located at: http://www.adelaide.edu.au/legals |
Appears in Collections: | Research Theses |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
01front.pdf | 677.1 kB | Adobe PDF | View/Open | |
02chapter1-3.pdf | 19.5 MB | Adobe PDF | View/Open | |
03chapter4-bib.pdf | 21.02 MB | Adobe PDF | View/Open | |
Permissions Restricted Access | Library staff access only | 4.13 MB | Adobe PDF | View/Open |
Restricted Restricted Access | Library staff access only | 134.95 MB | Adobe PDF | View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.