Object retrieval with large vocabularies and fast spatial matching. Philbin, J., Chum, O., Isard, M., Sivic, J., & Zisserman, A. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2007.
doi  abstract   bibtex   
In this paper, we present a large-scale object retrieval system. The user supplies a query object by selecting a region of a query image, and the system returns a ranked list of images that contain the same object, retrieved from a large corpus. We demonstrate the scalability and performance of our system on a dataset of over 1 million images crawled from the photo-sharing site, Flickr [3], using Oxford landmarks as queries. Building an image-feature vocabulary is a major time and performance bottleneck, due to the size of our dataset. To address this problem we compare different scalable methods for building a vocabulary and introduce a novel quantization method based on randomized trees which we show outperforms the current state-of-the-art on an extensive ground-truth. Our experiments show that the quantization has a major effect on retrieval quality. To further improve query performance, we add an efficient spatial verification stage to re-rank the results returned from our bagof-words model and show that this consistently improves search quality, though by less of a margin when the visual vocabulary is large. We view this work as a promising step towards much larger, "web-scale" image corpora. © 2007 IEEE.
@article{
 title = {Object retrieval with large vocabularies and fast spatial matching},
 type = {article},
 year = {2007},
 id = {c13a0854-90e0-33a8-9162-121f28795fc3},
 created = {2022-09-19T10:49:11.833Z},
 file_attached = {true},
 profile_id = {276016a7-2c9d-3507-8888-093db7c54774},
 group_id = {5ec9cc91-a5d6-3de5-82f3-3ef3d98a89c1},
 last_modified = {2022-09-19T10:50:33.691Z},
 read = {false},
 starred = {false},
 authored = {false},
 confirmed = {true},
 hidden = {false},
 folder_uuids = {02fb5526-03ff-44ad-8d5c-42bd496c3100},
 private_publication = {false},
 abstract = {In this paper, we present a large-scale object retrieval system. The user supplies a query object by selecting a region of a query image, and the system returns a ranked list of images that contain the same object, retrieved from a large corpus. We demonstrate the scalability and performance of our system on a dataset of over 1 million images crawled from the photo-sharing site, Flickr [3], using Oxford landmarks as queries. Building an image-feature vocabulary is a major time and performance bottleneck, due to the size of our dataset. To address this problem we compare different scalable methods for building a vocabulary and introduce a novel quantization method based on randomized trees which we show outperforms the current state-of-the-art on an extensive ground-truth. Our experiments show that the quantization has a major effect on retrieval quality. To further improve query performance, we add an efficient spatial verification stage to re-rank the results returned from our bagof-words model and show that this consistently improves search quality, though by less of a margin when the visual vocabulary is large. We view this work as a promising step towards much larger, "web-scale" image corpora. © 2007 IEEE.},
 bibtype = {article},
 author = {Philbin, James and Chum, Ondřej and Isard, Michael and Sivic, Josef and Zisserman, Andrew},
 doi = {10.1109/CVPR.2007.383172},
 journal = {Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition}
}

Downloads: 0