Exploiting Local and Global Patch Rarities for Saliency Detection. Borji, A. & Itti, L. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, Rhode Island, pages 1-8, Jun, 2012.
abstract   bibtex   
We introduce a saliency model based on two key ideas. The first one is considering local and global image patch rarities as two complementary processes. The second one is based on our observation that for different images, one of the RGB and Lab color spaces outperforms the other in saliency detection. We propose a framework that measures patch rarities in each color space and combines them in a final map. For each color channel, first, the input image is partitioned into non-overlapping patches and then each patch is represented by a vector of coefficients that linearly reconstruct it from a learned dictionary of patches from natural scenes. Next, two measures of saliency (Local and Global) are calculated and fused to indicate saliency of each patch. Local saliency is distinctiveness of a patch from its surrounding patches. Global saliency is the inverse of a patch’s probability of happening over the entire image. The final saliency map is built by normalizing and fusing local and global saliency maps of all channels from both color systems. Extensive evaluation over four benchmark eye-tracking datasets shows the significant advantage of our approach over 10 state-of-the-art saliency models.
@inproceedings{ Borji_Itti12cvpr,
  author = {A. Borji and L. Itti},
  title = {Exploiting Local and Global Patch Rarities for Saliency Detection},
  booktitle = {Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, Rhode Island},
  abstract = {We introduce a saliency model based on two key ideas.  The first one is considering local and global image
                  patch rarities as two complementary processes. The second one is based on our observation that for
                  different images, one of the RGB and Lab color spaces outperforms the other in saliency detection. We
                  propose a framework that measures patch rarities in each color space and combines them in a final
                  map. For each color channel, first, the input image is partitioned into non-overlapping patches and
                  then each patch is represented by a vector of coefficients that linearly reconstruct it from a learned
                  dictionary of patches from natural scenes. Next, two measures of saliency (Local and Global) are
                  calculated and fused to indicate saliency of each patch. Local saliency is distinctiveness of a patch
                  from its surrounding patches. Global saliency is the inverse of a patch’s probability of happening
                  over the entire image.  The final saliency map is built by normalizing and fusing local and global
                  saliency maps of all channels from both color systems. Extensive evaluation over four benchmark
                  eye-tracking datasets shows the significant advantage of our approach over 10 state-of-the-art
                  saliency models.},
  pages = {1-8},
  month = {Jun},
  year = {2012},
  review = {full/conf},
  type = {bu;mod},
  if = {2012 acceptance rate: 26.2%},
  file = {http://ilab.usc.edu/publications/doc/Borji_Itti12cvpr.pdf}
}

Downloads: 0