Saliency Using Natural Statistics (SUN)
In Kanan et al. (2009) we developed the SUN model of top-down visual saliency. As of early 2009, it is one of the best models for predicting what people will look at in natural images. However, the model knows nothing about human eye movements. SUN's predictive power is based on its ability to find the areas in a picture that are most likely to be the object it is searching for, just as people do. For the full details of the model's implementation and evaluation please refer to our paper.
To develop SUN we used the LabelMe dataset to train a classifier using features inspired by the properties of neurons in primary visual cortex. Torralba et al. (2006) gathered eye movement data from people who were told to look for particular objects (mugs, paintings, and pedestrians) in natural images. We used their data to evaluate how well SUN predicts the subject's eye movements when it is given the very same images, which SUN has never seen before. We compared our model to Torralba et al.'s Contextual Guidance Model, which is one of the few models with a comparable ability to predict human eye movements in natural images.