Advances in learning visual saliency: From image primitives to semantic contents

Qi Zhao, Christof Koch

Research output: Chapter in Book/Report/Conference proceedingChapter

4 Scopus citations

Abstract

Humans and other primates shift their gaze to allocate processing resources to a subset of the visual input. Understanding and emulating the way that human observers free-view a natural scene has both scientific and economic impact. While previous research focused on low-level image features in saliency, the problem of “semantic gap” has recently attracted attention from vision researchers, and higher-level features have been proposed to fill the gap. Based on various features, machine learning has become a popular computational tool to mine human data in the exploration of how people direct their gaze when inspecting a visual scene. While learning saliency consistently boosts the performance of a saliency model, insights of what is learned inside the black box is also of great interest to both the human vision and computer vision communities. This chapter introduces recent advances in features that determine saliency, reviews related learning methods and insights drawn from learning outcomes, and discusses resources and metrics in saliency prediction.

Original languageEnglish (US)
Title of host publicationNeural Computation, Neural Devices, and Neural Prosthesis
PublisherSpringer New York
Pages335-360
Number of pages26
ISBN (Electronic)9781461481515
ISBN (Print)9781461481508
DOIs
StatePublished - Jan 1 2014

Bibliographical note

Publisher Copyright:
© Springer Science+Business Media New York 2014.

Fingerprint

Dive into the research topics of 'Advances in learning visual saliency: From image primitives to semantic contents'. Together they form a unique fingerprint.

Cite this