computational cognitive science · compute feature maps for color, intensity, orientation at di...

33
Computational Cognitive Science Lecture 15: Visual Attention Chris Lucas (Slides adapted from Frank Keller’s) School of Informatics University of Edinburgh [email protected] 14 November 2017 1 / 28

Upload: others

Post on 12-Aug-2020

3 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Computational Cognitive ScienceLecture 15: Visual Attention

Chris Lucas(Slides adapted from Frank Keller’s)

School of InformaticsUniversity of [email protected]

14 November 2017

1 / 28

Page 2: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

1 IntroductionThe Visual Processing PipelineAttention and Visual Features

2 The Saliency ModelModel ArchitectureFeature MapsSaliency MapInhibition of Return

3 ResultsRobustness to NoiseStrengths and Limitations

Reading: Itti, Koch, and Niebur (1998).

2 / 28

Page 3: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

The Visual Processing Pipeline

The rest of the course will deal with human visual cognition. Wewill focus on high-level visual processing (not visual neuroscience):

Visual attention: How do we decide which parts of an imageto focus on?

Visual search: How do we search for a target in an image?

Object recognition: How do we identify objects in an image?

We will introduce computational models in all three domains.

3 / 28

Page 4: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

The Visual Processing Pipeline

When we view an image, we actually see this:

Imagefrom

http://eyetrackin

g.m

e/

Only the fovea, a small area in the center of the retina, is in focus.

4 / 28

Page 5: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

The Visual Processing Pipeline

In order to take in the whole image, we have to move our eyes:fixations (stationary periods) and saccades (rapid movements).

471

119 205

216

113

340

161

238

355250

154

Imagefro

mHen

derso

n,2003

How do we determine where to look? We need to work out whicharea are interesting, i.e., attract visual attention.

5 / 28

Page 6: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Visual Saliency

We attend to the areas that are visually salient. An area is salientif it stands out, is different from the rest of the image.

−→

The visual system computes a saliency map of the image, and thenmoves the eyes to the most salient regions in turn (Itti et al.,1998).

6 / 28

Page 7: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Visual Features

Saliency can be tested using visual search experiments: participantshave to find a target item among a number of distractors.

Examples for visual features that can make a target salient:

color;

orientation;

intensity.

Saliency can make the target pop out from its distractors if itdiffers in one of this features.

The pop-out effect doesn’t occur if the target is different from thedistractors in two aspects (conjunction target).

7 / 28

Page 8: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Visual Features

Pop-out because of color (Itti, 2007):

8 / 28

Page 9: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Visual Features

Pop-out because of orientation (Itti, 2007):

9 / 28

Page 10: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Visual Features

No pop-out: conjunction target (Itti, 2007):

10 / 28

Page 11: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Model Architecture

Itti et al.’s (1998) computational model of saliency:

compute feature maps for color, intensity, orientation atdifferent scales;

compute center-surround difference and apply a normalization;

combine the maps across scales into conspicuity maps;

saliency map is a linear combination of the conspicuity maps;

winner-takes-all operator predicts attended locations.

This model mainly works for free viewing. In the next lectures wewill talk about models that can account for search data.

11 / 28

Page 12: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Model Architecture

A Model of Saliency-Based Visual Attention

—A visual attention system, inspired by the behavior and theneuronal architecture of the early primate visual system, is presented.

locations in order of decreasing saliency. The system breaks down the

computationally efficient manner, conspicuous locations to be analyzed

real time, despite the limited speed of the neuronal hardware avail-

12 / 28

Page 13: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Feature Maps

Feature maps are computed at nine spatial scales (1:1 to 1:256) bylow-pass filtering (blurring) and subsampling the image.

A center-surround operator is used to detect locations that standout from their surroundings:

this is implemented as the difference between finer andcoarser scales;

the center is a pixel at scale c ∈ {2, 3, 4};the surround is the corresponding pixel at scale s = c + d ,with d ∈ {3, 4};the across-scale difference between two maps is denoted as .

13 / 28

Page 14: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Intensity

At each spatial scale, a set of feature maps are computed based onthe red, green, and blue color values (r , g , b) of the pixels.

Intensity map: compute the intensity function I = (r + g + b)/3and then the intensity map using the center surround operator:

I(c , s) = |I (c) I (s)|

with c ∈ {2, 3, 4} and s = c + d , with d ∈ {3, 4}.

14 / 28

Page 15: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Color

Color maps: compute four color values R = r − (g + b)/2 for red,G = g − (r + b)/2 for green, B = b − (r + g)/2 for blue, andY = (r + g)/2− |r − g |/2− b for yellow.

Then compute color maps again using center-surround:

RG(c , s) = |(R(c)− G (c)) (G (s)− R(s))|

BY(c , s) = |(B(c)− Y (c)) (Y (s)− B(s))|

These are based on color opponencies (exist in the visual cortex).

15 / 28

Page 16: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Orientation

Orientation map: compute Gabor pyramids O(σ, θ) whereσ ∈ [0 . . . 8] is the scale and θ ∈ {0◦, 45◦, 90◦, 135◦} is thepreferred orientation.

Then compute color maps again using center-surround:

O(c , s, θ) = |O(c , θ) O(s, θ)|

In total, 42 feature maps are computed: six for intensity, 12 forcolor, and 24 for orientation.

16 / 28

Page 17: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Example

C I O

17 / 28

Page 18: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Saliency Map

Before we combine feature maps, a normalization operator N (·) isapplied, which promotes maps with a small number of strongpeaks, and suppressed maps with many similar peaks.

18 / 28

Page 19: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Saliency Map

The feature maps are combined into three conspicuity maps forintensity, color, and orientation at the same scale (σ = 4).

For intensity and color, we get:

I = ⊕4c=2 ⊕c+4

s=c+3 N (I(c , s))

C = ⊕4c=2 ⊕c+4

s=c+3 [N (RG(c , s)) +N (BY(c , s))]

where the ⊕ operator reduces each map to scale 4 and performspoint-by-point addition.

19 / 28

Page 20: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Saliency Map

For orientation, we first combine the six feature maps for a givenangle and then add them to get a single conspicuity map:

O =∑

θ∈{0◦,45◦,90◦,135◦}

N (⊕4c=2 ⊕c+4

s=c+3 N (O(c, s, θ)))

The overall saliency map is then computed by normalizing andaveraging the three conspicuity maps:

S =1

3(N (I) +N (C) +N (O))

Why do we normalize each conspicuity map separately? Similarfeatures compete strongly for saliency, while different onescontribute independently to saliency.

20 / 28

Page 21: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Example

21 / 28

Page 22: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Inhibition of Return

Now we can predict sequences of fixations from a saliency map:

the maximum of S is the most salient location, whichbecomes the focus of attention (FOA);

all other locations are ignored (inhibited);

then the saliency around the FOA is reset, so that the secondmost salient location becomes the new FOA.

The last property is crucial: it results in inhibition of return, so thatthe FOA doesn’t immediate return to the most salient location.

Itti et al. (1998) implement this using a winner-take-all neuralnetwork. This allows them to simulate fixation durations.

22 / 28

Page 23: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Inhibition of Return

23 / 28

Page 24: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Inhibition of Return

23 / 28

Page 25: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Inhibition of Return

23 / 28

Page 26: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Inhibition of Return

23 / 28

Page 27: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Robustness to Noise

Test the model by adding noise to the image, see if it is still ableto pick out salient locations correctly.

24 / 28

Page 28: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Robustness to Noise

Test the model by adding noise to the image, see if it is still ableto pick out salient locations correctly.

24 / 28

Page 29: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Robustness to Noise

Test the model by adding noise to the image, see if it is still ableto pick out salient locations correctly.

24 / 28

Page 30: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Evaluation

Evaluation reported by Itti et al. (1998):

saliency model can reproduce human performance in pop-outtasks (including conjunction target);

tested also on images of traffic signs, red soda cans, andemergency triangles (though no details given in the paper);

outperforms spatial frequency models.

No evaluation of saliency against eye-tracking data. However, thereis a lot of subsequent work on this topic, such as Borji, Sihite, andItti (2013).

25 / 28

Page 31: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Strengths and LimitationsStrengths:

simple feed-forward architectures generates complex behavior;

massively parallel implementation (biologically plausible);

very successful as model of early visual processing.

Weaknesses:

can only detect regions that are salient based on either color,intensity, or orientation;

other features (e.g., T junctions, line termination) orconjunctions of features are not accounted for in the model;

motion is important for saliency, but is not modeled;

the normalization function N (·) plays a crucial role withoutbeing theoretically well-founded;

no notion of object in the model (saliency is a property of apoint); but objectness crucial for human scene perception.

26 / 28

Page 32: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

Summary

Attention selects the part of the visual input which is fixatedand processed in detail;

attention is directed to visually salient areas in an image, i.e.,areas that are different from the rest of the image;

the saliency model is based on color, orientation, intensitymaps computed at various spatial scales;

center-surround differences are applied, and the mapsnormalized and combined into a single saliency map;

a winner-takes-all mechanism then predicts attended locations;

model is robust to noise and models human fixation behavior.

27 / 28

Page 33: Computational Cognitive Science · compute feature maps for color, intensity, orientation at di erent scales; compute center-surround di erence and apply a normalization; combine

References

Borji, A., Sihite, D. N., & Itti, L. (2013). Quantitative analysis of human-modelagreement in visual saliency modeling: A comparative study. IEEETransactions on Image Processing, 22(1), 55–69.

Henderson, J. (2003). Human gaze control in real-world scene perception.Trends in Cognitive Sciences, 7, 498–504.

Itti, L. (2007). Visual salience. Scholarpedia, 2(9), 3327.

Itti, L., Koch, C., & Niebur, E. (1998). A model of saliency-based visualattention for rapid scene analysis. IEEE Transactions on PatternAnalysis and Machine Intelligence, 20(11), 1254–1259.

28 / 28