• Shuffle
    Toggle On
    Toggle Off
  • Alphabetize
    Toggle On
    Toggle Off
  • Front First
    Toggle On
    Toggle Off
  • Both Sides
    Toggle On
    Toggle Off
  • Read
    Toggle On
    Toggle Off
Reading...
Front

Card Range To Study

through

image

Play button

image

Play button

image

Progress

1/45

Click to flip

Use LEFT and RIGHT arrow keys to navigate between flashcards;

Use UP and DOWN arrow keys to flip the card;

H to show hint;

A reads text to speech;

45 Cards in this Set

  • Front
  • Back

What is perceptual organization? What does it help us with?

Grouping




The organization of perceptual elements into larger units - Gestalt principles




We constantly see fragmented objects but are still able to extract objects from the background & integrate them into something that is useful


- Helps us navigate the environment

What are gestalt principles? And what are they?

Grouping rules


- Connectedness, common regions, trade-offs between different rules




1. Proximity


- Changing position of identical objects can cause automatic perceptual grouping




2. Similarity


- Assume a smoothness constraint


- Group similar objects


e.g. Colour, orientation, size




3. Common fate


e.g. 2 objects moving in same direction




4. Symmetry, parallelism




5. Continuity / Closure

How is competition between different gestalt principles controlled?

One option will win


Usually modulated by top-down control. Our experiences shape how we group objects (e.g. grouping of facial features into a face)

Who studied figure-ground segmentation?

E. Rubin, 1921

What is figure-ground segmentation?

Figure = Object


Ground = Surrounding area (e.g. background)




We complete surroundedness as a cue


e.g. Rubin's reversible face-vase figure




It is not just the sum of visual input but segmentation of the scene

What could properties of figure vs ground be?

Figure


1. Thing-like


2. Close to observer


3. Bounded by contour


4. Shape is defined by contour




Ground


1. Not thing-like


2. Further from observer


3. Extends behind contour


4. No shape at contour

What could cues of figure ground organization be?

1. Complete surroundedness


2. Size (figure tends to be smaller than background)


3. Orientation preference


4. Contrast


5. Symmetry


6. Convexity (curvature)


7. Parallelism

What is visual (amodal) completion?

When visual system completes occluded objects


- Usually automatically





What are factors influencing amodal completion?

1. Occlusion


2. Figural familiarity


3. Figural simplicity / "minimum principle"


4. Familiarity




Constrains the possible solutions to find the most plausible due to usually familiarity





What is example of modal completion?

Illusory contours

What is the involvement of V1?

Helps figure/ground segmentation


- via Horizontal connections, excitatory pyramidal cells & inhibitory interneurons


- Visual input is filtered through the receptive fields to the excitatory cells




- If mismatch of oriented elements, would expect this to activate a confused representation due to different orientation preferences


- BUT, it highlights edges as transition from one object to another occurs

How would you simulate connectedness across neurons?


What does this result in?

1. Self-excitation (sustaining the "vote" for a specific percept)




2. Inhibition (suppress other neuron's "vote")




3. Local excitation (supports the neighbours, assuming that they like similar regions, they will choose the same "vote")




Results in some areas being more active than others.


Tells the visual system there is something "special" about the changes in orientation.




This is the seed of figure-ground segmentation

Who tested edge detection & region filling in V1? How?

Self et al., 2013




1. Monkeys look to a particular region where stimulus is displayed.


It is positioned so that the receptive field "sees" the center of the stimulus, edge of the stimulus or only the background




2. Record different layers of V1 (input, lateral connections, feedback connections etc.) & measure the timing of responses after stimulus onset

What did Self et al., 2013 find in edge detection & region filling in V1? What does this suggest?

Horizontal connections implement edge detection


(occured ~70ms after onset, with stronger activity in upper layer 4 & superficial areas)




Feedback connections implement region filling


(~100ms after onset, input arrives in layers 1, 2 & 5)




Suggests that some kind of object processing is already occurring at level of V1

What else could V1 be used for? Who studied this?

"Tracing" a visual route


Roelfsema

How was tracing of a visual route by V1 studied? What were the results?

1. Monkeys shown 2 objects with 2 curves to trace


2. Tell monkey to trace the first curve


3. Put a stimulus in a position such that the receptive field falls on the curve


4. Then tell monkey to trace the other curve, but leave receptive fields on 1st curve




Results:


1. Subtle, but measurable changes in how V1 will represent attended vs unattended object


2. Input is quickly transformed to take into account other factors

How does V1 & V2 respond to illusory contours? How was this tested? What does this suggest?

1. Stimulus moved across receptive field


2. Identical responses in V2 found for similar viewpoints, even if box covering stimulus had gaps on either side


3. Differences outside of the receptive field affect the response of the neurones. (e.g. if only half of the box around stimulus)


4. 40% of V2 neurons responded to illusory contours


5. End-stopped V1 cells induce "filling in" of missing contours




Suggests that "filling-in" gives rise to proto-objects

What is the hypothesis about V1 & V2 illusory contours?

There are neurons in V1 that are tuned to corners/ends of objects (end-stopped neurons)


Perhaps the integration of information across multiple neurons from V1 results in a "filling in" process in V2

What is role of V4 in illusory contours? Who studied this and how?

Cox et al., 2013




Test:


Activity of neurons with receptive fields on "blanks" of an illusory square surrounded by "pacman circles" or receptive fields on the "inducers" was analyzed


- Control was a flipped pacman stimuli (loss of illusory square)




Results:


1. If the receptive field was centered on the illusory surface, there was a greater V4 response than if there was no illusory square. (Respond to the illusion of a square)




2. Stronger representation of illusory square than when receptive field focused on the inducers. (Stronger for illusory stimulus than physical stimulus)





What are potential problems of the V4 illusory contour study?

Eye movements (attention drawn to physical stimuli while fixating on center) could result in higher activity

What is the different between Luminance and Lightness?

Luminance = Amount of light that reaches the eye


Is a physical quantity




Lightness = Perceived reflectance


is a psychological quantity




Perception is often independent of external illumination

What is lightness constancy?

1. Physical amount of light (luminance) doesn't necessarily determine what you see


- Percept is determined by relative reflectance





What is simultaneous contrast?

Physically identical strips of light look different based on its context/surroundings.


- Perceived lightness varies with background luminance


- For example, despite same reflectance, the same brightness of an object can be perceived differently depending on the background luminance

What does the visual system care more about? Local contrast or absolute luminance?

Local contrast

What is contrast gain control?

1. Contrast transfer characteristics adapt to the mean contrast in the input


- Shift dynamic range of representation based on illuminance


- i.e. bring down responses in a very bright environment in order to still get informative responses (avoid saturation)




2. 10^10-fold range of luminance levels is observed in natural scenes




3. Therefore it's better to respond more to local contrast than absolute luminance

What is shunting inhibition? How may it occur and what may it be useful for?

Silent inhibition as a mechanism for contrast gain control




Probably occurs via inhibitory interneurons and the conductance variation of inhibitory channels




Inhibitory neurons are in:


Perigeniculate nucleus (PGN) are feedback inhibited


Lateral geniculate nucleus (LGN) are feedforward inhibited by Retina

What is the normalization model?

Conductance of interneurons in modulated by many other cortical cells and firing rate is modulated by rectification?

What is relationship between mean firing rate of PGN/LGN vs contrast?

As contrast increases, mean firing rate increases & plateaus at around 3 contrast units

How does context affect perception of things? What was the experiment?

Even with similar inputs, context can change our perception




Edge type influences illumination match


(one target square affected by a shadow, and a nearby 'matching array' of 16 gray levels to match the gray-level of the target square)


e.g. if there are 2 contexts with matched local illuminance.




Border dividing 2 fields of illumination was made to appear either as a reflectance edge or an illumination edge by either concealing or revealing the larger context (e.g. showing a shadow producer or not)




1. Illumination edge


- Difference seems to arise from illumination


- Responses close to ratio match




2. Reflectance edge


- Difference looks like it arises from different material/colours


- Response is closer to luminance match




Thus, edge type influences lightness matching


- Luminance difference is ignored for illumination edge

What else influences lumination perception? Who studied this?

3D Structures - Adelson




3D cube with checkered pattern. Whether you look at edge with change of grey-scale or face of the cube with similar change affects your perception.



Local form cues modify transformation of luminances into perceived lightness


-> Features modulate perception of local brightness



What 2 things is luminance perception explained by?

1. Early visual circuitry


2. "inferential control" by higher-level knowledge about context

What is size constancy? Why may this occur?

The same perceived size despite different retinal image sizes


- Because our environment is 3D


- Must infer the 3D image from the 2D image on the retina & use depth cues to interpret relative sizes

Who measured perceived size and how? What was the result?

Holway & Boring, 1941




1. Subject sat with 2 walls (1 in front, 1 to the side, same visual angles)


2. Circles drawn on the walls of different sizes


3. Depth cues were either present (illumination of walls) or not




Results:


- When depth cues are present, correct size is inferred.


- When not, visual system relies on visual angle only and thus both stimuli perceived as same size irregardless of distance


- To perceive the stimuli as the same size, as the distance increases, the stimulus must be shrunk

What are some size constancy theories?

1. Compensation using "perceived distance"




2. Emmert's Law: Perceived size for constant retinal size depends on perceived distance




3. Proportionality hypothesis: Relative size is important




4. Texture gradients set an absolute spatial scale




5. Relative height


- Attribute size of objects based on how far away you think it is


- Further away, the bigger the visual angle


- Thus infer size based on visual angle the object subtends


- Height is relative to the horizon


- Angle ratio roughly equal to height ratio

What are some illusions that show size constancy?

1. Hallway illusion / Ames Room


- Perspective information suggests depth


- Stimuli are the same size, but depth cue suggests some stimuli are further. So the ones "further" away are perceived as larger


- Stimuli are only perceived the same size if the "further" ones are made smaller. Though then in reality, they aren't same size (different sized retinal image)




2. Ponzo illusion


- 2 horizontal lines on a vertical 'road' seem different sizes, but actually are the same


- No real depth perception, but depth system is still activated




3. Moon illusion


- Full moon near the horizon looks huge, whereas if its in the sky, looks smaller


- When there is context (e.g. mountains), you compare relative size of the moon & scaling factors & therefore perceive moon as huge


- This is because the perceived path of the moon is much closer to you than actual path (sky seems flattened)


- Use angle from horizon to determine size of object


What could underly some perceptual constancies in terms of receptive fields?

View-invariant neurons


e.g. Large view-invariant receptive fields & view-invariant object recognition

What are the major sources of information for depth perception?

1. Oculomotor


- Accomodation


- Convergence




2. Visual


- Binocular vs Monocular


- If monocular; static cues vs motion parallax


- If static cues; Interposition, size & perspective

What are oculomotor cues for depth? When do they work? How strong are these cues?

1. Monocular = Accomodation


2. Binocular = Convergence




3. Proprioceptive & motor command cues:


- Copy of eye muscle commands


- Measures of strain from organ tendons




Work for short distances (<5m)


Are weak cues

What are visual cues for depth?

1. Shading


- Where the shading is (above or below) affects our depth perception


- Usually assume illumination comes from above because of the sun (prior assumption).


- Therefore if shading comes from below, we assume the object is sticking out towards us.


- If shading from above, assume object is concave and going away from us




2. Perspective


- Linear (e.g. lines converging to a 'vanishing point')


- Non-linear (aerial) effects (e.g. atmospheric effects of haze --> What isn't in the haze is closer)




3. Texture gradients




4. Motion parallax


- Objects further away appear to move slower relative to you than those closer


- Based on relative motion


- Arises with & without ego-motion

How does stereovision help with vision?

1. Puts mathematical constraints on visual system


2. Images of left & right eye are slightly different


3. Disparity = spatial shift of corresponding points on the retina


- Crossed = nearer


- Uncrossed = further away from horopter


- Horopter = surface with disparity 0 on the retina (all points are mapped on corresponding points on retina)


4. When eyes converge, position of image on retina will correspond. When they diverge, it will not


- Distance between the eyes constrains this

How was stereovision measured and by who?

via a stereoscope


Wheatstone 1838

What is the correspondance problem? What is another name for it?

An ill-posed problem with no unique solution.


Marr-Pogio algorithm




"Are 2 pixels of two pictures corresponding to each other?"



What is the Marr-Pogio algorithm? Where in the brain does this occur?

1. Each neuron is interested in relative difference between image 1 & 2 composed of black & white pixels


- Black pixel can come from any point of the image.


- Other image also has many light sources


- Each neuron corresponds to a certain depth


- Some neurons vote for relative image disparity




2. Neurons compete through lateral interactions.


- Those that vote for same depth, support each other (self-excitation) and compete (inhibit) with neurons that support the other depth




3. Result is that the most likely depth emerges as the winner & neural networks converge on how far the image is




4. Occurs in V1 & V2 to sense image disparity









What is correspondance problem of motion based on?

Time-to-time rather than space-to-space

What are the neural mechanisms of image disparity?

Disparity selective neurons in V1 & V2 (some also in V3, MT & MST) are selective to specific disparities but not others (disparity-tuned)


- Some will be excited if 2 spots are at slightly different locations


- Others excited more if 2 spots at same location (zero disparity)




Neurons have different tuning characteristics (e.g. some are tuned for far, some for near, some for zero disparity)