Visual System Structure and Processing (6A)

Help Questions

MCAT Psychological and Social Foundations › Visual System Structure and Processing (6A)

Questions 1 - 10
1

A perception experiment presents a moving dot that travels left-to-right across a screen. In one condition, the dot briefly disappears behind an occluder and then reappears at a location consistent with a smooth trajectory; participants report a strong sense of continuous motion. In another condition, the dot reappears at an implausible location given its prior velocity; continuity reports drop sharply. Based on the scenario, which conclusion is most consistent with the principle that the visual system uses top-down expectations to resolve ambiguous input?

Continuity reports should be unaffected because motion perception depends only on retinal stimulation at each moment.

Continuity reports should depend mainly on binocular disparity cues, which are eliminated by the occluder in both conditions.

Continuity reports should be higher when reappearance matches prior motion because predictions about object persistence constrain interpretation under occlusion.

Continuity reports should increase when reappearance violates the prior trajectory because surprise enhances feature binding.

Explanation

This question tests understanding of top-down visual processing and how expectations influence perception of object continuity. The visual system uses predictive coding to resolve ambiguous situations like occlusion, where sensory input is temporarily absent. When an object disappears behind an occluder and reappears at a location consistent with its prior trajectory, the brain's prediction of smooth motion is confirmed, leading to strong continuity perception. When the object violates expected trajectory by reappearing at an implausible location, this prediction error disrupts continuity perception. Choice C correctly captures this principle of prediction-based interpretation, while choice B incorrectly suggests surprise enhances binding. A key check for top-down effects: when sensory evidence is ambiguous, prior expectations guide perception more strongly.

2

In a lab study of the visual pathway, participants view a bright flash presented only to the left visual field while fixating centrally. fMRI shows increased activity in the right primary visual cortex (V1). A subgroup with a lesion restricted to the optic chiasm shows reduced V1 activation compared with controls for the same stimulus, despite intact retinal responses. Which statement best reflects the visual pathway described?

The left visual field is represented in right V1 because each eye projects entirely to the contralateral cortex, and chiasm damage eliminates input from the left eye only.

The left visual field is represented in right V1 because temporal retinal fibers cross at the optic chiasm, and chiasm damage primarily disrupts temporal fiber crossing.

The left visual field is represented in right V1 because nasal retinal fibers cross at the optic chiasm, and chiasm damage disrupts this crossing.

The left visual field is represented in right V1 because information is rerouted through the superior colliculus before reaching the ipsilateral visual cortex when the chiasm is damaged.

Explanation

This question tests understanding of the visual pathway anatomy and how visual field information crosses at the optic chiasm. In the visual system, light from the left visual field strikes the nasal (medial) retina of the left eye and the temporal (lateral) retina of the right eye. The key anatomical principle is that nasal retinal fibers cross at the optic chiasm while temporal fibers remain ipsilateral, resulting in the left visual field being processed by the right hemisphere's V1. When the optic chiasm is damaged, the crossing nasal fibers are disrupted, reducing the signal reaching the contralateral cortex. Choice A correctly identifies this crossing pattern, while choice B incorrectly states that temporal fibers cross. To verify visual pathway organization, remember that nasal fibers cross and temporal fibers don't, ensuring each hemisphere processes the contralateral visual field.

3

In a color perception study, participants view a gray object under a bluish illuminant and then under a yellowish illuminant. The object’s reflected spectrum is manipulated so that the cone excitations at the retina are matched across the two illuminants, yet participants still report the object as “roughly the same gray” in both conditions. Based on the scenario, which conclusion is most consistent with color constancy as a perceptual process?

The reports imply that color perception depends primarily on binocular depth cues, which remain constant across illuminants.

The reports imply that perceived color is determined solely by absolute cone activation, so matching cone excitations guarantees identical perception.

The reports imply that the visual system discounts the illuminant using contextual inference, stabilizing perceived surface color across lighting changes.

The reports imply that rods dominate color judgments in bright light, overriding cone-based differences across illuminants.

Explanation

This question tests understanding of color constancy, a visual processing mechanism that maintains stable color perception despite changing illumination. Color constancy works by the visual system estimating and discounting the illuminant based on contextual cues, allowing perception of surface reflectance properties rather than raw cone activations. Even when cone excitations are matched across different illuminants, the brain uses surrounding context to infer the lighting conditions and maintain consistent object color perception. Choice B correctly identifies this illuminant discounting mechanism, while choice A incorrectly suggests color depends only on absolute cone activation. A key principle: color perception involves computational inference about surfaces and lighting, not just passive recording of wavelengths.

4

A visual attention task presents a target letter among distractors. When the target differs by a single basic feature (e.g., color), reaction time is nearly constant as the number of distractors increases. When the target is defined by a conjunction of features (e.g., color and orientation), reaction time increases with more distractors. Based on the scenario, which conclusion is most consistent with the underlying visual processing principle?

The pattern suggests that single-feature targets require binocular disparity computations that slow search with more distractors.

The pattern suggests that search slopes are determined only by retinal adaptation, not by attentional selection.

The pattern suggests that conjunction targets are detected preattentively, so distractor number should not matter.

The pattern suggests that conjunction targets rely more on serial, attention-demanding binding than single-feature targets.

Explanation

This question tests understanding of parallel versus serial visual search and feature binding mechanisms. Single-feature targets that differ in one basic dimension (color alone) can be detected through parallel, preattentive processing where the unique feature "pops out" regardless of distractor number. Conjunction targets requiring binding of multiple features (color AND orientation) demand serial, attention-dependent search where each item must be examined to bind features correctly. The increasing reaction time with more distractors for conjunction targets reflects this serial binding process. Choice A correctly identifies this distinction between parallel feature detection and serial conjunction binding, while choice C incorrectly claims conjunction search is preattentive. Remember that attention is required to bind features from different dimensions into coherent objects.

5

Participants view a classic illusion in which two horizontal line segments are physically equal in length, but one appears longer because of different arrowhead orientations at the ends (inward vs outward). When participants are instructed to judge length after prolonged exposure to the display, the illusion magnitude decreases slightly but does not disappear. Based on the scenario, which conclusion is most consistent with why the illusion deceives the brain?

The illusion persists because early visual processing incorporates contextual cues that bias size judgments even when observers know the segments are equal.

The illusion persists because the retina transduces inward arrowheads as shorter wavelengths than outward arrowheads.

The illusion persists because binocular disparity between the two lines forces the visual cortex to compute different physical lengths.

The illusion persists because the optic nerve inverts the image, causing systematic overestimation of the upper line segment.

Explanation

This question tests understanding of how contextual visual processing creates persistent illusions despite conscious knowledge. Visual illusions like the Müller-Lyer (arrowhead) illusion occur because early visual processing automatically incorporates contextual cues that influence size perception before conscious awareness. The visual system evolved to interpret 3D scenes where converging lines often indicate depth, causing systematic biases in 2D displays. Even when observers know the lines are equal, these automatic contextual computations in early visual areas continue to influence perception. Choice A correctly identifies this automatic contextual processing, while choice C incorrectly attributes the effect to wavelength transduction. Remember that many visual illusions persist because they exploit hardwired processing mechanisms that operate before conscious control.

6

Researchers test depth perception by having participants catch a foam ball tossed along the midline under two viewing conditions: (1) both eyes open, (2) one eye patched. The room lighting is constant, and the ball’s size and texture are unchanged. Participants’ timing errors increase substantially with one eye patched, especially when the ball is within arm’s reach. Which outcome related to depth perception would be expected?

Errors should increase only at far distances because retinal disparity grows as objects move farther away.

Errors should decrease with one eye patched because monocular cues become more salient and replace binocular convergence.

Errors should be unchanged because motion parallax is a binocular cue that remains intact with one eye patched.

Errors should increase with one eye patched because binocular disparity is reduced, disproportionately affecting judgments at near distances.

Explanation

This question tests understanding of binocular depth cues and their importance for near-distance perception. Binocular disparity arises from the slightly different images received by each eye, providing crucial depth information especially for objects within reaching distance. When one eye is patched, this disparity information is lost, forcing reliance on monocular cues like size and texture which are less precise for near objects. The increased errors at arm's reach specifically indicate loss of binocular convergence and disparity, which are most effective at close range. Choice A correctly identifies this relationship, while choice D incorrectly states disparity increases with distance (it actually decreases). Remember that binocular cues dominate depth perception for near objects, while monocular cues become relatively more important at far distances.

7

A patient can accurately point to a briefly flashed target in the left visual field but reports no conscious awareness of seeing anything there. Structural imaging shows damage to right V1 with relative sparing of subcortical visual structures. Based on the scenario, which conclusion is most consistent with the visual processing principle involved?

The behavior is most consistent with blindsight, in which non-geniculostriate pathways support visually guided action without conscious perception.

The behavior is most consistent with improved binocular disparity processing in the damaged hemisphere.

The behavior is best explained by enhanced color constancy, allowing accurate pointing without awareness.

The behavior is best explained by complete decussation at the optic chiasm, which preserves conscious vision despite V1 damage.

Explanation

This question tests understanding of blindsight and alternative visual pathways beyond the geniculostriate route. Blindsight occurs when damage to V1 eliminates conscious visual perception, but preserved subcortical structures like the superior colliculus and pulvinar can still guide visually-directed actions. These non-geniculostriate pathways bypass V1 and project to dorsal stream areas involved in action control, enabling accurate pointing without awareness. The patient's ability to point accurately despite no conscious vision in the affected field is the hallmark of blindsight. Choice B correctly identifies this phenomenon, while choice D incorrectly suggests complete decussation preserves conscious vision. A key principle: visual processing involves multiple parallel pathways, and action can be guided without conscious perception.

8

In a study of face perception, participants view upright faces or the same faces inverted. Accuracy in identifying subtle differences between two faces drops markedly for inverted faces, even though low-level features (luminance, size) are matched. Based on the scenario, which conclusion is most consistent with the principle that visual recognition can rely on configural processing?

Inversion should improve performance because it reduces reliance on top-down predictions that can bias perception.

Inversion should not affect performance because the retina encodes faces as a set of independent features regardless of orientation.

Inversion should impair performance because disrupting typical spatial relations among features reduces holistic/configural processing efficiency.

Inversion should impair performance only when binocular disparity is removed, because face identification is primarily a depth task.

Explanation

This question tests understanding of configural versus featural processing in face recognition. Face perception relies heavily on configural or holistic processing, where the spatial relationships among features (eyes, nose, mouth) are processed as an integrated whole rather than independent parts. Inversion disrupts this configural processing by placing features in unfamiliar spatial arrangements, forcing a less efficient feature-by-feature analysis. The marked performance drop for inverted faces despite matched low-level features demonstrates the importance of typical spatial configuration. Choice C correctly identifies this disruption of configural processing, while choice B incorrectly suggests faces are always processed as independent features. Remember the face inversion effect as evidence that some visual recognition depends on learned spatial templates.

9

Researchers present two stimuli: a high-contrast grating with thick bars and a high-contrast grating with very thin bars. Participants detect the thick-bar grating at lower light levels than the thin-bar grating, despite identical overall luminance. Based on the scenario, which conclusion is most consistent with how receptive field properties constrain visual processing?

Thin bars are detected more easily because larger receptive fields in early vision preferentially encode high spatial frequency detail.

Thick bars are detected more easily because early visual filters are more sensitive to lower spatial frequencies under reduced visibility.

Detection should be identical because spatial frequency is computed only in the optic nerve, not in the brain.

Thin bars are detected more easily because binocular convergence enhances resolution for high spatial frequencies at all light levels.

Explanation

This question tests understanding of spatial frequency channels and receptive field properties in early vision. The visual system contains multiple spatial frequency channels with different sensitivities, where larger receptive fields preferentially respond to lower spatial frequencies (thick bars) and smaller fields to higher frequencies (thin bars). Under reduced visibility conditions like low light, the visual system shows enhanced sensitivity to lower spatial frequencies, making thick-bar gratings more detectable. This reflects both the properties of early visual filters and adaptive mechanisms that prioritize coarse structure detection when fine detail is unavailable. Choice B correctly identifies this low spatial frequency advantage, while choice A incorrectly states large receptive fields prefer high frequencies. A transferable principle: visual sensitivity varies with spatial scale, and coarse features are detected more readily under degraded conditions.

10

In a study of peripheral vision, participants identify letters presented 15° from fixation. Accuracy improves when the letters are spaced farther apart, even though each letter’s size is unchanged. The researcher attributes the original impairment to crowding rather than acuity limits. Which result best supports this attribution?

Improvement occurs when flankers are removed or spaced out, suggesting nearby items interfere with feature integration in peripheral vision.

Improvement occurs only when both eyes are used, indicating binocular disparity resolves letter identity.

Improvement occurs only under colored lighting, indicating color constancy enhances peripheral acuity.

Improvement occurs only when stimuli are presented to the left visual field, indicating ipsilateral processing reduces interference.

Explanation

This question tests visual crowding in peripheral vision. Crowding impairs identification when flankers are close, due to feature integration errors beyond acuity limits. In this letter task, better accuracy with spacing supports crowding over resolution. Choice D correctly links improvement to reduced interference in periphery. Choice B fails by requiring binocularity, unrelated to crowding. For transfer, eccentricity effects: crowding worsens peripherally. Check by spacing: if isolated letters improve, crowding implicated.

Page 1 of 5