Operant Conditioning and Reinforcement Schedules (7C)

Help Questions

MCAT Psychological and Social Foundations › Operant Conditioning and Reinforcement Schedules (7C)

Questions 1 - 10
1

A study tests how quickly children clean up toys. In Group 1, a child gets a sticker after every 5 toys put away (fixed ratio 5). In Group 2, a child gets a sticker after the first toy put away after 2 minutes have passed (fixed interval 2 min). Assuming both groups value stickers similarly, which group is most likely to show a higher overall rate of toy pickup during the session?

Neither group, because reinforcement schedules do not affect response rate once a behavior is learned

Group 2, because fixed interval schedules produce the highest response rates

Group 1, because fixed ratio schedules directly reinforce higher response rates

Group 2, because reinforcement is delivered for every response once 2 minutes pass

Explanation

This question compares response rates between fixed ratio and fixed interval schedules in operant conditioning. Operant conditioning strengthens behaviors via reinforcements; fixed ratio (FR) rewards after a set number of responses, promoting high rates, while fixed interval (FI) rewards after time, often yielding lower overall rates with scalloping. Children in Group 1 get stickers on FR5 for toys, while Group 2 uses FI 2 minutes. The correct answer is B because FR directly ties reinforcement to response count, encouraging faster pickup to earn more stickers quickly. A distractor like A fails, as FI typically produces lower rates than FR, not higher. For a transferable check, compare if output is response-driven (FR) versus time-driven (FI). This applies to piecework pay versus hourly wages in productivity studies.

2

A coffee shop loyalty program gives a free drink after exactly 12 purchases (fixed ratio 12). Management changes the program so that a free drink is given after an unpredictable number of purchases averaging 12 (variable ratio). Which customer behavior change is most likely?

Customers will stop buying because the program change removes a negative reinforcer

Customers will show increasing purchases as the 12th purchase approaches, then pause after the free drink

Customers will buy less often because variable schedules reduce motivation compared with fixed schedules

Customers will buy coffee at a steadier rate because rewards are less predictable

Explanation

This question explores behavioral shifts from fixed ratio to variable ratio in consumer loyalty programs via operant conditioning. Operant conditioning uses schedules to maintain behaviors; fixed ratio (FR) often includes post-reward pauses, while variable ratio (VR) generates steady, high-rate responding due to unpredictability. The coffee program changes from FR12 to VR averaging 12 for free drinks. The correct answer is A because VR eliminates predictable pauses, encouraging more consistent purchases as each could trigger the reward. A distractor like B fails, as it reflects FR patterns with acceleration and post-reward lulls, not VR. For a transferable check, see if behavior steadies without post-reward drops, typical of VR. This is evident in variable-reward apps boosting user engagement.

3

A researcher trains mice to nose-poke for a sucrose drop. The mice are first reinforced on a fixed ratio 1 schedule (every nose-poke earns sucrose). The researcher then shifts to a fixed ratio 10 schedule (every 10th nose-poke earns sucrose) without changing sucrose amount. Which change in responding is most likely immediately after the shift?

Responding stops because sucrose now functions as punishment when delivered after 10 responses

Responding becomes timed to 10-second intervals because the schedule is now interval-based

An immediate increase in responding because fewer rewards per response always strengthens behavior

A temporary drop in responding because reinforcement is less frequent, followed by higher response rates once learned

Explanation

This question tests understanding of ratio strain and schedule thinning effects. When organisms are shifted from dense reinforcement (FR-1) to lean reinforcement (FR-10), they typically show ratio strain - a temporary disruption in responding as they adjust to the new requirement. Initially, mice may pause or show erratic responding because the expected reinforcement doesn't occur after single responses. However, once they learn the new contingency, response rates typically increase to maintain reinforcement frequency, often exceeding the original rate. Option B incorrectly suggests immediate increases, ignoring the adjustment period. Option C wrongly implies the schedule is interval-based when it remains ratio-based. The key principle is that abrupt increases in ratio requirements produce temporary disruptions before higher sustained rates emerge.

4

A dog is trained to sit. The trainer delivers a treat for the first correct sit that occurs after a variable amount of time has passed (sometimes 10 s, sometimes 45 s, averaging 30 s). This is a variable interval schedule. Which response pattern is most consistent with this schedule after training is established?

No sitting because the unpredictability of timing functions as punishment

Very high response rates with minimal pauses because each sit may be the one that earns the treat

Moderate, steady sitting attempts over time because reinforcement depends on time, unpredictably

A scalloped pattern: little sitting right after a treat, increasing sharply as 30 s approaches

Explanation

This question examines response patterns under variable interval schedules in operant conditioning. Operant conditioning reinforces behaviors through schedules; variable interval (VI) delivers rewards for the first response after an unpredictable time interval, typically producing moderate, steady responding without pronounced pauses. The dog's sitting is reinforced on a VI averaging 30 seconds, varying from 10 to 45 seconds. The correct answer is C because VI encourages consistent attempts over time, as the dog can't predict when the next opportunity arises. A distractor like B fails, as it describes fixed interval scalloping, not VI's steady pattern. To check transferably, observe if responses remain even without time cues, indicative of VI. This mirrors behaviors like checking email sporadically for unpredictable replies.

5

In a campus lab, participants play a simple computer game where pressing the spacebar can earn points exchangeable for a small gift card. The program delivers 10 points after every 20th spacebar press (responses between reinforced presses earn nothing). After 5 minutes, the researchers switch the program so that 10 points are delivered after an unpredictable number of presses averaging 20. Based on this change from a fixed ratio to a variable ratio schedule, which change in behavior is most likely to occur during the next 5 minutes?

A brief pause after each reward becomes more pronounced, with overall pressing slowing

Pressing stops because the schedule change functions as punishment for responding

Pressing decreases because rewards now occur after longer periods of time rather than more responses

Pressing becomes more steady and persistent, with fewer post-reward pauses

Explanation

This question tests understanding of operant conditioning reinforcement schedules, specifically the behavioral effects of switching from fixed ratio to variable ratio. Operant conditioning involves learning through consequences, with reinforcement schedules dictating when rewards follow behaviors; fixed ratio (FR) delivers after a set number of responses, while variable ratio (VR) delivers after an unpredictable number averaging a value. In this scenario, participants initially press on an FR20 schedule, producing high rates with post-reinforcement pauses, then switch to VR averaging 20. The correct answer is B because VR schedules eliminate predictable pauses, leading to steadier, more persistent responding as each press could yield the reward. A distractor like A fails as it describes FR patterns, where pauses after rewards are common, not VR. To check transferably, observe if behavior becomes more consistent without slowdowns after rewards, indicating a VR schedule. Remember, VR is effective for maintaining high response rates in activities like gambling or sales.

6

A researcher trains a rat to press a lever. Under a fixed interval 30-s schedule, the first lever press after 30 seconds has passed produces a food pellet; presses before 30 seconds do not. After several sessions, which behavior pattern is most consistent with this reinforcement schedule?

Rapid pressing only when a light is on, because the pellet is negative reinforcement

A high, steady rate of pressing with no predictable pauses after pellets

A burst of pressing immediately after a pellet, followed by a long pause until the next pellet

Low responding right after a pellet, then increasing responding as 30 seconds approaches

Explanation

This question assesses knowledge of fixed interval reinforcement schedules in operant conditioning. Operant conditioning shapes behavior via rewards or punishments, with schedules like fixed interval (FI) providing reinforcement for the first response after a set time has elapsed, often resulting in scalloped response patterns. Here, the rat's lever pressing is reinforced on an FI 30-second schedule, where presses before the interval ends yield nothing. The correct answer is C because FI schedules produce low responding immediately post-reinforcement, with acceleration as the interval nears, creating a scallop shape. A distractor like A fails as it describes variable ratio patterns, which lack time-based pauses and maintain high steady rates. For a transferable check, look for response rates that ramp up toward a predictable time, hallmark of FI. This pattern appears in real-life scenarios like checking mail near delivery time.

7

A language-learning app gives a user a “streak bonus” only when the user completes a lesson on the 7th day of a week (no bonus is given on days 1–6, even if lessons are completed). This is a fixed interval schedule with a 7-day interval. If the app wants to increase lesson completion on earlier days without changing the size of the bonus, which modification would be expected to produce more consistent daily studying?

Switch to a variable interval schedule where the first lesson after an unpredictable number of days earns the bonus

Keep the fixed interval but deliver the bonus only after an unpredictable number of lessons averaging seven

Remove the bonus entirely so studying becomes intrinsically motivated and therefore more frequent

Add a response cost so that missed days remove points, because punishment increases desired behavior

Explanation

This question evaluates strategies to modify fixed interval schedules for more consistent behavior in operant conditioning. Operant conditioning uses reinforcements to increase desired behaviors, with fixed interval (FI) schedules rewarding the first response after a fixed time, often leading to low early responding and a surge near the interval's end. The app's FI 7-day schedule likely results in procrastination, with lessons clustered on day 7. The correct answer is A because switching to variable interval (VI) promotes steady daily engagement, as the unpredictable timing encourages consistent checking or responding. A distractor like D fails, as removing reinforcement typically leads to extinction, not intrinsic motivation. To verify transferably, note if unpredictability in timing boosts regular behavior without end-loaded surges. This principle applies to improving habits like exercise through apps with random rewards.

8

In a classroom, a teacher calls on students who raise their hands. The teacher decides to reinforce hand-raising by giving verbal praise on a variable interval schedule (the first hand-raise after an unpredictable amount of time gets praise). Compared with praising on a fixed interval schedule of the same average length, which pattern is most likely under the variable interval schedule?

Hand-raising will cluster right before the expected praise time, with low responding after praise

Hand-raising will be relatively steady over time, without pronounced cycles tied to timing

Hand-raising will occur only after a specific number of questions have been asked

Hand-raising will decrease because praise is a punisher when delivered intermittently

Explanation

This question contrasts patterns under variable interval versus fixed interval schedules in operant conditioning. In operant conditioning, reinforcement schedules influence response timing; variable interval (VI) yields steady rates without cycles, unlike fixed interval (FI) with its scalloped increases near reinforcement time. The teacher praises hand-raising on VI, compared to FI of similar average. The correct answer is B because VI's unpredictability prevents timed surges, maintaining even responding throughout. A distractor like A fails, as it describes FI's clustering near expected times, not VI's consistency. To check transferably, note absence of time-linked patterns as a VI indicator. This applies to social media checking, where unpredictable notifications drive steady engagement.

9

A participant completes online surveys. Under a fixed ratio 1 schedule, they receive $0.50 after every survey. The payment system is changed to a fixed interval 10-min schedule: the first completed survey submitted after each 10-minute period earns $0.50, and additional surveys within the interval earn nothing. Based on this change, which behavior is most likely during each 10-minute interval?

Consistently high survey completion throughout the interval because each survey could be reinforced

Immediate cessation of responding because removing per-survey payment is the same as punishment

More surveys overall, because fixed interval schedules reinforce every response after a delay

Fewer surveys overall, with responding increasing as the end of the 10-minute period approaches

Explanation

This question investigates the effects of switching from fixed ratio to fixed interval on response patterns in operant conditioning. Operant conditioning schedules shape behavior; fixed ratio (FR) rewards per response, yielding high rates, while fixed interval (FI) rewards after time, producing scalloped patterns with lower early responding. The survey payment shifts from FR1 to FI 10 minutes. The correct answer is B because FI leads to fewer responses overall, with acceleration as the interval ends, as early efforts are unrewarded. A distractor like C fails, as FI does not reinforce every delayed response but only the first after the interval, often reducing total output. For a transferable check, look for end-loaded responding as FI's signature. This pattern occurs in timed tasks like waiting to check for a bus.

10

In a lab task, participants press a key to earn points exchangeable for money. The computer delivers 10 points after every 20th key press (fixed ratio 20) and provides no other feedback. After several minutes, the researcher temporarily pauses point delivery (extinction) without telling participants. Which response pattern is most likely immediately before and during extinction?

High-rate pressing with brief pauses after earning points, followed by a rapid decline when points stop

Moderate pressing that gradually increases over time, then slowly fades when points stop

Pressing mainly right after each point delivery, with long pauses that persist even when points stop

Low, steady pressing throughout, with little change when points stop

Explanation

This question tests understanding of fixed ratio schedules and extinction patterns. Fixed ratio (FR) schedules, where reinforcement occurs after a set number of responses, typically produce high response rates with brief post-reinforcement pauses. In this FR-20 schedule, participants press rapidly to reach the 20-press requirement, take a brief pause after earning points, then resume pressing. When extinction begins (reinforcement stops), FR schedules show rapid response decline because the predictable reinforcement contingency is broken. The correct answer describes this pattern: high-rate pressing with brief pauses during reinforcement, followed by rapid decline during extinction. Option A incorrectly suggests low, steady pressing, which contradicts FR's characteristic high rates.

Page 1 of 6