Introducing Statistics: Worry About Error
Help Questions
AP Statistics › Introducing Statistics: Worry About Error
A retail company A/B tests a new checkout design on a sample of website visitors and finds evidence of a higher purchase rate. The company considers rolling it out to all users immediately. Why is it important to consider potential error in this inference, given that a rollout could either increase revenue or unintentionally reduce sales for many customers?
Because Type I and Type II errors only apply to medical studies, not business decisions
Because if the new design worked in the sample, it must work for all future visitors with no exceptions
Because random assignment removes all uncertainty from the results
Because a confidence interval is the probability that the new design is better
Because the observed difference could be partly due to chance, and a wrong rollout decision could affect revenue and user experience at scale
Explanation
This question examines understanding of experimental error in business decision-making. The company must decide whether to roll out a new checkout design based on A/B test results. Option B correctly recognizes that the observed difference in purchase rates could be partly due to chance (sampling variability) rather than a true effect, and rolling out an inferior design could reduce revenue while missing a superior design means lost opportunities. Success in the sample doesn't guarantee success for all users (eliminating A), random assignment reduces bias but not sampling variability (eliminating C), confidence intervals quantify uncertainty about parameters not probabilities of superiority (eliminating D), and Type I/II errors apply to all hypothesis testing including business experiments (eliminating E). In high-stakes business decisions affecting revenue and user experience, companies must acknowledge that A/B test results are subject to sampling error and might not reflect true population effects.
A pharmaceutical company runs a clinical trial on a sample of patients and finds that a new drug appears to reduce symptoms more than a placebo. Regulators must decide whether to approve the drug for widespread use. Why is it important to consider potential error in the statistical inference from the trial when making this approval decision?
Because once a study is randomized, Type I error cannot occur
Because inference from a sample is uncertain, and approving an ineffective or unsafe drug (or delaying an effective one) can have major health consequences
Because considering error is unnecessary if the sample size is larger than 30
Because a statistically significant result proves the drug will work for every patient
Because the $p$-value is the probability that the drug is ineffective
Explanation
This question examines understanding of clinical trial error in drug approval decisions. Regulators must decide whether to approve a drug based on trial results showing apparent effectiveness. Option A correctly recognizes that inference from the trial sample is uncertain - approving an ineffective or unsafe drug (Type I error) exposes patients to risks without benefits, while rejecting an effective drug (Type II error) denies patients a helpful treatment. Randomization reduces bias but doesn't eliminate Type I error risk (eliminating B), statistical significance doesn't guarantee universal effectiveness (eliminating C), p-values measure evidence against the null not probabilities of ineffectiveness (eliminating D), and sample size requirements don't eliminate the need to consider error (eliminating E). In pharmaceutical regulation where decisions affect millions of patients, regulators must balance the risks of approving harmful drugs against denying beneficial treatments, recognizing that sample-based trial results might not reflect true drug effects.
A wildlife agency estimates the population size of an endangered species using a statistical method based on a sample of sightings and then decides whether to restrict hiking access to protect habitat. Why is it important to consider potential error in the estimate when making this decision, given tradeoffs between conservation benefits and public access?
Because the estimate could be too high or too low, and either unnecessary restrictions or insufficient protection could result from a wrong inference
Because an estimate from a sample is exact, so error is not a concern
Because sampling error and measurement error are the same thing and can be ignored
Because restricting access will automatically increase the sample size and remove uncertainty
Because statistical inference cannot be used for populations that move
Explanation
This question tests understanding of estimation uncertainty in conservation policy. The wildlife agency must decide on hiking restrictions based on a population estimate from sample sightings. Option B correctly identifies that the estimate could be too high (leading to unnecessary access restrictions) or too low (risking inadequate habitat protection), with both errors having consequences for conservation and public recreation. Estimates from samples are never exact due to sampling variability (eliminating A), restricting access doesn't affect the statistical uncertainty of past estimates (eliminating C), sampling error and measurement error are different concepts and neither can be ignored (eliminating D), and statistical inference can be applied to mobile populations using appropriate methods (eliminating E). Conservation decisions balancing species protection with public access must acknowledge that population estimates based on limited sightings contain uncertainty that could lead to over- or under-protection.
A manufacturer randomly samples 50 batteries from a day’s production and estimates the mean lifetime to decide whether to ship the entire batch. If the estimate is too high, customers may receive defective batteries; if too low, the company may discard good inventory. Why is it important to consider potential error when using the sample mean to make the shipping decision?
Because the sample mean can vary from sample to sample, and a wrong decision has financial and customer-safety consequences
Because random sampling eliminates all uncertainty, so the decision can be treated as a guarantee
Because the sample mean equals the population mean whenever the sample size is at least 30
Because sampling error and a Type I error are the same thing, so considering both would be double-counting error
Because the only important error is rounding error from the calculator output
Explanation
This question evaluates understanding of why error consideration is crucial when using sample means for business decisions in manufacturing. The decision context involves balancing the risks of shipping defective batteries, which could harm customers, against unnecessarily discarding good ones, which wastes resources. The correct choice is A, emphasizing that sampling variability causes sample means to fluctuate, potentially leading to erroneous shipping decisions with real consequences. Distractor B is misleading because while the central limit theorem applies for large samples, it doesn't make the sample mean exactly equal to the population mean; variability still exists. For a mini-lesson on uncertainty, sampling error arises because a sample is only a subset of the population, so its statistics are estimates that can deviate from true values by chance; acknowledging this helps avoid overconfidence in decisions based on samples.
A food safety inspector tests 25 randomly selected packages of bagged salad from a shipment and finds no contamination. The store considers selling the entire shipment immediately. If contamination is present but missed, customers could get sick; if the shipment is discarded unnecessarily, the store loses money and wastes food. Why is it important to consider potential error?
Because a small sample can miss rare contamination, so the decision should account for the chance of error and the serious consequences of a wrong conclusion
Because sampling error can be eliminated by testing only the cleanest-looking packages
Because finding no contamination in the sample guarantees the full shipment is contamination-free
Because the inspector should ignore uncertainty since random samples always produce the same results
Because error is only a concern when contamination is found, not when none is found
Explanation
This question in AP Statistics illustrates worrying about error in contamination testing, stressing its relevance to food safety and economic decisions. The store's plan to sell the shipment after no contamination in 25 packages risks customer illness if rare issues were missed, or unnecessary waste if overly cautious. Considering potential error is critical because small samples may not detect low-level problems, leading to serious consequences if inferences are wrong. Choice A distracts by asserting no findings guarantee safety, but sampling error allows for undetected issues. In a mini-lesson on uncertainty, absence in a sample doesn't prove absence in the population due to variability, and statistical power assesses detection reliability. This promotes informed risk management in inspections.
A manufacturing company tests a random sample of 60 batteries from a day’s production and finds 4 are defective. Management considers shutting down the line to recalibrate machines, which is costly; however, failing to recalibrate could send defective batteries to customers and create safety risks. Why is it important to consider potential error?
Because if 4 defects were found, exactly 4 defects exist in the entire day’s production
Because the sample defect rate could differ from the true defect rate, and either action (shutdown or no shutdown) has costly consequences if the inference is wrong
Because sampling error is the same as a machine defect, so recalibration always removes it
Because considering error is unnecessary unless the sample size is at least 1,000
Because a random sample eliminates all uncertainty about the day’s true defect rate
Explanation
This AP Statistics question focuses on worrying about error in sample-based defect rates, illustrating how uncertainty influences manufacturing decisions. Management's choice to shut down or continue production involves costs and safety risks if the sample of 60 batteries with 4 defects doesn't reflect the true rate. Considering potential error is key because sampling variability could mean the true defect rate is higher or lower, leading to faulty inferences with consequences like financial loss or customer harm. Choice B distracts by suggesting random samples remove all uncertainty, but they only provide estimates prone to variation. As a mini-lesson on uncertainty, sample statistics like defect rates vary from the population parameter due to chance selection, and statistical inference helps evaluate the likelihood of errors. This approach prevents overreactions or oversights in high-stakes scenarios.
A city council uses a random sample survey to estimate support for a proposed property-tax increase to fund road repairs. If they proceed when true support is below 50%, the measure may fail at the ballot and reduce public trust. Why is it important to consider potential error when using the sample estimate to decide whether to place the measure on the ballot?
Because any difference between the sample and population must be caused by biased questions, not chance
Because confidence intervals eliminate the possibility of an incorrect decision
Because if the sample size is at least 30, the estimate is guaranteed to be correct
Because the sample proportion is always exactly equal to the population proportion when the sample is random
Because the margin of error reflects possible sampling variability, and being wrong could waste money and damage credibility
Explanation
This question examines understanding of sampling variability in political polling contexts. The city council must decide whether to place a tax measure on the ballot based on a sample survey estimate. The correct answer (B) properly identifies that the margin of error reflects sampling variability - different random samples would yield different estimates of support. If the true support is below 50% but the sample estimate suggests it's above, the council might proceed with a doomed measure, wasting resources and damaging public trust. Choice A incorrectly claims sample and population proportions are always equal, while D wrongly suggests sample size alone guarantees accuracy. Statistical inference always involves uncertainty because we're using partial information (the sample) to make conclusions about the whole (all voters).
A university uses a random sample of applicants’ records to estimate whether a new admissions rubric changes the acceptance rate for first-generation students. If the inference is wrong, the university could adopt a rubric that unintentionally reduces access or reject a rubric that improves equity. Why is it important to consider potential error before acting on the inference?
Because considering error only matters after the policy has already been implemented
Because inference from sample data includes uncertainty, and a wrong conclusion could lead to harmful policy choices affecting students
Because if the sample was collected carefully, there is no need to consider the possibility of Type I or Type II error
Because any observed difference in sample acceptance rates must reflect a real change in the population
Because the acceptance rate is a percentage, and percentages cannot have sampling error
Explanation
This question assesses understanding of inference uncertainty in educational equity contexts. The university uses sample data to estimate whether a new admissions rubric affects acceptance rates for first-generation students. The correct answer (B) recognizes that inference from sample data includes uncertainty - the observed difference in sample acceptance rates might not reflect the true population change due to sampling variability. Type I error (adopting a rubric that doesn't actually improve equity) fails to help students who need support, while Type II error (rejecting a rubric that would improve access) misses an opportunity to increase educational equity. Choice A wrongly assumes any sample difference reflects real population change, while D incorrectly claims careful sampling eliminates error possibility. Statistical inference involves uncertainty, and when decisions affect student access to education, we must consider the serious consequences of potential errors.
A school district uses results from a sample of students to infer whether a new tutoring program increases average math scores. Ending the program when it truly helps could harm students; keeping it when it doesn’t help could waste funds. Why is it important to consider potential error when using the inference to decide whether to continue the program?
Because sampling and random assignment remove all uncertainty from conclusions
Because statistical conclusions can be wrong due to chance, and the costs of a wrong decision affect students and budgets
Because a $p$-value below 0.05 proves the program works for every student in the district
Because confidence levels measure how confident the researchers feel, not uncertainty in the process
Because if the sample mean is higher, the population mean must be higher by the same amount
Explanation
This question assesses understanding of statistical uncertainty in educational program evaluation. The school district uses sample data to infer whether a tutoring program improves math scores. The correct answer (C) recognizes that statistical conclusions can be wrong due to chance - even with proper randomization, the sample might not perfectly represent the true program effect. Type I error (concluding the program works when it doesn't) wastes funds, while Type II error (concluding it doesn't work when it does) harms students who would benefit. Choice A incorrectly interprets p-values as proof for every student, while B wrongly claims randomization eliminates uncertainty. Statistical inference involves making probabilistic statements, not certainties, and decision-makers must weigh the consequences of potential errors.
A bank uses a model based on a random sample of past loans to estimate the probability that a new applicant will default. The bank approves loans when the estimated default probability is below 5%. Approving risky loans can cause losses, while rejecting reliable borrowers can harm customers and reduce profits. Why is it important to consider potential error in the estimated probability when making approval decisions?
Because the estimate can be wrong due to sampling variability and model uncertainty, leading to costly approval or rejection mistakes
Because any wrong decision would be a Type I error, so the bank should ignore Type II errors
Because error matters only when the probability is above 50%, not near 5%
Because using a probability threshold eliminates the possibility of any decision error
Because the estimated probability is a guarantee of what will happen to this specific applicant
Explanation
This question tests awareness of error in model-based probability estimates for financial decisions like loan approvals. In the bank's context, errors could lead to losses from bad loans or missed profits from rejecting good ones, with a 5% threshold amplifying risks. B is correct, as it addresses sampling variability and model uncertainty that could make the estimate inaccurate, causing approval mistakes. Distractor A errs by treating the estimate as a personal guarantee, but probabilities are population-based and uncertain for individuals. Mini-lesson on uncertainty: Predictive models from samples inherit sampling error and may not capture all factors, so considering confidence in estimates prevents overreliance and supports better risk management.