From the ai output evaluation sample test
What does the "fluent-but-unfaithful summary" AOE scenario measure?
Note on framing: This is the aoe_sample_4 item-level explainer for the AOE (AI Output Evaluation) sample-test family. Construct-level coverage is in the aoe-evaluating-llm-output explainer; the canonical hallucination-detection item is documented in the aoe-hallucinated-citation explainer.
This scenario presents two summaries of the same source document. Summary A is fluent, well-structured, confident, and reads as authoritative — but introduces details that the source did not contain, drops a key qualifier the source emphasized, and reframes a tentative finding as a definitive one. Summary B is awkwardly phrased, hedged, and reads as under-confident — but every claim it makes is traceable to the source, and every qualifier the source used appears in the summary. The candidate must grade the two summaries on the AOE rubric. The scenario probes the fluent-and-wrong versus halting-and-right distinction, which is the single-most-important reflex an AOE evaluator must develop.
What this question tests
The item targets the ability to dissociate stylistic quality from epistemic quality when grading summarization output. LLM summarizers are trained to produce fluent prose; the training objective rewards readability and coherence, and only indirectly rewards faithfulness to the source. The result is a systematic bias: the model can produce a polished summary that loses or distorts source content, and the polish itself disguises the distortion. Evaluators who grade on fluency — even unconsciously — reward the wrong outputs.
The skill being measured is sometimes called summary faithfulness evaluation in the published NLP literature (Maynez et al. 2020 documented the prevalence of “intrinsic” and “extrinsic” hallucinations in abstractive summarization; Pagnoni et al. 2021’s FRANK benchmark catalogued the typology of faithfulness errors). The construct is well-defined and the evaluation skill is measurable: trained evaluators consistently identify unfaithful-but-fluent summaries that untrained evaluators miss. AOE measures whether a candidate performs at the trained-evaluator level rather than the naive-fluency-graded level.
Why this is the right answer (concrete worked example)
The correct grade is Summary B at value 4 or 5; Summary A at value 2 or below. The fluent-but-unfaithful summary fails on the dimension that matters for the use case (faithful representation of the source), and the awkward-but-faithful summary succeeds on that dimension despite stylistic friction.
A worked example illustrates the asymmetry. Suppose the source is a clinical-trial abstract that reads, “In a small, single-site study (n=42), we observed a non-significant trend toward reduced inflammation markers in patients receiving the intervention compared to placebo (p=0.08).”
Summary A renders this as: “A study found that patients receiving the intervention experienced reduced inflammation markers compared to placebo.” This summary is fluent and quotable. It is also unfaithful in three specific ways: it drops the “small, single-site” sample-size qualifier, it drops the “non-significant trend” hedge, and it converts a p=0.08 finding into a flat claim of effect. A reader of Summary A would believe the study established what it did not.
Summary B renders the same source as: “A small single-site study (n=42) reported a trend toward lower inflammation markers in the intervention group versus placebo, but the result was not statistically significant (p=0.08), so the effect should be considered preliminary.” This summary is clunky, repeats “study” structures, and reads as under-confident. It is also exactly faithful to the source — every qualifier is preserved, the statistical claim is correctly hedged, and the reader leaves with an accurate mental model of what the source actually said.
The strong AOE evaluator grades B above A because the use case for a summary is to represent the source accurately, and Summary A fails that use case while Summary B succeeds at it. The respondent who grades A above B has been seduced by fluency — the exact reflex that ships unfaithful summaries into production.
What the wrong answers reveal
The graded rubric has predictable failure patterns:
- Grading Summary A at value 4 or 5 (fluent-bias error). This is the most-common evaluator failure: the respondent reads the polished prose and pattern-matches it to “good summary.” The pattern is well-documented in summarization literature; trained evaluator panels consistently produce lower fluency-grade-to-faithfulness-grade correlations than untrained panels.
- Grading both summaries at value 3 (split-the-difference error). This option signals a respondent who notices that something differs between A and B but cannot articulate the dimension on which they differ. It is a hedge against having to commit to a faithfulness-over-fluency judgment.
- Grading Summary B at value 1 or 2 (anti-clunkiness error). A respondent who downgrades B because it reads awkwardly has confused stylistic polish with output quality. The output’s job is to represent the source; clunkiness is a remediable surface issue, while unfaithfulness is a structural one.
How the sample test scores you
In the AIEH 5-scenario AOE sample test, this item contributes one of five datapoints aggregated into the single aoe_quality score via the W3.2 normalize-by-count threshold. Graded scoring per summary, with the diagnostic target being whether the respondent grades Summary B strictly above Summary A.
Data Notice: Sample-test results are directional indicators only. The fluent-bias reflex is one of the hardest-to-train AOE failures; even respondents who grade this scenario correctly often grade fluent-but-unfaithful outputs incorrectly in unfamiliar domains. For a verified Skills Passport credential, take the full 40-scenario assessment.
See the scoring methodology for how AOE scores map onto the AIEH 300–850 Skills Passport scale.
Related concepts
- Faithfulness vs fluency in summarization. Maynez et al. 2020 distinguished “factuality” (does the summary contain facts that exist in the world?) from “faithfulness” (does the summary contain only facts present in the source?). AOE scenarios target faithfulness because that is the property the summarizer can be held responsible for; factuality depends on the source’s own grounding.
- Intrinsic vs extrinsic hallucination. Intrinsic hallucination is when the summary contradicts the source; extrinsic hallucination is when the summary adds claims not present in the source. The fluent-but-unfaithful scenario combines both — it drops a hedge (intrinsic) and reframes a finding (extrinsic).
- Reference-free evaluation methods. Newer faithfulness metrics (FactCC, QAGS, FRANK) attempt to grade faithfulness without a gold-standard summary; AOE evaluators provide the human-grader signal these automated metrics try to approximate.
- Calibration in summarization. Faithful summaries preserve the source’s calibrated language (“trend toward,” “non-significant,” “preliminary”); unfaithful summaries systematically over-confident the source. AOE evaluators flag de-hedging as a faithfulness failure, not a stylistic improvement.
For broader treatment of how AOE fits into role-readiness scoring, see the AI fluency in hiring overview, the assess page for the assessment workflow, and the learn library for AOE training material.
Sources
- Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? Proceedings of FAccT ‘21, 610–623.
- Maynez, J., Narayan, S., Bohnet, B., & McDonald, R. (2020). On Faithfulness and Factuality in Abstractive Summarization. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 1906–1919.
- Pagnoni, A., Balachandran, V., & Tsvetkov, Y. (2021). Understanding Factuality in Abstractive Summarization with FRANK: A Benchmark for Factuality Metrics. Proceedings of NAACL-HLT, 4812–4829.
- Schmidt, F. L., & Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology. Psychological Bulletin, 124(2), 262–274.