What does the "cross-functional disagreement" SJT scenario measure?

This is the third item-level explainer in the AIEH Situational Judgment family. Where the first SJT explainer covered teammate-feedback context and the second covered multi-stakeholder priority triage, this one targets the hardest variant of cross-functional work: the situation where engineering, design, and product genuinely disagree about whether a feature is ready to launch — and the respondent has to resolve the disagreement without flattening any one perspective.

What this question tests

This scenario — engineering says the build is stable enough to ship, design says the empty-state and error-state polish isn’t where it should be, and product says the customer- commit window closes Friday — measures judgment under genuine cross-functional disagreement on a launch criterion. Specifically, the item probes whether the respondent recognizes that:

  1. Each of the three positions reflects a real and different underlying constraint (engineering risk surface, design quality bar, product commitment cost), and treating any one as primary by default is the wrong frame.
  2. The work is to make the trade-off explicit, surface the reversibility of each path, and force a decision against a shared criterion the team has already agreed to — not to mediate emotionally or escalate to a manager.
  3. The disagreement is information; suppressing it (by over-deferring to the loudest or most senior voice) loses that information and produces worse launches than resolving it does.

Unlike personality items that capture stable dispositions across contexts, SJT scenarios capture situation-specific judgment — what a respondent does inside a specific multi-stakeholder pattern. Strong contributors hold all three constraints simultaneously and force the team to a decision; weaker ones tend to fail by collapsing the trade-off (picking one stakeholder’s frame as primary), deferring entirely (escalating to a manager who has less context), or splitting the difference in a way that satisfies none of the three constraints.

Why this is the right answer

The value-5 response handles the disagreement by doing four things in roughly this order: it names what each stakeholder is actually optimizing for, it converts the disagreement into a small number of decision-relevant questions, it proposes a default path that is explicitly reversible, and it commits to a decision-time so the disagreement doesn’t metastasize into delay-by-default. A worked example clarifies what this looks like in practice.

The respondent runs a 30-minute working session — engineering lead, design lead, and PM in the room — and frames it like this: “Engineering, what’s the specific failure mode you’re worried about if we ship now? Design, what specifically is unpolished, and what’s the user-visible impact of shipping without it? Product, what does missing the Friday window actually cost — is it the contractual commit, the marketing co-launch, or the internal momentum?” The framing forces each stakeholder to articulate the load-bearing concern rather than restate the surface position.

From there, the respondent surfaces reversibility: shipping to a 5% holdout on Friday with the unpolished empty-states is reversible (we can roll back if engineering’s failure mode appears, and design polish lands in the next sprint with no new code paths blocked); slipping the launch by two weeks for design polish is also reversible (the Friday commit becomes a Monday-the-22nd commit, customer relationship absorbs a small hit, marketing co-launch shifts) — but it has a different reversibility profile.

The decision then becomes a structured choice between two known-reversible paths against the team’s stated criteria. The respondent commits to the decision-time (“we decide this by 4pm today”), names who decides if the room can’t converge (usually the PM or the next-level manager, pre-agreed), and writes the decision down. That writing-it- down step is what makes the resolution durable — without it, the disagreement re-emerges in a Slack thread three days later and consumes another hour.

The published research on cross-functional team conflict (Jehn, 1995; De Dreu & Weingart, 2003) documents that task conflict (disagreement about what to do) is moderately productive when surfaced and resolved on content, while relationship conflict (disagreement about who is right) is uniformly destructive. Strong SJT respondents convert task conflict into a structured decision; weak ones let task conflict drift into relationship conflict, which is much harder to recover from.

What the wrong answers reveal

The graded option ladder catches three direction-of-failure patterns:

  • Engineer-defaults-to-PM (value 3). “I’ll let the PM call it — they own the launch criterion.” This response demonstrates partial competence — the respondent recognizes that someone has to decide and that PM is a defensible default — but skips the work of making the trade-off explicit before deferring. The PM ends up deciding without the structured disagreement-surfacing the value-5 path produces, which means the decision is worse-informed than it could be.
  • Split-the-difference (value 2). “Let’s ship Friday but only to internal users, and polish design over the weekend, and have engineering monitor closely.” This response signals conflict-aversion that produces a path satisfying none of the three constraints fully — internal shipping doesn’t honor the customer commit, weekend design work degrades sustainability, and “monitor closely” isn’t a real engineering plan. The split-the-difference pattern is a tell for low SJT; it looks reasonable in the moment and produces worse outcomes than either cleaner path.
  • Escalate-to-VP (value 1). “I’ll bring this to the VP to decide.” This response defers the cross-functional judgment that’s exactly what SJT is measuring. The VP has less context than the room does, will likely rubber-stamp whichever stakeholder presents most forcefully, and will (correctly) wonder why the room couldn’t decide. Escalation has a place — when the decision is genuinely above the room’s authority — but not as a default for cross-functional disagreement inside the room’s remit.

How the sample test scores you

In the AIEH 5-scenario Situational Judgment sample, this scenario contributes one of the five datapoints that aggregate into your single sjt_quality score. The W3.2 scoring fix normalizes by item count, so your score is the average of your five scenario values mapped onto a 1–5 scale, then bucketed into low (≤2), mid (≤4), or high (>4) for the directional result. The 1–5 binary scoring per item is mapped to low/mid/high level before reporting.

Data Notice: Sample-test results are directional indicators only. Five-scenario SJT samples are too few to be psychometrically valid; for a verified Skills Passport credential, take the full 40-scenario assessment.

The full 40-scenario assessment expands coverage across diverse cross-functional and individual-judgment contexts (stakeholder management, conflict mediation, escalation calibration, peer feedback) and produces a calibrated score on the AIEH 300–850 scale via the scoring methodology. For broader construct context, see skills-based hiring evidence and the hiring loop design overview.

  • Task vs relationship conflict. Jehn’s distinction — task conflict (about what to do) is productive when surfaced; relationship conflict (about who is right) is destructive. Strong SJT respondents keep disagreements in the task register.
  • Reversibility framing. Bezos’s “one-way doors vs two-way doors” frame — most launch decisions are two-way doors, and treating them as one-way produces unnecessary delay and unnecessary escalation. Naming the reversibility profile is core to the value-5 response.
  • Decision-time commitment. The discipline of naming when the decision will be made, even if the decision itself is deferred. Without a named time, disagreements drift; with one, they resolve.
  • Disagree-and-commit. The pattern of recording dissent, then committing to the decision once made. Strong cross-functional teams use this; weak ones re-litigate the decision after the fact.

For role pages where SJT is moderately to highly weighted, the hiring bias mitigation overview and the behavioral interview prep guide both treat cross-functional judgment as a foundational signal in structured evaluation.


Sources

  • McDaniel, M. A., Morgeson, F. P., Finnegan, E. B., Campion, M. A., & Braverman, E. P. (2001). Use of situational judgment tests to predict job performance. Journal of Applied Psychology, 86(4), 730–740.
  • Whetzel, D. L., & McDaniel, M. A. (2009). Situational judgment tests: An overview of current research. Human Resource Management Review, 19(3), 188–202.
  • Jehn, K. A. (1995). A multimethod examination of the benefits and detriments of intragroup conflict. Administrative Science Quarterly, 40(2), 256–282.
  • Schmidt, F. L., & Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology. Psychological Bulletin, 124(2), 262–274.

Try the question yourself

This explainer covers what the item measures. To see how you score on the full situational judgment family, take the free 5-question sample.

Take the situational judgment sample