Made with 🧠 and 🫀 by Youssef Bouksim

Back to library
🔬

Observer Expectancy Effect

Researchers unconsciously influence the outcomes of their own studies. The questions we ask, the way we present work for critique, the reactions we show when users struggle — all of it shapes what we observe. The data collected reflects not just what users think, but what the person collecting it expected them to think.

5 min readUser Research · Surveys · Design Critique

In 1963, psychologist Robert Rosenthal asked student experimenters to run rats through mazes. Half were told their rats were specially bred for maze intelligence. The other half were told their rats were ordinary. Both groups were working with the same standard lab rats — the labels were fabricated. But the “intelligent” rats genuinely performed better. The students, expecting more, handled them more carefully, trained them more patiently, and interpreted ambiguous outcomes more generously.

This is the observer expectancy effect: the unconscious influence of a researcher's hypothesis on the data they collect. It is not dishonesty. It operates below the level of conscious awareness, through thousands of small decisions — which observations to record, how to frame a question, when to intervene, what “hesitation” means.

The effect runs in two directions. The observer unconsciously interprets ambiguous data toward their hypothesis. And users, who are acutely sensitive to social cues, unconsciously respond to what the observer seems to want. Both effects compound.

✦ Key takeaways
✓
The designer of a feature should never moderate tests of it. Investment creates expectation, and expectation creates the effect. This is structural, not personal. The fix: separate design from moderation.
✓
Questions bake in expectations. “How much did you enjoy our new onboarding?” does not measure satisfaction. It measures social compliance with an assumed positive experience. Neutral phrasing produces more honest distributions.
✓
How you present work for critique determines what feedback you receive. “I've been working on this for three weeks” will receive validation. “Which of these two directions solves this problem better?” creates conditions for actual critique.
“The observer's expectations become a self-fulfilling prophecy — not through magic, but through the thousand small ways expectation shapes every decision made during data collection.”

In usability tests

The task prompt is the single most contaminated document in a typical usability test, and it is almost always written by the designer who built what is being tested. The rule is simple: describe what the user wants to accomplish, not what the design does. Any adjective describing the product, any framing that positions the experience as “new” or “improved” — all of it shapes what participants attend to.

Loaded

“We've just redesigned our checkout to make it faster and simpler. Please go ahead and purchase the running shoes using our new streamlined checkout.”

'Faster and simpler' signals the expected reaction. 'New streamlined' frames novelty as a quality. Participants who find it slow will doubt their own experience.

Neutral

“You have a pair of running shoes in your cart. Please complete the purchase using your saved Visa card. Talk through what you're doing as you go.”

Concrete goal. No adjectives about the design. No mention of a redesign. Participants approach it as users with a task.

The think-aloud instruction (“talk through what you're doing”) is neutral — it elicits behaviour without directing it. Even reactive sounds — “interesting,” “mmm,” audible note-taking when something goes wrong — are signals that participants read and respond to in real time.


In surveys

User surveys are where the observer expectancy effect reaches its largest scale, because the contamination is written into the question and then delivered to thousands of users simultaneously. The bias is invisible in the results — the numbers look credible, the sample size looks real — but they reflect the question's framing as much as the user's experience.

Loaded

“How much did you enjoy setting up your account?” / “Our new onboarding is designed to get you started in under 3 minutes. Did it meet that goal?”

The first assumes enjoyment. The second anchors to the team's target. The third presupposes something was liked.

Neutral

“How easy was it to get your account set up?” / “How long did account setup feel like it took?” / “What, if anything, was unclear during setup?”

Symmetric scale. No team goal embedded. 'If anything' signals that 'nothing' is a valid answer.

The audit is mechanical: read every question and ask whether there is a “correct” answer that a socially aware person would gravitate toward. If yes, the question is leading.


In critique sessions

Design critique is not immune to the observer expectancy effect — it may be its most common manifestation in product teams. The moment a designer discloses emotional investment, the review changes character. Reviewers shift from evaluating the work to managing the relationship.

Loaded

“I've been working really hard on this for three weeks and I think it's finally in a good place. I'd love to get your thoughts before I hand it off to dev.”

Discloses time investment. Uses 'love' and 'means a lot' -- emotional language that signals criticism is costly.

Neutral

“I have two directions. Option A keeps the current nav with visual updates. Option B collapses advanced settings behind a toggle. Which reduces cognitive load more? Feedback needed by EOD.”

No emotional context. Two named options. A specific criterion. A deadline. Reviewers engage with the design problem.

The structural fix is to always present at least two directions. When reviewers are choosing between options rather than validating a single one, they engage as decision-makers rather than as social actors managing a colleague's feelings.


Why it matters structurally

The observer expectancy effect cannot be eliminated through effort or awareness. A designer who knows about the bias and tries to moderate their own test neutrally will still moderate it less neutrally than a facilitator without investment in the outcome. The fix is structural: separate the person with the expectation from the person who designs the research instrument and collects the data.

✓ Apply it like this
→Never moderate usability tests of your own work. Swap moderation responsibilities with another designer, or run unmoderated sessions.
→Audit every survey question for assumed valence before sending. Replace emotional verbs (enjoy, love, like) with measurable ones (easy/difficult, clear/confusing).
→Present at least two directions in every critique session. Frame around a specific decision with a named criterion.
→Never disclose time investment when requesting critique. 'I spent three weeks on this' changes the room before anyone says a word.
✗ Common mistakes
→Task scripts that describe the design rather than the user's goal -- any adjective about the product contaminates what participants attend to.
→Surveys written by the team that built the feature -- the questions will reflect what that team hoped to find.
→Presenting a single direction with open-ended feedback requests -- positions reviewers as validators of a made decision.
→Moderator reactions during sessions -- leaning forward, saying 'interesting' when users struggle -- all broadcast expectations.

Rosenthal, R., & Fode, K. L. (1963). The effect of experimenter bias on the performance of the albino rat. Behavioral Science, 8(3), 183-189. Rosenthal, R., & Jacobson, L. (1968). Pygmalion in the Classroom. Orne, M. T. (1962). On the social psychology of the psychological experiment. American Psychologist, 17(11), 776-783.