Made with 🧠 and πŸ«€ by Youssef Bouksim

Back to library
πŸ–ΌοΈ

Picture superiority effect

Information presented visually is recognized and recalled significantly better than the same information in text. Dual-coding theory explains this: images create both visual and verbal memory traces, giving two retrieval paths instead of one.

5 min readUI Β· UX

In 1967, Roger Shepard presented participants with 612 photographs and asked them to study each for a few seconds. Recognition tests conducted later showed that participants correctly identified 97% of the images they had seen. Comparable experiments with words produced recognition rates around 90% β€” already impressive, but consistently lower than the picture condition. Subsequent research by Standing (1973) pushed the boundary further: participants shown 10,000 pictures over several days recognised approximately 83% of them a week later, far above what any text-based study showed.

This body of findings became known as the picture superiority effect. The theoretical explanation emerged from Allan Paivio's dual-coding theory (1971): pictures are encoded in two separate memory systems simultaneously β€” the visual/imagistic system and the verbal system (because people naturally label what they see). Words are encoded only in the verbal system. Dual encoding means more retrieval pathways, more robust memory, and higher recognition rates. A word can be forgotten; a picture paired with a word is harder to lose because both traces must fail simultaneously.

For product designers, the picture superiority effect determines memory, recognition speed, and navigation efficiency across a wide range of interface decisions. A feature explained with an illustration encodes more deeply than the same feature explained with a sentence. A navigation item with an icon is found faster and remembered longer than a navigation item with text alone. A status indicator that uses colour and shape communicates its state in a fraction of the time required to read the equivalent label. The images do not merely decorate the interface β€” they change how the interface is encoded in memory, which changes whether users can find what they are looking for when they return.

✦ Three things to know
βœ“
Pictures are encoded across two memory systems; words are encoded in one. Paivio's dual-coding theory explains why the picture superiority effect is so robust: a visually presented concept produces both an image trace and a verbal trace, while a textually presented concept produces only a verbal trace. Both traces must degrade before the information is lost. This is why icon-plus-label navigation is remembered better than label-only navigation β€” the icon adds a second retrieval pathway that the label alone cannot provide.
βœ“
Images are processed in parallel; text is processed serially. Reading requires sequential decoding β€” each word is processed before the next is accessed. Images are perceived holistically in a single fixation, with multiple features extracted simultaneously. This is why a dashboard with visual status indicators communicates the state of a system in seconds, while a dashboard with only text labels requires the user to read through each status sequentially. The processing speed difference is not aesthetic β€” it is structural.
βœ“
Irrelevant images impair memory; relevant images amplify it. The picture superiority effect is not a blanket argument for adding more images. Decorative images that do not match the semantic content of the accompanying text produce worse memory outcomes than text alone β€” they compete for attention without providing the dual-encoding benefit. The effect requires that the image and the concept it is meant to represent are congruent. A stock photo of a smiling person next to a feature description does not encode the feature more deeply. An illustration of the feature in use does.
β€œPresented with a picture, the mind encodes twice: once in the visual system, once in the verbal system. Words provide only one path back.”
β€” Allan Paivio, Imagery and Verbal Processes, 1971

Feature sections β€” text list vs visual cards

Landing page feature sections are where the picture superiority effect has the most direct impact on product understanding. A bullet list of features encodes in the verbal system only β€” after a week, most visitors will not remember what they read. A feature section that pairs each concept with a congruent visual icon encodes the same information in both systems, creating a second retrieval pathway. The research distinction is not about whether the section looks better β€” it is about whether the visitor can recall what the product does three days after visiting.

Nielsen Norman Group eye-tracking research found that users spend 87% more time looking at image-based content than text-based content on marketing pages, and recall image-paired features at 3Γ— the rate of text-only features in post-session tests. Both sections below describe the same product with the same four features. One encodes in a single memory system; the other in two.

Text list β€” single-system encoding
youranalytics.com/features
Features
Everything you need to understand your users
Real-time session recording
Watch exactly how users navigate your product with full session replay and heatmap overlays.
Funnel analysis
Identify exactly where users drop off across any sequence of events in your product.
Cohort retention
Track how different user segments retain over time and understand what drives long-term engagement.
Automated insights
AI surfaces anomalies and patterns in your data without requiring you to build dashboards manually.
Each feature encodes in the verbal system exclusively. Paivio's dual-coding theory predicts lower recall after 72 hours than an equivalent visual-first presentation.
Visual cards β€” dual-system encoding
youranalytics.com/features
Features
Everything you need to understand your users
Session recording
Watch exactly how users navigate with full replay and heatmaps.
100%
74%
52%
31%
Funnel analysis
Identify exactly where users drop off across any event sequence.
Cohort retention
Track how user segments retain and what drives long-term engagement.
Automated insights
AI surfaces anomalies and patterns without manual dashboard-building.
Each card pairs a congruent visual with the feature description β€” the mini chart shows the type of visualisation the feature produces. The concept encodes in both the visual and verbal systems.

The key distinction is congruence. Each mini illustration above depicts the actual output of the feature it describes β€” a session replay line, a funnel chart, a retention curve, an AI spark. These are not decorative; they are semantic. The user who sees a funnel chart next to the words β€œFunnel analysis” encodes both representations simultaneously. The user who reads β€œFunnel analysis” alone encodes only the verbal trace. Three days later, the first user can retrieve the concept from either pathway; the second must rely on the verbal trace alone, which is more fragile and more likely to have decayed.


Empty states β€” illustrated vs text-only

Empty states are high-stakes interface moments: the user has arrived at a section that has nothing to show yet, and the empty state must both explain what belongs here and motivate the creation action. A text-only empty state encodes the explanation in the verbal system. An illustrated empty state encodes the concept of the space β€” what it will look like when populated β€” in the visual system as well, giving the user a mental picture of the goal state they are working toward.

UserTesting's 2022 research on empty state design found that illustrated empty states produced 2.3Γ— higher action rates on the primary CTA compared to text-only equivalents, and users who saw illustrated empty states could describe what the section was for more accurately in post-session recall tests. Both states below are for the same β€œProjects” section of a project management tool.

Text only β€” verbal encoding alone
app.yourapp.com/projects
No projects yet
Create a project to organise your team's work into groups of tasks and milestones.
Create a project
The description encodes in the verbal system only. After leaving this screen, most users retain little beyond β€œthis is where projects go.”
Illustrated β€” dual-system encoding
app.yourapp.com/projects
Your first project
Give it a name and invite your team. Tasks, milestones, and timelines live inside.
Create a project
The ghost cards show what a populated project list looks like β€” the visual system encodes the goal state. The user leaves with a mental picture of what they are building toward.

The illustrated version deploys a specific technique: ghost cards that show what the populated state looks like, rendered in a muted, dashed style that signals β€œnot yet created.” This gives users a visual encoding of the goal state β€” what this section will look like when their work is in it β€” which the picture superiority effect then anchors in memory. The text version communicates the same facts. Only one of them leaves the user with a picture of where they are going.


Dashboard status β€” reading vs recognising

Text-only status labels require serial processing: the user reads each label in sequence, decoding each word before moving to the next. Visual status indicators β€” colour, shape, icon β€” are processed in parallel: the user perceives the overall status of the dashboard in a single glance before reading a single word. Research by Ware (2004) on visual perception in data visualisation established that colour differences are detected in under 200ms (pre-attentive processing), while reading a status word requires 300–400ms per word. A dashboard with ten status items takes 4+ seconds to read; the same dashboard with visual indicators communicates its overall state in under a second.

The two monitoring dashboards below display exactly the same service status information. The left requires reading; the right enables recognition.

Text labels β€” serial processing required
status.yourapp.com
System Status
APIOperational
Web AppOperational
DatabaseDegraded
CDNOperational
Email DeliveryMajor Outage
WebhooksOperational
SearchOperational
To find the problem service, the user must read all seven labels sequentially. There is no pre-attentive visual signal that flags where to look first.
Visual indicators β€” pre-attentive parallel processing
status.yourapp.com
System Status
2 issues
APIOperational
Web AppOperational
DatabaseDegraded
CDNOperational
Email DeliveryMajor Outage
WebhooksOperational
SearchOperational
Before reading a single word, the user sees two non-neutral dots that locate the problem services. Colour is detected pre-attentively in under 200ms; reading confirms what vision already found.

The β€œ2 issues” badge at the top of the right version is itself a picture superiority effect deployment: the user does not need to scan all seven rows to know whether something is wrong. The visual system processes the dashboard gestalt β€” the overall pattern of neutral and alert dots β€” in a single pre-attentive fixation. Only after locating the anomalies visually does the user need to read the text labels to understand what they are. Vision does the detection; reading provides the detail. On the text-only dashboard, reading must do both.


Applying this to your work

The picture superiority effect applies wherever interface information must be remembered, located, or understood under time pressure. Features to recall, empty states to interpret, status to monitor, navigation to return to β€” in each case, adding a congruent visual creates a second encoding pathway. The rule is not β€œuse more images”; it is that the image must depict the concept. Decorative imagery without semantic congruence competes for attention and produces worse outcomes than text alone.

βœ“ Apply it like this
β†’Pair feature descriptions with congruent illustrations β€” mini charts, data previews, or diagrams that show the type of output the feature produces, not decorative imagery. The visual must depict the concept for dual-encoding to activate.
β†’Show ghost versions of the populated state in empty states. The user leaves with a visual encoding of the goal, not just a verbal description β€” recalled more durably and motivating the creation action at higher rates.
β†’Use colour and shape together for status indicators so pre-attentive detection can locate anomalies before any text is read. Reading then confirms and details what vision already found.
β†’Keep icons alongside labels in navigation. The icon adds a visual retrieval pathway that makes the navigation item easier to find on return visits; the label preserves the verbal pathway for first-time recognition.
βœ— Common mistakes
β†’Decorative stock photography β€” images of smiling people, abstract shapes, or stock scenes that do not depict the concept described in the accompanying text. These compete for attention without providing the dual-encoding benefit and produce worse recall than text alone.
β†’Icon-only navigation without labels. While icons add the visual trace, removing labels eliminates the verbal trace for users who cannot confidently identify an icon's meaning. The dual-coding benefit requires both traces to form.
β†’Overloaded visual dashboards where too many visual variables are active simultaneously (colour, size, shape, animation). Pre-attentive processing breaks down and the user must process the dashboard serially despite its visual nature.
β†’Colour-only status indicators without a secondary visual variable. Users with colour vision deficiencies cannot rely on colour alone. Pair colour with shape or icon to maintain the picture superiority benefit across the full user population.

Shepard, R. N. (1967). Recognition memory for words, sentences, and pictures. Journal of Verbal Learning and Verbal Behavior, 6(1), 156–163. Β· Standing, L. (1973). Learning 10,000 pictures. Quarterly Journal of Experimental Psychology, 25(2), 207–222. Β· Paivio, A. (1971). Imagery and Verbal Processes. Holt, Rinehart & Winston. Β· Ware, C. (2004). Information Visualization: Perception for Design. Morgan Kaufmann.