Methodology Work Services Book Writing Make Me Think - chapters Methods Guide Notes Blog
The Complete Reference

Perception-First Design™

Source document for the methodology. ~100 academic citations across cognitive psychology, HCI, behavioral economics, design theory, STS, and ADHD cognitive neuroscience. 15 years of applied work across e-commerce, SaaS, consumer electronics, and entertainment.

Stefan Kovalik · Aurochs · Version 3.6 · April 2026

The Bouncer

I figured out my design methodology in a nightclub.

I was head of security at a 300-capacity venue. Staff of 15. Every night at 1:30 AM, same problem: lights slam on, DJ announces we’re closed, and I’m barking at 300 drunk people to get out. They hated it. I hated it. Fights happened. Stragglers lingered. The whole thing was hostile.

So I tried something different. At 1:10, I walked table by table. “Hey, last call’s coming up. Get any drinks you want. Thanks for coming out tonight.” That’s it. No urgency. No authority voice. Just information, a suggestion, and respect.

At 1:30, when the DJ made the announcement and the lights went up, half the venue was already walking out the door. No barking. No fights. And bar sales went up 30% in those last 20 minutes.

I didn’t know what to call that. I just knew it worked, and I knew why it worked. The old way treated people like obstacles. My way treated them like people with their own goals who’d cooperate if you gave them the right information at the right time with the right tone.

That’s Perception-First Design before I had words for it. Everything I’ve done in the 15 years since; revenue 3x, Costco pallets and Walmart shelves and Disney licensing, improv theaters and smart home brands. It all comes from that same instinct. Design for how people actually perceive and process the world, not how you wish they did.

The Thesis

The design industry has spent 15 years saying the same thing: users have limited attention, so don’t waste it. Lower friction. Reduce steps. Don’t make them think.

My version is different: users don’t think. Not until you make them.

Nobody is thinking. They’re on autopilot. The same way you catch a ball without calculating trajectory, or walk without planning each step; that’s how people interact with websites, products, brands, everything. Autopilot. Energy conservation. The brain handles the vast majority of processing unconsciously. Kahneman (2011) calls this System 1: the fast, automatic mode that runs most of cognitive life. System 2, slow and deliberate and conscious, only activates when something demands attention.

Contemporary cognitive science calls this predictive processing (Clark, 2013; Friston, 2010). The brain is constantly generating predictions about what it expects to encounter. When reality matches prediction: autopilot. When reality violates prediction: attention fires, System 2 kicks in. That prediction error is what I call the activation point.

The design challenge isn’t “don’t waste their limited attention.” It’s that effortful attention is dormant by default. The real question is: when you do make them stop and actually think, when you generate a prediction error that activates them from autopilot, what do they think about? How do they feel? And what do they do next?

That’s what I design for. Not the funnel. The activation.

The Industry SaysI Say
AssumptionAttention is scarce; conserve itEffortful attention is dormant; design the activation
GoalReduce friction to goalsCreate the right moments that make someone stop, think, and act
Failure modeUser runs out of patienceUser never engages; autopilots right past you
MetaphorBudget of attentionAutopilot with activation points

Steve Krug wrote Don’t Make Me Think. Great book. It gave me the vocabulary to talk to stakeholders early in my career. His scope was usability, and he stayed in it deliberately. But nobody’s thinking in the first place. Design for that bias, the bias of making someone stop and engage, instead of funneling them to an action with an affordance of attention they never gave you.

Why I See This

Two things about my brain that I didn’t choose but that turned out to be professional advantages.

I’m autistic. After my formal ASD diagnosis, my partner pointed out something I’d been missing: “Do you realize you’re not picking up on me not being interested at all in what you’re talking about, and you seem to be getting frustrated?” Around 2012, a friend I met at a Chicago meetup about emotions in video games suggested I try improv. I went in already analytical. I had impaired social instincts, not absent ones. I didn’t need to learn what a smile means. I had to learn what eye contact can mean, what a pause signals, what “said vs. unsaid” reveals. I learned it like you’d learn an instrument or ballet: theory, practice, application loops, over and over.

That means I can see and articulate the unconscious patterns that most designers apply intuitively but can’t explain. People interact with websites the same way they interact with other humans. We form first impressions on products the same way we do when someone walks into a room. It’s all unconscious, and we can’t help it.

I also have severe combined-type ADHD. Some experiences are literally painful for me because of how taxing they are on my attention and cognitive load. I’m acutely sensitive to friction, clutter, noise, confusion. The things that bad design inflicts on everyone, I feel at a higher volume. I design because I’m sensitive to these patterns. What most designers theorize about, I feel.

And the ADHD gave me something else: years of building mental prostheses. If it’s out of sight, it’s literally out of mind for me, so I keep post-it notes on my desk for today’s tasks. “Out of sight, out of mind” becomes: what MUST we show vs. what we want to show vs. what we could show. Then cut everything but what we must show.

Design is prosthesis for human cognitive functions and limitations. I know that because I’ve been building prostheses for myself my whole life. The curb cut effect: what I built for my own cognitive accessibility turned out to help everyone.

I’ve come to think of this as a curb-cut effect. The cognitive constraints I designed around — limited working memory, sensitivity to disfluency, impatience with unnecessary friction — are constraints every user has. Mine are just louder. A design that works for the constrained case tends to work better for everyone.

The five PFD layers draw from two largely distinct bodies of cognitive science — perception psychology and ADHD cognitive neuroscience — connected here for the first time. The citation lists share no duplicate entries.

The Conductor Model

I don’t build funnels. I conduct.

Think about a musician who’s reached unconscious mastery. They stop thinking about how things are done and focus entirely on how the music makes you feel. That’s the goal. Prime someone with a first impression, then conduct their thoughts and emotions through the experience. Design a mental waterfall.

There’s a flow of thought and perception: momentum, weight, cadence, tone, speed, context, priming, expectations. Like singing down a scale in front of a crowd and pausing before the last note. Everyone knows what comes next. That recognition, that pattern completion, is satisfying. That’s what a great CTA feels like: the only possible resolution to the experience that preceded it.

The User’s Mental Waterfall

The unconscious sequence of questions every visitor processes:

  1. What am I trying to accomplish here? → Intent
  2. Is there a clear path to that? → Navigation
  3. What’s the point of this site? → Value proposition
  4. For whom is this? → Do I belong here?
  5. What is it trying to tell me? → Messaging: ew or cool?
  6. What is it trying to get me to do? → Call to action
  7. How do I feel about doing it? → “Prove you’re worth my time” or “Tell me more!”
  8. Where’s the win-win? → Mutual value
  9. Does this give me what I want? → Delivery

The hero section is the activation point; the table visit at 1:10. If it answers questions 1 through 5 in the first second or two, the user is primed. Then you conduct them through 6 through 9 as they scroll. If the hero fails, it’s the lights slamming on.

My improv teacher gave me a formula: (Listen) × (Act + React). If Listen equals zero, if the user isn’t attending, then everything else multiplies by zero.

Methodology Siblings

PFD is a practitioner synthesis; the lineage isn’t novel. Three older traditions solve the same problem from different angles.

Kansei engineering (Nagamachi, 1995)

Developed at Hiroshima University in the 1970s, formalized in the mid-90s. Kansei = “psychological feeling, image of a new product.” The methodology maps consumer affective response onto measurable product attributes, then engineers toward the target feeling. This is the closest ancestor to PFD: both start with the pre-verbal response and work backward to design decisions. Kansei is PFD’s Japanese cousin. Anyone doing serious work in automotive, consumer electronics, or product design in Japan already knows it.

Gestalt psychology (Wertheimer, 1923; Metzger, 1936/2006)

The Berlin School’s empirical treatment of perception as organized wholes, not sums of parts. Proximity, similarity, closure, continuity, common fate, figure/ground. These aren’t decorative rules. They’re the first systematic description of how the visual system constructs meaning from raw input. PFD’s L1 (First Impression) and L2 (Processing Fluency) are applied Gestalt. The debt is direct.

Neuroaesthetics (Skov & Nadal, 2020; Leder, Belke, Oeberst & Augustin, 2004)

Modern empirical aesthetics, rooted in Vienna (Leder’s lab) and Copenhagen (Skov’s). Where Kansei engineers the feeling and Gestalt describes the perception, neuroaesthetics maps the neural substrate: amygdala, orbitofrontal cortex, reward pathways. Leder’s 5-stage aesthetic appreciation model (perception → implicit integration → explicit classification → cognitive mastery → evaluation) is a finer-grained L1 subroutine within PFD’s stack.

How PFD differs

PFD is a dependency stack. Upstream failures degrade everything downstream. Kansei maps feeling to attributes but is silent on ordering. Gestalt describes perception but not decision. Neuroaesthetics explains the brain but isn’t a working methodology for a Tuesday afternoon. What PFD adds is the ordering rule and the operational protocol. The sciences are shared; the stack logic is the practitioner contribution.

The Framework

Five layers as a dependency stack. Each depends on the layer below it. Fix them in order, because upstream failures degrade everything downstream. But think of it less as a stack and more as a composition. The foundation sets the key. Each layer adds another instrument. If the foundation is broken, the whole piece is noise.

Decision Architecture L4
Perception Bias Optimization L3
Processing Fluency L2
First-Impression Architecture L1
Cognitive Load Reduction Layer 0: Cognitive Load

In predictive processing terms (Clark, 2013; Hohwy, 2013): when a design matches the brain’s predictions at every layer, processing is fluid and unconscious. Autopilot. When a layer fails, it generates prediction error that demands conscious attention. Good design means each layer confirms predictions so the brain’s resources flow toward the decisions that matter, not toward deciphering the interface.

On the research base: The empirical benchmarks throughout (timing thresholds, working memory parameters, fluency effects) were measured primarily on WEIRD samples (Western, Educated, Industrialized, Rich, Democratic; Henrich, Heine & Norenzayan, 2010). The underlying cognitive architecture is expected to generalize; specific parameter values may vary. PFD is a practitioner synthesis, not an experimental program. The citations establish the mechanisms; the 15-year applied record tests them.
Layer 0: Cognitive Load

Cognitive Load Reduction

“Can’t perceive anything without the bandwidth to do so.”

Working memory holds about 3–5 chunks at a time, not the “7 plus or minus 2” you’ve seen quoted everywhere. Miller (1956) was describing the recurrence of the number 7 across disparate information-processing phenomena as a rhetorical observation, not proposing a single working memory limit. That nuance got lost. Cowan (2001, 2010) established the modern estimate: roughly 3–5 chunks when rehearsal and grouping strategies are prevented.

Sweller’s Cognitive Load Theory (1988) distinguishes intrinsic load (complexity inherent to the task) from extraneous load (complexity caused by poor design). You can’t change intrinsic load. But every unit of extraneous load you eliminate frees working memory for the decisions that actually matter.

This is L0 because without bandwidth, nothing else works. An overloaded visitor can’t form a first impression, can’t process your brand, can’t connect with your message, can’t find the trail. They’re just gone.

Dodson’s Interest-Based Nervous System model (IBNS) proposes a refinement: ADHD attention is regulated by interest, not importance. The PINCH model identifies five drivers that override cognitive load ceilings: Passion, Interest, Novelty, Competition, Humor. This expands L0 from “reduce cognitive load” to “reduce cognitive load OR increase intrinsic interest.” For ADHD users, and partially for everyone under low motivation, load reduction alone is insufficient when interest is absent. The constraint is not just bandwidth. It is bandwidth weighted by engagement.

Causal chain: Fewer choices → less mental effort → faster decisions → more conversions

In Practice

  • Reduce choices at every step. Audit every dropdown, radio button, and “which one?” moment.
  • Progressive disclosure: show what’s needed now, reveal the rest on demand.
  • Smart defaults: pre-select what 80% of users choose.
  • If you don’t need a form field to complete the transaction, delete it. If you need it later, ask later.
SSH: The site presented “very matter-of-fact statements and not much else.” Restructured content to answer user questions before they had to ask them. Freed up bandwidth so they could evaluate the product instead of deciphering the interface.

ClicSmart App: Started with complex user flows, narrowed down to 3 primary navigation tabs. Radical simplification.
Layer 1

First-Impression Architecture

“Your site has 50 milliseconds. That’s not a design constraint. That’s the physics of how your visitors’ brains work.”

Visual evaluation begins within 50ms, faster than conscious thought (Lindgaard et al., 2006; Tuch et al., 2012). Reinecke et al. (2013) at Google Research found effects at exposures as short as 17ms. The aesthetic-usability effect (Kurosu & Kashimura, 1995; replicated cross-culturally by Tractinsky, 1997) shows that attractive interfaces are perceived as more usable, even with identical functionality.

People form first impressions on products the same way they do when someone walks into a room. Heider & Simmel (1944) showed people attribute intentions and emotions to geometric shapes. Gray, Gray & Wegner (2007) formalized this as mind perception: Agency (capacity to think/act) and Experience (capacity to feel/sense). Users construct lightweight models of competence and care for websites; enough to pass or fail the same trust evaluation we run on humans.

The first impression IS the activation point. It’s the bouncer’s table visit. If it fails, Listen equals zero, and everything multiplies by zero.

Causal chain: Faces in hero → instant pattern recognition → positive emotional response → trust established before thinking begins

In Practice

  • The hero section is the thesis statement of the entire site. If the thesis is wrong, everything below it fails.
  • Faces exploit neonatal face preference (Johnson et al., 1991) and adult face “pop-out” in visual search (Hershler & Hochstein, 2005).
  • Visual quality must match price point. $20–40 tablets in wood cases? I created the perception of “smart home decor” and sold them for $100–200.
iO Theater: Hobbyist WordPress site. Cluttered, inconsistent, Google Maps in the hero. “Doesn’t look like a real theater.” Rebuilt with dramatic photography, clear event navigation, Bill Murray quote. Online ticket sales: 50% → 75%. Same shows. Different perception.
Layer 2

Processing Fluency

“If it’s easy to process, it feels true.”

Reber & Schwarz (1999) showed that identical statements in easy-to-read fonts are judged as more true. Alter & Oppenheimer (2009) showed fluency effects generalize across truth, confidence, liking, and trust. Dechêne et al. (2010) confirmed the illusory truth effect across a meta-analysis of 51 studies.

Consistency matters more than creativity. Lakoff & Johnson (1980) showed abstract thought is structured by metaphors rooted in bodily experience. Spence (2011) demonstrated cross-modal correspondences: people consistently match high-pitched sounds with small, bright objects. When sensory input is consistent and fluid across channels, the brain reads it as trustworthy.

Color consistency operates in perceptual space, not physical space. Near-miss color deviations (a brand blue 3% off) are disproportionately costly because they sit in the steep, high-sensitivity zone of the perceptual metric where prediction error is maximal (Bujack et al., 2022). Perceptually uniform spaces (OKLCH, CAM16-UCS) are L2 infrastructure.

Causal chain: Unified branding → feels professional → perceived as trustworthy → clears “would I buy from this?” bar
SSH 4-year arc: No brand system in 2020. By 2024: comprehensive typography system, seasonal palettes, Disney and Brookstone sub-brand mood boards, ClicSmart and myLuma sub-brands, all with family resemblance. Processing fluency is brand infrastructure that compounds. Revenue 3x.
Layer 3

Perception Bias Optimization

“It’s not for you. It’s for them.”

Users autopilot decisions, then construct rational explanations after the fact (Nisbett & Wilson, 1977). Kahneman (2011): System 1 makes the decision; System 2 constructs the explanation. Survey data captures the System 2 rationalization. Analytics captures the System 1 decision. The gap between them is where perception problems hide.

In predictive processing terms (Clark, 2013): design for what the brain predicts, not what users report. What users say in surveys is their System 2 reconstruction. What their brain predicted, and how the design matched or violated that prediction, is what actually drove behavior.

Ultimately what works in marketing isn’t “here’s what our product does.” It’s “here’s what having our product makes you feel.” The iPod wasn’t a better product than the Zune. It was about freedom of movement, self-expression, your library in your hands.

Causal chain: Emotional tagline → connects to real desire → resonates with target customer → action
SSH: Marketing focused on screen resolution, WiFi, app features. Workshopped through “what does this product actually DO for the user emotionally?” Discovery: connection, not technology. Tagline: “Stay connected, even when you’re apart.” Two years later, Aura Frames was using nearly identical messaging. When your competitor steals your tagline, you found the emotional core.
Layer 4

Decision Architecture

“They’re hunters looking for prey, and it’s my job to make a trail.”

Every design is a choice architecture. There is no neutral presentation of options. This builds on Thaler & Sunstein (2008), though a Bayesian re-analysis suggests nudge effects may be near-zero after correcting for publication bias (Maier et al., 2022). The structural principles (defaults matter, framing matters, fewer options at decision points reduce paralysis) remain well-supported.

Context qualifier: “Fewer options” applies strongest at high-stakes single decisions and unfamiliar purchase choices. It does NOT apply in expert tool palettes or workspace contexts, where option reduction actively harms productivity. The operative principle: reduce choices at decision points, not choices everywhere.

Barkley’s (1997) model adds another dimension for ADHD users: self-regulation failures occur at the point of performance, not at the point of planning. ADHD temporal processing collapses to “now and not now.” Future benefits have near-zero motivational weight. L4 must make the benefit immediate, not promised. Standard decision architecture assumes people weigh future consequences. For ADHD users, and to a lesser degree for everyone under cognitive load or time pressure, the trail must pay off at every step, not just at the destination.

Causal chain: Clear path → user finds goal → reduced friction → purchase completed
SSH homepage: Before: “didn’t answer any of the user’s questions, goals, or form a good impression.” Navigation reflected the company structure. After: rebuilt as a trail. Emotional hook → understands the product → social proof → clear path to purchase.

The Diagnostic

How I actually diagnose a site. Not a checklist; a process.

Step 1: Feel (Pre-Verbal)

Arrive at the page and let the emotional response fire before Wernicke’s area translates it into words. “This makes me feel X.” The feeling comes before the language. System 1 before System 2. My ADHD sensitivity is an advantage here; I feel cognitive load and friction more acutely than most users. The emotional response IS the diagnostic instrument.

Step 2: Unpack

“Why does this feel X?” Now I’m looking for the systematic rule that’s broken: don’t overload, be specific, get to your point, signal-to-noise ratio, everything has a purpose, this should feel X. Does it?

Step 3: Diagnose

SignalLikely Layer
High bounce + low time-on-siteLayer 2: First impression
High time + low conversionLayer 5: Decision architecture (or trust)
High add-to-cart + low checkoutLayer 1: Cognitive load (or price perception)
Good desktop + poor mobileLayer 3: Processing fluency
Stated behavior ≠ actual behaviorLayer 4: Perception bias

Step 4: Prescribe

Fix the lowest failing layer first. Always. Fixing decision architecture when the first impression is broken is like optimizing a trail that nobody enters.

The 5-Minute Perception Audit

1. Path Count (Cognitive Load)

Count every decision from landing to purchase. Compare to competitors. Higher count = losing to friction.

2. 5-Second Test (First Impression)

Show the homepage for 5 seconds. Can they state purpose, feel trust, credit card confidence?

3. Squint Test (Processing Fluency)

Blur your eyes. What stands out? If everything looks equally important, nothing is.

4. Analytics Gap (Perception Bias)

Compare what users say vs. what they do. The gap is where problems hide.

5. Trail Test (Decision Architecture)

Incognito mode. First-time visitor. Core offer in 3 clicks or search bar?

The Generative Protocol

The Diagnostic decodes failures. The Generative Protocol derives solutions. Evaluation and generation are different cognitive operations. Evaluating a proposal against PFD catches surface violations. Deriving a solution bottom-up from PFD’s constraints surfaces requirements that no existing proposal addresses.

Rule Zero: Do NOT propose any solution until all 5 layers are analyzed and requirements accumulated. Solution-first thinking contaminates the requirement space. You stop seeing what the layers demand and start seeing what the solution provides.
  1. State the Design Problem in terms of user experience, not missing features
  2. Work Each Layer Bottom-Up. For each: state the constraint, state the violation, derive the MUST requirement, label R[n]
  3. Accumulate Requirements. R1–R5, all non-negotiable. Lower layers win conflicts.
  4. Derive the Solution. What satisfies R1 AND R2 AND R3 AND R4 AND R5?
  5. Test Against Proposals. Which requirements does each proposal fail?
  6. Identify the Gap. Requirements no proposal satisfies = where the non-obvious solution lives

The Gate

Run before shipping. Each layer is pass/fail. If any layer below L4 fails, do not ship.

LayerGate QuestionPass Condition
L0Can a first-time user complete the primary task without getting stuck?Zero “where do I click?” moments
L1Does the first impression activate the right emotion?5-second test: users can state purpose + feel trust
L2Is every element consistent with brand and price point?Squint test passes. No rogue fonts, colors, spacing.
L3Does the design speak to what users respond to?Copy leads with outcomes. Analytics aligns.
L4Does the trail lead to conversion without manipulation?Core offer ≤ 3 clicks. Passes all three ethical tests.

The Communication Hierarchy

Music, humor, and stories: three modalities for changing perception, ranked by power.

Music (Most Powerful)

Music communicates emotion and bypasses rational processing entirely. Koelsch (2014) showed music modulates subcortical emotion structures rapidly, potentially prior to conscious evaluation. Juslin & Västfjäll (2008) identified six mechanisms for music-evoked emotion, the majority bypassing conscious processing. In design: typography rhythm, spacing cadence, visual flow, animation timing, scroll momentum. The “music” of a page controls pacing and emotional state.

Humor (Second)

Jokes change how people feel about a thing by pointing out absurdity and giving permission to think without rigidity. A joke is: setup creates expectation, punch reveals a pattern that either aligns with your self-image or doesn’t, and the anxiety between amplifies the payoff. Laughter is the anxiety being released. Three comedy escalation patterns map to UX: linear repetition (brand consistency), sequential escalation (funnel narrative), exponential escalation (hero → proof → testimonial → CTA).

Stories (Third)

Green & Brock (2000) demonstrated “narrative transportation,” where absorption into a story produces belief changes. Two meta-analyses confirm (van Laer et al., 2014; Braddock & Dillard, 2016). The story isn’t the events. It’s how someone changes from A to B. The user is the protagonist. “I have a problem” → “This solves my problem.” The hero is the inciting incident. The CTA is the climax.

Websites as Relationships

Users evaluate websites the same way they evaluate people. Same unconscious process, same red and green flags.

Relationship Red FlagWebsite EquivalentMind Perception
Doesn’t take appearance seriouslyLow-quality design, broken layoutsLow Agency (incompetent)
Lack of confidenceHedging language, unclear value propLow Agency
Too verboseWalls of text, burying the leadLow Agency
Non-specific / vagueGeneric claims, no proofLow Agency + Low Experience
Too good to be trueOver-promising without social proofSuspicious Agency
Acts like you’re lucky to be thereSelf-centered messaging, no empathyLow Experience

The trust-killer that operates before anyone reads a word: disrespect. A site that doesn’t treat users with respect, their time, their intelligence, their needs, triggers the same rejection as a person who acts like you’re lucky to interact with them.

The Responsibility

If we’re engineers of thoughts and emotions, we need a Hippocratic oath.

The distinction: The designer controls the perception layer: presentation, UX, visual hierarchy, emotional messaging, activation points. The organization controls the value layer: product quality, pricing integrity, fulfillment, service. PFD’s ethical mandate is to align perception with reality.

When the product is good and the perception is bad (the Simply Smart Home case) PFD closes the gap upward, bringing perception up to match genuine value. That’s not manipulation. That’s correction. The violation occurs in the opposite direction: inflating perception above reality.

The Three-Part Ethical Test

  1. The Alignment Test: Does this design bring perception closer to reality, or further from it? (Berdichevsky & Neuenschwander, 1999)
  2. The Sincerity Test: If the user fully understood what this design choice does, would they feel served or exploited? Consumers don’t object to persuasion. They object to persuasion perceived as insincere (Friestad & Wright, 1994; Campbell & Kirmani, 2000).
  3. The Golden Rule: Would I, as the designer, consent to being influenced by this technique if I were the user?

Dark patterns (Brignull, 2010; Gray et al., 2018), nagging, sneaking, forced action, are the cheap laughs of design. They convert short-term and destroy trust. PFD is the expensive laugh: genuine emotional resonance that compounds.

Relationship to Value Sensitive Design

PFD’s ethical framework sits within the HCI tradition of Value Sensitive Design (Friedman, Kahn & Borning, 2002). Knobel & Bowker (2011) extended this to ask how values get embedded into designed systems, not just studied after the fact, but actively inscribed through design decisions that privilege certain stakeholders, workflows, and outcomes over others.

PFD’s perception layer is exactly such an inscription point: every visual hierarchy, every default, every activation point encodes a judgment about whose attention matters and what they should do with it. Where VSD asks broad questions about which values a system embeds, PFD’s ethical mandate is narrower and operational: does this specific design bring perception closer to reality, for this user, in this context?

What PFD Doesn’t See

Every framework has blind spots: things it makes visible and things it occludes. Knobel’s concept of ontic occlusion (how one representation of reality blocks another from being seen) applies to PFD itself. Naming these isn’t a weakness. It’s a quality criterion.

Structural power dynamics. PFD’s ethical test says “align perception with reality.” But it doesn’t interrogate who decides what “reality” is. When SSH’s product page undersold a good product, the “reality” was clear: the product was better than the perception. But what about a company whose product is mediocre and whose “reality” is genuinely ambiguous? The alignment test assumes a knowable ground truth. In practice, “reality” is contested, and the designer choosing which reality to align with is itself a power move that PFD doesn’t surface.

Individual vs. collective. PFD models one user’s cognitive journey through the mental waterfall. It doesn’t model what happens when thousands of users are conducted through the same waterfall simultaneously. Individual perception optimization can produce collective harms: engagement loops, attention extraction at scale, homogenization of choice. PFD’s unit of analysis is the user. The effects that emerge at population scale are outside its frame.

The fluency trap. Processing fluency is L2’s core mechanism: “if it’s easy to process, it feels true.” That mechanism is value-neutral. It works for honest brands and dishonest ones. A fluent lie is more persuasive than a disfluent truth. PFD’s ethical tests catch deliberate manipulation, but they don’t address the structural problem: by optimizing for fluency, PFD makes it harder for users to engage the critical evaluation that would catch deception. The very thing that makes PFD effective (reducing prediction error, keeping users in autopilot through the desired path) is the same thing that can prevent them from stopping to question. This is ontic occlusion baked into the mechanism.

The designer’s own perception biases. PFD’s diagnostic starts with “Feel”: the designer’s pre-verbal emotional response as the primary instrument. But that instrument is calibrated by the designer’s own culture, class, neurotype, and aesthetic history. I name my neurodivergence as an advantage (heightened sensitivity to friction). But heightened sensitivity to friction could also mean under-sensitivity to other things: social warmth cues, cultural signifiers outside my experience, accessibility needs I don’t personally feel. The diagnostic instrument has its own biases, and PFD doesn’t yet have a protocol for calibrating them.

Limitations & ongoing research: PFD’s evidence base draws on strong lab research (processing fluency, 50ms impression formation, cognitive load theory) and practitioner observation across 15 years. The five-layer synthesis itself is practitioner-derived, not independently validated as a complete system. Cultural generalizability beyond Western markets remains an open question. I state this because intellectual honesty is a stronger authority signal than overclaiming.

The Results

Simply Smart Home

The problem: Template site. No unified branding. Feature-focused marketing. Premium price point with a presentation that said “knockoff.”

Every layer was failing: site didn’t answer basic questions (L0, cognitive load), no faces or warmth (L1), template at premium price (L2), feature specs instead of emotional connection (L3), homepage structured around company, not user goals (L4).

What I did: Unified brand system. Emotional messaging: “stay connected, even when you’re apart.” Hero sections with faces. Consistent collateral across every touchpoint. 4-year systematic brand infrastructure build.

  • Revenue 3x within a year
  • Disney licensing, Costco pallets, Walmart shelf space
  • Competitor stole the tagline two years later

iO Theater

“Doesn’t look like a real theater.” Full redesign: dramatic photography, clear event navigation, integrated payment. Online ticket sales 50% → 75%. Same shows. Different perception.

VacuumSealersUnlimited

Embedded design partnership, 5+ years. Revenue 4x running on infrastructure I built and maintain. The long game. Perception compounds.

What’s Teachable

3 Hours with Business Owners

Hour 1: Business Design. Your business IS a design. Feedback loops between your vertical, users, and staff. I’d give them the tools to see their business as a system.

Hour 2: People as Research. Your staff and clients are your greatest resources. Design problem-solving is really just a lens of empathy. “For whom is this?” gives you directionality that frameworks can’t.

Hour 3: Tactics. The 5-Minute Perception Audit. The dependency stack. The red flags. Application, communication, process.

The Teachability Line

The lens (empathy, for-whom, design-as-problem-solving) is fully teachable. The pattern library (hundreds of projects of instant recognition) takes years. The sensitivity (neurodivergent tuning fork for bad design) may not be teachable. But the diagnostic process (Feel → Unpack → Diagnose → Prescribe) can be taught as a substitute.

The Improv Rules

My design practice runs on principles I learned from improv classes. These aren’t metaphors. They’re operational tools.

Improv RuleDesign Application
Yes, andAccept what exists and build on it. Brick by brick, not castle at a time.
(Listen) × (Act + React)If you’re not listening, you multiply everything by zero. Same for users.
“If this is true, what else is true?”One insight cascades. If this is the brand voice here, it should be everywhere.
“What comes before / after?”User journey thinking. What preceded this page? What follows?
“It’s not for you, it’s for them”Protect the design from stakeholder preferences.
Rule of eventually“That’s horri-bly... good!” Subvert expectations for engagement.
Hold topics and returnProgressive disclosure. Introduce, develop elsewhere, callback at conversion.
“Familiar to you, novel to others”Design for their first encounter, not your hundredth.

The Dual Foundation

(“Foundation” here means evidentiary base: two non-overlapping citation pools. Not the L0 layer described above.)

PFD’s original 55 citations come from perception psychology: cognitive load theory, processing fluency, decision architecture, predictive processing. The Inverse Inflection research (2026) added 22 citations from ADHD cognitive neuroscience. Zero overlap between the two bodies. The v3.6 international expansion (April 2026) added 17 further citations across German, Austrian, Swiss, Japanese, Chinese, French, Dutch, Israeli, and Scandinavian research traditions. ~100 citations total, converging on the same design principles from distinct angles.

The Connection

Why do two separate literatures point to the same principles? Because they describe the same cognitive architecture from different angles. Perception psychology measures how the typical brain processes input: working memory limits, first-impression timing, fluency effects. ADHD neuroscience measures what happens when those same systems operate under tighter constraints: lower working memory ceiling, higher friction sensitivity, sharper temporal discounting. Same architecture, different parameters.

The five PFD layers map cleanly to ADHD neuroscience findings. Every layer has independent support from the ADHD literature, and the support is strongest at the layers that matter most: cognitive load (L0) and Processing Fluency (L2).

PFD-ADHD Layer Mapping

PFD LayerADHD Neuroscience SupportEvidence TierKey Citations
Cognitive Load (L0)WM deficits d=0.69–0.74. Context switching costs approach an hour for ADHD developers. Low perceptual load specifically benefits ADHD-symptomatic developers.Tier 1–2 (Strong)Meta-analyses; ICSE 2024; JetBrains 2023
L1Reduced latent inhibition means more sensory intake: a badly designed first impression hits harder. Wall of Awful: bad first impressions deposit emotional barriers. Rejection Sensitive Dysphoria amplifies mild negative affect into disproportionate avoidance.Tier 3 (Indirect)Carson et al. 2003; Mahan 2020; Dodson
L2Low perceptual load benefits ADHD developers. Flat associative hierarchies amplify pattern-break sensitivity. Neurodivergent users perceive LLM outputs as “overly neurotypical”: what is fluent for NT may be disfluent for ND.Tier 1–2 (Strongest extension)JetBrains 2023; Mednick 1962; Carik et al. 2024
L3The METR perception gap: developers thought they were 20% faster with AI but were 19% slower. Overconfidence amplification. The 9% competence penalty for visible AI assistance.Tier 2–3METR 2025
L4Barkley’s “point of performance”: self-regulation failures occur in the moment, not in planning. “Now and Not Now” temporal discounting means future benefits have near-zero motivational weight.Tier 1Barkley 1997

The Curb-Cut Argument

PFD was not designed for ADHD. But because ADHD involves the most constrained version of the cognitive architecture PFD designs for (lower working memory ceiling, higher friction sensitivity, sharper temporal discounting), designs that work for ADHD users satisfy the tightest constraints. Everyone benefits.

The principle is the same one behind curb cuts, closed captions, and automatic doors. Curb cuts were designed for wheelchair users but turned out to help parents with strollers, travelers with luggage, delivery workers with carts. Designs that reduce cognitive load for ADHD users reduce cognitive load for all users. The constrained case reveals what the general case tolerates but shouldn’t.

Cowan (2010) established that working memory holds 3–5 chunks. ADHD brains hit the ceiling of that range faster and harder, making violations visible sooner. A page that overloads an ADHD user’s working memory is overloading everyone’s. The ADHD user just notices first.

The research supports this directly. Kasatskii et al. (JetBrains, 2023) found that high perceptual load degrades ADHD-symptomatic developers’ performance faster than neurotypical developers’. Mark et al. (2008) measured context-switching recovery at 23 minutes for typical workers. For ADHD developers, Mark et al. (2024) found the cost approaches an hour. Same friction, amplified. Design for the amplified case and the typical case improves too.

The Diagnostic Instrument

I’m autistic and have severe combined-type ADHD. The autism means I learned social cognition analytically: theory, practice, application loops. I can see and articulate unconscious patterns that most designers apply intuitively but can’t explain. The ADHD means cognitive load and friction are literally painful for me. What most designers theorize about, I feel.

That sensitivity is a diagnostic instrument. When I arrive at a page, the emotional response fires before language catches up. System 1 before System 2. My ADHD sensitivity means I feel cognitive load and disfluency more acutely than most users. The feeling IS the diagnostic. The autism means I can then unpack exactly why, tracing the violation back to the systematic rule that’s broken.

I’ve come to think of this as a curb-cut effect applied to cognition. The cognitive constraints I designed around (limited working memory, sensitivity to disfluency, impatience with unnecessary friction) are constraints every user has. Mine are just louder.

The L0 Refinement

Dodson’s Interest-Based Nervous System (IBNS) proposes that ADHD attention is regulated by interest, not importance. The PINCH model identifies five drivers that override cognitive load ceilings: Passion, Interest, Novelty, Competition, Humor.

This means L0’s “reduce load” is incomplete. For ADHD users (and partially for everyone under low motivation), the correct formulation is “reduce load OR increase interest.” This is not a contradiction of L0. It is the specification of when load reduction alone is insufficient. When interest is absent, no amount of cognitive streamlining produces engagement.

On the dual citation base: The 55 perception citations and 22 ADHD citations share zero duplicate entries. Bridging them into a single actionable design framework is this methodology’s contribution. The underlying cognitive architecture is expected to generalize. The ADHD-specific parameters identify where the architecture operates under the tightest constraints.

Cognitive Constraint Design

The Inverse Inflection research surfaced a unifying meta-framework: Cognitive Constraint Design. The principle: cognitive constraints are design material, not obstacles to work around.

Most design approaches treat cognitive limitations as problems. Reduce choices because people can’t handle too many. Simplify because users are overwhelmed. The framing is defensive: protect the user from their own brain.

CCD inverts this. The constraint is the design input. Working memory holds 3–5 chunks? That’s not a limitation to accommodate. That’s a specification to build for. The constraint tells you what to build, not what to avoid.

DomainConstraintDesign Response
PFD (interfaces)Limited processing capacityReduce cognitive load, optimize fluency
ADHD-AI workflowsExecutive dysfunctionExternalize structure at point of performance
Spatial workflow toolsWorking memory limitsSpatial triggers externalize planning
AI writing assistantsCommunication pattern driftCorrections-as-training preserves authentic voice
Session memoryContext loss between sessionsHierarchical persistence survives crashes

Same principle, different surfaces. PFD applies Cognitive Constraint Design to interfaces. The five layers (L0 through L4) are a systematic treatment of perceptual and cognitive constraints as design specifications. Each layer names a constraint (bandwidth, first impression, fluency, bias, decision load) and prescribes a design response grounded in how that constraint actually operates.

CCD is the genus. PFD is a species. This distinction matters because it explains why PFD’s approach transfers across domains. The bouncer story, the improv rules, the diagnostic protocol: they all work because they’re applications of the same principle. Observe the constraint, design for it, test whether the design honors it.

The methodology page focuses on PFD as a design tool. This reference documents the broader intellectual framework. CCD is for researchers and practitioners who want to understand why PFD’s approach to interfaces also explains how to build session memory systems, voice preservation tools, and spatial orchestration canvases. The connecting thread is always: what constraint does the user face, and how do you design the system so the constraint becomes productive rather than destructive?

Practitioner Cognition

This section explains HOW PFD’s diagnostic produces insights that standard UX heuristics miss. It is not a different checklist. It is a different cognitive operation.

Bisociation

Koestler (1964) coined bisociation for the creative act of connecting two independent frames that don’t normally intersect. Standard association operates within a single frame. Bisociation jumps between frames to produce something neither frame contains alone.

PFD’s cross-domain structure (psychology + design + improv + neuroscience) is bisociative by architecture. The derivation protocol forces collision between constraint frames. When you work Layer 0 (cognitive load theory) and then Layer 3 (perception bias), the accumulated requirements R1 through R3 create a constraint space that neither field produces independently. The non-obvious solution lives in that collision.

Janusian Thinking

Rothenberg (1971) described Janusian thinking: the simultaneous conception of opposites as a creative process. PFD holds “reduce friction” AND “activate attention” simultaneously. The core thesis is literally Janusian: users don’t think, until you make them. Don’t make them think (L0). Make them think (activation point). Both are true. The resolution is not choosing between them but designing for the transition.

Deliberate and Spontaneous Modes

Dietrich (2004) mapped four creativity types to prefrontal function: deliberate/spontaneous crossed with cognitive/emotional. Creative insight involves both deliberate (prefrontal-guided, analytical) and spontaneous (defocused, associative) processing.

The PFD diagnostic protocol alternates between these modes. The Feel step is spontaneous: pre-verbal, emotional, System 1. The Unpack and Diagnose steps are deliberate: analytical, systematic, tracing violations back to specific layers. The protocol is not “first feel, then think.” It is “alternate between modes, and use each mode for what it does best.”

This is why the Feel step cannot be skipped. A pure checklist audit (deliberate mode only) catches surface violations but misses the emotional gestalt. A pure intuitive response (spontaneous mode only) identifies that something is wrong but cannot diagnose which layer is failing. PFD needs both.

Flat Associative Hierarchies

Mednick (1962) showed that creative people have flatter associative hierarchies: more distant conceptual connections are accessible. Where most people associate “table” with “chair” and stop, creative thinkers reach “table” to “periodic table” to “elemental structure” to “information architecture.”

Carson, Peterson & Higgins (2003) found eminent creative achievers were 7x more likely to have low latent inhibition: their brains let in more raw sensory and conceptual material. ADHD may produce flatter hierarchies via reduced latent inhibition through related mechanisms. The connection is indirect but convergent: ADHD involves reduced inhibitory filtering, creativity correlates with reduced inhibitory filtering, and PFD’s cross-domain derivation protocol rewards exactly the kind of distant associations that flat hierarchies produce.

The Communication Hierarchy

Three modalities for changing perception, ranked by power: Music > Humor > Stories.

Music communicates emotion and bypasses rational processing entirely (Koelsch, 2014; Juslin & Västfjäll, 2008; Huron, 2006). In design: typography rhythm, spacing cadence, visual flow, animation timing, scroll momentum. Humor changes how people feel about a thing by pointing out absurdity and giving permission to think without rigidity. Stories produce narrative transportation (Green & Brock, 2000; van Laer et al., 2014): absorption into a narrative produces belief changes.

The practitioner who understands all three modalities has more tools for changing perception than the practitioner who only understands information architecture. This is why PFD draws from improv, music theory, and narrative psychology alongside cognitive science. The communication hierarchy is not a bonus chapter. It is how the methodology gets delivered.

Evidence Tiers

Not all claims carry equal weight. This classification system governs what PFD states confidently, what it attributes carefully, what it frames as synthesis, and what it avoids. Every public claim should be traceable to one of these tiers.

Tier 1: Established Science

Well-replicated findings from peer-reviewed research, confirmed by multiple independent labs. PFD states these as facts.

FindingSource(s)Why Established
Working memory deficits in ADHD (d=0.69–0.74)Meta-analyses, Barkley modelDecades of convergent evidence, hundreds of studies
ADHD associated with stronger divergent thinkingWhite & Shah 2006/2011, Boot et al.Multiple independent labs, peer-reviewed, conceptual replications
DMN hyperconnectivity in ADHDCastellanos et al.One of the most replicated neuroimaging findings
Creative thinking requires DMN + ECN + Salience cooperationBeaty et al. (Three-Network Model)Multiple fMRI studies, established in cognitive neuroscience
7x rate of low latent inhibition in creative achieversCarson, Peterson & Higgins 2003JPSP (top-tier), widely cited
Flat associative hierarchies produce remote associationsMednick 196260+ years of citation, standard creativity psychology
DRD4/7R: opposite fitness outcomes by environmentEisenberg 2008Well-designed field study, conceptually replicated
Cognition extends into reliably accessed external toolsClark & Chalmers 1998One of the most cited papers in philosophy of mind
Context switching costs significant recovery timeGloria Mark (multiple), ICSE 2024Well-established productivity research
Self-regulation, not attention, is the core ADHD deficitBarkley modelDominant ADHD theoretical framework, decades of research

Public framing: “Research shows...” or state directly without hedging.

Tier 2: Strong Single Studies

Solid methodology, credible venues, but not yet independently replicated. PFD cites the specific source.

FindingSourceLimitation
ADHD is the only positive predictor of divergent thinking in engineering studentsTaylor 2020 (n=60)Single university, single lab
Emotional prompting improves LLM output ~11%Li et al. 2023 (EmotionPrompt)ArXiv preprint; 11% human eval is more defensible than 115% peak
25% higher AI satisfaction among neurodivergent workersUK Dept. for Business and TradeMeasures satisfaction, not performance
ND users perceive LLM outputs as “overly neurotypical”Carik et al. 2024Peer-reviewed (ACM), specific study context
Low perceptual load benefits ADHD-symptomatic developersJetBrains 2023Industry research, single study

Public framing: “A study by [X] found...” or “Research from [X] suggests...” Always name the source.

Tier 3: Reasonable Inferences

Logical connections drawn between established findings. The foundations are real. The bridge is untested. PFD presents these as synthesis, not established research.

InferenceEvidence BaseUntested Bridge
ADHD communication style is natural emotional promptingEmotionPrompt + IBNS (Dodson)Nobody has tested whether ADHD individuals produce better prompts
ADHD-AI error patterns are complementaryPNAS 2022 Bayesian frameworkZero studies measuring actual ADHD-AI error correlation
Centaur model generalizes beyond chessKasparov/ZackS 2005Chess has perfect information. Software development does not.
PFD disproportionately benefits ADHD usersWM deficits + L0/L2 researchPFD was not designed for ADHD. Connection is real but unmeasured.

Public framing: “Our synthesis suggests...” or “The evidence points toward, though this hasn’t been directly tested...”

Tier 4: Speculative or Corporate

Claims that lack methodology, independent verification, or peer review. PFD either avoids these or flags them explicitly as unverified.

ClaimSourceProblem
JPMorgan ND program: 90–140% higher productivityCorporate self-reportNo sample size, no comparison group, no methodology published
Deloitte “30% more productive” ND teamsIndustry reportNo peer review, no methodological detail
The “100+ sources” claim from popular ADHD-AI sitesSelf-reportCounts blog posts alongside peer-reviewed papers

Public framing: Do not use, or explicitly state: “Corporate reports claim...” Never present as research.

The governing principle: Credibility compounds. Every well-hedged claim builds trust. One overstated claim can undo it. The advantage of reading the primary literature is being the person who gets it right, not the person who gets caught overclaiming.

Accumulated Learnings

18 learnings from applying PFD across interfaces, spatial canvases, AI systems, pricing, ethics, and communication. These are not theoretical extensions. They emerged from practice and changed how the framework is applied.

#LearningDomainPFD Layer(s)
1Generative vs. EvaluativeCore methodAll
2Access vs. Signal MechanismsMarketing / UXL1–L4
3Pre-Send vs. Post-ResponseAI / CommunicationL1, L4
4Workspace vs. Product SeparationProduct designAll
5Recursive ValidationMeta-methodologyAll
6Infrastructure ≠ ActivationProduct designL1, L4
7Consistent Gap IdentificationCore methodL2
8Epistemic Asymmetry as L4 ProblemAI systemsL3, L4
9Temporal/Session ContinuitySpatial toolsGap (candidate L6)
10Constraints Are DistributionsAccessibilityL0
11Visual Channel Utilization AuditData viz / CanvasL2
12Route vs. Survey KnowledgeSpatial toolsL2, L4
13Keyboard Density Is L2, Not L4Expert workspacesL2
14Ethics Are Operational, Not StructuralEthicsEthics
15Experiential Self-ContradictionHero / landingL2, L3, L4
16Near-Miss Color AsymmetryDesign systemsL2
17Iterative Regression Is Deeper VisibilityForge / analysisAll
18Backend Mechanics as Frontend TiersPricing / SaaSL1–L4
#1: Generative vs. Evaluative

When analysts were given solutions first and asked to evaluate through PFD, they missed requirements entirely. When forced to derive bottom-up, the gaps emerged. PFD only generates non-obvious solutions when used bottom-up with constraint accumulation. Top-down evaluation refines but cannot escape the originating frame of the solution being evaluated. The derivation protocol exists because evaluation is not generation.

#2: Access vs. Signal Mechanisms

PFD’s layers distinguish between “how do I access content?” and “how does the user know where content exists?” L1/L2 handle signaling. L3/L4 handle access. Conflating them produces solutions that solve access without solving discovery. A CTA that’s easy to click (access) is different from a page that makes the CTA feel like the obvious next step (signal).

#3: Pre-Send vs. Post-Response

Standard UX says “show your sources after the response” (attribution). PFD derivation says “show what the AI will see before the user sends” (legibility). Attribution is an L3 debugging tool. Pre-send legibility is an L1+L4 schema and trust tool. What you show BEFORE the ask is higher-leverage than what you show AFTER. This generalizes to communication and copy: setup and framing outweigh explanation and justification.

#4: Workspace vs. Product Separation

When analyzing a specific workspace through PFD, recommendations naturally split between workspace-level fixes (rearrange, rename, add zones) and product-level features the tool should provide. The separation between “user organizes content” and “tool provides perceptual infrastructure” is a critical design boundary. PFD should be applied at both levels, but the intervention target must be identified correctly.

#5: Recursive Validation

When PFD is used to analyze a workspace that is ABOUT PFD, three levels operate simultaneously: PFD as framework, PFD applied to the workspace, the workspace as proof-of-concept for the tool. When all three levels are legible, the workspace becomes its own best case study. This recursive structure is a strength, not a liability, but only when intentionally designed.

#6: Infrastructure ≠ Activation

Infrastructure that exists without activation is worse than no infrastructure. It creates the illusion of completeness while users experience the original problem. PFD correctly distinguishes between “the mechanism exists” and “the user can perceive and use the mechanism.” This is a Layer 4 violation masquerading as a Layer 1 problem. Having the right product without the right perception layer produces the same failure mode.

#7: Consistent Gap Identification

When the same requirement (R2: pre-attentive content signaling) surfaced independently across four different analytical frames, it was not a preference. It was a genuine perceptual constraint. When running multiple derivations on the same problem, note which requirements appear consistently. Those are load-bearing.

#8: Epistemic Asymmetry as L4 Problem

When an AI system gathers context via invisible mechanisms (graph traversal, priority weighting, depth limits), the resulting epistemic asymmetry (AI knows more than user about its own context) is a Layer 3 violation. Users’ System 1 response to opaque behavior is distrust. The intervention is making the mechanism visible pre-send, not explaining it post-response. This generalizes to any expert/client asymmetry: show your reasoning process, not just your conclusions.

#9: Temporal/Session Continuity

PFD addresses what the user perceives NOW but not how they re-enter a workspace after 48 hours. Six of nine independent analytical personas identified this gap. A workspace visited after two days is a different perceptual object than the same workspace visited five minutes ago. PFD’s five layers assume continuous interaction and need supplementation for interrupted/resumed sessions. Candidate for formal L6 after testing.

#10: Constraints Are Distributions, Not Constants

PFD states working memory holds “~3–5 chunks” as a hard constraint. This is a population mean. Neurodivergent users, users under cognitive load, users with fatigue all shift the distribution. The constraint is real but the threshold varies. PFD’s constraints define the shape of the curve, not a fixed cutoff. Good design provides calibration mechanisms for users at the tails. This strengthens PFD by acknowledging that perception science describes populations, not individuals.

#11: Visual Channel Utilization Audit

Bertin’s 7 visual variables (position, size, lightness, texture, color, orientation, shape) provide a systematic audit checklist for design. Most UIs use 2–3 channels. Under-utilized channels mean information that COULD be signaled pre-attentively is hidden behind interaction. The derivation protocol should include a channel audit: “How many visual channels does this design use? Which are available but unused?”

#12: Route vs. Survey Knowledge

Siegel & White’s spatial knowledge theory: users develop route knowledge (sequential, egocentric, fragile) before survey knowledge (allocentric, configurational, robust). Canvas tools should support both stages. New users need route affordances (guided paths, breadcrumbs, landmarks). Experienced users need survey affordances (minimap, overview, configurable views). “Appropriate density” depends on which knowledge type the user has developed.

#13: Keyboard Density Is L2, Not L4

When multiple analysts flagged “too many keyboard shortcuts” in an expert tool, the initial diagnosis was L4 choice overload. After applying the Scheibehenne (2010) context qualifier (choice overload is near-zero in expert workspaces), the problem reframed to L2: users cannot predict what a keypress will do across cognitive modes. The fix is a keyboard mode indicator, not fewer shortcuts. Don’t reduce expert affordances. Signal which mode they operate in.

#14: Ethics Are Operational, Not Structural

PFD’s three ethical tests (Alignment, Sincerity, Golden Rule) are operational ethics: executable by one designer in a single review session. They don’t constitute structural ethics in the Value Sensitive Design sense (systematic processes that produce good outcomes regardless of individual practitioner values). Long-term client relationship accountability enforces alignment better than any framework clause could. Operational tests handle decision-making. Structural incentives handle accountability. When PFD scales beyond solo practice, structural ethics needs to be baked into the process, not just carried by the practitioner.

#15: Experiential Self-Contradiction

When a hero section’s thesis claims speed and simplicity (“Your customers decide in milliseconds. I design for that.”) but the hero itself takes 10+ seconds to parse due to competing visual elements, it disproves its own claim experientially. This is an L3/L4 failure, not an L1 failure: the visitor’s lived experience contradicts the promise being made. The resolution preserves complexity in the background (competence signal) while simplifying the foreground (single message, single CTA). A diamond on crushed velvet. The complexity is the bed, not the message.

#16: Near-Miss Color Asymmetry

Human color perception is non-additive (Bujack et al., 2022, PNAS). A brand blue that’s 3% off-target imposes disproportionate processing cost compared to a completely different color, because small deviations sit in the steep, high-sensitivity zone of the perceptual metric where prediction error is maximal. Near-miss color deviations are the highest-cost L2 violations, more disruptive than using a completely different color. Design-token linting should weight near-misses as more severe than far-misses. Color structure is hardware (universal geometry). Color meaning is software (cultural, context-dependent).

#17: Iterative Regression Is Deeper Visibility

When iterative PFD analysis shows a score DROP around rounds 3–6, the cause is almost never that fixes made things worse. Resolving surface-level issues (nav overload, font proliferation) unmasks deeper problems that were previously invisible. The dominant signal was drowning out subtler violations. The regression reflects updated understanding, not degraded quality. Three terminal states for iterative analysis: converged (done), impasse (approach won’t converge with current premise), abandoned (user stopped). The impasse state is the highest-value signal most iterative tools lack.

#18: Backend Mechanics as Frontend Tiers

When internal system state (subscription tiers, feature flags, scarcity counters) is surfaced as visible UI complexity, it converts a backend concern into a frontend cognitive load violation. Founding member pricing with counter badges, tier comparisons, and crossed-out prices triggers defensive pattern-matching against growth hacks. The user’s 50ms L1 read is “this is a marketing tactic,” which directly contradicts the intended signal of gratitude. The system needs tiers. The user needs a yes/no decision. These are different design surfaces at different layers. A rare 4-layer simultaneous violation from a single UI element.

Framework Comparison

PFD exists in a landscape of design frameworks. Understanding where it overlaps and where it differs helps practitioners decide when to use it and when something else serves better.

FrameworkWhat It DoesWhere PFD Differs
Laws of UX (Yablonski)Collection of psychological principles for designDescriptive catalog, not generative protocol. No dependency stack. No derivation method. Tells you what to know, not what to do first.
Fogg Behavior ModelBehavior = Motivation + Ability + PromptSingle-event model. PFD addresses the perceptual pipeline BEFORE behavior triggers. Fogg assumes the user is already attending.
Don’t Make Me Think (Krug)Reduce friction, make things obviousPFD’s Cognitive Load layer (L0). Krug covers L0 deliberately and well. PFD adds four more layers on top, plus a generative protocol and ethics framework.
Emotional Design (Norman)Visceral, behavioral, reflective processing levelsComplementary framing. Norman’s visceral ≈ PFD L1. PFD provides a dependency stack and generative protocol Norman doesn’t.
Nudge Theory (Thaler/Sunstein)Choice architecture shapes behaviorPFD L4. But Maier et al. (2022) shows nudge effects near-zero after correcting for publication bias. PFD contextualizes when choice architecture works and when it doesn’t.
Cognitive Constraint DesignMeta-framework: constraints as design materialPFD is an application of CCD to interface design. CCD is the genus. PFD is a species.

None of these frameworks are wrong. They are incomplete in different ways. Laws of UX catalogs principles without a generative method. Fogg models behavior but not perception. Krug covers L0 without the other four layers. Norman identifies processing levels without prescribing an order. Nudge Theory applies L4 without the dependency stack that determines when L4 interventions actually work.

PFD’s contribution is the dependency stack: five layers in a specific order, where each layer depends on the layer below it. Fix them in order, because upstream failures degrade everything downstream. This ordering principle is what separates PFD from frameworks that treat design principles as independent, parallel concerns. They are not parallel. They are stacked.

Make Me Think: Chapter Mapping

“Make Me Think: Perception-First Design for the Post-Usability Era” is a 12-part series that develops the framework through narrative, case studies, and citations. Each chapter maps to specific PFD layers and concepts.

Ch.TitlePFD Layer(s)What It Adds
1The BouncerOrigin (all layers)The nightclub story: how a security job became a design methodology
2They’re Already Not ThinkingCore thesisPredictive processing. Why most problems happen before conscious evaluation.
3The Tuning ForkDual foundationAutism as analytical social cognition. ADHD as friction sensitivity.
4Cognitive LoadCognitive Load (L0)Miller vs. Cowan, Sweller’s CLT, progressive disclosure, smart defaults.
5The 50-Millisecond VerdictL1First impressions, the iO Theater case study, visual-price coherence.
6The Feeling of TruthL2Processing fluency, the McGurk Effect, consistency compounding trust.
7The GapL3Perception bias. System 1/System 2 gap. Why a tagline change preceded revenue tripling.
8The TrailL4Decision architecture, path reduction, smart defaults.
9Feel, Unpack, Diagnose, PrescribeDiagnostic protocolEmotional primacy, embodied cognition, why feelings arrive before language.
10Music, Humor, StoriesCommunication hierarchyThree tools for changing perception, ranked by power.
11The OathEthics frameworkThree operational tests, structural ethics, cognitive deskilling.
12What I Don’t Know YetLimitationsWEIRD evidence base, the fluency trap, ontic occlusion, living methodology.

The series is available at aurochs.agency/writing/make-me-think/. Each chapter stands alone. Read them in any order, though the dependency stack (Chapters 4–8) makes more sense sequentially.

Open Questions

These are genuinely unresolved. A living methodology names its edges.

Cultural generalizability (partially resolved, v3.6). The evidence base is predominantly WEIRD (Western, Educated, Industrialized, Rich, Democratic). The underlying cognitive architecture is expected to generalize: predictive processing, working memory limits, first-impression formation. But the parameters are culturally shaped: which visual patterns signal trust, what counts as “too much” cognitive load, how fluency maps to credibility.

Recent cross-cultural research resolves part of this. Masuda & Nisbett (2001) showed East Asian observers attend to background and relational context while Western observers focus on foreground and focal objects. Chua, Boland & Nisbett (2005) replicated this with eye-tracking: different scanning patterns on the same scenes. Goh et al. (2010) found cultural differences in face-processing cortex activation. Ji, Nisbett & Su (2001) showed cultural differences in prediction direction (linear vs non-linear reasoning about change).

What this resolves: the architecture is universal; the calibration is cultural. PFD’s 5-layer stack holds across populations. But L1 parameters (what triggers trust) and L3 parameters (what predictive priors load automatically) must be validated against target-market users before cross-regional rollout.

What it doesn’t resolve: L2 processing-fluency effects across non-Latin typographic traditions, L4 decision-architecture priors across markets with different default authority structures, and the relative dependency-stack ordering in high-context vs low-context cultures. These stay open.

PFD meta-rule (cross-cutting): Layer requirements are universal. Layer calibration is cultural. Validate L1 and L3 against target-market users before rolling a design across regions. PFD cannot claim cross-cultural universality without this step.

Measurement. PFD lacks a validated measurement instrument. The 5-Minute Perception Audit is a practitioner heuristic: useful, fast, and directionally correct, but without inter-rater reliability data. A diagnostic tool where two PFD practitioners independently audit the same site and converge on the same layer diagnosis would strengthen the methodology from “expert intuition with a framework” to “teachable, repeatable diagnostic.” This is the gap between a clinical art and a clinical science.

Temporal dynamics. PFD describes the first encounter well: activation, first impression, mental waterfall. But user relationships with products evolve. At visit 100, the mental waterfall is different. Prediction errors that activated on first visit are now expected. Fluency that felt trustworthy can become invisible (users stop noticing). Does the 5-layer stack apply the same way to retention and long-term engagement as it does to acquisition? The current framework is weighted toward first contact.

Accessibility as justice vs. fluency. PFD frames accessibility through L0 (cognitive load reduction) and L2 (processing fluency). This produces good accessible design in practice. But disability justice scholars argue accessibility is a rights issue, not an optimization target. Framing it through “fluency” risks treating accessibility as something you optimize when the ROI justifies it, rather than something you guarantee because people deserve it. The framework’s utilitarian lens may not fully serve this.

AI-generated design. As AI tools generate more design artifacts, PFD’s emphasis on embodied sensitivity (the “neurodivergent tuning fork,” the pre-verbal Feel step) becomes harder to operationalize at scale. Can PFD’s diagnostic be automated? Can an AI run the Feel step? If not, the methodology has a ceiling on how far it scales beyond the individual practitioner. If yes, what’s lost when the diagnostic instrument is no longer a human body?

Appendix A: The Science

~100 citations across cognitive psychology, HCI, behavioral economics, design theory, STS, music cognition, narrative psychology, and ADHD cognitive neuroscience.

CitationYearFindingLayer
Heider & Simmel1944People attribute intentions/emotions to geometric shapesL1
Miller, G.A.1956“Magical Number Seven”: rhetorical, not literalL0
McGurk & MacDonald1976Visual lip movements integrate with auditory perceptionL2
Nisbett & Wilson1977People can’t accurately report why they made decisionsL3
Lakoff & Johnson1980Abstract thought structured by bodily metaphorsL2
Sweller1988Cognitive Load Theory (intrinsic vs. extraneous)L0
Johnson et al.1991Neonatal face preferenceL1
Sloboda1991Musical features trigger involuntary physical responsesComm.
Friestad & Wright1994Persuasion Knowledge Model: users object to insincerityEthics
Kurosu & Kashimura1995Aesthetic-usability effectL1
Salthouse1996Processing speed declines continuously from early adulthoodL0
Tractinsky1997Aesthetic-usability cross-cultural replicationL1
Sweller, van Merriënboer & Paas1998Formalized CLT taxonomyL0
Berdichevsky & Neuenschwander1999Golden Rule of persuasionEthics
Reber & Schwarz1999Processing fluency → perceived truthL2
Campbell & Kirmani2000Sincerity as function of perceived ulterior motivesEthics
Green & Brock2000Narrative transportation → belief changeComm.
Tractinsky, Katz & Ikar2000“What is Beautiful is Usable”L1
Cowan2001Working memory revised to ~4 chunksL0
Winkielman & Cacioppo2001Fluency generates positive affect without awarenessL2, L3
Friedman, Kahn & Borning2002Value Sensitive DesignEthics
Johnson & Goldstein2003“Do defaults save lives?”L4
Norman2004Emotional Design (visceral, behavioral, reflective)Bonus
Reber, Schwarz & Winkielman2004Processing fluency → aesthetic pleasureL2
Hershler & Hochstein2005Adult face “pop-out” in visual searchL1
Huron2006ITPRA theory: prediction/reaction are pre-consciousComm.
Lindgaard et al.200650ms visual appeal judgmentsL1
Tractinsky, Cokhavi et al.2006Consistent aesthetic judgments at brief exposuresL1
Gray, Gray & Wegner2007Two dimensions of mind perception: Agency + ExperienceL1
Juslin & Västfjäll2008Six mechanisms for music-evoked emotionComm.
Song & Schwarz2008Fluency and detection of misleading questionsL2
Thaler & Sunstein2008Choice architecture / Nudge theoryL4
Alter & Oppenheimer2009Fluency generalizes across truth, confidence, trustL2
Brignull2010Dark patterns taxonomyEthics
Cowan2010“The Magical Mystery Four”: WM refined to 3–5 chunksL0
Dechêne et al.2010Illusory truth effect meta-analysis (51 studies)L2
Friston2010Free-energy principleCore
Henrich, Heine & Norenzayan2010WEIRD caveat for benchmark generalizabilityNote
Scheibehenne, Greifeneder & Todd2010Choice overload effect near-zero in many conditionsL4
Kahneman2011Thinking, Fast and Slow: System 1 / System 2Core
Knobel & Bowker2011“Values in Design”: values inscribed through design decisions (CACM)Ethics
Owsley2011Age-related declines in visual processingL0
Spence2011Cross-modal correspondencesL2
Tuch et al.2012Replication of 50ms visual appealL1
Clark2013Predictive processing: brains match input to top-down predictionsCore
Hohwy2013The Predictive MindCore
Reinecke et al.2013Visual evaluation at 17ms (Google Research)L1
Koelsch2014Music modulates subcortical emotion structures (Nature Reviews Neuroscience)Comm.
van Laer et al.2014Narrative transportation meta-analysis (132 effect sizes)Comm.
Zak2015Narrative arcs trigger cortisol + oxytocin (mechanism undemonstrated)Comm.
Braddock & Dillard2016Narrative persuasion meta-analysisComm.
Gray et al.2018Dark patterns taxonomy: nagging, obstruction, sneakingEthics
Sweller, van Merriënboer & Paas2019CLT reconceptualizedL0
Mierop et al.2020Intranasal oxytocin: ~90% false positive rateNote
Bujack et al.2022Perceptual color space is non-Riemannian (PNAS)L2
Maier et al.2022Nudge effects near-zero after publication bias correctionL4
v3.6 International Expansion (April 2026)
Wertheimer, M.1923Gestalt foundations — perception as organized wholesL1, L2
Metzger, W.1936 / 2006Comprehensive Gestalt empirical treatment (English trans. 2006)L1, L2
Mori, M. / MacDorman & Kageki1970 / 2012Uncanny valley — near-human-but-not-quite triggers eeriness before conscious evaluation (trans. 2012)L1
Nagamachi, M.1995Kansei engineering — systematic methodology mapping consumer affective response to product attributesSibling
Masuda & Nisbett2001Cultural attention differences — East Asian observers attend to background/relational; Western to foreground/focalL1, L3
Ji, Nisbett & Su2001Cultural differences in prediction direction (linear vs non-linear reasoning about change)L4
Leder, Belke, Oeberst & Augustin20045-stage aesthetic appreciation model (perception → implicit integration → explicit classification → cognitive mastery → evaluation)L1
Chua, Boland & Nisbett2005Eye-tracking replication of cultural attention differencesL3
Pessiglione et al.2007Subliminal incentive processing — reward cues below conscious threshold drive motor behaviorL1, L3
Hassin, Bargh, Engell & McCulloch2009Implicit working memory — WM capacity spent on unconscious processing, not just conscious attentionL0
Hertwig & Erev2009Description-experience gap — decisions from description underweight rare eventsL4
Goh, J.O.S. et al.2010Cultural differences in face-processing cortex activation (FFA)L1
Trope & Liberman2010Construal-level theory — psychological distance shapes concrete vs abstract linguistic processingL4
Forster, Leder & Ansorge2013Subjective felt fluency drives aesthetic liking; objective fluency alone insufficientL2
Hassin, R.R.2013“Yes It Can” — unconscious performs every high-level cognitive function previously assumed to require awarenessL3, Core
Seckler, Heinz, Forde, Tuch & Opwis2015Visual characteristics (color, typography, layout coherence) drive trust judgments directlyL3
Skov & Nadal2020Neuroaesthetics as hedonic valuation — neural substrates of aesthetic experienceL1, Sibling
ADHD Cognitive Neuroscience
Barkley, R.1997Self-regulation, not attention, is the core ADHD deficit. Executive function failures occur at the point of performance.L0, L4
Castellanos, F.X.2002ADHD involves altered default mode network dynamicsL0
Carson, Peterson & Higgins20037x rate of low latent inhibition in high-functioning creative achieversL2
Dodson, W.2005Interest-Based Nervous System: ADHD motivation responds to PINCH (Passion, Interest, Novelty, Competition, Hyperurgency)L0, L3
White, H. & Shah, P.2006ADHD associated with stronger divergent thinking on multiple measuresL2
White, H. & Shah, P.2011ADHD linked to higher real-world creative achievementL2
Kounios & Beeman2009Pre-problem brain state predicts whether solution arrives via insight or analysisL0
Beaty et al.2014Three-network model of creative cognition (default + executive + salience)L0
Mahan, B. (Wall of Awful)2020Emotional barrier accumulated from repeated executive function failuresL4
Taylor, J.2020Twice-exceptional individuals in professional settingsL0
Eisenberg et al.2008DRD4/7R allele associated with better nutritional status in nomadic Ariaal pastoralists but worse in settled — opposite fitness outcomes by environmentL2/L0
Koestler, A.1964Bisociation: the creative act of holding two independent, normally incompatible frames of reference simultaneouslyL0
Mednick, S.A.1962Flat associative hierarchies produce more remote associations; basis for the Remote Associates Test (RAT)L2
Dietrich, A.2004Four creativity types mapped to prefrontal function: deliberate/spontaneous x cognitive/emotionalL0
Clark, A. & Chalmers, D.1998Extended Mind Thesis: cognitive processes extend beyond brain/body when external tools reliably couple with internal cognitionCore
Rothenberg, A.1971Janusian thinking: actively conceiving two or more opposite concepts simultaneously as a creative processL0
Kasparov, G.2017Centaur model: weak human + machine + better process outperforms strong computer alone (Deep Thinking, PublicAffairs)L4
Mark, G. et al.2008Cost of interrupted work: average 23 minutes 15 seconds recovery time after interruption (CHI 2008)L0
Steyvers, M. et al.2022Bayesian human-AI complementarity: emerges when human and AI error patterns are sufficiently independent, even when AI outperforms humans individually (PNAS)L4
Li, C. et al.2023EmotionPrompt: emotional stimuli in prompts improve LLM output; 10.9% average improvement on human evaluation, up to 115% on edge-case benchmarks (AAAI 2024)L3
Kasatskii, V. et al. (JetBrains)2023Perceptual load in IDE environments differentially affects developers with ADHD symptoms; low load shows speed benefits that differ from neurotypical developersL2/L0
Carik, B. et al.2024Neurodivergent users perceive LLM output as “overly neurotypical”; community-driven workarounds documented (ACM PACMHCI / CSCW2, peer-reviewed)L2/L3
On provenance: Several PFD concepts were independently developed through embodied practice (improv, security work, 15 years of client work) before encountering the corresponding literature. The improv notebook (2012–2013) contains a practitioner articulation of the core predictive processing operation (“comparing mental representation with reality”) that predates Clark’s landmark 2013 paper.

On methodology: This document was formalized with AI assistance (Claude, Anthropic). The underlying methodology, case studies, and practitioner insights are the author’s own work. AI was used for structuring, identifying relevant literature, and iterative refinement. Citations were separately audited for accuracy. The framework’s ideas predate the AI assistance; the formalization does not.

Appendix B: Glossary

Perception barrierAn unconscious objection that prevents conversion. Not a product, price, or demand problem.
Fence-sitterSomeone with intent who doesn’t convert due to perception barriers.
Mental waterfallThe conducted sequence of thoughts and emotions from first impression through conversion.
Activation pointThe moment a user switches from autopilot to active engagement.
Listen Multiplier(Listen) × (Act + React). If the user isn’t listening, everything multiplies by zero.
Cognitive loadTotal mental effort. Intrinsic (task) + extraneous (design).
Processing fluencySubjective ease of processing. Higher fluency = more positive evaluations.
Perception biasSystematic gap between what users say, what stakeholders believe, and what actually happens.
Decision architectureThe structure of choices as presented. Every structure is a nudge.
Empathy modelingEmbodying the traits and context of a target demographic for directionality.
Hardware layerShared cognitive architecture: attention, memory, autopilot, mind perception. Universal.
Software layerCultural conventions: style preferences, expectations, literacy. Varies by audience.
Mind perceptionUsers construct lightweight models of competence (Agency) and care (Experience) for websites.
Predictive processingBrain generates predictions about input. Matched = autopilot. Violated = attention fires.
Mental prosthesisScaffolding for cognitive functions and limitations. Good design IS prosthesis for the brain. The curb cut effect.
Cognitive Constraint DesignA proposed meta-framework: the principle that cognitive constraints are design material, not obstacles. PFD is the first formal application to interface design.
Curb-Cut EffectWhen designs optimized for constrained users improve experience for all users. Named after sidewalk curb cuts designed for wheelchair users that benefit everyone.
Dual FoundationPFD’s evidence base draws from two largely distinct citation pools — perception psychology and ADHD cognitive neuroscience — with zero citation overlap.
IBNS (Interest-Based Nervous System)Dodson’s clinical model describing ADHD motivation as driven by interest, novelty, competition, and urgency rather than importance or reward.
Prediction ErrorThe mismatch between what the brain expects and what it encounters. The fundamental mechanism underlying first impressions and trust judgments.
Convergent DiscoveryWhen multiple independent practitioners arrive at the same insight without knowledge of each other. Ten or more ADHD-AI builders discovered the same collaboration patterns independently, confirming environmental fitness.
Point of PerformanceBarkley’s (1997) principle: self-regulation failures in ADHD occur in the moment of action, not during planning. Design must provide support where the decision happens, not where the instruction was given.

Use It

Now you have the framework. What comes next depends on what you need.

Learn it. The Make Me Think series walks through each layer with case studies, citations, and the stories behind the principles. This reference page is yours to come back to. Cite it, teach it, build on it.

Use it in your tools. The framework and AI skill are open source on GitHub. Free to use with attribution.

Apply it. If something isn’t converting and you want the methodology applied to a specific problem, that’s what I do. E-commerce, SaaS, consumer products. 15 years of it. Services has the details.

Ask me. Genuinely. If you read this far, I’ll read your email.

← Back to the methodology page