The Sam Reich Problem: What a Comedy Game Show Teaches Us About Building Deception That Actually Works

Defensive cyber deception explained through Game Changer: how honeypot design, choice architecture, and flow theory combine to build deception environments that actually work against real attackers.

There is a moment in nearly every episode of Game Changer — Dropout’s delightfully unhinged panel show — where you can watch a very smart person walk confidently in the wrong direction.

Game_Changer_logo.svg

They’re not being reckless. They’re being reasonable. They’re reading the room, applying heuristics that have served them well their entire lives, following what looks like the obvious next step. And somewhere off-camera, Sam Reich is watching them do it with the particular serenity of someone who already knows how the story ends.

That expression on his face — warm, patient, ranging from giddy to bemused — is the face of a person who designed an environment and is now watching another person live inside it exactly as intended.

I want to talk about that expression. Because I think it’s the face you want to be wearing when you’re running a deception operation.


What Game Changer Is, for the Three People Who Haven’t Seen It

Game Changer is a game show where the players don’t know the rules. That’s the whole bit. Every episode, three comedians — some of the sharpest improvisers working anywhere — show up and participate in a completely new game whose mechanics, scoring criteria, and objectives are hidden from them until they’re already playing. Sam Reich opens every episode the same way: “The only way to learn is by playing, the only way to win is by learning, and the only way to begin is by beginning.” Then he steps back and watches.

The players have to deduce the game from the inside. They probe, they test, they observe what earns points and what doesn’t, they build mental models and revise them in real time. They are smart, experienced, adaptive — and they are completely at the mercy of a man who thought about all of this long before they walked in the room.

Polygon described the show, memorably, as “like Saw, but instead of people captured by a serial killer, it’s improv comics captured by video producers within the trap of their own performing instincts.”

I’ve been thinking about that description for months, because it is — accidentally — one of the most precise articulations of what good cyber deception actually does that I’ve ever encountered. The attacker isn’t captured by your firewall. She’s captured by her own methodology: the mental model of “what a real network looks like” that she carries into every engagement, the heuristics baked into her tooling, the instincts that have made her effective everywhere else. Your job, as the designer of a deceptive environment, is to build something that fits that model just closely enough that she commits to it fully.

And then you let the performing instincts do the rest.


The Oldest Game

Before we get into networks and honeypots and kill chains, I want to take a step back — because one of the biggest mistakes in how we talk about defensive deception is framing it as an advanced technique. Something novel. Something that requires special justification.

Deception is not novel. Deception is ancient. Deception is, in a very real sense, what cognition is for.

It shows up in viruses manipulating cell machinery. It shows up in orchids that produce fake pollen to trick bees. It shows up in the anglerfish, dangling a bioluminescent lure in the permanent dark of the abyssal zone while everything with eyes and hunger swims toward what looks like a meal and is, in fact, a mouth. The anglerfish didn’t invent that strategy. Evolution iterated on it for hundreds of millions of years because it works — not because predators are stupid, but because cognition operates on expectation, and expectation can be shaped.

What makes the anglerfish’s lure convincing isn’t that it looks exactly like a small fish under a microscope. It’s that it looks enough like a small fish to the specific sensory system of the specific prey in the specific context that the prey commits to the approach. The deception doesn’t need to be perfect. It needs to be sufficient for the receiver’s decision threshold.

This is the first thing Sam Reich understands that a lot of security practitioners don’t: a convincing decoy is not a property of the decoy itself. It’s a property of the relationship between the decoy and the observer’s expectations. A honeypot doesn’t need to pass a Turing test. It needs to pass an attacker’s recon heuristics.

Those are very different design problems.


Theory of Mind, or: You Have to Be Willing to Live in Someone Else’s Head

Here’s the honest version of what it takes to deceive someone: you have to model their mind accurately enough to predict what they’ll believe.

Deception, at its philosophical core, is the intentional act of convincing another of a falsehood — not merely causing false beliefs, but achieving a genuine dynamic of convincing. That distinction matters. A tripwire that logs a connection doesn’t deceive anyone. A tripwire that an attacker walks past thinking it’s a legitimate service, while her entire mental model of your environment reconfigures around the wrong premise — that’s a deception. The difference is whether you’ve done the theory-of-mind work.

To craft a convincing lie, you need to hold the truth in one hand and the target’s worldview in the other, and figure out exactly where the gap between them can be papered over. Does the lie cohere with everything else the target already knows? Does it match her priors about how the world works? Where are the seams, and will she notice them?

This is cognitive labor. Real cognitive labor. And it’s the part that gets skipped most often when organizations deploy deception technology, because dropping a honeypot on the network is easy. Understanding your specific threat actors well enough to make the honeypot convincing to them specifically is hard. It requires knowing what they’re hunting, what their tooling looks like, what their mental model of a “real network” is, and what progress feels like to them.

Sam Reich, before he designs a game, does his research. He surveys his cast. He knows who will jump out of a plane and who won’t. He knows who has a competitive streak that can be weaponized and who has a people-pleasing reflex that can be redirected. He doesn’t design a generic game — he designs a game that will work on these specific people in this specific moment. The environment is tailored to the player.

Your threat-actor profile is your casting call. Build the environment for who’s actually coming.


”It’s Always Falling Apart a Little Bit” — The Doctrine of Convincing Imperfection

Here is where Sam Reich’s instincts diverge most sharply from conventional security thinking, and it’s worth dwelling on.

sam-reich-funny.jpeg

Sam has admitted, with characteristic candor, that Game Changer does almost no playtesting. Things go wrong on the day. Buzzers don’t work. Rules are ambiguous at the edges. The show, as he puts it, “feels like it’s always falling apart a little bit. Just a little bit, which is a part of the fun, you know?”

Security people hear this and feel a familiar anxiety. Don’t you want everything to work perfectly? Don’t you want the environment buttoned up? The answer, for a deception operation, is: absolutely not.

A perfect decoy is a dead giveaway. Real target networks are messy. They have a server running an OS two versions behind because nobody got around to patching it. They have a file share with an ambiguous name that turns out to be someone’s personal backups from 2019. They have the fingerprints of human beings all over them — misconfigured services, legacy debt, residual artifacts of decisions made by people who’ve since left the company. A pristine environment, one where every service is correctly configured and every hostname is perfectly consistent and there is no evidence that anyone actually works here, will feel wrong to any attacker who’s spent real time in real networks. It tickles the part of her brain that’s always scanning for the too-good-to-be-true.

The literary concept of verisimilitude is useful here. Verisimilitude isn’t accuracy — it’s internal consistency. It’s the quality that makes a fictional world feel real regardless of how implausible its premise is. You get there through three things: consistency (don’t contradict yourself where the audience is looking), specificity (concrete details, not vague generalities), and human truth (make the behavioral logic ring true). When you need to bend reality, surround the implausible element with thoroughly grounded details.

Translated into honeypot design: you don’t need a perfect replica of a production environment. You need an environment that doesn’t contradict itself anywhere an attacker thinks to look. A file created before the OS was installed. A service version that doesn’t match the banner. A domain-joined machine whose last logon is three years ago with no explanation. Each of these is a pebble in the shoe — a tiny friction that breaks flow and triggers suspicion. The goal isn’t perfection. It’s coherent imperfection. An environment that feels like a real network that someone actually works in.

Just slightly falling apart. Just a little bit.


”How Is It Fun in Minute Thirty?” — Sustaining Engagement Across a Full Engagement

Game Changer’s writers return obsessively to one question when designing new episodes: “The hardest part of coming up with a good idea is not ‘how is it fun in minute one?’ It’s ‘how is it fun in minute 30?’” Sam has described this in terms of rate of change — how quickly the episode evolves to sustain attention.

It’s the right question. Almost nobody asks it about deception environments.

Most honeypots are designed to be convincing in the first thirty seconds of contact. They present an attractive surface, invite enumeration, catch the initial probe. And then — if a skilled attacker stays longer — they run dry. There’s nothing else to find. The depth isn’t there. And an attacker who stops finding things to find starts doing something far more dangerous: she starts auditing the environment itself. She starts looking for the edges of the set. She may even break the fourth wall.

The concept of flow state, developed by psychologist Mihaly Csikszentmihalyi, describes a condition of optimal engagement where a person is so absorbed in a task that they lose track of time and self-consciousness. It’s characterized by a balance between skill and challenge, clear goals, and immediate feedback — the sense that you’re making progress, that your actions are producing results, that there’s always something more around the corner. Games are specifically engineered to induce this state. So are good deceptive environments.

An attacker in flow is not questioning her environment. She is living in it. She’s chasing the next discovery, following the trail, convinced she’s making progress toward something real. The moment flow breaks — the moment the environment stops rewarding her curiosity — she surfaces. She starts thinking about where she is rather than where she’s going.

Csikszentmihalyi also distinguishes between what game designers call microflow — the moment-to-moment rewards that keep a player engaged — and macroflow, the sustained arc of challenge across a full session. Game designers talk about breadcrumbs: small discoveries that each feel satisfying in isolation and also point toward the next one. A file with an interesting name. A credential that almost works. A service that seems slightly misconfigured in a way that suggests something valuable is just behind it.

In deception technology terms, honeytokens and canary documents and fake credentials aren’t just tripwires. They’re microflow rewards. The breadcrumbs that keep the attacker playing. Design them like collectibles in a game: obviously valuable, just reachable enough, tantalizingly real. Make sure they lead somewhere. The deception should have somewhere to go no matter how long the attacker stays.

The macroflow question — how is this still interesting in minute thirty? — is the difference between a set and a game. Sets get recognized. Games get played.


The Ratfish Problem — When the Attacker Suspects the Frame

vic-jacob-lobster.jpg

At some point in Game Changer’s run, the cast became meta-aware. They’d been surprised in so many novel ways that they started treating every Dropout production as a potential deception environment. They entered shoots looking for the game inside the game, questioning the framing of ordinary instructions. The show’s own success had created a problem: the players were suspicious of the frame itself.

Sam’s response was not to try harder to hide. It was to lean into it. The meta-awareness became part of the design.

This maps directly to one of the more counterintuitive findings in empirical deception research. A controlled experiment involving over 130 professional red teamers found that deploying decoys and explicitly telling attackers that deception was present had the greatest measurable impact on attacker behavior — greater than decoys alone, and far greater than no deception at all.

Sit with that for a moment. Telling an attacker you’re running deception increases the deception’s effectiveness. Not because she stops looking, but because now she can’t stop looking. Every service is a potential fake. Every credential is a potential canary. Every file with an interesting name might be a trap. The cognitive load of operating under that kind of uncertainty is enormous — and under cognitive load, automatic heuristic-driven decision-making takes over. Behavioral researchers call the behaviors that leak out under this pressure cognitive leakage: the unconscious actions that reveal internal state because the conscious mind is too busy managing uncertainty to monitor them.

An attacker under meta-uncertainty moves slower. She second-guesses her tooling. She makes choices she’d never make in a clean environment. And in making those choices, she reveals herself.

This is the Ratfish principle. Once an attacker knows your organization deploys deception technology, you’ve introduced uncertainty into every action she takes. The uncertainty is not a side effect of the defense. The uncertainty is the defense.


The “Bingo” Episode — Nested Observation and the Architecture of Watching

Season 6 of Game Changer contains what I think is the show’s most technically sophisticated episode, and I want to describe it carefully because it maps almost perfectly onto what well-designed deception infrastructure should look like.

“Bingo” works like this: there are players on stage playing a visible game. Unknown to them, a second group in a green room earns points by tracking the quirks and behaviors of the visible players. Unknown to them, a third group earns points based on the green room players’ interactions with each other and with the film crew. Sam Reich, above all of it, watches everything.

Three nested games. Three nested observation layers. Each layer convinced it is playing the only game, while being observed by a layer it cannot see.

This is what deception-in-depth looks like when it’s built correctly. The attacker enumerating your decoy network is the visible game — she’s the person on stage. Your telemetry stack, logging her tooling and exfil patterns and C2 callbacks, is the green room. Her own infrastructure — the C2 servers and staging environments she believes are hidden, but which your threat intelligence pipeline has already fingerprinted through her probes — is the third layer. And the analyst watching all of it? That’s Sam.

The key insight from “Bingo” is that each layer of observation is possible only because the layer below it is convinced it’s the only player. A contestant who knew about the green room would behave completely differently. An attacker who knew the full extent of your telemetry would behave completely differently. The deception doesn’t just obscure the network — it obscures the observation of the network.

The technical literature describes this as a deception fabric: multiple techniques complementing each other across every layer of the deception stack — network, system, software, data. No single technique is sufficient on its own. An attacker who identifies your honeypot and evades it walks directly into your honeytoken. An attacker who avoids honeytokens lights up your network telemetry. An attacker who keeps her traffic clean reveals her infrastructure through threat intelligence correlations. Each evasion reveals a different face. The nested structure means there is no safe path through — only paths with different costs and different amounts of observable behavior.

Sam never tips his hand. He sits at the center of all the nested games and watches. That’s the position you want to be in.

sam-laughing-subtitle.avif


Choice Architecture, or: Making the Wrong Door the Obvious Door

There’s a behavioral economics concept every deception operator should internalize, and it goes by the unglamorous name of choice architecture. Richard Thaler and Cass Sunstein, in their book Nudge, defined it this way: a nudge is any aspect of the choice architecture that alters behavior in a predictable way without forbidding options or significantly changing incentives.

The Schiphol Airport example is famous: etching a fly on the urinal reduced spillage by 80%. Nobody told anyone not to miss. Nobody changed the rules or the incentives. They gave the automatic brain something to aim at, a reason to engage. The choice architecture did the work.

For deception design, the insight is this: you don’t force an attacker toward your decoy. You make the decoy the path of least resistance — the lower-effort target, the more obviously promising lead, the door that’s just slightly more open than the ones beside it. Thaler and Sunstein observed that people default to whatever requires the least cognitive effort, and that this tendency is so reliable it can be leveraged at government scale to reshape organ donation rates and retirement savings behavior.

An attacker moving through your network is, in behavioral economics terms, a decision-maker under significant time pressure, operating on incomplete information, relying heavily on heuristics. She is not making fully rational choices. She’s making satisficing choices — the first option that meets the threshold of “good enough” given her current constraints. Your job is to be that option.

In practice: real assets should be slightly harder to find, slightly less obviously configured. Decoy assets should be slightly easier — but not so easy they feel planted. The gradient has to feel natural. Like someone just happened to leave something interesting in a directory that was just a little too easy to navigate to. Like the credentials in the password manager that’s just slightly less locked down than the others.

The path of least resistance should lead exactly where you want her to go. Like setting a trail of apple slices through the forest and then acting surprised when the deer shows up exactly where you laid them.


”Second Best” and the Limits of Optimization

One of Game Changer’s recurring structural moves is awarding points for unexpected outcomes — not the best performer, but whoever came second, or whoever was most surprising, or whoever the audience liked for reasons contestants can’t predict. In one episode, the scoring mechanic explicitly rewards whoever performed second-best, much to Brennan Lee Mulligan’s chagrin.

The effect on contestants is both predictable and illuminating: they lose the ability to optimize. They can’t simply try hard and expect it to work, because the reward structure doesn’t map onto effort the way they’ve been trained to expect. They have to hold competing hypotheses about what the game actually values, in real time, while playing the game.

This is a useful model for what happens when a skilled attacker encounters a well-designed deceptive environment. If the environment consistently rewards her expected behavior — find credentials, credentials are valid, escalate — she builds a confident mental model and moves through it efficiently. She’s in flow, but it’s your flow, the predetermined path. She’s doing exactly what you expected.

The more interesting design question is: what does your environment do when she deviates? Does the game hold together if she comes in sideways? A deception environment that only maintains coherence along the anticipated attack path will shatter the moment a careful adversary decides to stress-test it. The best environments are structurally improvisational — like a good Game Changer episode, they have somewhere plausible to go regardless of what direction the player moves, because the underlying logic is consistent even when the surface varies.

An environment that can only sustain deception along one path isn’t an environment. It’s a corridor with a painted backdrop at the end.


The Technical Stack, Read Through Sam’s Eyes

Let me try to translate all of this into something you can actually use, because theory is only as good as what it generates in practice.

Layer one — receiver profiling. Before you build anything, model your attacker. What is she hunting — credentials, data, infrastructure access? What does her typical kill chain look like? What would feel like progress to her? What does she assume a real network looks like? This is Sam surveying his cast before the season starts. The environment you build will only be as convincing as this model is accurate. A generic honeypot is a game designed for no one in particular, and any attacker who’s done this before will feel that genericness the moment she touches it.

Layer two — verisimilitude engineering. Every element of the deceptive environment needs to be internally consistent from the attacker’s vantage point. Hostnames, service banners, file timestamps, directory structures, last logon timestamps, user activity patterns — all of these together contribute to the gestalt sense of does this feel real? One contradiction is a pebble in the shoe. Two and she’s auditing. Build the texture of a lived-in environment: the legacy cruft, the human fingerprints, the slightly-fallen-apart quality of something people actually use. The goal isn’t a perfect replica. It’s an environment where nothing catches on anything.

Layer three — choice architecture. Design the topology so that the decoy is the path of least resistance at every decision node. The real domain controller should be slightly harder to find than the fake one. The real credential store should have slightly higher barriers than the honeytoken. You’re not hiding real assets behind impenetrable walls — you’re making them marginally less obvious than the fakes. The gradient has to feel natural. An attacker who feels like she’s being funneled will notice. An attacker who feels like she just keeps finding the obvious move will keep moving.

Layer four — the breadcrumb loop. Design a progression. The first discovery leads to a second, which leads to a third. Each is more apparently valuable than the last. Each is plausible given what came before. The fake database should have plausible-looking data in it. The fabricated user accounts should have browser history. The honeytoken should feel like it belongs to a real person who made a real decision to store it where they did. Each breadcrumb triggers a detection event — but more importantly, the sequence creates a narrative that keeps the attacker committed to the environment.

Layer five — the nested observation fabric. Your telemetry doesn’t live only on the decoy. It reaches as far as the attacker’s reach: her tooling, her exfil patterns, the infrastructure she’s running. The observation layer is designed so that every attempt to evade one detection mechanism lights up another. Each evasion is itself observable. The nested structure means there is no truly safe path through — only paths with different costs and different faces.

Layer six — deliberate meta-disclosure. Tell her you have deception technology. Not in a way that reveals specifics — but in a way that introduces pervasive uncertainty. A security posture statement. An industry reputation. A public blog post about your deception program. The uncertainty is free and it compounds. Every action she takes now carries an extra cognitive tax. Cognitive load breaks flow and produces leakage. Leakage produces intelligence.


The Ethical Weight of the Game

I want to sit with something before we close, because a post that goes this deep into the theory of deception without acknowledging its moral texture would be incomplete.

The tools I’ve been describing are powerful. The same cognitive machinery that lets defenders protect networks through misdirection is the machinery behind phishing, social engineering, and fraud. Deception, as a philosophical category, is held to a higher moral standard than manipulation precisely because of how completely it recruits the target’s own reasoning against her. Convincing someone of a falsehood isn’t the same kind of wrong as nudging someone toward a choice — it’s a violation of the epistemic relationship itself.

Defensive deception operates in a morally serious space. It requires clarity about purpose, proportionality in deployment, transparency within the organization about what’s being done and why, and genuine attention to where the line sits between defensive intelligence-gathering and overreach. These aren’t questions that have clean answers, and I don’t think the community has finished working through them.

What I do think is that how you hold these tools matters as much as the tools themselves. A defender who thinks carefully about the ethics of deception will also design better deception — because the same rigor that makes you examine the moral implications is the rigor that makes you think carefully about what you’re actually doing to a human mind, and why it works.

Sam Reich, for his part, has said explicitly that he thinks carefully about consent in how he designs games for his cast. He’s described his discomfort with live audience episodes precisely because post-production editing provides a safety valve for performers that a live format doesn’t. Even in a game show, the ethics of designing environments for other people to inhabit isn’t a question you get to skip.


The Reveal

Every episode of Game Changer ends the same way: the rules come out. The scoreboard makes sense. The things that seemed arbitrary crystallize into a coherent system. And the contestants — some delighted, some distraught — see the whole shape of the game they’ve been playing for the last thirty minutes.

They were inside a designed environment the entire time. Every door they opened, Sam already knew where it led. Every choice that felt like their own was shaped by a structure they couldn’t see.

A deception operation’s reveal looks different — written up in a threat intelligence report rather than aired on a streaming service — but the epistemological structure is identical. The attacker was inside a designed environment. Every probe she sent returned information you’d curated. Every credential she found was one you’d placed. Every pivot she made followed a path you’d made easier than the alternatives. You watched all of it from behind the scoreboard.

A blocked intrusion attempt tells you something tried to get in. A well-run deception operation tells you who, how, with what, toward what goal, and how long they were willing to sustain it. It tells you the shape of the threat in intimate, behavioral detail. That’s the intelligence. That’s the reveal.

The goal was never the gotcha. The goal was always the picture.


One More Thing Sam Said

In an interview about Game Changer’s production, Reich described what makes a game idea ready to shoot: “One is we don’t [know]. And we do it anyway.”

There’s something honest in that I want to hold onto. Deception operations, like game show episodes, are not fully playtestable. You build the best environment you can with your best model of your threat actor, and then you run it, and then you learn. The attacker will do something you didn’t predict. The environment will hold or it won’t. You revise.

The goal is not to build the perfect deceptive environment before you start. The goal is to build one that’s good enough to generate intelligence, and then to let the intelligence make the next one better.

That’s the loop. That’s the game.

The only way to learn is by playing. The only way to win is by learning. And the only way to begin is by beginning.


This post draws on research in interpersonal deception theory (Buller & Burgoon, 1996), behavioral economics (Thaler & Sunstein, 2008), flow theory (Csikszentmihalyi, 1990), empirical cyber deception research (Ferguson-Walter et al., 2021, USENIX Security), and the literature on honeypot design and deception-in-depth (surveyed in “Three Decades of Deception Techniques in Active Cyber Defense,” 2021). The Game Changer citations are from publicly available interviews and episode descriptions; the show is on Dropout.tv and you really should watch it. The best episode is “Official Cast Recording.”