Interaction design asked how a person uses a tool. Behavior design asks what the tool does: when it acts, when it waits, when it suggests, when it stays silent. As software stops being purely reactive, behavior is the unit you actually ship.
Two earlier chapters mapped the materials. The product layers in Chapter 2 — state, memory, permissions, timing, confidence, orchestration, recovery, visibility — name what is beneath the interface. The negotiation loop in Chapter 3 names how the product moves between user and system over time. This chapter is about the active design decisions the builder makes on top of both: which questions to answer when a feature has the option to act on its own.
From clicks to choices
A button is an interaction. You press it. The system responds. The loop closes in a single beat. Most of the design vocabulary you grew up with — affordance, target size, label clarity, micro-interaction — is built around that beat.
A behavior is different. Autosave is a behavior. The way a notification batches itself instead of firing six times in a minute is a behavior. The photos app that surfaces a “memory” from a year ago — picking the trip and not the funeral, picking the friend and not the ex, picking the morning the user is most likely to want it and not the moment they are walking into a difficult meeting — is a behavior.
Each of those is the system making a decision over time, on its own initiative, with no click as the immediate trigger.
The shift from interaction to behavior is the shift from designing one moment to designing a stance. The button’s job is to react well. The behavior’s job is to act well — sometimes by acting, sometimes by waiting, sometimes by deferring, sometimes by undoing.
The visible part of an interaction is often enough to understand its shape. The visible part of a behavior rarely is.
Products that feel intelligent rarely have a single feature you can point at. The intelligence is distributed across choices the system is making moment by moment, most of which never surface as a screen.
That is why behavior has become its own design material.
The layers beneath the interface produce behavior
A product lives beneath the interface: in state, memory, permissions, timing, confidence, orchestration, recovery, and visibility.
Those layers are not static architecture. They are what produce behavior.
State decides what is possible. Memory decides what carries forward. Permissions decide who has authority. Timing and confidence shape how boldly the system moves. Recovery and visibility decide whether that movement can be trusted after it happens.
A product behaves because those layers interact.
This is why behavior design is not the same as adding “smart” features. A product can have AI inside it and still behave badly. It may generate useful output while interrupting too often, acting without authority, hiding its assumptions, or asking the user the same question forever.
Behavior is not intelligence alone. Behavior is intelligence shaped by product judgment.
The questions behavior asks
There is a working set of questions a behavior designer asks for any system worth shipping. Each one looks small. Each one is the seed of a long argument.
What should the system notice?
Notice is the front edge of behavior.
The autosave that knows about pause length. The spam filter that knows about a sender’s history with you. The calendar that notices you keep declining a recurring meeting. The finance tool that notices a merchant appears every month but is categorized differently each time.
Anything the system does not notice, it cannot act on. Anything it notices that it did not need to is surveillance dressed as feature.
This is the first discipline of behavior design: deciding what the system has the right to observe.
What should it infer?
Inference is what the system does with what it noticed.
The expense category guessed from the merchant. The thread the half-finished message belongs to. The meeting intent inferred from a vague sentence. The “you usually wake up around seven” that lets an alarm be soft instead of jarring.
Inference is where confidence enters the work. The system has to know how sure it is before it acts. It has to know the difference between a fact, a pattern, a preference, a guess, and a coincidence.
A product that cannot tell those apart will behave with the wrong posture.
What should it do?
Acting is the part teams over-design and under-decide.
Most “the system should…” sentences in a spec are wishful. The real question is which actions the system takes on its own, which it stages for review, and which it escalates.
It may need to save, suggest, send, archive, pre-fill, block, or ask someone else. Each option changes the product’s posture.
A system that always acts is rude. A system that never acts on its own is not behaving; it is waiting for a button you forgot to draw. The craft is in deciding where initiative belongs.
What should it ask first?
Asking is expensive.
Every confirmation dialog is a small admission that the system did not trust itself to act. Sometimes that admission is respect, not weakness.
Destructive actions, irreversible payments, legal commitments, public posts, sensitive messages, anything a user would feel betrayed by — these often deserve explicit confirmation.
But many confirmation dialogs are not respect. They are uncertainty pushed onto the user. A good behavior asks rarely, asks well, and remembers the answer.
It does not ask because the team was afraid to decide. It asks because the user’s authority matters.
What should stay visible?
Visibility is what lets a user trust a behavior enough to keep letting it happen.
The “saved” timestamp. The reason an email was filtered. The list of files an assistant just touched. The explanation for why a transaction was categorized. The note that a suggestion is based on last month’s pattern, not a rule.
A behavior that hides what it did has to ask for trust every time. A behavior that shows what it did can earn trust once and reuse it.
Visibility does not mean flooding the interface with system noise. It means making the right part of the behavior legible at the moment trust depends on it.
What should be reversible?
Reversibility is the safety net that lets a behavior take more risk.
An auto-archive that can be restored in one click is often a stronger product than an archive that asks first and lives forever. A draft rewrite that can be compared against the original lets the system be more useful. A bulk action with clear undo is less frightening than a product that asks for confirmation at every step because recovery is weak.
The cost of a wrong action is the time, attention, and anxiety it takes to undo it. Cheap undo lets the system do more. Expensive undo forces the product to ask, block, and hesitate.
Recovery is not what happens after behavior design fails. It is what makes better behavior possible.
When should it back off?
Backing off is the part most products never learn.
The photos app that stops surfacing memories from the same week after the user dismisses two in a row. The notifications that quiet themselves in the hour after you mute one. The model that retreats to deterministic behavior when its confidence drops below a threshold. The assistant that stops offering summaries for a thread after you repeatedly ignore them.
A behavior that does not know how to back off becomes noise. A behavior that backs off well becomes a colleague.
This is one of the clearest differences between a product that is merely active and a product that feels intelligent.
Intelligence is not only knowing what to do. It is knowing when the help has stopped helping.
These seven questions do not give you a product. They give you a way to argue about one. And that is already a great deal, because many weak products are not weak because the team lacked ideas. They are weak because the team never had the right argument.
Example: the feature request becomes a behavior spec
Imagine a product team writes a small feature request:
“Suggest a meeting preparation summary before important calls.”
As an interaction, the request looks simple. Add a card before the meeting. Put the summary in it. Let the user open the source notes. As behavior, the request becomes more serious.
The system has to notice that a meeting is approaching, that the meeting is important enough to prepare for, and that there is relevant material to summarize. It has to infer what counts as relevant instead of treating every nearby document as context.
Then it has to decide what to do. Should it generate the summary automatically, or wait until the user opens the meeting? Should it ask before including private notes? Should it back off after the user ignores the summary three times?
A behavior spec would not stop at “show summary card.”
It would say:
- Trigger: 30 minutes before a meeting marked external or high priority.
- Inputs: calendar metadata, linked account notes, open tasks, and explicitly attached documents.
- Confidence: summarize automatically only when at least two reliable sources exist.
- Visibility: show the source documents used and mark missing context.
- Control: let the user exclude a source before the summary is generated.
- Correction: let the user remove an item and mark why it was irrelevant.
- Backoff: stop preparing summaries for a meeting type after repeated dismissal.
The artifact is not a longer feature spec. It is a behavior spec: a compact account of the system's attention, authority, confidence, visibility, correction path, and retreat condition.
The difference is not polish. The difference is that the second version designs the system’s stance. It gives the team something concrete to argue about before the screen starts making the behavior feel decided.
Restraint is half the craft
Most behavior decisions are decisions about what not to do. The autosave that holds for a beat instead of firing on every keystroke. The transaction that does not auto-categorize when the merchant string is ambiguous. The smart reply that does not appear in a thread whose emotional weight the system has no right to flatten. The photos app that does not surface a memory when the system cannot tell whether the people in it are still in the user’s life. The system that does not act on a low-confidence inference, even though it could, even though acting would feel impressive.
Restraint is hard. Not because the design call is subtle, but because two structural forces work against it on every team.
Restraint does not demo well. A feature that does something is easy to show. A feature that wisely does nothing in the right context is invisible until the context arrives, and then is invisible because nothing went wrong. Teams that cannot tell the difference between “the system stayed quiet by design” and “the system stayed quiet because nothing was built there” will reliably under-invest in restraint.
Restraint also requires a confidence model. To not act, the system has to know it has not earned the right to act. That is a real engineering shape: uncertainty propagated end to end, surfaced honestly to the part of the system that decides whether to step forward.
Most products lack that shape. The ones that have it tend to feel calmer than the ones that do not.
Restraint is what makes visible behavior trustworthy.
A system that acts every time it could is a system the user has to override every time they disagree. A system that acts only when it should leaves the user with energy to do their actual work.
This is one reason behavior design is more than interaction design with AI added. It is a discipline of initiative and restraint.
The argument inside the craft
Behavior design becomes real when a team can argue about it precisely.
Two senior people can disagree about whether a system should infer or ask, and the disagreement can be productive. It surfaces what each of them believes about authority and the cost of being wrong.
Should the assistant archive automatically, or stage the action?
Should the product ask every time, or only after a failed recovery?
Should a repeated correction become a preference, or remain local to the session?
Should low confidence be shown explicitly, or expressed through softer behavior?
Those are not decoration questions. They are product questions with technical and ethical weight.
That kind of disagreement is not process noise. It is the craft becoming explicit.
Modern software is not only a sequence of screens. It is a pattern of behavior. And patterns of behavior deserve design.
Framework
Behavior Design Questions
A working list of questions to ask when designing behavior rather than interaction.
01 — Notice
What does the system observe, and what should it ignore?
Behavior begins with attention. Deciding what the system notices is also deciding what kind of power it has.
02 — Infer
What meaning does the system derive, and how confident is it?
The product must distinguish between facts, patterns, preferences, guesses, and uncertainty.
03 — Act
What can the system do without explicit instruction?
Initiative is useful only when it is bounded by confidence, authority, and recovery.
04 — Ask
Where does the system need consent, confirmation, or human judgment?
Asking should be a designed act of respect, not a substitute for product judgment.
05 — Reveal
What must stay visible for trust to hold?
The user needs to understand enough of the behavior to keep granting the system room to act.
06 — Reverse
How does the user recover from a wrong action?
Reversibility determines how much risk the behavior can safely take.
07 — Back off
When should the system stop helping?
A good behavior knows when its assistance has become interruption.
The framework is simple because the arguments are not. Use it whenever a feature starts to behave on the user’s behalf.
Behavior Design Questions
A working list of questions to ask when designing behavior rather than interaction.
What does the system observe, and what should it ignore?
What meaning does the system derive, and how confident is it?
What can the system do without explicit instruction?
Where does the system need consent, confirmation, or human judgment?
What must stay visible for trust to hold?
How does the user recover from a wrong action?
When should the system stop helping?
Behavior spec for the feature you restated
Commit the artifact this chapter produced. The portfolio strip in Chapter 11 reads back what you have written here.
Trust Is Now a Product Layer
Once a system begins to behave, trust stops being cosmetic.