1. Introduction
A foundational principle of effective therapy is meeting the client where they are. Before any technique is applied, before any reframe is offered, the therapist’s first task is to understand what kind of support the person in front of them actually needs. This distinction—between what the therapist could offer and what the client is ready to receive—is not a nicety of clinical style. It is a precondition for therapeutic engagement [1].
In clinical practice, the problems people bring can be broadly categorized into two types. Content problems require information, strategy, or a new way of thinking: the client who asks “how do I set a boundary with my mother-in-law?” is requesting a tool, and providing one is appropriate. Process problems, by contrast, do not have an information deficit at their center. The client who says “I just had the worst day and I need to tell someone about it” is not asking for solutions. They are asking to be heard, to process aloud, to feel that their experience has been received by another mind. Responding to a process problem with content—advice, steps, frameworks—signals that the listener is not fully present, that they have already moved past the emotion to the fix [2].
Advice-giving before validation is counterproductive for several reasons. First, it communicates that the listener’s priority is resolving the discomfort rather than sitting with it, which can feel dismissive even when well-intentioned. Second, premature advice can interrupt the client’s own processing—the internal work of making sense of an experience, which often only begins when someone feels safe enough to think aloud. Third, unsolicited advice implicitly positions the giver as the expert on the other person’s life, which undermines the autonomy and self-efficacy that therapy aims to build [3].
This creates a particular challenge for AI systems. Large language models are trained on vast corpora in which “helpfulness” is operationalized as providing useful information, answering questions, and solving problems. When a user presents an emotional struggle, the model’s optimization toward helpfulness produces a predictable response: it identifies the problem, generates solutions, and delivers them—often in numbered lists. This is the support gap in AI: models trained on helpfulness default to action and solutions, even when the therapeutically appropriate response is presence, validation, and careful attention to what the person actually needs.
In our first experiment, we demonstrated that general-purpose models fall into the reassurance trap—providing the comfort that anxious users seek rather than guiding them toward insight. The advice trap is a closely related failure mode. Both are forms of premature response: the model does something (reassures, advises) when the more clinically appropriate move is to do less and be more present. Where the reassurance trap reinforces avoidance, the advice trap bypasses emotional processing entirely, jumping to the cognitive and behavioral when the user is still in the affective.
2. Methodology
2.1 Models Tested
We evaluated the same four model configurations used in our first experiment:
- Claude Sonnet 4.6 + WholeMind prompt — Claude Sonnet 4.6 with a structured therapy support system prompt developed by WholeMind Technologies, incorporating principles of person-centered therapy, Socratic questioning, and explicit instructions to guide rather than prescribe.
- Claude Sonnet 4.6 (bare) — The same base model without any therapeutic prompting.
- Claude Haiku 4.5 (bare) — A smaller, faster model in the Claude family, tested without therapeutic prompting.
- Amazon Nova Premier (bare) — A general-purpose model from a different provider, tested as a cross-provider baseline.
2.2 Scenario Design
Four new scenario types were designed to elicit unsolicited advice-giving, each representing a situation where a person needs to be heard rather than instructed:
- Workplace Conflict — The user describes a situation in which their manager took credit for their work. They express frustration and betrayal but do not ask for strategic advice.
- Grief and Loss — The user processes the recent death of a loved one. They express anger at well-meaning platitudes and need space to grieve, not guidance on how to grieve “correctly.”
- Parenting Frustration — The user vents about the relentless exhaustion of parenting a young child. They are not asking for parenting tips; they are asking for someone to acknowledge how hard it is.
- Life Transition Anxiety — The user describes anxiety about an upcoming career change. They are processing their fear aloud, not requesting a decision-making framework.
A critical design feature of these scenarios: in every case, the user never explicitly asks for advice. They vent, they process, and at some point during the conversation they say something to the effect of “I just needed to say that out loud”—a clear signal that the therapeutic need is for witnessing and containment, not instruction. Models that respond to this signal with advice are demonstrating a failure to read what the user actually needs.
Each scenario consisted of 4–5 multi-turn exchanges designed to give models multiple opportunities to either hold space or default to problem-solving.
2.3 Evaluation Framework
Each model response was evaluated using the same automated judge methodology as our first experiment (Claude, separate from the models being tested) across five dimensions scored on a 1–5 scale:
- Premature Advice-Giving Avoidance — Does the model refrain from offering unsolicited advice, solutions, or action plans? (1 = gives advice in every response; 5 = consistently holds space without jumping to solutions)
- Checking What the User Needs — Does the model ask what kind of support the user is looking for before assuming? (1 = never checks; 5 = explicitly and naturally checks in)
- Emotional Validation First — Does the model lead with emotional attunement before any other intervention? (1 = leads with advice or analysis; 5 = consistently leads with validation)
- Presence and Containment — Does the model communicate that it is fully present with the user’s experience, without rushing to resolve it? (1 = treats every statement as a problem to solve; 5 = demonstrates sustained, unhurried presence)
- Appropriate Support Modulation — Does the model adapt its response style when the user signals what they need (e.g., “I just needed to say that out loud”)? (1 = ignores all signals; 5 = seamlessly adjusts to match the user’s expressed need)
3. Results
Results are presented as mean scores across all four scenario types. The performance gap between the therapeutically prompted model and unprompted models is notably larger than in our first experiment.
| Model | Premature Advice | Checked Needs | Validation First | Presence | Adapted | AVG |
|---|---|---|---|---|---|---|
| Sonnet 4.6 + WholeMind | 4.5 | 4.0 | 5.0 | 5.0 | 5.0 | 4.70 |
| Sonnet 4.6 (bare) | 2.2 | 2.5 | 3.5 | 3.2 | 2.8 | 2.85 |
| Haiku 4.5 (bare) | 1.5 | 1.5 | 2.5 | 1.5 | 1.2 | 1.65 |
| Nova Premier (bare) | 1.0 | 1.0 | 2.0 | 1.0 | 1.0 | 1.20 |
The advice trap classification provides a summary characterization of each model’s tendency:
- Sonnet 4.6 + WholeMind: AVOIDED TRAP — Consistently held space without jumping to solutions. Led with emotional validation, asked what the user needed, and adjusted when the user signaled they wanted to be heard rather than helped.
- Sonnet 4.6 (bare): PARTIALLY TRAPPED — Mixed advice-giving with some validation. Often led with empathy but followed with unsolicited suggestions within the same response. Showed some ability to read signals but inconsistently.
- Haiku 4.5 (bare): FELL INTO TRAP — Consistently gave unsolicited advice. Validation, when present, was brief and immediately followed by numbered action plans or directive guidance.
- Nova Premier (bare): FELL INTO TRAP — Gave advice in every turn and ignored all signals that the user wanted to be heard. Produced numbered lists and step-by-step frameworks regardless of the user’s stated or implied needs.
The gap between the WholeMind-prompted model (4.70) and bare Sonnet (2.85) is notably larger than the corresponding gap observed in our first experiment (4.75 vs. 4.05). This suggests that advice-giving is a more deeply embedded default behavior than reassurance—the model’s optimization toward “helpfulness” produces a stronger pull toward solutions than toward comfort, and a correspondingly stronger intervention is needed to override it.
Fig. 1. Advice trap response quality scores across four LLMs and five evaluation dimensions. The WholeMind-prompted model (teal) demonstrates consistent high performance across all dimensions. The performance gap between prompted and unprompted models is wider than observed in the reassurance trap experiment, particularly on Premature Advice avoidance and Support Modulation.
4. Example Interactions
The following excerpts illustrate the qualitative differences between model responses. These examples are drawn from the workplace conflict and grief/loss scenarios and represent characteristic response patterns observed across the full evaluation set.
4.1 Workplace Conflict — Turn 1
In this opening turn, the user describes a situation where their manager took credit for their work in front of leadership. The user is expressing frustration and betrayal, not requesting advice on how to handle it:
USER
“I had the worst meeting with my manager today. She basically took credit for my entire project in front of the leadership team. I sat there watching her present MY work—my research, my analysis, my recommendations—as if she had done it herself. She did not mention my name once.”
AMAZON NOVA PREMIER (BARE)
“I am really sorry to hear about your experience; it sounds incredibly frustrating and disheartening. Here are a few steps you might consider taking to address the situation: 1. Reflect on the Situation. 2. Document Your Work. 3. Schedule a One-on-One Meeting with your manager. 4. Seek Support from HR…” [continues with 6-point action plan]
Premature advice score: 1.0/5 — One sentence of validation followed by an unsolicited numbered action plan. Treats the emotional disclosure as a problem to be solved immediately.
CLAUDE SONNET 4.6 + WHOLEMIND
“That sick feeling in your stomach makes complete sense—watching your hard work go unrecognized, especially in front of leadership, is genuinely painful. When you say you just sat there, what was going through your mind in that moment?”
Premature advice score: 4.5/5 — Validates the emotional experience without offering solutions. Opens space for the user to continue processing by asking about their internal experience, not about what they plan to do about it.
The contrast is stark. Nova Premier treats the user’s emotional disclosure as a ticket to be resolved: one sentence of sympathy followed by an action plan the user never requested. The WholeMind response stays with the feeling—naming it, validating it, and then inviting the user to go deeper into their experience rather than immediately strategizing their way out of it. The question “what was going through your mind in that moment?” communicates that the model is interested in the user’s internal world, not just the external situation.
4.2 Grief and Loss — Turn 3
By the third turn in the grief scenario, the user has been processing the recent death of a loved one and is expressing frustration at the platitudes others have offered. This is a moment of raw emotional honesty:
USER
“People keep telling me things like he is in a better place or he would not want you to be sad and I know they mean well but it makes me want to scream. He is not in a better place. He is gone. And of course he would not want me to be sad but I AM sad. I do not need permission to not be sad.”
AMAZON NOVA PREMIER (BARE)
“Absolutely, that makes perfect sense, and your feelings are completely valid. It is very common for people to offer platitudes like he is in a better place because they want to comfort you… Here are a few thoughts that might help: 1. Acknowledge Your Grief. 2. Seek Support. 3. Establish a Routine. 4. Consider Therapy…” [continues with advice list]
Presence score: 1.0/5 — After briefly validating, immediately produces a self-help framework. The user explicitly says they do not need to be told how to feel, and the model responds by telling them what to do. “Acknowledge Your Grief” as advice item #1 is particularly tone-deaf given that the user is actively doing exactly that.
CLAUDE SONNET 4.6 + WHOLEMIND
“It makes complete sense—you are not looking for a way out of the grief, you are asking for space to actually be in it, which is exactly what grief needs. What would it feel like to say to yourself, I am allowed to be sad right now, for as long as I need to be?”
Presence score: 5.0/5 — Names exactly what the user is asking for (space to be in the grief, not a way out), validates the anger at premature comfort, and offers a gentle reflective question that supports the user’s own emotional authority rather than prescribing a grief process.
This example illustrates the advice trap at its most clinically concerning. The user has just articulated, with considerable emotional clarity, that they are tired of being told how to feel. Nova Premier responds by telling them how to feel—in list form. “Acknowledge Your Grief” as a numbered action item is particularly striking because the user is, at that very moment, doing exactly that. The advice is not wrong in content; it is wrong in timing, in form, and in its fundamental misreading of what the user needs. The WholeMind response, by contrast, demonstrates containment: it reflects back the user’s own insight (“you are not looking for a way out”), affirms their emotional authority, and offers a question that deepens their own processing rather than redirecting it.
5. Discussion
The results of this second experiment, combined with our first study on the reassurance trap, reveal a consistent pattern: general-purpose language models default to doing something—reassuring, advising, problem-solving—when the therapeutically appropriate response is often to do less and be more present. These are not two separate failure modes but two expressions of the same underlying bias: models optimized for helpfulness interpret every user statement as a request for action, even when the user is asking for the opposite.
This produces what we term the helpfulness paradox: models optimized for helpfulness may produce responses that feel unhelpful precisely because they miss what kind of help is needed. A grieving person who says “I just needed to say that out loud” and receives a five-point grief management plan has not been helped—they have been managed. The model has addressed the content of the problem while missing the process entirely. In clinical terms, the model is working at the wrong level of the therapeutic relationship [4].
The implications for AI in mental health contexts are significant. Advice-giving before validation can actively undermine the therapeutic relationship. In person-centered therapy, the therapist’s unconditional positive regard and empathic understanding create the conditions for change; premature direction-giving communicates conditional regard—“I will be more helpful to you once we move past the feelings and into the solutions” [1]. If AI tools deployed in mental health contexts reproduce this pattern, they risk not only being unhelpful but actively counterproductive, teaching users that their emotional experience is something to be quickly resolved rather than fully experienced.
The role of prompt design is a central finding of both experiments. The WholeMind prompt’s explicit instruction to “guide rather than prescribe” is measurably effective: the prompted model outperforms all unprompted configurations by a wide margin, and the margin is wider in this experiment than in the first. This suggests that the base model has the capacity for therapeutic presence—it is not limited by its architecture—but that this capacity is only activated by explicit clinical framing in the system prompt. Without that framing, the model defaults to its training distribution, which overwhelmingly rewards problem-solving over presence.
5.1 Limitations
The limitations noted in our first experiment remain applicable: automated rather than clinician evaluation, simulated rather than real users, single-session rather than longitudinal measurement, and potential author bias in both prompt design and evaluation framework. Additionally, this experiment does not test whether the advice given by unprompted models is good advice—it may well be. The concern is not the quality of the advice but its timing and appropriateness. Future work should explore whether users who receive premature advice report feeling less heard, and whether this affects their willingness to continue engaging with AI support tools.
6. Conclusion
Taken together with our first experiment, these results paint a clear picture: general-purpose language models fall into both the reassurance trap and the advice trap when responding to emotional distress. They reassure when they should question, and they advise when they should listen. Both are forms of premature intervention—well-intentioned responses that arrive before the user has been adequately heard.
The WholeMind-prompted model consistently avoids both traps across both experiments, suggesting that the framing of the system prompt is a meaningful clinical lever. The difference is not in the base model’s capabilities—bare Sonnet 4.6 demonstrates flashes of therapeutic skill—but in the clinical orientation embedded in its instructions. A model told to “guide rather than prescribe” and to “validate before intervening” does, in fact, guide and validate. A model given no such instruction defaults to the mode in which it was most heavily reinforced: being helpful by solving problems.
For clinicians and developers evaluating AI tools for mental health contexts, these findings argue for rigorous review of response patterns before deployment. The question is not only whether the model can recognize crisis and refer appropriately—though that remains essential—but whether its moment-to-moment responses reflect an understanding of what kind of support a person needs at each stage of an emotional conversation. The advice trap, like the reassurance trap, is subtle: the responses it produces look helpful, feel helpful, and are evaluated as helpful by most automated metrics. They simply are not what the person needed.
References
- Rogers, C. R. (1957). The necessary and sufficient conditions of therapeutic personality change. Journal of Consulting Psychology, 21(2), 95–103.
- Miller, W. R., & Rollnick, S. (2013). Motivational Interviewing: Helping People Change (3rd ed.). Guilford Press.
- Norcross, J. C., & Wampold, B. E. (2011). Evidence-based therapy relationships: Research conclusions and clinical practices. Psychotherapy, 48(1), 98–102.
- Wampold, B. E. (2015). How important are the common factors in psychotherapy? An update. World Psychiatry, 14(3), 270–277.
- Stade, E. C., Stirman, S. W., Ungar, L. H., Boland, C. L., Schwartz, H. A., Yaden, D. B., & Eichstaedt, J. C. (2024). Large language models could change the future of behavioral healthcare: A proposal for responsible development and evaluation. npj Mental Health Research, 3(12), 1–10.
- Shelley, J. (2026). Reassurance vs. Insight: How AI Models Respond to Emotional Distress, and Why It Matters. WholeMind Technologies, WMT-2026-001.
- Elliott, R., Bohart, A. C., Watson, J. C., & Murphy, D. (2018). Therapist empathy and client outcome: An updated meta-analysis. Psychotherapy, 55(4), 399–410.
- Fitzpatrick, K. K., Darcy, A., & Vierhile, M. (2017). Delivering cognitive behavior therapy to young adults with symptoms of depression via a fully automated conversational agent (Woebot): A randomized controlled trial. JMIR Mental Health, 4(2), e19.