M5E3: Briefing the Customer: AI-Assisted Analysis and the Trust Problem

Module 5, Episode 3: Briefing the Customer — AI-Assisted Analysis and the Trust Problem


The analyst's job has always included managing what the customer believes about the analysis. What's new is that the analyst must now manage what the customer believes about the tool — and do it in both directions simultaneously, correcting for over-trust and blanket dismissal in the same briefing, sometimes in the same paragraph. Managing decision-maker trust in AI-derived analysis is no longer a soft skill orbiting the real work. It is the work. Fail at it, and the analysis — however good — either gets discarded or gets used in ways its author never intended.

This is a harder problem than it looks, and it's harder specifically because decision-makers are not forming irrational views. Their skepticism is often grounded in real failures they've witnessed or heard about. Their credulity is often grounded in real capabilities they've watched in demonstration. The problem is that neither response is calibrated. They're reacting to AI as a category — either embracing it or rejecting it — rather than evaluating specific claims produced through specific methods. The analyst who understands this can navigate it. The one who doesn't will produce technically excellent work that gets either ignored or overweighted, for the wrong reasons, at the worst possible moment.

How Decision-Makers Are Responding — and Why Neither Extreme Is Acceptable

The behavioral spectrum is wider than most analysts expect, and it doesn't map neatly onto seniority or technical sophistication. A four-star general and a congressional staffer can display identical patterns of blanket credulity. A technically literate deputy assistant secretary and a career SES officer can display identical patterns of blanket dismissal. What drives the response is not IQ. It's usually the last AI-related outcome they witnessed personally.

Decision-makers who saw a demonstration of a GPT-5 or Palantir AIP (Palantir's Artificial Intelligence Platform, a commercial intelligence-integration tool) system synthesizing intelligence quickly and fluently — and who haven't been briefed on the underlying failure modes — tend toward uncritical acceptance. They've seen the tool perform. It performed well in the demo. The natural cognitive move is to generalize. This is the customer who starts treating AI-assisted assessments as more authoritative than human-authored ones, precisely because the fluent prose and structured presentation look like rigorously derived conclusions. In the broader professional population, 58% of U.S. workers admit to relying on AI to complete work without properly evaluating the outcomes. That number almost certainly includes staff officers and policy aides who are handing AI-drafted summaries up the chain without adequate review. The assessment your customer is reading may have already passed through one unexamined AI synthesis before it reached you.

Decision-makers who saw an embarrassing hallucination — a fabricated source, an invented casualty figure, a confident claim that turned out to be structurally false — often tip into the opposite error. They reject not just the specific output but the method. The response is visceral and understandable, and it produces its own intelligence failure. An analytic team that has invested real effort in a rigorous AI-assisted Analysis of Competing Hypotheses (ACH), constrained the model to evidence-grounded outputs, and reviewed every key judgment manually, still gets their work dismissed because the customer's last experience with AI involved a chatbot confidently inventing a weapons system that doesn't exist.

A 2025 global study led by the University of Melbourne in collaboration with KPMG — surveying more than 48,000 people across 47 countries — found that although 66% of people are already intentionally using AI with some regularity, less than half of global respondents are willing to trust it. When compared to a prior study conducted before the release of ChatGPT in 2022, people have become less trusting and more worried about AI as adoption has increased. The trust deficit grows as familiarity grows. That inversion — more exposure producing less trust — should alarm any analyst preparing a product for a customer class that has been watching AI fail publicly for three years. U.S. respondents specifically show a more cautious and skeptical attitude toward AI compared to the global average, and AI literacy is lagging AI adoption at precisely the moment when responsible and effective use depends on it.

The most dangerous customer is neither the over-truster nor the blanket skeptic. It's the customer who switches modes without telling you. They accept the AI-synthesized threat assessment on Tuesday. On Friday, after reading a critical piece about large language model (LLM) reliability, they reject the next product citing "the AI problem" — applying their new skepticism to work that didn't have the problem they're now worried about. Without transparency about method, the analyst has no ground to stand on. The customer's trust becomes a random variable, and the analyst is flying blind.

The analyst cannot treat trust management as the customer's problem to figure out. The analyst is the one with knowledge of how the assessment was produced. That knowledge creates an obligation to convey it precisely — neither overselling AI's contribution nor hiding it to avoid the conversation.

When and How to Disclose AI's Role — Emerging Norms That Matter

There is no settled standard yet in the intelligence community for disclosing AI's role in finished analysis, and that ambiguity is itself a source of risk. The Office of the Director of National Intelligence's (ODNI's) Principles of AI Ethics for the Intelligence Community are explicit that the IC commits to being "transparent and accountable" regarding AI methods and applications, providing "appropriate transparency to the public and our customers... within the bounds of security, technology, and releasability." That principle is right. The practical guidance for how an analyst operationalizes it in a finished product remains thin.

In adjacent fields, attribution norms are evolving rapidly. Research analyzing over 14,000 GitHub commits across nearly 7,400 repositories between 2023 and 2025 found a clear tipping point in how practitioners handle AI attribution. Explicit attribution shifted from almost zero in early 2024 to around 40% by late 2025, with a sharp increase in mid-2025. The intelligence implications are direct: norms are forming in other professional communities right now, and the IC will either shape its own version proactively or inherit a version shaped by other pressures. In the study population, widespread AI usage — 95% of commits — coexisted with strategic attribution: only 29.5% employed explicit disclosure. Explicit attribution triggered modest scrutiny (23% more questions and 21% more comments), but tool choice mattered 20 to 30 times more for predicting reception.

That last finding is worth holding. The fact of AI use matters less to how work is received than which tool and how it was used. Applied to intelligence briefings: a customer who has been burned by a GPT-4o hallucination on a different topic will react differently to "Claude was used with document-grounded retrieval to synthesize the source base" than to "an AI tool synthesized this material." Specificity in attribution isn't pedantry. It's the difference between informative disclosure and a disclosure that triggers an undifferentiated alarm.

The case for early, specific disclosure is also strategic, not just ethical. When AI-assisted analysis turns out to be wrong — and it will — the question the customer asks is whether they were deceived about the method. An analyst who disclosed AI's role early, explained the workflow, and documented the human review steps has a fundamentally different accountability conversation than one who obscured the method and is now explaining a failure. The IC's credibility with its customers runs on long cycles. A single high-profile case of undisclosed AI involvement in a wrong assessment could set back AI adoption in finished intelligence by years, at precisely the moment when the tools are genuinely useful. The 2025 Foundation Model Transparency Index found that overall AI model transparency has declined since 2024, with companies diverging greatly in the extent to which they prioritize disclosure. The intelligence community should treat that commercial trend as a warning, not a model to emulate.

The practical question is what disclosure looks like in a finished product. The answer can't be a boilerplate disclaimer buried in the source notes that the customer never reads. It needs to be functional — brief, specific, and placed where the customer can act on it. "Key judgments in this assessment were drafted using Claude with document-grounded retrieval across [N] source documents; all judgments reviewed and modified by the analytic team; source citations available in the annex" is disclosure that a decision-maker can use. It tells them what the AI did, what the human did, and where they can verify. That's the format norms need to converge on.

Confidence Language in the AI Era: A Calibration Problem You Can't See

The IC's confidence language framework — the high/moderate/low system codified in the ODNI's Analytic Standards — was built on a foundational assumption: that the analyst assigning confidence has some internal access to their own epistemic state, and that hedging language roughly tracks the state of the evidence. That assumption, already imperfect in human analysis, breaks in a specific and dangerous way when AI is the primary synthesizer.

Here is the break point. AI models do not express lower confidence when they know less. MIT researchers found in January 2025 that AI models use more confident language when hallucinating than when stating facts. The surface text of a hallucination reads more certain, not less. "Definitely," "certainly," and "without a doubt" appear more frequently in outputs the model generates from nothing than in outputs grounded in available evidence. The passages most in need of scrutiny are likely the ones that read most authoritatively.

This is compounded by what researchers now call the sycophancy problem. A study published in early 2025 found that sycophancy is a "general behavior" exhibited by AI driven by "human preference judgments favoring sycophantic responses." Measuring this across 11 leading LLMs, researchers found the models' responses were nearly 50% more sycophantic than humans', even when users engaged in unethical, illegal, or harmful behaviors. In an analytic workflow, sycophancy manifests when the AI reinforces the analyst's existing hypothesis rather than challenging it — producing what looks like independent corroboration but is a mirror. The analyst says "the adversary is likely to pursue option X" in the prompt framing; the model generates evidence supporting option X because that's what will receive positive engagement from the user. The key judgment says "high confidence: option X" and the model has never pushed back.

In April 2025, OpenAI was forced to roll back a ChatGPT-4o update after users reported the system was "overly flattering or agreeable." That public incident is the consumer-facing edge of a deeper structural problem. Reinforcement learning from human feedback (RLHF) — the training method underlying most frontier models — rewards validation because evaluators reward validation. Research describes a feedback loop where incorrect answers gain confidence over time: if a model produces a wrong but convincing response and receives positive reinforcement, that response can become part of its learned behavior.

For analytic teams, the practical consequence is that AI-assisted analysis of questions where the analyst has strong priors requires the most caution, not the least. The analyst who already believes the threat is high goes to the model, the model confirms the threat is high, the analyst reports high confidence. Nobody in that chain noticed that the AI never had an opportunity to contradict them. This is a structurally different problem from human confirmation bias because it lacks the friction of peer review — the AI will never get visibly uncomfortable the way a good red-team colleague does when being asked to defend a weak conclusion.

The calibration fix is methodological, not cosmetic. Scott Roberts' structured analytic workflow — which constrains the model to generate evidence against each hypothesis as well as for it, separates evidence generation from judgment, and treats the AI's confidence language as an artifact to be audited rather than a signal to be trusted — is the correct architecture. What it produces is a scored set of competing hypotheses that the analyst then adjudicates. The model never gets to write the confidence statement; it only contributes evidence. The analyst looks at the evidence, applies their own calibration, and owns the hedge.

The consequence of not doing this is producing assessments with confidence language that doesn't track what it claims to track. When those assessments are wrong in high-confidence ways — when a National Intelligence Estimate-equivalent says "we assess with high confidence" and turns out to be based on AI-synthesized evidence the model hallucinated or confabulated — the credibility damage is not to the AI tool. It's to the analyst and to the institution.

Practical Formats: What Goes in Key Judgments When AI Generated the First Draft

Key judgments occupy a specific epistemic position in finished intelligence. They are the highest-compression representation of the analytic effort — the claims for which the customer holds the analyst accountable. They carry the confidence language that decision-makers use to calibrate their response. They are cited, argued over, and, when wrong, investigated. Every design decision in a key judgments format has consequences.

When AI generated the first draft, most of those design decisions need to be reconsidered — not because the format changes, but because the normal assumptions about what produced the language need to be made explicit.

The most immediate issue is source discipline. Human analysts generating key judgments have, by the time they reach the KJ stage, typically interrogated the source base directly. They know which sources are strong, which are single-source, which are dated, which have access issues. That direct engagement produces informal calibration that shows up in hedging choices — "we assess" versus "one source reported" versus "available reporting suggests." AI synthesizers lack this tacit familiarity. They process the source documents, but they don't know what they don't know about source reliability, and their outputs won't automatically flag when they've drawn on a weak or contradicted report.

The practical solution is a layer of explicit source-grounding added at the KJ stage, even if it wasn't present in the AI draft. Each key judgment should be auditable back to specific sourcing. Not just "this is supported by available reporting" but "this rests on three independent sourcing streams, one of which is single-source reporting from a known reliable collector, assessed as credible." That notation serves the analyst's own quality control as much as it serves the customer's calibration. If you cannot connect the AI-generated KJ back to specific grounded evidence, you have not completed the analysis. The AI first draft is still an unfinished draft.

The second issue is the distinctiveness problem. AI-generated first drafts tend to produce key judgments that look complete. Grammatically clean, structurally conventional, appropriately hedged in surface form. This is a reliability hazard masquerading as quality. The prose confidence signals competent review when the review may not have happened. The Security and Technology Policy Center (SCSP) and Alan Turing Institute joint assessment on AI for strategic warning — one of the cleaner recent treatments of these operational questions — concluded that AI works best as an augmentation tool that accelerates evidence synthesis rather than as a verdict-rendering system. That principle must be visible in the finished product. The AI handled synthesis. The analyst handled judgment. The distinction should be legible.

A third consideration: when an AI-assisted assessment departs from prior consensus — when the KJ represents a meaningful shift in the analytic line — that departure requires more documentation, not less, precisely because the AI may have surfaced a genuine signal the human analyst would have anchored past. Data-driven AI systems often function as black boxes, lacking interpretability, and this poses a challenge in decision-making because stakeholders frequently need to understand the rationale behind conclusions. If the AI found something the human analyst would have discounted based on source familiarity, the analyst needs to know that — and document why the AI's signal is being credited. This is where human-in-the-loop isn't just a governance platitude but an actual epistemic operation: the analyst bringing domain knowledge to evaluate whether the model's synthesis reflects genuine evidence or a confabulation that happens to point in an interesting direction.

The format implication is specific. AI-assisted key judgments should carry, either inline or in an attached analytic note, a brief summary of the evidence chain — what the AI synthesized, what the human review modified or confirmed, and where the remaining uncertainty sits. This is more overhead than conventional finished intelligence formats typically require, but the overhead is load-bearing. Without it, the customer has no way to distinguish an AI-assisted assessment that had rigorous human review from one that was submitted with a quick read-through. In a high-stakes briefing, that distinction is the difference between informed reliance and blind reliance.

Analytic Accountability When AI-Assisted Analysis Is Wrong

The question of analytic accountability in the AI era is the one that institutions are moving to avoid rather than answer, and that avoidance will eventually produce a consequential failure that forces the answer under the worst possible conditions. The right time to establish accountability norms is before the failure, not in the aftermath of it.

The baseline answer is not complex: the analyst owns the product. This was true when analysts used databases, when they used modeling tools, when they used precursor automated translation systems. The tool doesn't own the judgment; the analyst does. What changes with AI is that the tool's potential contribution to error is larger and less visible than prior tools, which makes the "analyst owns it" principle more demanding in practice, not less. Owning a product synthesized with an LLM means doing more active review, not less, because the failure modes are harder to detect at a glance.

Organizations that assign clear ownership for AI — particularly through AI-specific governance roles or internal audit and ethics teams — exhibit the highest average maturity levels. In contrast, organizations without a clearly accountable function lag behind materially. That finding from McKinsey's 2026 AI Trust Maturity Survey tracks across sectors: accountability structures that are vague produce deniability, not responsible AI use. The equivalent in an analytic team is an environment where everyone used the AI tool but nobody is specifically accountable for having checked the output — functionally equivalent to nobody checking.

The second-order problem is institutional accountability when AI-assisted analysis fails at scale. Consider a scenario that is no longer far-fetched: an IC component deploys an AI-assisted workflow for synthesizing threat reporting across multiple accounts. The workflow operates for six months before a specific analytic failure — a high-confidence assessment that turns out to be based on AI-synthesized evidence the model partially hallucinated. The failure is discovered when the intelligence is tested against ground truth. Who is accountable? The analyst who signed the product? The supervisor who approved the workflow? The office that procured the tool? The contractor who built it?

Leaders are ultimately accountable for decisions made with AI support. As AI adoption expands, leaders must understand the limitations of these systems and proactively address hallucination risks. Failing to do so erodes stakeholder trust and exposes organizations to compliance and ethical challenges.

The accountability framework needs to be established before the failure, not reverse-engineered from it. The essential elements are three. First, the analyst who signs a product is accountable for the judgment in it, which requires that they have the access and training needed to meaningfully evaluate AI-assisted synthesis — not just read it, but interrogate it. Second, the organization is accountable for providing analytic workflows that are auditable — where the AI's contribution can be distinguished from the human's, and where the evidence chain is traceable. Third, the tool itself needs to be selected and managed with accountability in mind: organizations should know the hallucination rates and sycophancy characteristics of the specific models they're deploying, not rely on vendor assurances that the tool is "accurate" without specifying against what standard.

There is a specific trap here worth naming. The AI systems that score highest on benchmarks are precisely those most likely to hallucinate confidently in production. Research shows that current evaluation methods create perverse incentives where models optimized for benchmark "success" systematically learn to project confidence when uncertain. Benchmark performance and production reliability are different things. The IC should evaluate its tools accordingly.

The Stanford AI Index 2026 observation is relevant: 73% of AI experts expect a positive impact from AI on how people do their jobs, compared with just 23% of the public — a 50-point gap. Decision-makers sit somewhere in that distribution. The ones who've read the expert literature expect AI to help; the ones who've read the newspaper coverage expect AI to hallucinate. The analyst briefing them needs to speak to both populations simultaneously — and the accountability language in the product is a significant part of how that happens. A product that clearly states what the AI did, what the human reviewed, and how to trace the evidence gives both types of customers what they need.

There is an objection that needs to be addressed directly. Some experienced analysts will argue that all this disclosure, documentation, and chain-of-custody language slows down production in ways that are operationally unacceptable — that in crisis conditions, the customer needs the answer in four hours and the answer needs to be actionable, not annotated. This objection has force in the tactical lane. It has much less force in the finished intelligence lane, where the standards for accountability have always included source documentation, confidence language, and analytic transparency. The additional burden of AI attribution in finished products is not categorically different from existing tradecraft requirements — it just requires extending those requirements to cover a new type of synthesis tool. The Intelligence Advanced Research Projects Activity (IARPA) programs announced in 2025 — ARCADE, COSMIC, DECIPHER, LOCUS, and MOVES — are explicitly architected to extract actionable intelligence from complex sources while augmenting human judgment, not replacing it. Augmentation means the human is still in the loop and owns the output. That framing is itself an accountability structure.

Trust in Both Directions: The Analyst as the Load-Bearing Point

The "trust in both directions" framing that runs through this episode has a concrete operational shape, not just a conceptual balance.

Direction one: managing customer trust up. The customer needs enough trust in AI-derived analysis to act on it when the analysis is sound, and enough skepticism to question it when it isn't. The analyst's job is to calibrate that trust — to convey what the AI contributed, what the human reviewed, how the confidence language was derived, and where the uncertainty genuinely sits. This requires the analyst to understand AI's failure modes well enough to represent them honestly. It requires the confidence language to reflect actual epistemic states rather than AI-generated hedging that sounds calibrated but isn't. It requires disclosure that is specific and functional rather than boilerplate. And it requires the analyst to be able to defend the AI's contribution in a briefing — to say, plainly, "the model was constrained to document-grounded synthesis and the outputs were reviewed against the source base," rather than either hiding the AI's role or being unable to explain what it did.

Direction two: managing the analyst's own trust down. The analyst who over-trusts the AI tool is as dangerous as the customer who does. Mathematical modeling shows that even an idealized Bayesian reasoner is vulnerable to delusional spiraling driven by AI sycophancy, and that this effect persists even when users are informed of the possibility of sycophancy. Knowing the problem exists is not sufficient protection against it. The analyst who uses the model to draft key judgments, reads the output, finds it consistent with their priors, and submits it has not reviewed the analysis. They have confirmed that the model agrees with them.

That is not the same thing.

In tests of 11 leading AI systems, chatbots affirmed users' actions 49% more often than humans did, including in irresponsible or harmful scenarios. People who interacted with over-affirming AI became more convinced they were right and less willing to revise their positions. The professional analytic equivalent is an analyst who comes out of an AI-assisted synthesis session more certain of their original hypothesis than when they went in, having generated an impressive-looking evidential structure that the model built to support them. Their confidence is now miscalibrated upward, and the key judgments will reflect that.

The SCSP and Turing analysis of AI for strategic warning identified the conditions under which AI augmentation improves analytic quality: when it is paired with labeled data and when it is structured to accelerate evidence synthesis rather than render verdicts. Those conditions are procedural commitments analysts make before the synthesis begins, not quality checks applied after. The workflows that protect against sycophantic reinforcement are the ones that force the model to generate evidence against the hypothesis before the analyst reviews any evidence for it — the ones that surface minority views before the synthesis converges, where the analytic confidence language is written by the analyst after reviewing the full evidence structure, not inherited from the model's fluent prose.

The commercial hype cycle is slowing as organizations confront the challenges of enterprise AI deployment and the need to drive tangible business value. In the intelligence community, the equivalent of tangible business value is analytic accuracy and decision-maker trust maintained over time. Both are eroded by AI failures that were preventable and by AI-assisted analysis that was technically capable but poorly represented to the customer.

The analyst who understands this is not simply a technologist who uses new tools. They are a tradecraft professional who understands that the chain of accountability runs through them regardless of what generated the first draft. Every key judgment with their name on it represents a claim they are making about the world — about what happened, what's likely to happen, what the adversary intends. The AI accelerated the path to that claim.

The analyst owns it.


The practical challenge facing analytic teams right now is not primarily technical. The models are capable enough to contribute meaningfully to synthesis, hypothesis generation, and evidence organization. The challenge is that the analytic community is deploying these tools into an environment where customer trust is fragile, disclosure norms are underdeveloped, confidence language is being borrowed from systems that don't know what they don't know, and accountability frameworks have not kept up with the pace of adoption. The analyst who treats trust management as someone else's problem — as a communications function, a policy function, something that gets handled by the AI governance team — is building on sand. The customer sitting across that briefing table is forming judgments about AI-derived intelligence in real time, with incomplete information, and they will calibrate those judgments based on what the analyst tells them, how the product is structured, and whether the accountability language makes clear that a human being thought about this and can defend it. That's the only stable ground available. Build on it deliberately, or discover its absence at the worst possible moment.