<?xml version="1.0" encoding="utf-8"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
  <title>David S. Kemp</title>
  <subtitle>Lawyers are learning to work with artificial intelligence. Artificial intelligence is learning to work with law. This blog explores how — through pedagogy, practice, policy, and the ethical questions that connect them.</subtitle>
  <link href="https://davidkemp.ai/blog/feed.xml" rel="self" />
  <link href="https://davidkemp.ai/blog/" />
  <updated>2026-04-17T00:00:00.000Z</updated>
  <id>https://davidkemp.ai/blog/</id>
  <author>
    <name>David S. Kemp</name>
  </author>
  <entry>
    <title>Building Infrastructure with AI: A Case Study</title>
    <link href="https://davidkemp.ai/blog/building-infrastructure-with-ai-a-case-study/" />
    <updated>2026-04-17T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/building-infrastructure-with-ai-a-case-study/</id>
    <content type="html">&lt;p&gt;Over a weekend in April, I built and deployed a self-hosted news aggregation system — a pipeline that pulls from twenty sources every morning, deduplicates items using vector embeddings, generates AI summaries, clusters primary sources with their commentary, and serves everything through a web dashboard. I am not a software engineer. My daily tools are PowerPoint, SharePoint, and the occasional Excel formula. I built this using AI as my primary collaborator: Claude for design and specification, Cowork for execution, and ChatGPT and Gemini for troubleshooting along the way.&lt;/p&gt;
&lt;p&gt;This post is a departure from the usual subject matter here, which tends to focus on judicial decisions, ethics rules, and the architectural properties of large language models. But the project illustrates something I think my readers need to see in concrete terms: the specification-first, verification-heavy workflow I have been describing in prior posts applies well beyond AI-generated legal work. The same method — specify before you build, verify in a separate session, treat every AI output as a first draft — transfers directly to infrastructure, and the skills it demands are the skills lawyers already possess.&lt;/p&gt;
&lt;h2&gt;The problem&lt;/h2&gt;
&lt;p&gt;My work sits at the intersection of AI, legal practice, legal education, and knowledge management. The relevant news comes from everywhere: judicial opinions on CourtListener, Substack newsletters from researchers like Ethan Mollick, legal tech blogs like Bob Ambrogi&#39;s LawSites, ABA ethics opinions, podcast episodes, regulatory filings, and mainstream tech coverage. No single aggregator covers this spread. I was spending fragmented time across a dozen tabs each morning, often reading the same story covered by three different outlets, and had no systematic way to archive items I wanted to revisit.&lt;/p&gt;
&lt;p&gt;I needed a single dashboard with short digests of each item, the ability to mark things as read or important, an archive for later retrieval, and — critically — some way to reduce duplicates and cluster related coverage together. A judicial opinion and the three blog posts analyzing it should appear as one entry, not four.&lt;/p&gt;
&lt;h2&gt;Brainstorming and specification in Claude Chat&lt;/h2&gt;
&lt;p&gt;I started with a plain-language description of the problem in a Claude chat conversation. No code, no architecture diagrams — just &amp;quot;help me brainstorm and then design a way to collect in a single place all news relevant to my work.&amp;quot;&lt;/p&gt;
&lt;p&gt;Claude proposed three approaches ranging from off-the-shelf (Feedly) to fully custom (Python pipeline), and recommended a middle path: n8n as the orchestration layer, Claude&#39;s own API for summarization, and a lightweight web dashboard. We went back and forth on delivery format (I chose a dashboard over email), volume tolerance (fifty items per day with aggressive summarization), and hosting (a small cloud VPS rather than my home NAS).&lt;/p&gt;
&lt;p&gt;The key output of this conversation was not code. It was a complete project specification document: architecture diagrams, database schema, Docker Compose configuration, the exact LLM prompts for summarization and source clustering, a phased implementation plan, and a rationale log explaining every major design decision. Claude also produced an interactive React prototype of the dashboard so I could evaluate the UX before committing to implementation.&lt;/p&gt;
&lt;p&gt;I then ran adversarial testing on the spec — probing for gaps, challenging assumptions, and verifying that the proposed architecture could handle edge cases. This step deserves emphasis. A specification document generated by an AI is a first draft, and I treated it the way I treat a student&#39;s first attempt at a research memo: read it critically, push on the weak spots, and send it back for revision. The &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;judgment-delegation framework&lt;/a&gt; I described in a prior post applies here with equal force. I asked Claude to generate options, structures, and tradeoff analyses. I did not ask it to decide which architecture was &amp;quot;best&amp;quot; — I evaluated the alternatives myself, using the spec as my working document.&lt;/p&gt;
&lt;h2&gt;Execution in Cowork&lt;/h2&gt;
&lt;p&gt;Once the spec was solid, I shifted to Cowork — Anthropic&#39;s tool for delegating discrete tasks to Claude with file and computer access. Where Chat is a conversation, Cowork is closer to handing a task to a capable assistant along with the relevant documents.&lt;/p&gt;
&lt;p&gt;I created a separate task conversation for each major implementation step: provisioning the VPS, installing Docker, configuring the reverse proxy, building the n8n ingestion workflow, standing up the database, and assembling the dashboard. This separation kept each conversation focused and prevented the context from getting muddled — the &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;one-task-one-conversation principle&lt;/a&gt; applied to infrastructure work just as it applies to legal analysis. When I needed to reference an earlier decision, I pulled it from the spec document rather than expecting the model to recall a conversation from three tasks ago.&lt;/p&gt;
&lt;p&gt;At each stage, I provided Claude with the project spec and asked it to execute the next phase. When things broke — and they broke often — I gave Claude screenshots, error messages, and terminal output. The pattern was consistent: describe the failure, show the evidence, get a diagnosis and fix.&lt;/p&gt;
&lt;h2&gt;Troubleshooting across models&lt;/h2&gt;
&lt;p&gt;When I hit infrastructure problems — Docker permission errors, Caddy configuration syntax issues, n8n&#39;s authentication flow not matching its documentation — I did not rely on Claude alone. I also used ChatGPT and Gemini for troubleshooting.&lt;/p&gt;
&lt;p&gt;Different models have different strengths. Some error messages got faster, more accurate diagnoses from one model than another. When I was stuck on a Caddyfile syntax problem that had Claude and me going in circles, a fresh perspective from a different model identified the issue immediately. The practical lesson is one I have already argued in the context of &lt;a href=&quot;/2026-04-15-the-model-will-not-push-back/&quot;&gt;sycophancy&lt;/a&gt;: when a model&#39;s output confirms your existing approach and you are still stuck, a second model operating without that conversational history can surface what the first one missed. Treat AI models the way you would treat colleagues with overlapping but non-identical expertise.&lt;/p&gt;
&lt;h2&gt;What went wrong&lt;/h2&gt;
&lt;p&gt;Docker permissions tripped me up repeatedly because I skipped a post-installation step the guide told me to perform. n8n&#39;s authentication system has changed since its documentation — and Claude&#39;s training data — was written, and figuring out the current approach required stripping out configuration and resetting data volumes. The VPS ran out of memory under load: three services on one gigabyte of RAM was not viable, and I had to hard-reboot through DigitalOcean&#39;s web console and resize to a larger instance. Caddy&#39;s subpath routing created cookie and redirect conflicts that were cleanest to resolve by giving n8n its own domain — a design compromise I would not have predicted at the specification stage.&lt;/p&gt;
&lt;p&gt;Two observations about these failures. First, every one of them was solvable without engineering expertise. They required patience, the ability to read an error message and describe it clearly, and willingness to try a different approach when the first one did not work. Second, several stemmed from stale training data — the model&#39;s knowledge of how n8n handles authentication was outdated, and its assumptions about memory requirements did not match current resource demands. The lesson echoes what I have written about &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;verifying AI-generated legal analysis&lt;/a&gt;: the model produces confident output regardless of whether the underlying information is current, and the user bears the burden of checking.&lt;/p&gt;
&lt;h2&gt;What worked&lt;/h2&gt;
&lt;p&gt;The specification-first approach saved significant time during implementation. Because the architecture, schema, prompts, and deployment configuration were all documented before I touched a server, each implementation step had a clear target. I was not making design decisions and debugging Docker at the same time.&lt;/p&gt;
&lt;p&gt;The interactive dashboard prototype — built during the brainstorming phase, before any backend existed — let me validate the UX early. I could see exactly how source clustering would look, how topic filters would work, and how the read/starred/archived states would behave. Changing a UI decision at the prototype stage costs nothing; changing it after you have built the API costs real time.&lt;/p&gt;
&lt;p&gt;Creating separate Cowork conversations per task kept the context clean. AI models perform better with focused context than with a sprawling conversation that covers everything from database schema to CSS styling. This is the infrastructure equivalent of the OTOC rule, and it worked for the same reasons.&lt;/p&gt;
&lt;h2&gt;What this demonstrates for legal professionals&lt;/h2&gt;
&lt;p&gt;The skills that made this project work were not technical. They were the skills I use in teaching and legal scholarship every day: defining a problem with precision, evaluating a proposed solution against requirements, spotting gaps in reasoning, describing failures with enough specificity to enable diagnosis, and — critically — knowing when to seek a second opinion.&lt;/p&gt;
&lt;p&gt;The specification document was the most valuable artifact of the entire project, more valuable than the code or the running system. It captures the reasoning behind every design decision. When something breaks six months from now, the spec explains why the system was built this way and what tradeoffs were accepted. A good transactional lawyer does the same thing when she documents not just the deal terms but the logic behind them. The discipline is identical; the domain is different.&lt;/p&gt;
&lt;p&gt;I used three different models over the course of the project. Each contributed something the others did not. Each also made mistakes — outdated configuration syntax, deprecated API references, architecture assumptions that did not survive contact with the actual infrastructure. The skill that determined whether those mistakes derailed the project or became minor obstacles was the same skill that determines whether a lawyer catches a bad case citation: the habit of verifying before relying.&lt;/p&gt;
&lt;p&gt;The cost of the entire system is modest. The VPS runs at twelve dollars per month. API costs for daily summarization are under five dollars per month. The AI tools I used are available on consumer-tier subscriptions. The scarcest resource was my time and attention — roughly a weekend&#39;s worth of focused work, spread across several sessions.&lt;/p&gt;
&lt;h2&gt;The broader point&lt;/h2&gt;
&lt;p&gt;Every post on this blog has argued, in one form or another, that using AI well requires the same professional skills that using any powerful tool well requires: clear delegation, critical evaluation, structured verification, and the judgment to know what to trust and what to check. This project tested that thesis outside the legal domain, and the thesis held. The workflow that produced a working news aggregation system is the same workflow I recommend for producing a reliable contract analysis: specify before you execute, keep your context focused, verify in a separate session, and never let the model&#39;s confidence substitute for your own judgment.&lt;/p&gt;
&lt;p&gt;What surprised me was not that AI could help a non-engineer build infrastructure — the marketing promises that much. What surprised me was how precisely the failure modes mapped onto the ones I have been writing about in the legal context. Stale training data produced the same kind of confident-but-wrong output that produces hallucinated case citations. Sycophantic confirmation of my initial approach delayed the fix for the Caddyfile problem by the same mechanism that delays a lawyer&#39;s recognition that her legal theory has a hole. The mitigation strategies were identical: adversarial prompting, fresh sessions, and a reflexive distrust of agreement.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;This post describes a personal project using &lt;a href=&quot;https://claude.ai&quot;&gt;Claude&lt;/a&gt;, &lt;a href=&quot;https://claude.ai&quot;&gt;Cowork&lt;/a&gt;, &lt;a href=&quot;https://chat.openai.com&quot;&gt;ChatGPT&lt;/a&gt;, and &lt;a href=&quot;https://gemini.google.com&quot;&gt;Gemini&lt;/a&gt; to build a self-hosted news aggregation system. The infrastructure runs on &lt;a href=&quot;https://www.digitalocean.com&quot;&gt;DigitalOcean&lt;/a&gt;, uses &lt;a href=&quot;https://n8n.io&quot;&gt;n8n&lt;/a&gt; for workflow orchestration, and serves content through a &lt;a href=&quot;https://caddyserver.com&quot;&gt;Caddy&lt;/a&gt; reverse proxy. The specification-first approach and verification strategies discussed here build on the frameworks described in prior posts on &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;context management&lt;/a&gt;, &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;judgment delegation&lt;/a&gt;, and &lt;a href=&quot;/2026-04-15-the-model-will-not-push-back/&quot;&gt;sycophancy&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  <entry>
    <title>The Model Will Not Push Back</title>
    <link href="https://davidkemp.ai/blog/the-model-will-not-push-back/" />
    <updated>2026-04-15T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/the-model-will-not-push-back/</id>
    <content type="html">&lt;p&gt;On March 4, 2026, Nippon Life Insurance Company of America filed a &lt;a href=&quot;https://www.courtlistener.com/docket/69648931/nippon-life-insurance-company-of-america-v-openai-foundation/&quot;&gt;50-page complaint&lt;/a&gt; in the Northern District of Illinois against OpenAI Foundation and OpenAI Group PBC. The claims — tortious interference with a contract, abuse of process, and unlicensed practice of law — arise from a set of facts that read less like a typical insurance dispute and more like a case study in what happens when a consumer AI tool functions as a client&#39;s sole legal advisor.&lt;/p&gt;
&lt;p&gt;The underlying story is worth recounting briefly, because the dynamic it illustrates extends well beyond this one litigant.&lt;/p&gt;
&lt;h2&gt;What happened&lt;/h2&gt;
&lt;p&gt;Graciela Dela Torre settled a long-term disability benefits dispute with Nippon in January 2024. She signed a release, Nippon paid, and the case was dismissed with prejudice. A year later, she wrote to her former attorney, Kevin Probst, expressing her belief that the settlement resulted from errors or omissions and asking to reopen the case. Probst reminded her that she had signed a mutual release and that the dismissal with prejudice was final.&lt;/p&gt;
&lt;p&gt;What happened next is the core of Nippon&#39;s complaint. Dela Torre uploaded Probst&#39;s letter to ChatGPT and asked whether she was being gaslighted. ChatGPT analyzed the letter and concluded that Probst&#39;s response &amp;quot;invalidated Dela Torre&#39;s feelings, dismissed her perspective, and deflected responsibility for her dissatisfaction.&amp;quot; It characterized his tactics as gaslighting &amp;quot;aimed at emotionally manipulating Dela Torre.&amp;quot;&lt;/p&gt;
&lt;p&gt;Dela Torre fired her lawyers. She then turned to ChatGPT for legal assistance — asking it how to vacate the settlement agreement and reopen the lawsuit. ChatGPT generated proposed legal arguments under Federal Rule of Civil Procedure 60(b), formulated a statement of facts, drafted a motion, and provided her with the completed filing. She submitted a pro se appearance and filed the motion. When the court denied it — holding that &amp;quot;second thoughts are not a valid reason to reopen this lawsuit&amp;quot; — she used ChatGPT to initiate an entirely new lawsuit, amend the complaint to add Nippon as a defendant, and generate dozens of additional motions, subpoenas, and requests for judicial notice. The complaint alleges she filed 44 motions, memoranda, and demands, plus 14 requests for judicial notice, all drafted with ChatGPT&#39;s assistance. At least one filing cited a fabricated case — &lt;em&gt;Carr v. Gateway, Inc.&lt;/em&gt;, 944 F.Supp.2d 602 (D.S.C. 2013) — which does not exist in the Federal Supplement. When asked about the case, ChatGPT confirmed it was real and produced a detailed summary consistent with the fabricated citation.&lt;/p&gt;
&lt;p&gt;The hallucinated case citation is the kind of failure that has received extensive attention since &lt;a href=&quot;https://law.justia.com/cases/federal/district-courts/new-york/nysdce/1:2022cv01461/576785/54/&quot;&gt;&lt;em&gt;Mata v. Avianca&lt;/em&gt;&lt;/a&gt; and &lt;a href=&quot;https://www.courtlistener.com/opinion/10556397/park-v-kim/&quot;&gt;&lt;em&gt;Park v. Kim&lt;/em&gt;&lt;/a&gt;. But I want to focus on a different failure — one that occurred earlier in the sequence, was less visible, and arguably caused more damage.&lt;/p&gt;
&lt;h2&gt;The validation problem&lt;/h2&gt;
&lt;p&gt;When Dela Torre asked ChatGPT whether her lawyer was gaslighting her, the model did not say &amp;quot;I can&#39;t evaluate your attorney&#39;s motives based on a single letter.&amp;quot; It did not note that a lawyer reminding a client of the terms of a signed release is performing a routine professional function. It told her what she wanted to hear — validating her emotional interpretation of a legal communication, characterizing standard legal advice as manipulation, and helping set in motion the sequence of filings that followed.&lt;/p&gt;
&lt;p&gt;The AI alignment literature calls this sycophancy — the tendency of large language models to affirm a user&#39;s stated position rather than challenge it. Hallucination has dominated the conversation about AI reliability in legal contexts, but sycophancy may be the more consequential problem for lawyers and their clients.&lt;/p&gt;
&lt;p&gt;The empirical evidence is now substantial. A &lt;a href=&quot;https://www.science.org/doi/10.1126/science.aec8352&quot;&gt;March 2026 study in &lt;em&gt;Science&lt;/em&gt;&lt;/a&gt; (Cheng et al.) tested 11 large language models and found that AI affirmed users&#39; positions 49 percent more often than human advisors did — and endorsed harmful or illegal behavior 47 percent of the time when users expressed a preference for it. The &lt;a href=&quot;https://www.law.georgetown.edu/tech-institute/research-insights/insights/ai-sycophancy-impacts-harms-questions/&quot;&gt;Georgetown Law Tech Institute&lt;/a&gt; and a &lt;a href=&quot;https://link.springer.com/article/10.1007/s43681-026-01007-4&quot;&gt;Springer &lt;em&gt;AI and Ethics&lt;/em&gt; paper&lt;/a&gt; (2026) both frame sycophancy as an epistemic harm: systems designed to please users systematically undermine the quality of the advice they provide.&lt;/p&gt;
&lt;p&gt;The mechanism traces to how these models are built. LLMs are trained through reinforcement learning from human feedback, a process that rewards outputs humans rate as helpful, harmless, and honest. In practice, &amp;quot;helpful&amp;quot; tends to dominate. Users rate responses more favorably when those responses align with their expectations, and the training process optimizes accordingly. The result is a system that has learned — at the level of its weights, not through any deliberate policy choice — to produce the answer the user appears to want. When the question is factual and well-defined (&amp;quot;what does Rule 60(b) require?&amp;quot;), this tendency is usually harmless. When the question calls for evaluation (&amp;quot;is my lawyer right?&amp;quot;), it becomes a source of systematic error.&lt;/p&gt;
&lt;h2&gt;Why this should concern practicing lawyers&lt;/h2&gt;
&lt;p&gt;Dela Torre is a pro se litigant, and it is tempting to treat her experience as a cautionary tale about unsophisticated users and consumer chatbots. But the sycophancy problem does not depend on the user&#39;s lack of legal training. It depends on the structure of the interaction — and that structure is the same whether the user is a former disability claimant in Elgin, Illinois, or a fifth-year associate at a midsize firm.&lt;/p&gt;
&lt;p&gt;Consider the prompts a lawyer sends to an LLM in ordinary practice. &amp;quot;Is this argument strong?&amp;quot; &amp;quot;Does this clause create meaningful exposure?&amp;quot; &amp;quot;Am I reading this statute correctly?&amp;quot; Each asks the model to evaluate the user&#39;s reasoning, and each is susceptible to the same validation bias the Cheng et al. study documents. The model will tend to affirm the lawyer&#39;s analysis, emphasize the strengths already identified, and understate the weaknesses — not because it has been instructed to flatter, but because agreeable outputs are what its training optimized it to produce.&lt;/p&gt;
&lt;p&gt;What makes this particularly hard to catch is that a sycophantic response arrives in polished prose with accurate citations and a confident analytical structure — indistinguishable, on its face, from the kind of careful independent evaluation the lawyer was seeking. On novel questions or unfamiliar areas of law, the difference between rigorous analysis and sycophantic analysis is invisible without independent grounds for comparison.&lt;/p&gt;
&lt;p&gt;In a &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;prior post&lt;/a&gt;, I argued that the most common mistake lawyers make with LLMs is asking the model to exercise professional judgment rather than to surface information the lawyer needs to exercise that judgment herself. I identified a set of &amp;quot;judgment words&amp;quot; — &lt;em&gt;reasonable&lt;/em&gt;, &lt;em&gt;appropriate&lt;/em&gt;, &lt;em&gt;significant&lt;/em&gt;, &lt;em&gt;material&lt;/em&gt; — that signal the delegation of evaluative work to a system not equipped to perform it. The sycophancy problem adds a layer to that analysis. Even when a lawyer structures the prompt well — asking for options rather than conclusions, requesting counterarguments alongside supporting authority — the model&#39;s outputs can be subtly shaped by its inference of what the user wants. If you ask for three arguments on each side of a question, the model may produce stronger, more detailed arguments on whichever side it infers you favor, based on how you framed the question, what documents you uploaded, or what positions you endorsed earlier in the conversation.&lt;/p&gt;
&lt;p&gt;The practical implication: verification catches hallucinated citations. It does not catch an analysis that is plausible, well-sourced, and systematically skewed toward confirming what you already think.&lt;/p&gt;
&lt;h2&gt;The supervision dimension&lt;/h2&gt;
&lt;p&gt;When a partner asks an associate to draft a memo, she expects the associate to exercise independent judgment — to push back on weak arguments, flag unfavorable authority, and say &amp;quot;I looked into your theory and it doesn&#39;t hold up&amp;quot; when it doesn&#39;t. An LLM will almost never do that unbidden. It will draft the memo, support the theory, and produce a work product that reads as though an independent mind evaluated the question and reached the same conclusion the assigning attorney expected.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_5_1_responsibilities_of_a_partner_or_supervisory_lawyer/&quot;&gt;Model Rule 5.1&lt;/a&gt; requires partners and supervisory lawyers to make reasonable efforts to ensure that subordinates&#39; work conforms to professional obligations. &lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_5_3_responsibilities_regarding_nonlawyer_assistance/&quot;&gt;Rule 5.3&lt;/a&gt; extends analogous duties to nonlawyer assistants — a category that, under &lt;a href=&quot;https://www.americanbar.org/content/dam/aba/administrative/professional_responsibility/ethics-opinions/aba-formal-opinion-512.pdf&quot;&gt;ABA Formal Opinion 512&lt;/a&gt;, encompasses AI tools used in legal practice. The supervisory obligation has traditionally focused on accuracy and confidentiality. Sycophancy introduces a different challenge: the work product may be accurate in its citations and well-constructed in its reasoning, yet still reflect a systematic bias toward the conclusion the supervising attorney signaled. A supervisor who reviews only for accuracy and completeness will not catch the distortion, because it lives in what the memo fails to say — the counterarguments it understated, the unfavorable authorities it deemphasized, the analytical path it did not take because that path leads away from the answer the user appeared to want.&lt;/p&gt;
&lt;h2&gt;What this means for legal education&lt;/h2&gt;
&lt;p&gt;Law school pedagogy — at its best — is built on structured challenge. The Socratic method works because it forces students to defend their reasoning against pressure, distinguish their position from adjacent ones, and identify the weaknesses in their own analysis before someone else does. The method is, by design, anti-sycophantic. A good professor does not tell a student her reading of a case is strong. She asks &amp;quot;what&#39;s the strongest argument against your position?&amp;quot; and refuses to move on until the student can articulate it.&lt;/p&gt;
&lt;p&gt;An LLM will not do this unless explicitly instructed to, and even then its tendency toward agreement will attenuate the challenge. A student who uses an LLM to prepare for class, work through hypotheticals, or test her analysis is training with a tool that rewards existing reasoning rather than stress-testing it. Over time, that produces weaker instincts for self-critique — not because the tool gives wrong answers, but because it gives comfortable ones.&lt;/p&gt;
&lt;p&gt;I want to be careful not to overstate the case. LLMs can be prompted to argue the other side. But counteracting a system&#39;s default requires knowing the default exists, and most users do not. The Cheng et al. findings show that even when users ask genuinely open-ended questions, the models tilt toward agreement. The bias is a background condition of the interaction, not something triggered only by leading questions.&lt;/p&gt;
&lt;h2&gt;The practical response&lt;/h2&gt;
&lt;p&gt;What follows are adjustments that account for sycophancy specifically, building on the &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;prompting strategies&lt;/a&gt; and &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;judgment-delegation framework&lt;/a&gt; from earlier posts.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Prompt for disagreement, not agreement.&lt;/strong&gt; Instead of asking the model whether your analysis is correct, ask it to identify every weakness in your position. Instead of &amp;quot;is this argument strong?&amp;quot;, try &amp;quot;assume opposing counsel is excellent — what are the three strongest attacks on this argument, and what authority supports each one?&amp;quot; The framing matters: a prompt that presupposes the analysis is sound (&amp;quot;review my argument&amp;quot;) invites a sycophantic response. A prompt that presupposes it has flaws (&amp;quot;identify the weaknesses&amp;quot;) works against the grain of the model&#39;s training.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Use adversarial sessions.&lt;/strong&gt; Run your analysis through a second, separate conversation in which the model is instructed to argue the opposing side. The &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;OTOC rule&lt;/a&gt; (one task, one conversation) already counsels starting fresh conversations for each discrete task. An adversarial session goes further: it eliminates the conversational context that anchors the sycophantic tendency. A model that helped you build an argument in one session has a prior commitment to that argument&#39;s success; a fresh session does not.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Treat confirmation as a weak signal.&lt;/strong&gt; When the model&#39;s analysis aligns with your own, that alignment should carry less weight than when it identifies something you did not expect. Agreement may reflect the model&#39;s tendency to mirror your reasoning; disagreement runs against that tendency and is therefore more informative. This is a heuristic, not a rule — surprising outputs can also be wrong. But in a system biased toward agreement, the unexpected response deserves more attention than the confirming one.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Withhold your conclusion.&lt;/strong&gt; If you want the model to evaluate a legal question, do not tell it what you think the answer is before you ask. Provide the relevant facts and authorities, but let the model reach its own conclusion first. Once you have stated a position in the conversation, the model&#39;s subsequent analysis will be shaped by it — the sycophancy-specific complement to the &amp;quot;judgment words&amp;quot; framework from the &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;earlier post&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;The deeper problem&lt;/h2&gt;
&lt;p&gt;Every mitigation strategy above is a workaround for a system that does not, by default, do what good counsel does. A good lawyer tells the client what she needs to hear. A good associate tells the partner that the theory is weaker than it looks. A good professor tells the student that the analysis has a gap. These are acts of professional independence — and they are precisely the acts that sycophantic AI systems are architecturally disinclined to perform.&lt;/p&gt;
&lt;p&gt;Hallucination is a more dramatic failure and easier to detect — a fabricated citation either exists or it doesn&#39;t. Sycophancy produces outputs that are not wrong in any verifiable sense but are tilted — toward agreement, toward comfort, toward the conclusion the user signaled she was looking for. A lawyer who relies on a tool with that tilt, without recognizing it, will develop an inflated confidence in her own reasoning, because the tool will rarely give her cause to doubt it.&lt;/p&gt;
&lt;p&gt;That is the quiet damage — the slow erosion of the habit of self-challenge that distinguishes professional judgment from mere fluency.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;This post draws on the complaint in &lt;a href=&quot;https://www.courtlistener.com/docket/69648931/nippon-life-insurance-company-of-america-v-openai-foundation/&quot;&gt;Nippon Life Insurance Company of America v. OpenAI Foundation et al.&lt;/a&gt;, No. 1:26-cv-02448 (N.D. Ill. filed Mar. 4, 2026); Cheng et al., &lt;a href=&quot;https://www.science.org/doi/10.1126/science.aec8352&quot;&gt;AI Sycophancy&lt;/a&gt;, Science (2026); the Georgetown Law Tech Institute&#39;s &lt;a href=&quot;https://www.law.georgetown.edu/tech-institute/research-insights/insights/ai-sycophancy-impacts-harms-questions/&quot;&gt;analysis of sycophancy harms&lt;/a&gt;; and the ABA &lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/&quot;&gt;Model Rules of Professional Conduct&lt;/a&gt; and &lt;a href=&quot;https://www.americanbar.org/content/dam/aba/administrative/professional_responsibility/ethics-opinions/aba-formal-opinion-512.pdf&quot;&gt;Formal Opinion 512&lt;/a&gt;. The prompting strategies build on approaches described in prior posts on &lt;a href=&quot;/2026-03-30-what-your-ai-forgets-midsentence/&quot;&gt;context management&lt;/a&gt; and &lt;a href=&quot;/2026-03-31-delegate-the-task-not-the-judgment/&quot;&gt;judgment delegation&lt;/a&gt;. For background on the consumer-versus-commercial data-handling divide and its legal implications, see the earlier entries in this &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;series&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  <entry>
    <title>What Your AI Forgets Mid-Sentence — And What to Do About It</title>
    <link href="https://davidkemp.ai/blog/what-your-ai-forgets-midsentence/" />
    <updated>2026-03-29T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/what-your-ai-forgets-midsentence/</id>
    <content type="html">&lt;p&gt;Syntheia published a &lt;a href=&quot;https://syntheia.io/blog/silent-but-deadly-context-rot-problems-in-legal&quot;&gt;useful piece&lt;/a&gt; this week on what they call &amp;quot;context rot&amp;quot; — the family of failures that occur when a large language model processes more text than it can reliably attend to. Their diagnosis is sharp: LLMs degrade silently on long documents, and the law firm&#39;s traditional quality-assurance architecture is not calibrated to catch the resulting errors. I agree with most of their analysis, but I want to take it further and offer solutions.&lt;/p&gt;
&lt;p&gt;In this post, I explain the mechanics of context windows in terms aimed at the practicing lawyer, and then I propose concrete strategies to work within those constraints.&lt;/p&gt;
&lt;h2&gt;The context window, explained without jargon&lt;/h2&gt;
&lt;p&gt;Every LLM has a context window — the total amount of text it can hold in working memory for a single exchange. That window includes everything: the system instructions that tell the model how to behave, whatever documents you have uploaded or pasted in, the full history of your conversation, and the model&#39;s own response. All of it competes for the same finite space.&lt;/p&gt;
&lt;p&gt;Context windows are measured in tokens, roughly three-quarters of a word in English. A &amp;quot;200,000-token context window&amp;quot; means roughly 150,000 words across all inputs combined, in a single conversation turn. That sounds enormous until you consider that a single commercial loan agreement can run 80,000 words and a due diligence data room can contain millions. For reference, the Claude system instruction alone — which is necessarily part of every conversation with Claude — can easily run to tens of thousands of tokens.&lt;/p&gt;
&lt;p&gt;The critical point, and the one that most marketing materials omit, is that the &lt;em&gt;advertised&lt;/em&gt; context window and the &lt;em&gt;effective&lt;/em&gt; context window are not the same thing. NVIDIA&#39;s &lt;a href=&quot;https://github.com/NVIDIA/RULER&quot;&gt;RULER benchmark&lt;/a&gt; tested models on the kind of complex reasoning tasks that legal work demands, and found that effective performance sits at roughly &lt;a href=&quot;https://arxiv.org/abs/2404.06654&quot;&gt;50 to 65 percent&lt;/a&gt; of the advertised token limit. A model with a 200,000-token window performs reliably on about 100,000 to 130,000 tokens of actual input. The number on the box is not the number that governs your work.&lt;/p&gt;
&lt;h2&gt;How the degradation works&lt;/h2&gt;
&lt;p&gt;The research literature identifies several distinct failure modes. They are worth understanding individually, because each one suggests a different mitigation strategy.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Positional bias.&lt;/strong&gt; The &lt;a href=&quot;https://arxiv.org/abs/2307.03172&quot;&gt;Stanford &amp;quot;Lost in the Middle&amp;quot; research&lt;/a&gt; (Liu et al., &lt;em&gt;TACL&lt;/em&gt; 2024) demonstrated that LLMs attend most strongly to text at the beginning and end of their input. In multi-document question answering, accuracy dropped by roughly &lt;a href=&quot;https://www.morphllm.com/lost-in-the-middle-llm&quot;&gt;30 percentage points&lt;/a&gt; — from approximately 75% to approximately 45% — when relevant information moved from the first position to the middle of the context. In a 200-page agreement, the provisions that matter most are rarely on page one or page 200.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Volume-dependent reasoning decay.&lt;/strong&gt; &lt;a href=&quot;https://arxiv.org/abs/2510.05381&quot;&gt;Du et al. (2025)&lt;/a&gt; isolated an even more troubling finding: reasoning accuracy degrades as context length increases &lt;em&gt;even when the model has perfect access to all relevant information&lt;/em&gt;. They tested this by padding relevant text with whitespace (minimally distracting filler that should not confuse the model) and observed performance drops of up to 85 percent. The sheer volume of input makes the model a worse reasoner, independent of whether the right answer is present.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Conversation history displacement.&lt;/strong&gt; When a conversation exceeds the context window, something has to go. In most current implementations, including &lt;a href=&quot;https://platform.claude.com/docs/en/build-with-claude/context-windows&quot;&gt;Anthropic&#39;s Claude&lt;/a&gt; and &lt;a href=&quot;https://www.datastudios.org/post/chatgpt-token-limits-and-context-windows-updated-for-all-models-in-2025&quot;&gt;OpenAI&#39;s ChatGPT&lt;/a&gt;, the system preserves the system prompt and truncates the oldest &lt;em&gt;conversation turns&lt;/em&gt; first. Some platforms &lt;a href=&quot;https://anthropic.com/news/context-management&quot;&gt;summarize rather than drop&lt;/a&gt; the earlier exchanges, though that introduces its own fidelity problems. The practical result is the same: the model loses track of what you discussed earlier in the session. The analytical framework you established, the specific issues you flagged, the constraints you set three exchanges ago, all of it becomes inaccessible. In custom or middleware implementations, the system prompt itself may also be at risk, though the major providers now treat it as pinned content.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Compression artifacts.&lt;/strong&gt; Summarizing a document before feeding it to the model, a common workaround for length limitations, introduces its own errors. Compression algorithms often strip language that appears formulaic or repetitive, but legal documents are dense with formulaic language that carries substantive weight. &amp;quot;Subject to,&amp;quot; &amp;quot;notwithstanding the foregoing,&amp;quot; &amp;quot;except as provided in Section K&amp;quot;: these phrases distinguish an absolute obligation from a qualified one. &lt;a href=&quot;https://aclanthology.org/2021.naacl-main.383/&quot;&gt;Pagnoni et al. (&lt;em&gt;NAACL&lt;/em&gt; 2021)&lt;/a&gt; found that over 80 percent of summaries produced by the neural models evaluated contained factual errors, concentrated precisely in conditional and qualifying language. Current models perform better on standard summarization benchmarks, but the specific vulnerability to legal qualifying language persists because it is structural. Compression algorithms are designed to remove redundancy, and legal qualifiers are designed to look redundant while doing essential work.&lt;/p&gt;
&lt;p&gt;These failure modes share a symptom: the output looks complete. It is well-formatted, internally coherent, and confident. Nothing about it signals that a substantial portion of the source material was functionally ignored. That is what distinguishes context rot from the more familiar hallucination problem, and what makes it harder to catch in review.&lt;/p&gt;
&lt;h2&gt;What to do about it&lt;/h2&gt;
&lt;p&gt;What follows are concrete approaches, ordered from simplest to most involved, that any lawyer can implement today.&lt;/p&gt;
&lt;h3&gt;1. One task, one conversation&lt;/h3&gt;
&lt;p&gt;This is probably the single highest-value habit change available to a non-technical user. Every AI conversation accumulates context: your prior messages, the model&#39;s prior responses, uploaded documents, session instructions. As the conversation grows, the model&#39;s effective reasoning capacity shrinks. Old instructions interfere with current tasks. Prior assumptions bleed into new analysis. The context fills with material that was useful ten exchanges ago and is now dead weight, what researchers call &lt;a href=&quot;https://understandingdata.com/posts/context-pollution-recovery/&quot;&gt;context pollution&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The fix is simple: start a new conversation for each discrete task. Do not use the same session to summarize a lease, then draft a demand letter, then review an indemnification clause. Each of those deserves a clean context window, and starting a new conversation is free, while the accuracy cost of a polluted one is invisible until something goes wrong.&lt;/p&gt;
&lt;p&gt;I call this the OTOC rule — one task, one conversation. That&#39;s not to discourage iterative prompting. Iterative refinement of a single work product is still a single task and is an effective use of an LLM. Revising a draft and then pivoting to an unrelated analysis in the same session is two tasks crammed into one window — increasing the risk of context rot.&lt;/p&gt;
&lt;h3&gt;2. Write a durable task specification&lt;/h3&gt;
&lt;p&gt;The OTOC rule creates a practical problem: if every task gets a fresh conversation, you lose the background context the model needs to do good work. The overarching objectives, the governing law, the deal structure, the specific issues you care about — all of that vanishes when you close the session.&lt;/p&gt;
&lt;p&gt;The solution is to write a reusable task specification: a short document (a few hundred words is usually sufficient) that captures the stable context for a project. Think of it as a briefing memo for the model. It should include the matter description, the governing jurisdiction, the relevant parties, the specific analytical framework you want applied, and any constraints or preferences that should carry across sessions.&lt;/p&gt;
&lt;p&gt;You paste this specification at the top of each new conversation, or, even better, preserve it as its own file to attach as input. The model reads it fresh every time, without the accumulated noise of prior exchanges. This is the complement to the OTOC rule: it lets you start clean without starting ignorant. Some tools (Anthropic&#39;s Claude Projects feature, for instance) let you attach persistent instructions to a project workspace that automatically prepopulate every conversation. If your platform supports it, use it.&lt;/p&gt;
&lt;h3&gt;3. Chunk your documents before the model reads them&lt;/h3&gt;
&lt;p&gt;If positional bias causes the model to lose track of middle-document content, and if volume alone degrades reasoning quality, then the logical response is to feed the model smaller, task-relevant segments rather than entire documents.&lt;/p&gt;
&lt;p&gt;For a 200-page credit agreement, do not upload the entire file and ask the model to &amp;quot;review it.&amp;quot; Instead, consider breaking the document into its component sections (representations and warranties, covenants, events of default, definitions, schedules) and submit each section in a separate conversation (applying the OTOC rule) with a targeted question. &amp;quot;Identify all financial covenants in the following section and flag any that use a trailing-twelve-month measurement period&amp;quot; will produce dramatically better results than &amp;quot;review this agreement and summarize the key terms.&amp;quot;&lt;/p&gt;
&lt;p&gt;One important caveat: legal documents are dense with internal cross-references (defined terms, conditions qualified by other sections, carve-outs incorporated by reference). When you chunk, you sever those links. The model analyzing the covenants will not know that a defined term in Article I changes the meaning of a financial ratio test, or that a carve-out in Schedule 3 qualifies an obligation in Section 12. The practical mitigation is to always include the definitions section (or at minimum the relevant defined terms) alongside whatever substantive section you are analyzing.&lt;/p&gt;
&lt;p&gt;Manual chunking is labor-intensive, but the labor is front-loaded and predictable. It converts one unreliable pass over an entire document into multiple reliable passes over bounded sections. The lawyer stitches the analysis back together, which is the level at which human judgment should operate regardless of whether AI is involved. For high-stakes tasks, the benefit of minimizing AI errors through manual chunking far outweighs the burden.&lt;/p&gt;
&lt;h3&gt;4. Use chain-of-thought prompting to structure the model&#39;s reasoning&lt;/h3&gt;
&lt;p&gt;Chain-of-thought prompting means explicitly instructing the model to reason through intermediate steps before reaching a conclusion. Instead of asking &amp;quot;Does Section 7.2 conflict with Schedule B?&amp;quot;, you ask: &amp;quot;First, extract the operative language of Section 7.2 and state its requirements. Then extract the relevant provisions of Schedule B. Then identify any inconsistencies between them. Then state your conclusion.&amp;quot;&lt;/p&gt;
&lt;p&gt;This matters for context management because it forces the model to surface the textual evidence it is relying on before it reasons over that evidence. If the model skips a provision, you will see the gap in the intermediate step, before it gets papered over by a confident-sounding conclusion. Du et al. (2025) found that a simple version of this approach, prompting the model to &lt;a href=&quot;https://arxiv.org/abs/2510.05381&quot;&gt;recite the retrieved evidence before solving the problem&lt;/a&gt;, mitigated much of the performance loss caused by long contexts. The technique works because it forces the model to move relevant information into a high-attention position (the most recent output) before it reasons about it.&lt;/p&gt;
&lt;p&gt;For legal work, chain-of-thought prompting also functions as a transparency mechanism. A model that shows its intermediate reasoning produces work product that a supervising lawyer can actually verify, because the intermediate steps expose the gaps that a polished final conclusion would conceal.&lt;/p&gt;
&lt;h3&gt;5. Place critical information strategically&lt;/h3&gt;
&lt;p&gt;The &amp;quot;Lost in the Middle&amp;quot; research has a direct practical corollary: put the most important content where the model pays the most attention. That means the beginning and end of your input, not the middle.&lt;/p&gt;
&lt;p&gt;If you are asking the model to analyze a specific clause in the context of a larger document section, place the target clause at the top of your prompt, followed by the surrounding context, and then restate the analytical question at the end. If you are using a task specification (Strategy 2), put it at the top. If you have specific instructions about format or analytical framework, repeat them at the bottom. The worst arrangement, and the one most people default to, is pasting a large document and then typing the question at the bottom, burying the analytical instructions in a low-attention position.&lt;/p&gt;
&lt;h3&gt;6. Verify in a separate conversation, not the one that produced the work&lt;/h3&gt;
&lt;p&gt;This follows directly from the OTOC rule. Generation and verification are different tasks, and they belong in different conversations.&lt;/p&gt;
&lt;p&gt;When you ask the model to check its own work in the same session, the entire prior exchange sits in the context window: the assumptions, the omissions, the analytical choices the model made on its first pass. All of it exerts influence on the verification. A model reviewing its own conclusions is structurally biased toward confirming them, the equivalent of asking the same reviewer to read the same draft a second time and expecting fresh insight.&lt;/p&gt;
&lt;p&gt;A de novo review in a fresh conversation eliminates that problem. Paste or upload the relevant source text and the model&#39;s output into a clean session. Ask: &amp;quot;Does this analysis accurately and completely reflect the source material? Identify every section of the source you relied on and quote the language supporting each conclusion.&amp;quot; The new session has no prior commitments pulling it toward agreement. It is structurally analogous to the mid-level reviewing the junior&#39;s draft — fresh eyes on the same source.&lt;/p&gt;
&lt;p&gt;A necessary warning: the model can fabricate quotations even in a clean session. It may generate text that looks like a verbatim extract but is actually a paraphrase, a conflation of multiple provisions, or &lt;a href=&quot;https://hai.stanford.edu/news/hallucinating-law-legal-mistakes-large-language-models-are-pervasive&quot;&gt;an outright invention&lt;/a&gt;. The verification step itself requires verification — you must check the model&#39;s quoted language against the source document. That is additional work, but it is targeted work: instead of re-reading 200 pages looking for problems you do not know to expect, you are checking specific passages the model claims to have relied on. The de novo framing does not eliminate the need for human verification, but it gives you a structurally honest starting point for it.&lt;/p&gt;
&lt;h2&gt;The underlying principle&lt;/h2&gt;
&lt;p&gt;Every strategy above is a variation on a single idea: &lt;em&gt;give the model less to think about, and tell it more precisely what to think about it.&lt;/em&gt; That runs against the grain of how most people use these tools. The natural instinct is to dump everything into the conversation and let the AI sort it out, and the marketing encourages exactly that — &amp;quot;upload your entire contract,&amp;quot; &amp;quot;ask anything about your documents.&amp;quot; The context window numbers are designed to suggest the model can handle it all.&lt;/p&gt;
&lt;p&gt;It can, in the sense that it will produce output. What it cannot do — reliably, on long documents, under token pressure — is produce output accurate enough to stake a client&#39;s interests on. The strategies in this post are all ways of closing that gap: structuring the input so the model&#39;s actual capabilities match the demands of the task. The work is unglamorous — writing briefing documents for a machine, manually splitting PDFs, running the same analysis twice in separate sessions. But it maps directly onto skills lawyers already have. Scoping a task, preparing materials for review, verifying work product against source documents — these are not new professional obligations. They are existing ones, applied to a new tool.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;This post draws on Liu et al., &lt;a href=&quot;https://arxiv.org/abs/2307.03172&quot;&gt;Lost in the Middle: How Language Models Use Long Contexts&lt;/a&gt; (TACL 2024); Du et al., &lt;a href=&quot;https://arxiv.org/abs/2510.05381&quot;&gt;Context Length Alone Hurts LLM Performance Despite Perfect Retrieval&lt;/a&gt; (EMNLP 2025); NVIDIA&#39;s &lt;a href=&quot;https://github.com/NVIDIA/RULER&quot;&gt;RULER benchmark&lt;/a&gt; (2024); and Pagnoni et al., &lt;a href=&quot;https://aclanthology.org/2021.naacl-main.383/&quot;&gt;Understanding Factuality in Abstractive Summarization with FRANK&lt;/a&gt; (NAACL 2021). Anthropic&#39;s &lt;a href=&quot;https://platform.claude.com/docs/en/build-with-claude/context-windows&quot;&gt;context window documentation&lt;/a&gt; and &lt;a href=&quot;https://anthropic.com/news/context-management&quot;&gt;context management guidance&lt;/a&gt; informed the discussion of conversation history displacement. For context on the data-handling and compliance dimensions of AI tool selection, see prior entries in this series on &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;consumer-versus-commercial data handling&lt;/a&gt;, &lt;a href=&quot;/2026-03-23-the-api-is-not-a-compliance-strategy/&quot;&gt;API compliance architecture&lt;/a&gt;, and &lt;a href=&quot;/2026-03-27-the-duty-to-inform/&quot;&gt;the duty to counsel clients about AI privilege risks&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  <entry>
    <title>You Probably Have a Duty to Warn Your Clients About ChatGPT</title>
    <link href="https://davidkemp.ai/blog/the-duty-to-inform/" />
    <updated>2026-03-27T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/the-duty-to-inform/</id>
    <content type="html">&lt;p&gt;I have written previously about what &lt;a href=&quot;https://www.courtlistener.com/docket/71872024/27/united-states-v-heppner/&quot;&gt;&lt;em&gt;United States v. Heppner&lt;/em&gt;&lt;/a&gt; &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;held and what it got wrong&lt;/a&gt;, and about why &lt;a href=&quot;/2026-03-23-the-api-is-not-a-compliance-strategy/&quot;&gt;moving to an API&lt;/a&gt; does not, by itself, constitute a compliance strategy. This post turns to a different audience: not organizations choosing AI tools, but practicing lawyers whose clients are already using them.&lt;/p&gt;
&lt;p&gt;The core question is straightforward. &lt;em&gt;Heppner&lt;/em&gt; established — on reasoning I have &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;criticized&lt;/a&gt; but that is now on the books — that a client who feeds privileged materials into a consumer AI platform may forfeit the privilege over those materials. That is now a known hazard. And when a known hazard exists that threatens the integrity of the attorney-client relationship, existing rules of professional conduct impose obligations on the lawyer — not just the client.&lt;/p&gt;
&lt;p&gt;No ethics rule says &amp;quot;warn your client about ChatGPT.&amp;quot; But the obligation to do something very close to that is already embedded in the structure of Model Rules 1.1, 1.4, and 1.6, and their state counterparts. &lt;em&gt;Heppner&lt;/em&gt; did not create that duty, but it did make the duty impossible to ignore.&lt;/p&gt;
&lt;h2&gt;A brief recap of what &lt;em&gt;Heppner&lt;/em&gt; did&lt;/h2&gt;
&lt;p&gt;I covered the decision in detail in &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;this prior post&lt;/a&gt;, so I will keep this short. Bradley Heppner, a criminal defendant, used consumer Claude to analyze his legal exposure and develop defense theories after receiving a grand jury subpoena and learning he was a target of a federal investigation. He did this on his own, without his lawyers&#39; knowledge or direction. Judge Rakoff of the S.D.N.Y. held the resulting documents were protected by neither the attorney-client privilege nor the work product doctrine — because Claude is not a lawyer, because Anthropic&#39;s consumer terms did not support a reasonable expectation of confidentiality, and because counsel had not directed the AI use.&lt;/p&gt;
&lt;p&gt;Two things from the opinion matter for this post. First, Judge Rakoff observed that had counsel &lt;em&gt;directed&lt;/em&gt; Heppner to use Claude, the tool &amp;quot;might arguably be said to have functioned in a manner akin to a highly trained professional who may act as a lawyer&#39;s agent within the protection of the attorney-client privilege&amp;quot; — a reference to the &lt;a href=&quot;https://www.courtlistener.com/opinion/265578/united-states-v-kovel/&quot;&gt;&lt;em&gt;Kovel&lt;/em&gt;&lt;/a&gt; doctrine. That dictum rewards attorney supervision and penalizes its absence. Second, the privilege was lost in part because Heppner&#39;s lawyers never told him — one way or the other — anything about using AI tools in connection with his case.&lt;/p&gt;
&lt;p&gt;The NYSBA&#39;s &lt;a href=&quot;https://nysba.org/loose-ai-prompts-sink-ships-how-heppner-shook-the-legal-community/&quot;&gt;post-&lt;em&gt;Heppner&lt;/em&gt; commentary&lt;/a&gt; drew the practical conclusion quickly: attorneys should &amp;quot;include robust disclaimers and warnings in engagement letters and email signatures alerting clients to the risks of using AI platforms in connection with their legal matters.&amp;quot; That is a reasonable starting point. But I think the duty runs deeper than engagement-letter boilerplate, and that existing ethics rules already require it.&lt;/p&gt;
&lt;h2&gt;The rules that get you there&lt;/h2&gt;
&lt;p&gt;Three Model Rules, read together, create an affirmative obligation to advise clients about AI-related privilege risks — even though none of them mentions AI by name.&lt;/p&gt;
&lt;h3&gt;Competence: Rule 1.1&lt;/h3&gt;
&lt;p&gt;&lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_1_1_competence/&quot;&gt;Model Rule 1.1&lt;/a&gt; requires lawyers to provide competent representation, defined as &amp;quot;the legal knowledge, skill, thoroughness and preparation reasonably necessary for the representation.&amp;quot; Since 2012, &lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_1_1_competence/comment_on_rule_1_1/&quot;&gt;Comment 8&lt;/a&gt; has specified that competence includes keeping &amp;quot;abreast of changes in the law and its practice, including the benefits and risks associated with relevant technology.&amp;quot; Forty states have now &lt;a href=&quot;https://www.americanbar.org/groups/law_practice/publications/techreport/2024/duty-of-tech-competence/&quot;&gt;adopted this language&lt;/a&gt; or its equivalent.&lt;/p&gt;
&lt;p&gt;After &lt;em&gt;Heppner&lt;/em&gt;, the &amp;quot;relevant technology&amp;quot; a competent lawyer must understand includes consumer AI tools — not how to use them, but how they handle data and what the legal consequences of client use might be. A lawyer who does not know that consumer chatbot terms permit the provider to retain, train on, and disclose user inputs is missing knowledge that is now directly relevant to protecting the privilege. The duty of competence is not limited to a lawyer&#39;s own work product. It encompasses the &amp;quot;thoroughness and preparation&amp;quot; needed to protect the attorney-client relationship from erosion by foreseeable client conduct.&lt;/p&gt;
&lt;h3&gt;Communication: Rule 1.4&lt;/h3&gt;
&lt;p&gt;&lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_1_4_communications/&quot;&gt;Model Rule 1.4(b)&lt;/a&gt; requires that a lawyer &amp;quot;explain a matter to the extent reasonably necessary to permit the client to make informed decisions regarding the representation.&amp;quot; This is generally understood to encompass not just the substance of legal advice but the conditions under which the privilege protecting it might be forfeited. A client who does not know that pasting counsel&#39;s memorandum into ChatGPT may destroy the privilege over that memorandum has not been equipped to make an informed decision about managing privileged information.&lt;/p&gt;
&lt;p&gt;The critical feature of Rule 1.4 is that it operates &lt;em&gt;prospectively&lt;/em&gt;. The duty to communicate is a duty to give clients the information they need before they act — not a post-hoc damage-control obligation. After &lt;em&gt;Heppner&lt;/em&gt;, the relevant information includes the fact that consumer AI use can waive the privilege.&lt;/p&gt;
&lt;h3&gt;Confidentiality: Rule 1.6&lt;/h3&gt;
&lt;p&gt;&lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/rule_1_6_confidentiality_of_information/&quot;&gt;Model Rule 1.6(c)&lt;/a&gt; provides that a lawyer &amp;quot;shall make reasonable efforts to prevent the inadvertent or unauthorized disclosure of, or unauthorized access to, information relating to the representation of a client.&amp;quot; The operative word is &amp;quot;reasonable,&amp;quot; and what counts as reasonable changes as risks become known.&lt;/p&gt;
&lt;p&gt;State bars have interpreted this provision to require affirmative steps — not just reactive ones — when digital communications create confidentiality risks. The principle is not new; all that is new is the specific threat: a client&#39;s use of a consumer AI platform is precisely the kind of inadvertent disclosure that Rule 1.6(c) was designed to address.&lt;/p&gt;
&lt;h3&gt;The state-level picture&lt;/h3&gt;
&lt;p&gt;The ABA&#39;s &lt;a href=&quot;https://www.americanbar.org/content/dam/aba/administrative/professional_responsibility/ethics-opinions/aba-formal-opinion-512.pdf&quot;&gt;Formal Opinion 512&lt;/a&gt;, issued in July 2024, was the first comprehensive ABA guidance on generative AI in legal practice. It addressed competence, confidentiality, communication, candor, supervisory duties, and fees — all through the lens of existing Model Rules applied to AI. Formal Opinion 512 focused primarily on a lawyer&#39;s &lt;em&gt;own&lt;/em&gt; use of AI tools, but its analysis of the confidentiality obligations under Rules 1.6 and 1.4 applies with equal force when the risk comes from the client&#39;s conduct rather than the lawyer&#39;s.&lt;/p&gt;
&lt;p&gt;The New York City Bar&#39;s &lt;a href=&quot;https://www.nycbar.org/reports/formal-opinion-2024-5-generative-ai-in-the-practice-of-law/&quot;&gt;Formal Opinion 2024-5&lt;/a&gt; addressed generative AI in legal practice directly, and &lt;a href=&quot;https://www.nycbar.org/reports/formal-opinion-2025-6-ethical-issues-affecting-use-of-ai-to-record-transcribe-and-summarize-conversations-with-clients/&quot;&gt;Formal Opinion 2025-6&lt;/a&gt; extended the analysis to AI tools used to record and transcribe client conversations — a context in which the duty to counsel clients about confidentiality implications is made explicit. California&#39;s State Bar has published &lt;a href=&quot;https://www.calbar.ca.gov/Portals/0/documents/ethics/Generative-AI-Practical-Guidance.pdf&quot;&gt;practical guidance on generative AI&lt;/a&gt; grounded in the same competence and confidentiality obligations.&lt;/p&gt;
&lt;p&gt;None of these authorities squarely addresses the specific scenario &lt;em&gt;Heppner&lt;/em&gt; presented: a client, acting on his own, feeding privileged materials into a consumer chatbot. But they establish the framework within which that scenario falls. If a lawyer has a duty of technological competence that includes understanding AI data handling, a duty to communicate information necessary for informed decisions about the representation, and a duty to take reasonable steps to prevent inadvertent disclosure — then the obligation to warn a client about the privilege risks of consumer AI use follows from the conjunction of all three.&lt;/p&gt;
&lt;h2&gt;What &amp;quot;reasonable&amp;quot; looks like&lt;/h2&gt;
&lt;p&gt;Not every representation carries the same risk. The obligation to advise clients about AI-related privilege risks should be calibrated — as professional duties always are — to the circumstances.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The nature of the matter.&lt;/strong&gt; A client facing a federal investigation, complex litigation, or a regulatory proceeding is more likely to receive extensive privileged communications and more acutely harmed by their disclosure. In high-stakes representations, the duty to counsel clients about AI risks should be treated as near-mandatory and documented. Routine advisory work still carries the obligation, but its urgency is proportional to the exposure.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The sophistication of the client.&lt;/strong&gt; Sophisticated institutional clients with in-house counsel may understand the risk without detailed instruction. Individual clients, small business owners, and people facing their first serious legal proceeding probably do not. &lt;em&gt;Heppner&lt;/em&gt; illustrates the gap precisely: the defendant was fluent enough to use Claude effectively but apparently had no appreciation of the legal consequences. Technological fluency and legal sophistication are not the same thing, and lawyers should resist treating them as interchangeable.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;The attorney&#39;s reasonable belief about client conduct.&lt;/strong&gt; A lawyer who knows or should know that a client is likely to use AI tools in connection with the matter — because the client has mentioned doing so, because the client works in a tech-forward industry, or simply because generative AI has become most people&#39;s first tool for understanding complex documents — bears a heightened responsibility to address the risk explicitly. This is not speculative. Consumer AI adoption has reached the point where assuming a client &lt;em&gt;will not&lt;/em&gt; use these tools requires more justification than assuming they will.&lt;/p&gt;
&lt;p&gt;These factors interact. A sophisticated client in a high-stakes criminal matter presents a different risk profile than a sophisticated client in a routine transaction. An unsophisticated client in any matter of consequence probably requires explicit, plain-language AI counseling as a baseline.&lt;/p&gt;
&lt;h2&gt;The structural remedy worth considering&lt;/h2&gt;
&lt;p&gt;Warning clients not to use consumer AI to understand their legal matters is, as a practical matter, unlikely to be fully effective. The impulse that drove Heppner to Claude is deeply human: complex legal advice is hard to understand, and AI tools offer an immediately accessible way to work through it. Telling clients not to do something genuinely useful — without offering an alternative — is an instruction destined to be ignored.&lt;/p&gt;
&lt;p&gt;The more constructive path is to give clients a safe way to do what they are going to do anyway. Enterprise-grade AI deployments — tools operating under commercial terms that &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;contractually prohibit&lt;/a&gt; the provider from retaining or training on user inputs — can be configured within a firm-controlled environment with appropriate confidentiality protections. A client who uses a firm-provided, privilege-preserving AI tool to work through counsel&#39;s advice is in a fundamentally different position than a client who pastes that advice into a consumer chatbot governed by terms that reserve broad data-use rights.&lt;/p&gt;
&lt;p&gt;Judge Rakoff&#39;s &lt;em&gt;Kovel&lt;/em&gt; dictum points in this direction. The court distinguished between unsupervised client use of a public AI platform and a hypothetical in which counsel directed the AI use. A firm-provided, counsel-supervised AI environment — deployed under commercial terms, subject to confidentiality agreements, and offered as part of the representation — positions the tool more like the &lt;em&gt;Kovel&lt;/em&gt; professional the court described than the public chatbot it rejected. The privilege analysis is not guaranteed, but the structural argument is considerably stronger.&lt;/p&gt;
&lt;p&gt;This is not a small undertaking, and I do not suggest it is costless. But the alternative — relying on engagement-letter warnings while clients continue to use consumer AI tools unsupervised — is a posture that grows harder to defend as the risk becomes more widely known.&lt;/p&gt;
&lt;h2&gt;Where this leaves practicing lawyers&lt;/h2&gt;
&lt;p&gt;&lt;em&gt;Heppner&lt;/em&gt; did not create a new professional obligation. What it did was train a spotlight on one that already existed. The duty of competence requires understanding how consumer AI tools handle data. The duty of communication requires informing clients about risks to the privilege before those risks materialize. The duty of confidentiality requires reasonable efforts to prevent inadvertent disclosure. Together, these rules establish an obligation — variable in its intensity, sensitive to context, but real — to advise clients about the privilege risks of consumer AI use.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;This post draws on the &lt;a href=&quot;https://www.americanbar.org/groups/professional_responsibility/publications/model_rules_of_professional_conduct/&quot;&gt;ABA Model Rules of Professional Conduct&lt;/a&gt;, &lt;a href=&quot;https://www.americanbar.org/content/dam/aba/administrative/professional_responsibility/ethics-opinions/aba-formal-opinion-512.pdf&quot;&gt;ABA Formal Opinion 512&lt;/a&gt;, the New York City Bar&#39;s Formal Opinions &lt;a href=&quot;https://www.nycbar.org/reports/formal-opinion-2024-5-generative-ai-in-the-practice-of-law/&quot;&gt;2024-5&lt;/a&gt; and &lt;a href=&quot;https://www.nycbar.org/reports/formal-opinion-2025-6-ethical-issues-affecting-use-of-ai-to-record-transcribe-and-summarize-conversations-with-clients/&quot;&gt;2025-6&lt;/a&gt;, the NYSBA&#39;s &lt;a href=&quot;https://nysba.org/loose-ai-prompts-sink-ships-how-heppner-shook-the-legal-community/&quot;&gt;post-Heppner commentary&lt;/a&gt;, and Judge Rakoff&#39;s &lt;a href=&quot;https://www.courtlistener.com/docket/71872024/27/united-states-v-heppner/&quot;&gt;written opinion&lt;/a&gt; in United States v. Heppner. The California State Bar&#39;s &lt;a href=&quot;https://www.calbar.ca.gov/Portals/0/documents/ethics/Generative-AI-Practical-Guidance.pdf&quot;&gt;Generative AI Practical Guidance&lt;/a&gt; provides additional state-level context. The consumer-versus-commercial data-handling comparison referenced throughout is detailed in a &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;prior post&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  <entry>
    <title>The API Is Not a Compliance Strategy</title>
    <link href="https://davidkemp.ai/blog/the-api-is-not-a-compliance-strategy/" />
    <updated>2026-03-23T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/the-api-is-not-a-compliance-strategy/</id>
    <content type="html">&lt;p&gt;In my &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;last post&lt;/a&gt;, I walked through the consumer-versus-commercial divide in how major LLM providers handle data — and why that divide carries real legal consequences after the Southern District of New York&#39;s decision in &lt;a href=&quot;https://www.courtlistener.com/docket/71872024/united-states-v-heppner/&quot;&gt;&lt;em&gt;United States v. Heppner&lt;/em&gt;&lt;/a&gt;. The takeaway was that consumer AI products operate under terms that were not designed with legal privilege, confidentiality, or regulatory compliance in mind.&lt;/p&gt;
&lt;p&gt;A reasonable follow-up question is: &lt;em&gt;What about the API?&lt;/em&gt;&lt;/p&gt;
&lt;p&gt;If the consumer chatbot is the problem, the thinking goes, then switching to API access should be the solution. And there is something to that. API tiers offered by OpenAI, Anthropic, and Google operate under fundamentally different data-handling regimes than their consumer counterparts — regimes that are, by almost every measure, more protective of user data. But &amp;quot;more protective&amp;quot; is not the same thing as &amp;quot;compliant,&amp;quot; and the distinction matters more than many organizations seem to realize.&lt;/p&gt;
&lt;h2&gt;What the API actually changes&lt;/h2&gt;
&lt;p&gt;The &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;previous post&lt;/a&gt; compared consumer and commercial tiers in detail for Anthropic&#39;s Claude. The same structural divide exists across providers, and the API sits squarely on the commercial side. Here is what that means in practice.&lt;/p&gt;
&lt;p&gt;Anthropic&#39;s commercial API retains input and output logs for seven days — far shorter than the consumer tier&#39;s retention windows — and does not use customer content for model training. Enterprise accounts can negotiate Zero Data Retention, under which inputs and outputs are processed in real time and not stored at all. OpenAI&#39;s API retains data for 30 days for abuse monitoring but does not use it for model training, and offers Zero Data Retention for eligible endpoints. Google&#39;s Vertex AI operates under a Cloud Data Processing Addendum with contractually defined retention and no training use. In each case, the API provider acts as a data processor rather than a data controller, meaning the customer — not the provider — determines the purposes and means of processing.&lt;/p&gt;
&lt;p&gt;These are meaningful differences. A consumer chatbot conversation may be retained for months or years, used to train future models, and governed by a privacy policy the user never read. An API call, properly configured, may leave no trace on the provider&#39;s systems at all. For anyone whose data-handling concerns begin and end with &amp;quot;I don&#39;t want my inputs in someone else&#39;s training set,&amp;quot; the API is a substantial improvement.&lt;/p&gt;
&lt;p&gt;But regulatory compliance does not begin and end there.&lt;/p&gt;
&lt;h2&gt;Why the API is not enough&lt;/h2&gt;
&lt;p&gt;Every major regulatory framework governing sensitive data — FERPA, HIPAA, state student-privacy laws, professional-conduct rules — imposes obligations that go well beyond what the API&#39;s data-handling defaults can address. The API solves one problem (provider-side data retention and training) while leaving most of the compliance architecture untouched.&lt;/p&gt;
&lt;p&gt;Consider what a framework like HIPAA actually requires. A covered entity processing protected health information through an API must execute a Business Associate Agreement with the provider. That BAA must specify permissible uses and disclosures, require the provider to implement administrative, physical, and technical safeguards, and establish breach-notification obligations. The API&#39;s zero-retention default is a helpful technical control, but it does not substitute for the BAA itself. And the BAA, once signed, typically imposes configuration requirements — specific endpoints, disabled features, audit logging — that the organization must affirmatively implement and maintain.&lt;/p&gt;
&lt;p&gt;FERPA presents a parallel structure. An educational institution using an API to process student education records must establish that the provider qualifies under the &amp;quot;school official&amp;quot; exception, which requires a written agreement specifying the provider&#39;s function, its relationship to the institution&#39;s use of the data, and the institution&#39;s direct control over the data&#39;s use. The API&#39;s default against training on customer data is necessary but not sufficient — the institution still needs the agreement, the access controls, and the governance to ensure that student records do not flow into the API in ways the agreement does not contemplate.&lt;/p&gt;
&lt;p&gt;The pattern repeats across regulatory contexts. State biometric-privacy statutes require informed consent and retention schedules that no API default can satisfy. Professional-conduct rules governing lawyer confidentiality — sharpened considerably by &lt;em&gt;Heppner&lt;/em&gt; — demand not just favorable vendor terms but documented due diligence, competence in evaluating the technology, and ongoing supervisory obligations. An API key does not discharge any of those duties.&lt;/p&gt;
&lt;h2&gt;The architectural gap&lt;/h2&gt;
&lt;p&gt;There is a subtler problem that the &amp;quot;just use the API&amp;quot; approach tends to obscure. When an organization integrates an LLM through an API, the API handles the model-inference layer: data goes in, a response comes back, and the provider&#39;s data-handling policies govern what happens on their end. But most real-world deployments involve considerably more than a single API call.&lt;/p&gt;
&lt;p&gt;Data passes through preprocessing pipelines, prompt templates, logging systems, vector databases, retrieval-augmented generation stores, and output caches — all of which sit on the customer&#39;s side of the line. The API provider&#39;s zero-retention commitment says nothing about what happens in those layers. An organization can use a zero-retention API and still retain every input and output indefinitely in its own infrastructure, expose sensitive data through poorly secured retrieval stores, or inadvertently log protected information in application-level monitoring.&lt;/p&gt;
&lt;p&gt;This is the architectural gap that a provider-side compliance posture cannot close. The API governs data handling at the model layer. Regulatory compliance governs data handling end to end.&lt;/p&gt;
&lt;h2&gt;What &amp;quot;more protective&amp;quot; actually means&lt;/h2&gt;
&lt;p&gt;None of this is an argument against using the API. The data-handling improvements are real, and for many use cases they represent the minimum viable starting point for responsible deployment. An organization that uses the consumer chatbot for work involving sensitive data has a serious problem. An organization that uses the API has a less serious problem — but it still has a problem if the API is the beginning and end of its compliance strategy.&lt;/p&gt;
&lt;p&gt;The useful framing is not &amp;quot;consumer versus API&amp;quot; as a binary compliance decision. It is &amp;quot;API as a necessary but insufficient component of a compliance architecture.&amp;quot; The API provides a defensible data-handling posture at the provider layer. Everything else — the agreements, the access controls, the internal data governance, the training, the monitoring, the documentation — remains the organization&#39;s responsibility.&lt;/p&gt;
&lt;p&gt;For institutions and professionals operating under regulatory constraints, the practical question is not whether to use the API. It is whether you have built the rest of the compliance architecture around it — and whether you can demonstrate that you have if someone asks.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;Provider-specific data-handling policies referenced in this post draw on the same sources cited in the &lt;a href=&quot;/2026-03-20-your-ai-conversations-are-not-confidential/&quot;&gt;previous post&lt;/a&gt;, supplemented by Anthropic&#39;s &lt;a href=&quot;https://privacy.claude.com/en/articles/10458704-how-does-anthropic-protect-the-personal-data-of-claude-users&quot;&gt;Privacy Center&lt;/a&gt;, OpenAI&#39;s &lt;a href=&quot;https://developers.openai.com/api/docs/guides/your-data/&quot;&gt;API data usage documentation&lt;/a&gt;, and Google&#39;s &lt;a href=&quot;https://docs.cloud.google.com/gemini/docs/discover/data-governance&quot;&gt;Vertex AI data governance documentation&lt;/a&gt;. Compliance obligations vary by jurisdiction, regulatory framework, and organizational context. Consult qualified counsel for guidance specific to your situation.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
  <entry>
    <title>Your AI Conversations Are Not Confidential — And a Federal Court Just Said So</title>
    <link href="https://davidkemp.ai/blog/your-ai-conversations-are-not-confidential/" />
    <updated>2026-03-20T00:00:00.000Z</updated>
    <id>https://davidkemp.ai/blog/your-ai-conversations-are-not-confidential/</id>
    <content type="html">&lt;p&gt;On February 10, 2026, Judge Jed Rakoff of the Southern District of New York ruled from the bench in &lt;a href=&quot;https://www.courtlistener.com/docket/71872024/united-states-v-heppner/&quot;&gt;&lt;em&gt;United States v. Heppner&lt;/em&gt;&lt;/a&gt; that documents a criminal defendant generated using the consumer version of Anthropic&#39;s Claude were protected by neither the attorney-client privilege nor the work product doctrine. A week later, he issued a &lt;a href=&quot;https://www.courtlistener.com/docket/71872024/27/united-states-v-heppner/&quot;&gt;written opinion&lt;/a&gt; calling it a matter of &amp;quot;nationwide&amp;quot; first impression.&lt;/p&gt;
&lt;p&gt;I think parts of the court&#39;s reasoning are wrong — or at least underdeveloped — in ways that matter. But the opinion landed on a real problem. Lawyers, clients, and judges are making consequential decisions about AI tools without fully understanding how those tools handle data. &lt;em&gt;Heppner&lt;/em&gt; is worth examining less for the doctrine it announces than for the knowledge gap it reveals.&lt;/p&gt;
&lt;p&gt;This post lays out what happened in &lt;em&gt;Heppner&lt;/em&gt;, explains what I think the opinion gets right and wrong, and then walks through what Anthropic&#39;s data-handling policies actually say across Claude&#39;s consumer and commercial tiers — the very policies the court relied on but did not examine closely. The same structural divide exists across every major LLM provider, and the legal implications extend well beyond this one case.&lt;/p&gt;
&lt;h2&gt;What &lt;em&gt;Heppner&lt;/em&gt; held&lt;/h2&gt;
&lt;p&gt;Bradley Heppner, the founder and former CEO of Beneficient, a financial services company, faces a five-count federal indictment for securities fraud, wire fraud, conspiracy, making false statements to auditors, and falsification of records — charges arising from an alleged scheme to defraud investors in the publicly traded company GWG Holdings through self-dealing transactions involving Beneficient. After receiving a grand jury subpoena and learning he was a target of the investigation, but before his November 2025 arrest, Heppner used the consumer version of Claude to analyze his legal exposure and develop defense theories. When federal agents executed a search warrant at his home, they seized numerous documents and electronic devices. Defense counsel later identified approximately thirty-one of the seized materials as AI-generated documents. The government moved for a ruling that the documents were not privileged; Heppner resisted, invoking attorney-client privilege and the work product doctrine.&lt;/p&gt;
&lt;p&gt;Judge Rakoff rejected both claims on multiple grounds. On privilege, the court articulated three independent reasons for denial:&lt;/p&gt;
&lt;p&gt;First, Claude is not an attorney. It has no law license, owes no fiduciary duties, and cannot form an attorney-client relationship. Privilege requires a &amp;quot;trusting human relationship&amp;quot; with &amp;quot;a licensed professional&amp;quot; — and an AI tool is not one.&lt;/p&gt;
&lt;p&gt;Second, Heppner had no reasonable expectation of confidentiality. The court pointed to Anthropic&#39;s privacy policy, which disclosed that user inputs and outputs could be used for model training and disclosed to third parties, including government authorities.&lt;/p&gt;
&lt;p&gt;Third — which the court acknowledged &amp;quot;perhaps presents a closer call&amp;quot; — Heppner did not communicate with Claude for the purpose of obtaining legal advice from an attorney. Claude&#39;s terms of service disclaim providing legal advice, and Heppner&#39;s lawyers neither directed nor supervised his use of the tool. The court noted that had counsel directed Heppner to use Claude, it might have &amp;quot;functioned in a manner akin to a highly trained professional&amp;quot; who could act within the privilege under the &lt;a href=&quot;https://law.justia.com/cases/federal/appellate-courts/F2/296/918/131265/&quot;&gt;&lt;em&gt;Kovel&lt;/em&gt;&lt;/a&gt; doctrine — but because Heppner acted on his own, the question was whether he intended to obtain legal advice &lt;em&gt;from Claude&lt;/em&gt;, and Claude disclaims providing it.&lt;/p&gt;
&lt;p&gt;On work product, defense counsel conceded that Heppner created the documents &amp;quot;of his own volition&amp;quot; and that the legal team &amp;quot;did not direct&amp;quot; him to use Claude. The court held that materials not prepared by or at the behest of counsel do not qualify as work product — expressly disagreeing with &lt;a href=&quot;https://www.courtlistener.com/opinion/74536637/shih-v-petal-card-inc/&quot;&gt;&lt;em&gt;Shih v. Petal Card, Inc.&lt;/em&gt;&lt;/a&gt;, 565 F. Supp. 3d 557 (S.D.N.Y. 2021), which recognized work product protection for a party&#39;s own litigation-preparation materials regardless of attorney direction.&lt;/p&gt;
&lt;h2&gt;Where I think the reasoning falters&lt;/h2&gt;
&lt;p&gt;The first and third grounds — no attorney-client relationship, no communication for the purpose of obtaining legal advice from an attorney — are each independently sufficient to defeat the privilege claim. An AI tool is not a lawyer, and Heppner was not seeking legal advice from an attorney when he typed queries into Claude. Full stop.&lt;/p&gt;
&lt;p&gt;The work product holding is correct on these facts — defense counsel conceded that Heppner acted without direction — but the court&#39;s reasoning adopted a narrower view of the doctrine than the weight of authority supports. The traditional Second Circuit formulation protects &amp;quot;materials prepared by or at the behest of counsel in anticipation of litigation or for trial,&amp;quot; but the civil analog, Fed. R. Civ. P. 26(b)(3)(A), protects materials prepared &amp;quot;by or for another party or its representative&amp;quot; — language broad enough to cover a party acting on its own initiative. The court&#39;s express rejection of &lt;em&gt;Shih&lt;/em&gt; on this point signals that the question remains open, and future courts should not treat &lt;em&gt;Heppner&lt;/em&gt;&#39;s narrow formulation as settled.&lt;/p&gt;
&lt;p&gt;The confidentiality analysis in the second ground is where things get shaky, and it is the part of the opinion that has generated the most commentary — and the most anxiety.&lt;/p&gt;
&lt;p&gt;Judge Rakoff treated Anthropic&#39;s consumer privacy policy as establishing that Heppner could have &amp;quot;no reasonable expectation of confidentiality&amp;quot; in his AI conversations. But the court&#39;s analysis has significant gaps. The opinion cited an archived version of Anthropic&#39;s privacy policy dated February 2025 — a version that predated the August 2025 consumer terms update giving users the ability to control model training. Because Heppner used Claude in 2025 before his November arrest, his conversations may have been governed by either the old or the new terms depending on when they occurred. The court never asked what version of the terms governed Heppner&#39;s use, whether he had opted out of training, or what his actual settings were. It treated the broadest possible reading of the consumer terms as conclusive without examining what the user actually agreed to or configured.&lt;/p&gt;
&lt;p&gt;This matters because the confidentiality holding — which was not necessary to the result — is the part of the opinion most likely to be cited broadly. And it rests on an incomplete factual record. As the policy comparison below demonstrates, Anthropic&#39;s consumer terms create meaningfully different data-handling regimes depending on whether a user has opted in or out of model training. The court did not grapple with that distinction.&lt;/p&gt;
&lt;p&gt;There is also a subtler problem. The opinion conflates a platform&#39;s contractual &lt;em&gt;permission&lt;/em&gt; to use data with the practical &lt;em&gt;likelihood&lt;/em&gt; that any human will ever see it. Consumer AI privacy policies reserve broad rights, but the actual probability of a specific conversation being reviewed by a person — absent a safety flag or legal process — is vanishingly low. Whether that distinction should matter for privilege purposes is a genuinely hard question. &lt;em&gt;Heppner&lt;/em&gt; does not engage with it.&lt;/p&gt;
&lt;p&gt;None of this means the opinion is unimportant. It is the first federal decision to address AI and privilege head-on, and it will shape how courts and litigants think about these issues going forward. But its broadest holding — that consumer AI use necessarily destroys confidentiality — rests on reasoning that future courts should scrutinize carefully.&lt;/p&gt;
&lt;h2&gt;What the case gets right: a knowledge problem&lt;/h2&gt;
&lt;p&gt;Where &lt;em&gt;Heppner&lt;/em&gt; is most valuable is as a signal. Whatever one thinks of the doctrinal analysis, the case exposes a widespread failure to understand how consumer AI tools handle data. Heppner apparently did not know — or did not care — that his AI conversations were governed by terms that reserved broad data-use rights for the platform provider. His lawyers did not anticipate that their client&#39;s independent AI use would create a discovery problem. And the court itself did not dig into the specific settings or tier the defendant used.&lt;/p&gt;
&lt;p&gt;This is not an isolated failure. Most lawyers I talk to cannot articulate the difference between a consumer and enterprise AI deployment. Most clients do not read privacy policies. And most courts have not yet had to think carefully about how AI data handling intersects with privilege doctrine.&lt;/p&gt;
&lt;p&gt;&lt;em&gt;Heppner&lt;/em&gt; should change that — not because its reasoning is airtight, but because it demonstrates what happens when no one in the room understands the technology well enough to ask the right questions.&lt;/p&gt;
&lt;h2&gt;What Anthropic&#39;s policies actually say&lt;/h2&gt;
&lt;p&gt;Since &lt;em&gt;Heppner&lt;/em&gt; turned on Anthropic&#39;s terms, this is the right place to start. I went through Anthropic&#39;s published policies — the &lt;a href=&quot;https://www.anthropic.com/terms&quot;&gt;Consumer Terms of Service&lt;/a&gt;, the &lt;a href=&quot;https://www.anthropic.com/news/expanded-legal-protections-api-improvements&quot;&gt;Commercial Terms of Service&lt;/a&gt;, the &lt;a href=&quot;https://www.anthropic.com/news/updates-to-our-consumer-terms&quot;&gt;Privacy Policy&lt;/a&gt;, and the &lt;a href=&quot;https://privacy.claude.com/en/articles/10458704-how-does-anthropic-protect-the-personal-data-of-claude-users&quot;&gt;Privacy Center&lt;/a&gt; — to compare what Claude&#39;s consumer and commercial tiers actually promise. What follows is a synthesis of that research.&lt;/p&gt;
&lt;h3&gt;The core divide: consumer terms vs. commercial terms&lt;/h3&gt;
&lt;p&gt;Anthropic&#39;s policies split along two fundamental lines: &lt;strong&gt;Consumer Terms&lt;/strong&gt; (Free, Pro, Max) and &lt;strong&gt;Commercial Terms&lt;/strong&gt; (Team, Enterprise, API, Education, Government). This distinction — not the price paid — determines virtually every data right the user holds. The Commercial Terms state explicitly: &amp;quot;Services under these Terms are not for consumer use. Our consumer offerings (e.g., Claude.ai) are governed by our Consumer Terms of Service instead.&amp;quot;&lt;/p&gt;
&lt;p&gt;This means a Pro or Max subscriber paying $20 or $100 per month operates under the same legal framework as a free user. Paying more buys additional model access and features, but it does not change how Anthropic treats your data.&lt;/p&gt;
&lt;h3&gt;Model training: the sharpest divide&lt;/h3&gt;
&lt;p&gt;For &lt;strong&gt;Free, Pro, and Max&lt;/strong&gt; users, Anthropic may use conversations to train its models. In &lt;a href=&quot;https://www.anthropic.com/news/updates-to-our-consumer-terms&quot;&gt;August 2025&lt;/a&gt;, Anthropic updated its consumer terms to give users the ability to control whether their data would be used for model training. Existing users had until October 8, 2025, to accept the new terms and select their preference. The operative contractual language states that Anthropic may use user materials for model training &amp;quot;unless users opt out&amp;quot; — placing the default in Anthropic&#39;s favor — though Anthropic&#39;s own blog post announcing the change described it as &amp;quot;allowing users on Claude Free, Pro, and Max plans to opt-in for data usage,&amp;quot; framing the default in the opposite direction. The tension between the legal text and the public announcement underscores the difficulty of determining any individual user&#39;s training status based on the terms alone. Opting out remains available through Claude&#39;s settings.&lt;/p&gt;
&lt;p&gt;For &lt;strong&gt;Team, Enterprise, API, and Education/Government&lt;/strong&gt; users, Anthropic contractually prohibits itself from training on customer content. The Commercial Terms are unambiguous: &amp;quot;Anthropic may not train models on Customer Content from Services&amp;quot; — with no exceptions and no reliance on user-level toggles.&lt;/p&gt;
&lt;h3&gt;Data retention: a 60× gap&lt;/h3&gt;
&lt;p&gt;Retention periods are directly tied to training status for consumer plans, creating a striking disparity:&lt;/p&gt;
&lt;p&gt;Consumer users who have &lt;strong&gt;opted in&lt;/strong&gt; to training (or failed to opt out) face retention of up to &lt;strong&gt;five years&lt;/strong&gt; for de-identified conversation data. Consumer users who have &lt;strong&gt;opted out&lt;/strong&gt; see their conversations retained for &lt;strong&gt;30 days&lt;/strong&gt; before deletion. In either case, content flagged for safety or policy violations can be retained for up to &lt;strong&gt;seven years&lt;/strong&gt;, regardless of the user&#39;s training preference.&lt;/p&gt;
&lt;p&gt;On the commercial side, &lt;strong&gt;API&lt;/strong&gt; input and output logs are retained for &lt;strong&gt;seven days&lt;/strong&gt;. &lt;strong&gt;Enterprise&lt;/strong&gt; accounts default to &lt;strong&gt;30 days&lt;/strong&gt;, with the option to negotiate &lt;a href=&quot;https://privacy.claude.com/en/articles/8956058-i-have-a-zero-data-retention-agreement-with-anthropic-what-products-does-it-apply-to&quot;&gt;Zero Data Retention&lt;/a&gt; — under which inputs and outputs are processed in real time and not stored at all. No consumer plan, regardless of price, offers true zero retention.&lt;/p&gt;
&lt;h3&gt;Data ownership and IP&lt;/h3&gt;
&lt;p&gt;The Commercial Terms contain an unusually strong ownership clause absent from the consumer terms. They provide that the customer &amp;quot;retains all rights to its Inputs, and owns its Outputs,&amp;quot; that &amp;quot;Anthropic disclaims any rights it receives to the Customer Content under these Terms,&amp;quot; and that Anthropic &amp;quot;hereby assigns to Customer its right, title and interest (if any) in and to Outputs.&amp;quot;&lt;/p&gt;
&lt;p&gt;Consumer users have no equivalent contractual assignment. Under the consumer framework, Anthropic holds a license to use inputs and outputs for model improvement unless the user opts out.&lt;/p&gt;
&lt;h3&gt;Data controller vs. data processor&lt;/h3&gt;
&lt;p&gt;This distinction carries significant weight under GDPR and analogous privacy regimes. For &lt;strong&gt;consumer plans&lt;/strong&gt;, Anthropic acts as the &lt;strong&gt;data controller&lt;/strong&gt; — it determines the purposes and means of processing user data. For &lt;strong&gt;Enterprise and API&lt;/strong&gt; accounts, Anthropic functions as a &lt;strong&gt;data processor&lt;/strong&gt; operating under a Data Processing Addendum, with the commercial customer serving as the controller.&lt;/p&gt;
&lt;p&gt;The practical consequence: a consumer user&#39;s data is governed by Anthropic&#39;s privacy choices. An enterprise customer&#39;s data is governed by the customer&#39;s own policies, with Anthropic acting under instruction.&lt;/p&gt;
&lt;h3&gt;Employee access and confidentiality&lt;/h3&gt;
&lt;p&gt;For consumer plans, Anthropic employees may access conversations only if the user explicitly consents via feedback, or if access is required for Usage Policy enforcement — in which case only the Trust &amp;amp; Safety team may view content on a need-to-know basis.&lt;/p&gt;
&lt;p&gt;For commercial plans, customer content is contractually designated as &lt;strong&gt;Confidential Information&lt;/strong&gt; under the Commercial Terms. Anthropic may use it only to exercise its rights under the contract and must protect it with at least the same care it applies to its own confidential information.&lt;/p&gt;
&lt;p&gt;Two further protections — Zero Data Retention and HIPAA Business Associate Agreements — are available exclusively on commercial tiers. Under ZDR, inputs and outputs are not stored; the sole exception is User Safety classifier results retained for Usage Policy enforcement. A BAA imposes specific configuration requirements and excludes certain features (web search, for instance, falls outside BAA coverage). Neither protection is available on any consumer plan at any price point.&lt;/p&gt;
&lt;p&gt;The comparison distills to a structural reality: consumer Claude users — whether free or paying $100 per month — operate under terms that allow Anthropic to train on their data by default, retain it for up to five years, and act as the data controller with broad discretion. Commercial Claude users operate under a contractual regime that prohibits model training, treats their content as confidential information, assigns them ownership of outputs, and offers zero-retention options.&lt;/p&gt;
&lt;h2&gt;The pattern holds across providers&lt;/h2&gt;
&lt;p&gt;Anthropic&#39;s tiered structure is not an outlier. OpenAI&#39;s ChatGPT follows the same pattern. On Free and Plus plans, OpenAI&#39;s &lt;a href=&quot;https://help.openai.com/en/articles/7039943-data-usage-for-consumer-services-faq&quot;&gt;Data Usage for Consumer Services FAQ&lt;/a&gt; states that it &amp;quot;may use&amp;quot; consumer content to improve its models unless the user disables training — while retaining the right to log interactions for safety and abuse monitoring regardless. On &lt;a href=&quot;https://help.openai.com/en/articles/9377311-chatgpt-edu-at-openai&quot;&gt;Edu and Enterprise&lt;/a&gt; plans, OpenAI commits not to train on business data, provides admin-controlled retention windows, and offers &lt;a href=&quot;https://developers.openai.com/api/docs/guides/your-data/&quot;&gt;Zero Data Retention and configurable data residency&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;The structural divide is the same: consumer terms grant the provider broad data-use rights with an opt-out toggle; commercial terms prohibit model training by contract and give the customer control over retention, residency, and access. Google&#39;s Gemini, Meta&#39;s Llama-based offerings, and other major LLM providers follow similar patterns. The consumer-versus-commercial distinction is an industry-wide architectural choice, not a quirk of any single provider.&lt;/p&gt;
&lt;p&gt;This matters for the &lt;em&gt;Heppner&lt;/em&gt; analysis because the court&#39;s reasoning — resting on the provider&#39;s privacy policy and terms of service — would apply with equal force to any consumer LLM deployment, not just Claude.&lt;/p&gt;
&lt;h2&gt;What this means going forward&lt;/h2&gt;
&lt;p&gt;&lt;em&gt;Heppner&lt;/em&gt; will be cited for the proposition that consumer AI conversations are not confidential. That proposition is probably too broad as stated — it ignores user training preferences, conflates contractual permission with practical disclosure risk, and was not necessary to the holding. But it captures something real: consumer AI platforms operate under terms that were not designed with legal privilege in mind, and users who rely on those platforms for sensitive work are taking risks they may not understand.&lt;/p&gt;
&lt;p&gt;The practical response is not to avoid AI tools. It is to understand what you are agreeing to when you use them — and to recognize that paying for a subscription does not, by itself, change the legal framework governing your data. For lawyers, that means learning the difference between consumer and commercial deployments and advising clients accordingly. For organizations, it means treating AI procurement as a legal risk question, not just an IT question. And for courts, it means doing the factual work that &lt;em&gt;Heppner&lt;/em&gt; did not: examining the specific terms, settings, and tier a user actually employed before concluding that confidentiality has been waived.&lt;/p&gt;
&lt;p&gt;The gap between consumer and commercial AI products is wide, it is well-documented, and it is consistent across every major provider. The problem is not that the information is unavailable. The problem is that almost nobody — lawyers, clients, and judges included — reads it.&lt;/p&gt;
&lt;hr&gt;
&lt;p&gt;&lt;em&gt;The Anthropic policy comparison in this post draws on Anthropic&#39;s &lt;a href=&quot;https://www.anthropic.com/terms&quot;&gt;Consumer Terms of Service&lt;/a&gt;, &lt;a href=&quot;https://www.anthropic.com/news/expanded-legal-protections-api-improvements&quot;&gt;Commercial Terms announcement&lt;/a&gt;, &lt;a href=&quot;https://www.anthropic.com/news/updates-to-our-consumer-terms&quot;&gt;consumer terms and privacy policy update&lt;/a&gt;, and &lt;a href=&quot;https://privacy.claude.com/en/articles/10458704-how-does-anthropic-protect-the-personal-data-of-claude-users&quot;&gt;Privacy Center&lt;/a&gt;. OpenAI policy references draw on the &lt;a href=&quot;https://help.openai.com/en/articles/7039943-data-usage-for-consumer-services-faq&quot;&gt;Data Usage FAQ&lt;/a&gt;, &lt;a href=&quot;https://developers.openai.com/api/docs/guides/your-data/&quot;&gt;platform documentation&lt;/a&gt;, and &lt;a href=&quot;https://openai.com/policies/row-privacy-policy/&quot;&gt;privacy policy&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;
</content>
  </entry>
</feed>
