Incident story · AI data exposure
The associate just wanted to summarize tax documents. Consumer ChatGPT retains for training.
The managing partner of a 22-CPA tax and advisory firm in the West got the email on a Tuesday in late March — the busiest two weeks of the year. A long-time client, a small-business owner, asked a question that did not make sense: had her tax return information been "used to train an AI" somehow? Three hours later the firm understood what had happened. An associate, behind on a backlog of advisory engagements, had spent two weeks pasting entire client tax document sets into the consumer version of ChatGPT to summarize. Consumer-tier retains content for training. The associate had thought she was using the firm's enterprise account. She had not been. Approximately 3,800 individual client records — Social Security numbers, EINs, prior-year returns, banking details — had sat in OpenAI's consumer training pipeline for two weeks before any opt-out was applied.
At a glance
The shape of the incident
What happened
Two weeks of pasting, sixteen days before anyone noticed
Mid-March — the backlog. An advisory associate, two years into her career, was working through a backlog of small-business engagements. Each one involved reading three to five years of returns plus current-year working papers plus QuickBooks exports — and synthesizing a memo. She had heard about ChatGPT being able to summarize long documents. She opened a browser tab, signed into ChatGPT with her personal Gmail account (the firm's SSO did not cover consumer AI), and began uploading.
What she thought she was using. The firm had purchased ChatGPT Enterprise the previous fall for the partners. The associate had attended the rollout meeting. She remembered seeing the firm's logo on the sign-in screen. She did not realize her browser had defaulted to her personal account, not the firm tenant. The interface looked identical. The firm's IT-managed devices did not block consumer ChatGPT at the network or identity layer.
What ChatGPT's consumer tier does. Consumer ChatGPT — the Free, Plus, and Team tiers as they existed at the time — retains uploaded content and chat history for model improvement unless the user has explicitly turned off training in settings. The associate had never opened settings. Every document she uploaded over two weeks was added to a content pool eligible for future model training.
Two weeks of uploads. Approximately 47 client engagements. Several hundred individual tax returns spanning three years. QuickBooks exports including bank account numbers and EINs. Prior-year W-2 and 1099 detail. K-1s with high-net-worth distribution information. Approximately 3,800 individual taxpayer records implicated when later scoped.
Day 16 — the client call. A small-business owner client called to ask about her engagement memo. The associate had referenced specific details about her prior-year return that the client had not raised in their current-year meeting. The client's husband, who worked in tech, asked his wife to ask the question directly: was AI involved in preparing this memo, and if so, what AI. The client called the managing partner.
Hours 1 to 8 — the investigation. The managing partner asked the associate. She showed him her workflow. He saw the personal Gmail sign-in. He called the firm's outside counsel before he called anyone else.
Days 1 to 7 — scope determination. Through counsel, the firm requested a data-export from OpenAI under the consumer account's user data rights. The export confirmed two weeks of uploads. The retroactive opt-out from training was applied — but only forward-looking. OpenAI confirmed in writing that the prior uploads had been in the training-eligible pool for the full two weeks before opt-out.
Day 21 — IRS, state AG, and client notification. The firm filed notification with the IRS under tax preparer rules, with state attorneys general in the eight states where affected clients resided, and with every individual client. The notification named ChatGPT directly. The firm offered two years of credit monitoring and identity protection.
Days 21 to 90 — the exodus. The firm's A-tier clients — high-net-worth families and small-business owners with the most to lose — began moving. 31 percent of A-tier clients had transitioned to another firm by day 90. The remainder stayed but with explicit AI-use restrictions written into engagement letters. B-tier and C-tier retention was relatively unaffected.
What it cost
The bill, itemized
A-tier client churn — 31% departed within 90 days
$2,400,000Trailing-12-months revenue impact. The firm's top tier of clients (high-net-worth and small-business owners) left fastest because they had the most to lose.
Client notification + credit monitoring
$420,000~3,800 individuals across 8 states. State-by-state notification compliance, mailed notices, call center, two years of credit + identity monitoring.
DFIR + AI exposure forensic investigation
$180,000Determine exactly what was uploaded, what tier of ChatGPT was used, the retention status at the time of upload, and the post-incident opt-out scope.
State AG + IRS notification + legal defense
$240,000Eight state attorney general notifications, IRS notification under tax preparer rules, breach counsel through the response period.
AI governance program + enterprise tier rebuild
$95,000Block consumer AI at identity layer, deploy enterprise AI tier with DLP, AI use policy rebuild, mandatory training, quarterly attestation.
Cyber insurance premium increase + new exclusions
$215,000Premium up 165% YoY. New AI use exclusions added unless firm could prove enterprise-tier with executed terms.
Cyber insurance recovery
($650,000)Carrier covered investigation, notification, and partial defense. Excluded the majority of revenue loss citing AI use exclusions in the renewal terms (not in the prior year's policy).
Net to the firm
~$2,900,000Two partners deferred draws for 14 months. One partner left the firm. Trailing reputational impact still being felt 22 months later.
What we did
EFROS-style response — what an engagement looks like
First call — stop the bleeding, preserve evidence. We immediately block consumer ChatGPT, Claude, Gemini, and Perplexity at the firm's identity layer and at the network edge. We preserve the associate's browser history, the firm's SSO logs, and the OpenAI consumer account export through counsel. We do not delete the consumer account — that is evidence.
Days 1 to 5 — scope. We work with the associate to reconstruct every upload over the two-week window. We cross-reference against client engagement files to identify which clients were touched, what specific data was in each upload, and which states have notification obligations based on client residency. The scope output goes directly to breach counsel.
Week 2 — notification strategy. Tax preparer notification rules and state AG breach notification rules apply in different windows. We help the firm prioritize: IRS notification within seven days where required, state AG notification according to each state's timeline (some 30 days, some shorter), client notification as soon as the scope is firm enough to support honest disclosure. The notification names ChatGPT — naming the service is what regulators expect.
Weeks 3 to 8 — AI governance program rebuild. We deploy enterprise AI tiers with full audit logging, contractual no-training terms, and DLP for sensitive data classes (SSNs, EINs, account numbers). We rebuild the firm's AI use policy with specific named tools approved or blocked. Every employee signs the policy with attestation that they have completed associated training. We add a quarterly attestation cycle.
Months 2 to 9 — client retention. The managing partner spends months on the phone with A-tier clients. We provide the technical evidence package — what we have done, what the new controls are, what AI tools are now blocked, what is in writing with vendors. Most clients who stay want to see the policy and the enterprise contract terms. We help the firm produce those for client diligence.
Ongoing. Quarterly AI use attestation. Monthly review of AI traffic at the network edge. Annual AI use policy refresh. The firm's engagement letters now name AI use restrictions explicitly — the clients who stayed wrote it in themselves.
What you should take from this
Five things to do this week
- 01
ChatGPT Plus, ChatGPT Free, and ChatGPT Team are not the enterprise version. Only ChatGPT Enterprise and the OpenAI API offer the contractual terms a regulated industry needs — no training on customer data, audit logging, admin control. An employee with a personal Plus subscription using it for work is the same as printing client files and dropping them in a public trash can.
- 02
AI use policy must exist before the first employee opens a browser tab. Not after the incident. The policy needs to say what is allowed, what is blocked, who pays for enterprise access, and what the disciplinary consequence is for unauthorized use.
- 03
Identity-layer blocking is the only control that works. Telling people 'do not use consumer AI for client data' is not a control. Blocking the consumer tier at the corporate identity layer — so they cannot sign in with their work account — is a control. The associate in this story used her personal Gmail, but the block would have at least made the policy real.
- 04
Cyber insurance is rewriting AI exclusions in real time. Many policies issued before 2025 covered AI exposure by default. Many policies issued after exclude it unless you can prove enterprise-tier deployment with executed contract terms. Read the next renewal carefully.
- 05
Client trust in a CPA firm is built on the assumption that financial data does not leave the firm. The moment that assumption is violated, A-tier clients — the ones with the most at stake — move first and fastest. The firm in this story lost a third of its top tier within 90 days, not because the regulator forced anyone to leave but because clients lost faith.
The 60-second self-check
Three yes/no questions
If any answer is no — or any answer is "I think so" — you have the same exposure profile as the firm in this story.
1. If an employee opens ChatGPT, Claude, or Gemini on their work computer right now using their personal account, would your systems stop them — or would they be able to paste a client document and click submit?
If the answer is 'we tell them not to,' that is not a control.
2. Do you have a written AI use policy — signed by every employee — that names which AI tools are approved, which are blocked, and what happens if someone uploads client data to a non-approved tool?
If you cannot produce the signed policy in two minutes, it does not exist for compliance purposes.
3. If a client called tomorrow asking 'has any of my data been entered into any AI tool by your firm,' could you produce an answer with logs and audit evidence — within 48 hours?
Enterprise AI tiers log this. Consumer tiers do not. If your firm is on consumer, you cannot answer this question at all.
What this would cost you
Three private numbers, none of them require talking to a salesperson.
Related incident stories
Two more patterns owners ask about
Orthopedic specialty clinic
Insider data theft — $1.8M settlement
A departing medical assistant spent her last two weeks copying patient records to personal cloud. 4,200 patients notified. Eighteen-month HHS-OCR audit.
Regional logistics firm
Supply-chain compromise — $8M ARR lost
The dispatch software vendor pushed a routine update containing a backdoor. Customer contracts on a dark-web forum. The largest shippers did not renew.
Names, locations, and identifying details changed. Numbers represent typical ranges from EFROS engagements; specific cases vary. Nothing on this page is legal advice.