On this page
- TL;DR
- What ChatGPT Enterprise actually provides
- What ChatGPT Enterprise does NOT do
- 1. Your conversations still live on OpenAI infrastructure
- 2. OpenAI employees can (under controlled circumstances) access it
- 3. US legal process applies
- 4. Breach risk is non-zero
- 5. Connectors extend the data perimeter
- When ChatGPT Enterprise is the right call
- When ChatGPT Enterprise is NOT the right call
- The real question: what's your threat model?
- 1. What would be bad if this leaked?
- 2. Who are you defending against?
- 3. What does compliance actually require?
- The alternatives
- Option 1: BYOK with local-first client
- Option 2: Self-hosted open-source models
- Option 3: EU/UK-sovereign providers
- Option 4: Azure OpenAI with private deployment
- The decision tree
- The uncomfortable truth
- How to actually evaluate a vendor
- The summary
TL;DR
- ChatGPT Enterprise offers real upgrades over consumer ChatGPT: no training on your data, SOC 2, encryption, SSO, DLP hooks.
- But it's not what most buyers assume: your conversations still live on OpenAI's servers, subject to their retention policy, subpoenas, and employee access.
- For teams that need genuine data isolation (healthcare, legal, regulated industries), BYOK with a local-first client or self-hosted open-source models is often a better fit.
- The right architecture depends on your threat model. This post helps you figure out which one you need.
What ChatGPT Enterprise actually provides
ChatGPT Enterprise at the time of writing:
- No training on your data. This is the headline. OpenAI won't train future models on your org's conversations.
- SOC 2 Type 2 certification. Real compliance assurance.
- Encryption in transit and at rest. Table stakes, but verified.
- SAML SSO and SCIM. Standard enterprise identity controls.
- Admin console, audit logs. Governance tooling.
- Data Processing Addendum (DPA). Legal assurances around data handling.
- Custom data retention. You can set how long conversations are retained (down to minimum allowed by abuse monitoring).
This is a genuinely good product for many organizations. It's also not the same thing as "your data stays with you."
What ChatGPT Enterprise does NOT do
Here's where buyers often misunderstand:
1. Your conversations still live on OpenAI infrastructure
Every message, every attachment, every response exists in OpenAI's databases for at least the minimum abuse-monitoring retention period (typically 30 days, extendable up to longer for flagged content). Even with "no training," the data exists — it's just not used for model training.
2. OpenAI employees can (under controlled circumstances) access it
OpenAI's policies permit employee access for:
- Abuse monitoring
- Legal compliance
- Debugging with customer consent
- Trust & safety investigations
This is standard industry practice. But it's not "no one sees it."
3. US legal process applies
If OpenAI is subpoenaed (by US authorities, by civil litigants, by grand juries), your conversations are discoverable. Data in the US is subject to US law — including CLOUD Act, FISA, and regular subpoena processes. This is true for every US-based AI provider.
4. Breach risk is non-zero
OpenAI has had security incidents. Every company has. The more data is stored, the more material a potential breach exposes.
5. Connectors extend the data perimeter
The Connectors feature lets ChatGPT Enterprise read your Google Drive, GitHub, Gmail, Salesforce, etc. This creates a widening data exposure surface — those services now have access granted to your ChatGPT tenant.
When ChatGPT Enterprise is the right call
Be honest with yourself. It's a good fit when:
- Your data is not exceptionally sensitive. Internal docs, strategy memos, general product work — OK.
- You need broad adoption with low friction. ChatGPT is what your team already knows.
- Compliance needs are covered by SOC 2 + DPA. Most corporate buyers' needs, actually.
- You want OpenAI's support SLA and product polish.
- You're willing to pay the per-seat premium for convenience (typically $25-60/seat/month).
For a marketing team at a B2B SaaS company? Perfect. For a consulting firm writing client work? Probably fine. For a legal team handling privileged material? Read on.
When ChatGPT Enterprise is NOT the right call
Healthcare (PHI): HIPAA has real teeth. While OpenAI can sign BAAs for enterprise customers, the retention and access model is still "US-hosted cloud." Many healthcare organizations prefer self-hosted or BYOK with explicit control over every byte.
Legal (privileged communications): Attorney-client privilege can be compromised by third-party access. Conservative firms won't put privileged material into any cloud AI, even Enterprise.
Financial services (trade secrets, non-public info): Regulators increasingly demand strict data-flow controls. "We pay OpenAI for enterprise tier" doesn't cut it for all audits.
Defense / intelligence / government: Obviously different threat model.
Non-US sovereignty: EU, UK, and other jurisdictions increasingly disfavor US-hosted AI for sensitive work due to US surveillance law applicability.
Extremely sensitive R&D: IP worth $100M+ shouldn't sit on a third party's servers, period.
The real question: what's your threat model?
Don't evaluate "privacy" abstractly. Ask three questions:
1. What would be bad if this leaked?
- "Embarrassing but survivable" → ChatGPT Enterprise is fine.
- "Regulatory fines" → need stricter controls.
- "Existential (trade secrets, patient data, national security)" → self-host or BYOK with local-first only.
2. Who are you defending against?
- Random cybercriminals: SOC 2 + encryption is enough.
- Insider threats at the AI provider: you need data isolation.
- State actors / subpoenas: need non-US hosting or full local control.
3. What does compliance actually require?
- Read your DPA, HIPAA BAA, or regulatory framework carefully. "Data stays in the US" and "data stays with us" are different sentences.
The alternatives
Option 1: BYOK with local-first client
You get API keys from OpenAI, Anthropic, Google, etc. You use a local-first BYOK client (NovaKit for example) where your conversations live in your own browser, your keys are encrypted locally (AES-256-GCM), and API calls go directly to the provider.
What this buys you:
- Conversation history on your own device, not a third-party's database.
- Same underlying models — GPT-4o, Claude Opus 4 — via API.
- API-tier data policies (no training, short retention) apply, not consumer-tier.
- Zero telemetry on your usage going to the client maker.
What it doesn't fix:
- The model calls themselves still hit OpenAI/Anthropic API servers; US legal process still applies to API calls.
- You still have to trust the provider.
Best for: Individuals and small teams who want a meaningful privacy upgrade without self-hosting.
Option 2: Self-hosted open-source models
Use Llama 3.3 70B, Qwen 2.5 72B, or DeepSeek on your own hardware (or a dedicated GPU cloud you control).
What this buys you:
- Zero external API calls. Your data never leaves your infrastructure.
- Full auditability.
- Compliance with even the strictest sovereignty requirements.
What it costs you:
- Hardware (~$15-30k for a server running a 70B model at production speeds, or ~$3-5k/month of cloud GPU).
- Operational complexity (updates, monitoring, scaling).
- Quality: open-source is good but still not GPT-5 or Claude Opus 4 at the frontier.
Best for: Regulated industries, large enterprises with existing infra teams, anyone where cost of a data breach >>> cost of hosting.
Option 3: EU/UK-sovereign providers
Mistral (France), Aleph Alpha (Germany), and UK-sovereign AI offerings host in-region with jurisdiction-specific legal protections.
What this buys you:
- Data residency in EU/UK.
- No US CLOUD Act / FISA exposure.
- Compliance with stricter EU regulations (AI Act, GDPR).
Trade-offs:
- Slightly fewer model options.
- Sometimes higher cost per token.
- Models are good but generally a half-step behind frontier US labs.
Option 4: Azure OpenAI with private deployment
Microsoft's Azure OpenAI Service lets you deploy GPT models in your own Azure subscription, in specific regions, with stricter data controls than OpenAI's direct API.
What this buys you:
- Microsoft's compliance stack (HIPAA BAA, FedRAMP High, DoD IL-5, etc.).
- Regional data residency.
- Enterprise integration with the Microsoft ecosystem.
Trade-offs:
- Still ultimately running on a cloud provider.
- Usually newer models arrive on Azure weeks/months after OpenAI's direct API.
- Cost structure is opaque and usually higher.
The decision tree
Q: Is your data subject to HIPAA, attorney-client privilege, or export controls?
- Yes → Self-hosted open-source or Azure with proper contracts.
- No → continue.
Q: Do you have hard sovereignty requirements (EU-only, no US)?
- Yes → Mistral or another EU-sovereign provider; or self-host.
- No → continue.
Q: Do you have a dedicated infra team?
- Yes → Self-host is an option, but not mandatory.
- No → BYOK or ChatGPT Enterprise.
Q: Is per-seat pricing acceptable?
- Yes → ChatGPT Enterprise is fine (if threat model allows).
- Prefer pay-as-you-go → BYOK.
Q: Do you want conversations on your local devices only?
- Yes → BYOK with a local-first client.
- Cloud sync is fine → Either works.
The uncomfortable truth
"Privacy" in AI is a spectrum, not a checkbox. ChatGPT Enterprise is more private than consumer ChatGPT — really, meaningfully. It is less private than BYOK with a local-first client. BYOK with a local-first client is less private than self-hosted open-source. Self-hosted open-source is less private than pen and paper.
Pick the point on the spectrum your threat model requires. Don't buy "enterprise" and assume you've solved it. Don't self-host for ego when enterprise would cover your real risks.
How to actually evaluate a vendor
Five questions to ask (and get in writing):
- Where is our data stored, physically and logically?
- What's the default retention, and can we override it?
- Who at your company can access our data, and under what conditions?
- What's your subpoena / legal compliance process?
- What happens to our data if we churn?
Any vendor who can't answer these cleanly doesn't deserve your sensitive data.
The summary
- ChatGPT Enterprise is good for most corporate use cases. Not for everything.
- Know your threat model before picking a tier.
- BYOK + local-first gives real privacy gains at a lower cost for individuals and small teams.
- Self-host when your data is too sensitive for any third party.
- "Enterprise" doesn't mean "stays with you." Read the DPA.
NovaKit is a BYOK, local-first AI workspace — conversations stay in your own browser, API keys are encrypted locally (AES-256-GCM), and nothing goes through our servers. For teams that want enterprise-grade privacy without enterprise-grade data exposure.