
When employees use ChatGPT at work, every prompt they type potentially shares company data with an external service. While enterprise AI plans have stronger data protections, the risk of data leakage is real — and one careless prompt can expose customer information, trade secrets, or confidential business data.
This guide explains the specific risks and practical steps to prevent data leakage.
An employee pastes a customer complaint email (including the customer's name, account number, and order details) into ChatGPT to draft a response. The customer's personal data is now processed by an external service.
Over multiple prompts, an employee shares enough context about a confidential project — team names, financial targets, strategic plans — that the accumulated information constitutes a confidential briefing.
A developer pastes proprietary source code into ChatGPT for debugging help. The code may contain algorithms, API keys, or business logic that constitutes trade secrets.
With consumer-tier AI products, user prompts may be used to improve the model. This means sensitive data could theoretically influence future outputs visible to other users. (Enterprise plans typically exclude data from training.)
The first defence against data leakage is a clear data classification system. Every piece of information in your company falls into one of these categories:
Information that is already publicly available or intended for public distribution.
AI Rule: Can be freely used with any AI tool.
Information that is not confidential but is meant for internal use only.
AI Rule: May be used with approved enterprise AI tools only (not free-tier consumer products).
Information that could harm the company or individuals if disclosed.
AI Rule: Must be anonymised before use. Remove all identifying details (names, numbers, dates). Use only with approved enterprise AI tools.
Information that must never enter any external AI system.
AI Rule: NEVER enter into any AI tool, under any circumstances.
Consumer-tier AI products (free ChatGPT, free Claude) have different data handling practices than enterprise plans. Key differences:
| Feature | Consumer/Free | Enterprise |
|---|---|---|
| Data used for training | Often yes | Typically no |
| Data retention | Extended | Limited/configurable |
| Admin controls | None | Full |
| Usage monitoring | None | Audit logs |
| Data processing agreement | None | Available |
| Compliance certifications | Limited | SOC 2, ISO 27001 |
Every employee who uses AI tools must understand:
Before pasting any text into an AI tool, check for and remove:
When data leakage occurs (or is suspected):
The Personal Data Protection Act requires organisations to protect personal data and obtain consent for its use. Inputting personal data into AI tools without proper safeguards may constitute a breach. Penalties can reach S$1 million per breach.
Malaysia's Personal Data Protection Act similarly requires organisations to safeguard personal data. Sharing personal data with AI services may violate data processing principles if proper consent and safeguards are not in place.
A company with effective AI data protection:
The landscape of ChatGPT data leakage prevention shifted dramatically between early 2024 and March 2026, driven by three converging developments: OpenAI's enterprise architecture updates, regulatory enforcement actions, and the emergence of dedicated interception technologies.
Enterprise API Controls versus Browser-Based Usage. Organizations that relied solely on acceptable use policies discovered through incident reports that browser-based ChatGPT sessions remained the primary exfiltration vector. OpenAI introduced Team and Enterprise workspace tiers with data retention opt-outs and administrative conversation logging, but these controls only apply when employees use sanctioned accounts. Shadow usage through personal subscriptions continues to bypass organizational safeguards entirely.
DLP Gateway Solutions. Dedicated proxy tools now inspect prompts before they reach external language model endpoints. Nightfall AI, Microsoft Purview (formerly Compliance Manager), Zscaler GenAI Security, and Harmonic Security each intercept outbound requests and scan for sensitive patterns including personally identifiable information, source code fragments, financial projections, and intellectual property markers. Nightfall's classification engine uses context-aware detection trained on healthcare records, legal documents, and engineering codebases, achieving approximately ninety-two percent precision according to their published benchmark from September 2025.
Regulatory Enforcement Precedents. Italy's Garante temporarily suspended ChatGPT operations in March 2023, and subsequent GDPR enforcement guidance from the European Data Protection Board (EDPB Opinion 28/2024) established that submitting personal data into generative models constitutes processing under Article 4(2). South Korea's Personal Information Protection Commission (PIPC) issued similar interpretive guidance in January 2025, requiring organizations to conduct data protection impact assessments before deploying conversational AI tools.
Effective prevention combines technical controls with procedural safeguards across four layers:
Enterprise-grade prevention architectures incorporate Symantec DLP, Microsoft Purview Information Protection, and Nightfall AI detection engines performing real-time lexical and regex pattern matching against sensitive data taxonomies including PII, PHI, and PCI-DSS cardholder attributes. Organizations operating across Cyberjaya, Changi Business Park, and Batam Industrial Estate implement tokenization gateways through Protegrity, Voltage SecureData, or Thales CipherTrust platforms ensuring plaintext credentials never traverse egress boundaries. CASB (Cloud Access Security Broker) configurations through Netskope, Zscaler, and Palo Alto Prisma enforce inline inspection policies calibrated against MITRE ATT&CK tactics and OWASP LLM Top Ten vulnerability classifications, generating tamper-evident forensic telemetry satisfying ISO 27701 privacy information management system attestation requirements.
Yes, if employees input sensitive information into AI tools. The risks include: direct input of personal data, accumulation of confidential context across prompts, and exposure of intellectual property. Enterprise AI plans provide stronger protections, but employee training and data classification are essential safeguards.
ChatGPT Enterprise is significantly safer than consumer/free versions. Data is not used for model training, retention is configurable, admin controls are available, and SOC 2 compliance is maintained. However, even with Enterprise, employees must follow data classification guidelines — do not input restricted data (PII, credentials, source code).
Immediately stop the session, document what was shared, and report to IT Security within 1 hour. If personal data was involved, assess PDPA notification requirements. Then update safeguards to prevent recurrence — this may include additional training, technical controls, or policy updates.