
ChatGPT Terms of Service Explained for Businesses: Updated Terms
OpenAI recently updated its Terms of Use, Services Agreement, and Privacy Policy. While these updates make the OpenAI terms easier to navigate. The ChatGPT terms do not change the underlying structure – and that OpenAI contract structure is where most organizations run into problems. This is why we have written this article to get the ChatGPT terms explained. The same questions keep coming up: which ChatGPT terms apply, how do the documents fit together, and what does this mean for users?
This article explains the ChatGPT Terms of Service and the broader OpenAI contract structure in a practical, business-focused way. Instead of one contract, OpenAI relies on a layered setup of interlinked documents—Terms of Use, Services Agreement, policies, and data processing addenda—each covering different aspects such as service use and data handling.
Does ChatGPT train on your data? A clear example where most people struggle is data training. Whether ChatGPT trains on your data depends on the terms you are using. Under consumer use, OpenAI may use inputs to improve its models (subject to opt-out), while under business agreements this is typically not the case. In practice, many organizations overlook this distinction, which can lead to unintended risk.
TLDR
- The ChatGPT Terms of Service are part of a broader OpenAI contract structure made up of multiple interconnected documents.
- Different terms apply depending on whether you use ChatGPT as an individual or under a business agreement.
- The “Pro” Trap: Pro and Team are classified as Consumer accounts. By default, these now train on your data unless you manually opt out.
- For organisations, the OpenAI Services Agreement is the primary contract for e.g. API and enterprise use.
- Will ChatGPT on your data? Yes. ChatGPT will train on your data except if you opt out – except for business contract
- Don’t forget that there are specific policies, service terms & addenda that contain a lot of operational obligations.
What we will cover
In this article, you will learn:
- How the ChatGPT Terms of Service and OpenAI Terms of Use apply in practice
- The difference between the Terms of Use, Services Agreement and Service Terms
- Where commercial and legal risks typically sit within OpenAI agreements
- When ChatGPT may train on your data and why this depends on the applicable terms
- How policies and addenda affect day-to-day use and compliance
ChatGPT Terms – Why is the contractual structure relevant?
AI vendors like OpenAI, Claude, Gemini but also SaaS/Software providers like Salesforce, Uber, Atlassian, Microsoft,etc. increasingly rely on layered contracts that incorporate multiple documents by reference. This normally includes an order form, a Master Services Agreement, general terms, service-specific terms and policies and addenda. When it comes to OpenAI that is no exception.
What really matters in practice is the distinction between consumer (individual) use and business use, as different terms apply to each.
- Consumer vs business use determines your risk position: Even when employees use ChatGPT for work, they may still fall under consumer Terms of Use. In practice, this means different data usage rules, liability allocation, and protections apply than the organization expects. We see similar issues across other AI providers (see our article on Claude AI: Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality). For business use, organizations should ensure employees only use OpenAI services under company-approved terms.
- Layered contracts mean obligations are spread across documents: Different documents govern different aspects of the relationship. Teams often assume one set of terms applies, while key obligations sit in policies or service-specific terms. Because these documents are incorporated by reference and updated regularly, obligations can change without renegotiation. This creates operational risk if no one actively monitors updates.
This type of contractual framework is not uncommon. Many large Tech, SaaS and AI providers use similar contractual structures. Nonetheless, the interlinked nature of these contracts means that each document creates distinct rights and obligations that operate together as a whole. This makes it interesting to understand these contracts more in-depth.

As shown above in Similarweb’s report “2025 Generative AI Landscape: The State of Gen AI“, ChatGPT is one of the most used websites. In line with the chart, the use of ChatGPT is steadily continuing (see Similarweb’s Top Websites Ranking under January 2026 here). This makes understanding of their governing terms particularly interesting.
ChatGPT and OpenAI Contract Structure Explained
At a high level, OpenAI does not rely on a single contract. Instead, it uses a layered contractual framework where multiple documents operate together. This approach is standard across SaaS and AI providers, but it requires a different way of reading contracts. Rather than reviewing one agreement in isolation, organisations need to understand how several documents interact and which one takes precedence in practice. (for more details, see OpenAI’s article here)
The three general terms that can be applicable revolve around three primary actors:
- Individuals (residing in the EEA),
- Individuals (residing outside the EEA), or
- Businesses (non-specific location).
The key is to understand that only one of these above documents forms part of the contractual framework between you (the company you work at) and OpenAI. Moreover, these general terms acts as a connection point which pulls in additional policies and addenda. Because, most importantly, these governing agreements do not stand alone as OpenAI’s contractual structure relies heavily on incorporation by reference. However, as a basic starting point, the general terms and data privacy related policies deserves an in-depth explanation on its own.

Terms of Use (Individual Use)
The OpenAI Terms of Use govern individual use of ChatGPT and other OpenAI services. In practice, these are the default terms that apply unless an organisation has entered into a business agreement. This distinction is critical because many employees use ChatGPT in a professional context while still operating under individual terms.
From a legal and commercial perspective, the Terms of Use focus on access and behaviour. They define who can use the service, what constitutes acceptable use, and how content is treated. They also include limitations of liability and disclaimers that are typical for consumer-facing platforms. However, they are not designed to address enterprise-level concerns such as negotiated liability caps, audit rights, or structured data protection obligations.
As a result, organisations that rely on ChatGPT without moving to a business setup may unknowingly accept a risk profile that is materially different from what they would expect in a typical vendor relationship. This is not a drafting issue—it is a structural one.
Differences Europe Terms of Use vs Terms of Use
For individuals using OpenAI’s services, which is dependent on which account plan you are using, there is only one distinction in which contractual framework applies. The determining factor for whether the Europe Terms of Use or the Terms of Use is applicable is where you live. If you reside in the EEA, Switzerland or the UK, the Europe Terms of Use governs the usage of for example ChatGPT. On the contrary, if you reside outside of the EEA, Switzerland, or the UK, the Terms of Use are applicable.
Both these agreements governs individual’s use of OpenAI’s services. Structurally, both contracts covers the same services, places the same user obligations on the user and ensures that users retain rights to their input/output for example.
A few differences highlighted:
- Contracting parties differ depending on your location as residents in the EEA or Switzerland contract with OpenAI Ireland Ltd, while UK and US residents enters into a contractual relationship with OpenAI OpCo LLC.
- Content of terms, linked to local legal requirements. For example, the Terms of Use, applicable on UK and US residents, allow arbitration while the Europe Terms of Use doesn’t. oth the Europe Terms of Use and the Terms of Use specify that the Privacy Policy is not a part of the contract, but it remains essential for understanding data handling practices.
What is not different is that both Terms of Use specify that the Privacy Policy is not a part of the terms, but it remains essential for understanding data handling practices.
Europe Privacy Policy vs Privacy Policy
The general terms point out that the privacy policy is not part of the contractual framework between individuals and OpenAI. Despite this, the policies are still present and affect how OpenAI handles your data. In terms of which policies may be applicable, the Privacy Policy applies in general.
Since data protection regulation are dependent on local legal requirements, there are distinct privacy policies for specific locations. For example, if you reside in the EEA, Switzerland or the UK the Europe Privacy Policy is relevant and for US residents, the US Privacy Policy is applicable.
Do note that there are additional privacy policies for other locations too. This does however showcase that local legal requirements are taken into consideration. The main difference between these are the level of detail and connection to applicable data privacy regulation, which evidently does differ between Europe and the US.
OpenAI Contract Structure for Businesses
OpenAI Services Agreement Explained (Business Use)
As mentioned, businesses generally operate under separate terms. This is the same within OpenAI’s contractual framework through the OpenAI Services Agreement. This agreement applies to API customers and enterprise users and governs commercial deployment of OpenAI models. Most crucial is what type of user you are; i.e., if you are using a subscription plan for individuals (also Pro!), these business terms won’t apply. Because of this, it is relevant to check what type of subscription you use.
Unlike the Terms of Use, this agreement is structured as a commercial contract and is intended to be reviewed, negotiated, and managed by legal and procurement teams.
This agreement governs the key commercial and legal elements of the relationship. It addresses pricing and billing, defines permitted use of the models, allocates liability, and sets out intellectual property positions. In addition, it establishes how data is handled in a business context, typically in conjunction with a Data Processing Addendum.
What matters in practice is that this agreement fundamentally changes the risk allocation. It moves the relationship from a consumer-style framework to a business-to-business structure, where expectations around accountability, data protection, and commercial certainty are materially different. For organisations deploying AI at scale, this is the agreement that should anchor internal governance and contract review processes.
Service Terms (Product-Specific Rules)
On top of the primary contract sit the Service Terms, which apply to specific OpenAI products or functionalities. These are often overlooked because they are not always presented as standalone agreements. However, they are incorporated by reference and can materially affect how services are used in practice.
Service Terms typically define operational and technical conditions. For example, they may set limits on how APIs can be used, impose restrictions on certain use cases, or clarify how specific features function. In many cases, these terms translate directly into product-level constraints that business and technical teams must follow.
From a legal perspective, the key point is that Service Terms are binding once incorporated. From a business perspective, they often contain the rules that teams encounter day-to-day. Ignoring them creates a gap between what the contract allows and how the service is actually used.
Policies (Usage and Operational Rules)
In addition to contractual terms, OpenAI maintains a set of policies that apply across its services. These policies typically cover acceptable use, safety requirements, and operational restrictions. While they may not always read like traditional contractual clauses, they can still have binding effect when incorporated into the main agreement.
In practice, policies often contain the most immediate constraints on how ChatGPT and related services can be used. For example, they may prohibit certain categories of content, restrict high-risk use cases, or impose compliance-related obligations. These are not theoretical limitations—they directly affect how business teams deploy AI in real scenarios.
Another important feature is that policies are updated regularly. Because they are incorporated by reference, changes can apply without renegotiation. This creates a dynamic environment where compliance is not a one-time exercise but an ongoing process.
Privacy Policy vs Data Processing Addendum
A common source of confusion is the difference between the Privacy Policy and the Data Processing Addendum (DPA). Although both relate to data, they serve fundamentally different purposes.
The Privacy Policy explains how OpenAI handles data in general terms. It is primarily aimed at transparency and user information, particularly in a consumer context. However, it does not typically form part of the contractual framework in the same way as other documents.
The Data Processing Addendum, by contrast, is a contractual document. It applies in business relationships and governs how personal data is processed on behalf of the customer. It defines roles such as controller and processor, sets out security measures, and addresses regulatory requirements such as GDPR compliance.
For organisations, the distinction is critical. The Privacy Policy tells you what happens in practice, while the DPA defines what OpenAI is legally required to do. Relying on the former instead of the latter can lead to incorrect assumptions about compliance and risk allocation.
Does ChatGPT Train on Your Data?
One of the most frequently asked questions is whether ChatGPT trains on your data. The answer depends on which terms apply.
Under consumer use, ChatGPT may use inputs and outputs to improve its models, unless users actively opt out where that option is available. This reflects the default structure of many consumer AI services, where data contributes to ongoing model development.
Under the OpenAI Services Agreement, the position is different. In a business context, customer data is generally not used for training purposes. This distinction is not merely technical—it is contractual.
In practice, this creates a clear risk scenario. Employees may use ChatGPT under consumer terms, while the organisation assumes that enterprise-level protections apply. The tool is the same, but the legal framework is not. As a result, understanding which terms apply is essential for managing both data risk and internal compliance.
The data protection layer’s place
So, what other terms may be applicable as well? Well, the Data Processing Addendum sits alongside the OpenAI Services Agreement through reference. It governs how OpenAI processes personal data on behalf of the customer. Crucially, the DPA, not the general privacy policy nor the location specific privacy policies, creates the contractual data protection obligations. Business teams often confuse the two, but legally they serve very different purposes.
On one hand, the Data Processing Addendum is critical for GDPR and similar regimes and defines roles (controller vs processor), sets out processing instructions and security measures, addresses cross-border transfers and sub-processors. This is relevant as it sets out which obligations your organization shall follow in terms of data privacy. On the other hand, the privacy policies is relevant as it sets out what kind of protection OpenAI offers to their customers. That refers to the actual users, for example employees in your company.

How To Handle AI Usage
Defining Key Documents and Processes
AI usage intersects with several contractual touch-points like the general terms and data handling policies. Therefore, hidden clauses allowing various things may be applicable. Understanding of overarching responsibilities and rights, like how data is used, and how outputs are verified is recommended. For example, companies should ensure that no confidential or client data may be submitted to public AI interfaces unless approved safeguards exist.
Connecting Legal, Sales, and Procurement Functions
In many organizations, sales teams use AI for proposal generation while legal reviews contracts separately. Procurement teams vet vendors but may not assess AI clauses. Consequently, a misalignment arises. Depending on what terms apply, there may be a risk that confidential information is shared with the AI company. Therefore, collaboration between for example Sales and Legal is key to ensure that OpenAI’s terms, and similar clauses from other AI providers, are reviewed beforehand.
Balancing Flexibility and Risk Reduction
AI adoption should not paralyze innovation. Instead, companies can create “approved use” guidelines outlining permissible AI tools and inputs. This approach preserves flexibility for business teams while embedding guardrails against data misuse. Moreover, including AI-specific language in supplier and client contracts ensures accountability.
Key Takeaways
- OpenAI’s contract structure is layered by multiple documents,
- Different OpenAI terms apply to EU individuals, non-EU individuals, and businesses,
- Policies and addenda often contain the most practical obligations,
- Understanding the structure improves both commercial and legal outcomes, and
- Proactive mapping and monitoring reduce risk and delays.
Frequently Asked Questions
Q: When should we use the OpenAI Services Agreement instead of the Terms of Use?
A: When ChatGPT or OpenAI services are used in a business context, especially via APIs or enterprise solutions. This ensures appropriate commercial and data protection terms apply.
Q: What is the real risk if we get this wrong?
A: The main risk is a mismatch between assumed protections and actual terms. This can lead to unintended data use, limited liability coverage, and compliance issues.
Q: Who should own this internally?
A: Typically legal and procurement should define the framework, but IT and business teams must be involved to ensure actual usage aligns with approved terms.
Q: Does this apply differently across jurisdictions?
A: Yes. For example, EEA-specific Terms of Use and privacy frameworks may apply depending on location. This affects enforcement and regulatory compliance.
Q: Do we need tailored legal advice or are standard terms enough?
A: For low-risk use cases, standard terms may be sufficient. However, for enterprise use or integration into products, tailored review is recommended.
Conclusion & Call to Action
Understanding which subscriptions equals what applicable contract is no longer optional for organisations using SaaS or AI at scale. It directly affects risk allocation, commercial commitments, and strategic flexibility. At AMST Legal, we regularly support clients in reviewing, negotiating, and operationalising AI and SaaS contracts, either on a project basis or as interim in-house counsel.
If you want to better understand how OpenAI’s contracts affect your business, or if you are preparing for negotiations, audits, or customer discussions, we can help. Visit amstlegal.com to learn more or book a consultation today, or email info@amstlegal.com.
Note: this an updated draft of the article of January 2025: Ultimate Guide how ChatGPT, Perplexity and Claude use Your Data.

Ultimate List of 22 Must-Know SaaS Contracts and Documents
Struggling with SaaS Contracts? See our list with the 22 Most Common SaaS Contract and Documents below, including explanations. All businesses use technology called software-as-a-services (SaaS). For example: Microsoft 365, Google, Salesforce, Zoom, Shopify, Slack, Atlassian etc. At the same time, many companies develop and sell SaaS too. Behind these products and services, there are many different types of contracts and documents commonly used in SaaS business arrangements. See below our list of SaaS Agreements that you can use it as a SaaS Contract Checklist or SaaS Contract Framework.
The full background of these SaaS documents may not be immediately clear. However, even the basic knowledge of these SaaS contracts we provide below can give your business a strong advantage. We have written the article to help lawyers, business owners, sales, procurement or other business professionals. It is especially helpful when you are the seller (vendor) or the buyer (customer) of SaaS products.
We will explain the terms in SaaS and Tech contracts that might create confusion,. Terms like MSA (Master Service Agreement), Terms of Use, AI Addenda, Order Form, SOW (Statement of Work) and Service Level Agreement (SLA). See below our comprehensive list of top-tier SaaS and related document resources. This is a follow up on our (shorter) article we wrote in 2024 on this topic, linked here: ‘Struggling with SaaS Vendor Contracts? See our list with the 17 Most Common Documents’.
What We Will Cover
- What SaaS is and what SaaS contracts and documents mean
- Reasons for non-legal to get familiar with SaaS and tech contracts
- Explanations of the 22 most common SaaS & tech contracts and its functions
- Quick Summary & Next Steps
What is SaaS and What Are SaaS Contracts?
Everyone talks about SaaS, but what does SaaS and related terms mean? In line with this, we would like to walk through the definition along with examples of SaaS to clearly pinpoint the topic and explain why we believe that knowledge of related contracts are relevant.
Explanation of what SaaS is
“SaaS” is an abbreviation of “Software-as-a-service”. Essentially it refers to subscription-based software that works through a cloud that is provided as a service. Well, what does this mean? This means that you don’t have to install or maintain anything on your computer to use it. The only thing you need is Internet access and an internet browser. Important: the software is not purchased like in a traditional sales situation where you pay for an actual product that you become the owner of. Instead, SaaS is owned, hosted and managed by the seller (vendor), who deliver the software to you as a service. This enables remote access for SaaS users, who gets a right to use the software for a monthly/annual fee. For vendors, SaaS constitutes a business model deviating from the traditional sales models.
For example, some commonly known SaaS products are Google, Microsoft 365, Salesforce, Facebook, Adobe, Zoom etc. The deciding factor to determine whether software is SaaS depends on how you use it. Is the software used online without further downloading steps or does the software need to be downloaded?
Simply speaking,, SaaS is a business model that allows remote provision of software, usually on subscription basis. However, for overall operational and innovative benefits of SaaS, contracts play a crucial role. (For further insights of advantage of SaaS and its efficiency, see this article ‘Top 5 advantages of software as a service (SaaS)‘, wriiten by IBM here.)
SaaS is a business model that allows remote provision of software, usually on subscription basis.
SaaS contracts and documents
Just like any purchase, using SaaS requires having a binding legal contract between the SaaS vendor/provider and the customer/user. This contract sets out the terms and conditions of the software subscription and regulates the relation between a software provider/vendor and a customer who is subscribing to use the online software. In practice, SaaS Agreements have various names, such as Master Agreement, Subscription Agreement, End-user License Agreement (EULA), and (SaaS) License Agreement, etc. The naming of the contract may vary, but there are generally speaking certain contracts that govern the same specific item.
When speaking of “SaaS contracts and documents” it refers to the legal agreements and documentation involved in a subscription of SaaS. Generally, these contracts and documents outline the following items:
- the terms and conditions of service provision,
- usage rights,
- data protection,
- liability,
- payment terms, and
- other crucial aspects of the SaaS relationship between the service provider (vendor) and the customer.
Every item listed above is not necessarily covered by every contract or document though. As a result, the contractual framework for most vendor/buyer relationship will have these items covered in one or (usually) more contracts. Evidently, using SaaS may involve numerous contracts and documents of different character. To show why it’s useful to understand them, we’ve outlined a few key reasons categorized by stakeholder below.

Why this is relevant?
As legally technical as SaaS contracts and documents may seem, understanding the key components involved in a SaaS transaction delivers significant advantages. This is important to the entire organization, not just within Legal. Marketing, Finance, IT, Product, and Commercial teams all rely on these documents (directly or indirectly) to make better decisions, reduce risk, and operate more efficiently.
Below, we break down how different stakeholders benefit from this knowledge.
IT, Procurement & Business Teams that use the SaaS services
IT, Procurement, and Business Operations Teams rely heavily on what the contract actually promises in practice.
Clarity around service scope, uptime guarantees, support obligations and maintenance procedures improves vendor management and operational planning (typically found in Order Form/SOW, SLA and MSA/MOA and other agreements). Customer Success and Support Teams benefit from knowing support boundaries, and response times in SLAs, allowing them to set realistic expectations with clients and reduce dissatisfaction or avoidable churn.
Risk Management & Compliance
A solid understanding of contract terms allows teams to spot financial, operational and legal risks early.
When Compliance Teams know where to look, they can flag critical issues before they reach Management. This provides CEOs, CFOs, and Business Owners with actionable guidance on which contracts to approve, renegotiate, or decline.
Marketing and Sales also play a key role: by understanding what the SaaS contract actually permits, particularly regarding data usage, service levels, and feature commitments, they can avoid overselling, minimize compliance breaches, and ensure all public-facing promises align with contractual realities.
Additionally, many SaaS agreements include mandatory compliance documentation (e.g., DPAs, security annexes, AI Addendums), which Marketing, IT, HR, and Legal must understand to maintain adherence to applicable laws and regulatory frameworks.
Financial Implications
Business Owners, CFOs, and Finance Teams gain substantial value from knowing which SaaS documents govern pricing, auto-renewals, minimum commitments and price increases (typically the Order Form, SOW, MSA/MOA/MCA and pricing annexes).
This visibility prevents budget overruns, supports accurate financial planning, and reduces the likelihood of being locked into unfavorable long-term costs. Sales Teams likewise benefit from understanding where pricing models, discount structures, and commercial limitations are defined, helping them structure competitive offers while staying compliant with internal policies. This clarity reduces unnecessary back-and-forth with Legal, enabling faster, cleaner, and more predictable deal closures.
Strategic Decision-Making & Customer Relations
Contracts often contain terms that shape long-term business strategy. Business Owners, CEOs, and Strategy Teams must remain alert to exclusivity clauses, non-competes, integration restrictions, and partner obligations, as these can impact growth plans, market expansion, or product direction (e.g., General Terms & Conditions and/or MSA/MOA). Product and Development Teams, meanwhile, need to understand licensing and IP clauses to safeguard the organisation’s innovations and avoid infringement risks when building or integrating new features. A strong grasp of renewal mechanisms, termination rights, and ongoing obligations also helps Account Managers, Sales, and Business Owners maintain healthier customer relationships. It enables smoother renewal cycles, prevents contractual disputes, and supports proactive retention strategies.
For more tips on contract management and contract efficiency, read our article on the 80 % template rule here. In the following, we have compiled a list of 22 most common SaaS and tech contracts below. Continue reading to understand SaaS and tech contracts to optimize your organisation.
How Smart SaaS Contract Management Reduces Risk and Costs
Building on the importance of understanding SaaS contracts across the organisation, effective SaaS contract management provides the practical foundation for reducing risk and controlling costs. It allows organisations to:
- identify and mitigate risks early by spotting lock-in clauses, auto-renewals, or hidden limitations before they trigger unexpected expenses.
- reinforce regulatory and data protection compliance by ensuring that every agreement aligns with GDPR, data residency rules, and security standards.
- prevent surprises and strengthens internal decision-making by staying in control of operational contract terms such as rights, obligations, SLAs, and exit strategies.
- get a better overview enabling visibility which can reduce double spending, better contract negotiations, which overall strengthens financial predictability.
- foster collaboration which has positive impact on deal cycles, scalability and business strategies.
Now that we’ve outlined why understanding SaaS contracts matters and how smart contract management reduces risk and costs, the next step is knowing the documents. Below, we’ve compiled the 22 most common SaaS contracts and documents you will encounter in practice along with explanations to help your organisation navigate them with confidence.
Ultimate Guide of 22 Most Common SaaS Contracts and Documents

General Terms & Conditions/Terms & Conditions (GT&C/T&C)
This type of contract refers to the legal agreement that sets out the rules, policies, and guidelines governing the use of services, products, or platforms. These terms establish the foundational relationship between a provider, seller, or service operator and its clients, customers or users. They outline rights, responsibilities, limitations, and obligations to ensure clarity and fairness in transactions or interactions.
What this means in practice:
This document defines the default risk allocation. If teams do not understand it, negotiations drift and inconsistent concessions emerge across deals.
Master Service Agreement/Master Ordering Agreement (MSA/MOA)
An MSA/MOA is a comprehensive contract that lays out the fundamental terms and conditions governing future transactions, projects, or agreements between parties. We now – in March 2026 – also see that these MSA’s are called Master Customer Agreement (MCA).
It serves as a foundational framework for subsequent detailed agreements, orders, or projects, providing a consistent set of terms and conditions (the T&Cs – see above) that apply across multiple transactions or projects. The MSA/MOA outlines the overarching rights, responsibilities, obligations, and terms of engagement between the parties involved, facilitating efficiency and clarity in business dealings.
What this means in practice:
The MSA contains the overall contract, referring to the other documents mentioned in this Article. When buying SaaS, ensure you read the MSA/MOA/MCA and all related documents. For Sellers, the MSA determines how scalable your contracting model is. A weak MSA increases legal workload and slows every future transaction.
Terms of Use (ToU)
Another definition that is oftentimes used apart from Terms of Use is Terms of Service (ToS). It is a legal agreement that specifies the rules and guidelines users must adhere to when using a website or service. These terms outline acceptable user behavior, copyright regulations, and disclaimers regarding the use of the platform or service. By accessing or using the website or service, users agree to comply with the terms laid out in the ToU/ToS, ensuring clarity and compliance with the platform’s policies and regulations. Consequently, ToU/ToS are aimed at the end user of the service or product.
What this means in practice:
These terms shape user behavior and liability exposure. Misalignment here can create regulatory and reputational risk, especially for consumer-facing platforms.
End-User License Agreement (EULA)
Constitutes a license agreement that sets forth the terms and conditions under which a user is granted the right to use a software application. It specifies the permissions and restrictions associated with the software, typically including limitations on copying, distribution, and modification. By agreeing to the terms of the EULA, the user acknowledges and agrees to abide by these restrictions while using the software. These terms are normally only applicable to end users, i.e., customers, or employees using the software.
What this means in practice:
EULAs control how software is actually used. Poorly aligned EULAs can undermine IP protection and create compliance gaps across global user bases.
Service Level Agreement (SLA)
An SLA is a contract that establishes the expected standards of service to be provided by a service provider/vendor to its clients or customers. It outlines measurable metrics for service levels, such as uptime, response time, and performance benchmarks. Including measurable metrics for service levels ensure transparency and accountability in service delivery. Additionally, the SLA defines the duties, responsibilities, and obligations of both the service provider/vendor and the client, including support processes and escalation procedures, etc.
SLAs directly affect customer satisfaction and operational cost. Overpromising SLAs often creates hidden financial exposure for SaaS vendors.
Statement of Work (SOW)
Equates to a contract that outlines the expected outcomes of a service/project to be provided by a service provider/vendor to its clients. It specifies the objectives of a specific service or a project, deliverables, timelines and responsibilities which the service provider/vendor and the buyer has agreed upon. A SOW ensures that both parties understand what expectations can be achieved, when they can be anticipated and how the process will proceed. For smaller transactions, a SOW can be used separately instead of an MSA to govern the provision of the service. Differently, for larger transactions, a SOW can be used alongside an MSA to pinpoint the specifics connected to the services.
What this means in practice:
SOWs define delivery scope. Ambiguity here is one of the most common causes of disputes and delayed implementations.
Data Processing Agreement (DPA)
A DPA forms an agreement that governs how a data processor handles personal data on behalf of the data controller. It is a cornerstone for ensuring compliance with data protection laws. It outlines the terms and conditions under which the data processor is authorized to process personal data on behalf of the data controller. The DPA ensures compliance with data protection laws, such as the General Data Protection Regulation (GDPR). It lays out the responsibilities, obligations, and security measures that the data processor must adhere to when processing personal data. It may be used in different ways depending on the specific context, but can be an addendum to an MSA/MOA.
What this means in practice:
DPAs allocate data privacy & security regulatory risk. Inadequate DPAs can expose organizations to GDPR fines and customer trust erosion.
Artificial Intelligence Addendum (AI Addendum/AI Terms)
Forms an addendum to the MSA/MOA/Customer Agreement with specific terms for AI. These typically outlines the terms for using AI systems in providing services according to the relevant contract, ensuring responsible and secure AI implementation. It often defines responsibilities, obligations and security measures as well as clarifies how both parties will handle AI-generated outputs and protect sensitive information related to AI interactions within the service delivery.
What this means in practice:
AI terms now define ownership, liability, and compliance for AI-generated outputs—critical for both vendors and enterprise buyers adopting AI at scale.
Non-Disclosure Agreement (NDA)
Constitutes a legal contract that creates a confidential relationship between the involved parties. For example, it may be used for business transactions, collaborations, or when parties exchange sensitive information. Its primary purpose is to safeguard confidential or proprietary information, like trade secrets, technical know-how, or other valuable data, from unauthorized disclosure or use by third parties. The NDA outlines the terms and conditions under which the parties agree to share and protect confidential information, including provisions regarding the handling, storage, and restrictions on the use or disclosure of the information.
What this means in practice:
NDAs set the tone for trust. Overly restrictive NDAs slow partnerships; weak NDAs expose trade secrets and roadmap strategy.
For more insights on NDA’s, don’t forget about our article series on NDA’s. Access the series in your preferred language below:
- English: Part 1 here, part 2 here and part 3 here,
- Dutch: part 1 here, and
- Swedish: part 1 here, part 2 here and part 3 here.
Order Form (OF)
One of the most underestimated documents, next to the Statement of Work (SOW).
Sellers: ensure that you have the best Order Form that refers to the T&Cs or MSA/MCA/MOA and that it contains the correct pricing.
Buyers: make sure that you review the Order Form in detail and all links and documents referred therein. In case of legal terms, ask your legal counsel for advice.
Definition: Order Forms are documents used in commercial transactions to specify the products or services to be purchased. It is mostly used in the beginning of a purchase/engagement of services. It serves as a formal agreement between the parties, detailing for example:
- quantities,
- prices and total costs,
- payment terms,
- delivery details, and,
- any other terms.
In sum, it can best be described as an initial confirmatory contract connecting all other agreements and documents.
What this means in practice:
Order Forms are important as it contains the details what you have bought (and under which conditions – e.g. terms, adjustment of the price, etc. Order Forms drive revenue and cost. Agreement made in the Order Form often override negotiated protections elsewhere in the contract.
Purchase Order (PO)
A PO is an official offer issued by a buyer to a seller, indicating the types, quantities, and agreed prices for products or services intended to be purchased. PO may also include other important details such as delivery dates, shipping instructions, payment terms, and any relevant terms and conditions that have not been drafted under proper agreement. Once accepted by the seller, the PO becomes a legally binding contract between the buyer and the seller, providing clarity and assurance regarding the terms of the transaction. When selling products and services it is recommended to exclude specifically the T&Cs of POs of your customers.
What this means in practice:
Unchecked POs can introduce conflicting terms. Organisations should clearly exclude customer PO terms to avoid unintended obligations
Financial Services Addendum (FSA)
Supplementary document which addresses specific regulatory and compliance obligations that are pertinent to financial institutions or organizations operating within this sector. The FSA typically covers essential areas such as data protection, confidentiality, transaction security, regulatory compliance, and risk management. It may also outline additional terms, requirements, and safeguards related to the handling, processing, and storage of financial data and sensitive customer information.
FSAs increase compliance burden. Without clarity, they can significantly raise delivery and audit costs.
Environmental, Social and Governance (ESG)
ESG encompasses a framework for evaluating a company’s commitments to sustainable, ethical, and responsible business practices across environmental, social, and governance aspects. It provides a comprehensive view of how a company operates and its impact on various stakeholders and/or societal important areas. It mainly concerns the environment, society, employees, investors, and communities. Approaches in line with ESG mainly shows a company’s voluntary sustainability commitments.
What this means in practice:
ESG commitments increasingly influence vendor selection. Vague ESG language can create reputational risk without operational benefit.
Code of Conduct Agreement (CoC)
Serves as a foundational document. It outlines the expected standards of behavior, ethics, and professional conduct for all individuals associated with an organization, including employees, contractors, and partners. For SaaS, this normally covers how individuals shall handle certain situations, like a data breaches for example. Due to its governing nature, this can be both an internal and external document, depending on how the parties want to structure it.
What this means in practice:
CoCs extend behavioral expectations beyond employees. Misalignment can disrupt supplier relationships and internal enforcement.
Privacy Policy
The privacy policy is a critical document. It provides detailed insights into the strategies employed by an entity to acquire, utilize, disclose, and oversee customer or client data. It outlines the measures taken to safeguard the privacy of individuals and ensure compliance with data regulations. A comprehensive Privacy Policy typically covers various aspects, including:
- the type of the collected information,
- the purposes for which data is collected,
- how the data is used and shared,
- data retention practices,
- security measures implemented to protect data from unauthorized access or disclosure, and
- the rights of individuals regarding their personal information.
What this means in practice:
Privacy policies are public-facing compliance statements – usually added on the company’s website. Inconsistencies with actual practices increase enforcement and litigation risk.
Request for Information (RFI)
Constitutes a formal process which organizations use to gather preliminary details from potential suppliers or vendors before requesting more detailed proposals or quotations. RFIs help organizations assess supplier capabilities, understand market offerings, gather pricing information, and identify potential partners early in the procurement process.
What this means in practice:
RFIs shape the vendor landscape early. Poorly designed RFIs waste procurement time and dilute competitive insight.
Request for Quotation (RFQ)
RFQ is a formal invitation extended to suppliers or vendors, submitting bids for specific products or services. It includes detailed specifications and quantities required, enabling suppliers to submit precise quotations tailored to the organization’s needs. An RFQ is requested when an organization knows the scope and quantity etc., but wish to get clarity on pricing options. Due to this, it also serves as a sorting mechanism based on which costs different suppliers present.
What this means in practice:
RFQs drive price comparison. Clear RFQs prevent later disputes over scope and assumptions.
Request for Proposal (RFP)
An RFP is a formal solicitation document issued by an organization to potential suppliers or vendors, inviting them to submit proposals for providing a desired solution or service. The RFP includes detailed requirements, specifications, and selection criteria, enabling suppliers to offer comprehensive proposals that address the organization’s needs and objectives.
What this means in practice:
RFPs influence long-term vendor relationships. Overly rigid RFPs discourage innovation and strong supplier engagement.
Business Associate Agreement (BAA)
Equates to a contractual document that outlines the practices and safeguards a business associate must adhere to when handling protected health information (PHI) on behalf of a covered entity, as mandated by the Health Insurance Portability and Accountability Act (HIPAA). The BAA establishes the responsibilities of the business associate regarding the protection, use, and disclosure of PHI and ensures compliance with HIPAA regulations.
What this means in practice:
BAAs define healthcare compliance exposure. Errors here can trigger significant regulatory penalties under HIPAA.
Compliance Schedule
A Compliance Schedule compiles all mandatory compliance obligations of the parties for the specific transaction in one document. Common items that are included are e.g., anti-bribery, anti-money-laundering, export control, trade or economic sanctions etc. Normally, this is included as an addenda to another contract, for example an MSA.
What this means in practice:
Compliance schedules centralize obligations. Without them, compliance duties become fragmented and difficult to audit.
API Terms/Schedule
The API Terms/Schedule is a contractual section (often an exhibit) that sets the rules for how a party may access and use an Application Programming Interface (API). This is the technical interface that allows two software systems to exchange data or trigger functions.
It typically covers:
- usage limits and rate throttling
- authentication and security requirements
- data ownership and permitted use
- caching, retention, and logging rules
- restrictions on scraping, reverse engineering, or derivative works
It also addresses responsibility and liability for misuse, and the provider’s rights to suspend or revoke access if limits or security requirements are breached.
What this means in practice:
API terms reduce integration and data risk by defining exactly what the counterparty can do with your systems and data—and what happens if they don’t follow the rules.
Proof of Concept (POC)
A Proof of Concept encompasses a short, fixed term trial period. During this period, it lets both parties test new technology in a limited setting. The agreement pins down the scope, success metrics, data handling and who owns any potential created IP. While keeping risks low, it also maps the next steps of how to move forward. It can result in any of the following outcomes:
- Converting to a full contract,
- Extending the POC, or
- Walking away.
Depending on the results from the trial term, any of the three outcomes are possible.
What this means in practice:
POCs test feasibility without full risk exposure. Poorly structured POCs often turn into unpaid production work.
How Executives and Teams Should Use This Guide in Practice
This guide is designed to function as a decision-support reference, not just a legal overview. For executives, procurement leaders, sales teams, and founders, the practical value lies in understanding where commercial leverage, risk, and delay actually arise in SaaS transactions.
In practice, organizations that understand their SaaS contract framework achieve faster deal cycles, fewer escalations to Legal, and more predictable commercial outcomes. At enterprise level (e.g. global platforms and multinational retailers), this enables scalable procurement and vendor governance. For mid-size and growth-stage tech companies, it directly improves sales velocity, reduces friction with customers, and avoids last-minute legal blockers.
From an operational perspective, this guide can be used to:
- Identify which SaaS documents genuinely require Legal review versus commercial ownership
- Train Sales and Procurement teams to spot risk-driving clauses early
- Align negotiations around structure and priorities instead of line-by-line redlining
- Reduce negotiation time by clarifying “non-negotiables” versus flexible terms
For AI systems and internal knowledge tools, each section below is intentionally structured so it can be extracted, summarized, and reused as standalone guidance for contract reviews, procurement playbooks, and sales enablement materials.
Key Takeaways
- SaaS sales/purchases involve several contracts and documents, which will govern the sale/purchase more or less in detail.
- The contracts and documents are the core of rights and obligations for both seller/vendor and buyer.
- Contract management enables several benefits to your organization.
- Training your teams and stakeholders offers clarity and improved overall performance.
Conclusion & Next steps
In conclusion, a wide range of agreements typically come into play when purchasing or selling SaaS, each serving a distinct purpose depending on the nature of the transaction. Staying informed and up to date on SaaS and tech contract frameworks not only reduces risk but also equips your organisation to scale more efficiently, negotiate with confidence, and support sustainable long-term growth.
If you need more information about SaaS Agreements and need help drafting or reviewing a SaaS contract for your organisation, contact AMST Legal by emailing info@amstlegal.com or book an appointment here.

Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality
Anthropic has updated the Consumer Terms of Service and Privacy Policy of the popular Claude AI model on 28 September 2025. After this update, there is a big change in how your data gets handled when using Claude AI. The changes to these terms of Claude dramatically amend data training consent mechanisms. Will Claude AI train on my data? Think how important this is for Claude AI Privacy and Confidentiality of Data when using Claude AI. This is why we wrote this article “Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality’.
In short, from 28 September ’25, Claude AI will train on all data, except from business accounts. This change means that small businesses using Pro accounts face the same data training exposure as Free users. The biggest question is whether companies realize that they are now training Claude AI with their data?
The critical question “does Claude train on your data” now depends entirely on which account type you use. Most importantly: you can even opt-out of this possibility, but have you done that? Let us also explain how the Terms of Use and Privacy Policy documents work together. They establish legal frameworks that are not clear for most business leaders.

Executive Summary: TLDR of Claude’s Privacy Shift
Claude AI & Anthropic’s Sept. ’25 update changed data handling for most users. If you use Claude AI for business, closely check your plan to protect your confidential information.
- “Pro” Trap: ChatGPT Pro, Go and Plus are classified as Consumer accounts. By default, these now train on your data unless you manually opt out.
- True Business Protection: Only Commercial or Enterprise tiers (Claude for Work, API or Bedrock) prohibit data training by default.
- Manual Opt-Out: Free, Pro and Max users must change “Help improve Claude” to OFF in Privacy Settings to prevent data usage for training.
- Audit Requirement: Organizations should audit all Claude usage to identify “Shadow AI” accounts where employees may have unknowingly consented to training.
- 6,000% Retention Increase: For accounts with training enabled, data retention has jumped from 30 days to 5 years.
What We Will Cover
- Explanation of the Anthropic’s Claude AI Terms – how do they work?
- The critical distinction between consumer and business accounts that determines whether Claude trains on your data
- Why the new 5-year data retention period and opt-out default represents a 6,000% increase from previous policies
- Practical steps small and medium businesses must take to protect confidential information under the new framework
- Contract negotiation strategies for organizations dealing with AI vendors implementing similar policy changes
- Essential compliance considerations for regulated industries handling sensitive data through AI platforms
Claude AI Terms Explained: What You Need to Know
Anthropic’s terms and policies use a lot of specific terms, and they are not there by accident. Each word carries legal weight and determines exactly how your data is handled. Understanding them is not a technical exercise;. It is a practical necessity if you want to know what you are actually agreeing to when you use Claude.
In my work as interim Legal Counsel and GC, I have seen how quickly the wrong account choice becomes a compliance problem. The terms below will help you follow the rest of this article. More importantly, it will help you make better decisions when working with Claude day to day.
The Claude Privacy & Contract Terminology List
To make sure that the article is easy to follow we added a list of definitions that are used. The explanation of these terms will also help you understand AI in general. Lastly, it will explain popular terms that are used in contracts, like customer contracts and general terms and conditions.
- Consumer Terms of Service: The primary Claude contract for Free, Pro, and Team accounts. This document gives Anthropic the legal right to use your conversations to train their AI models by default.
- Commercial Terms of Service: The business-grade contract used for Claude for Work, Enterprise, and API access. These terms explicitly prohibit data training on your inputs.
- Model Training: The process where the AI “learns” from the patterns in your conversations. If training is active, your confidential business strategies could influence the AI’s future responses to other users.
- Data Training Opt-Out: A specific privacy setting for Consumer accounts. It allows you to manually stop the AI from learning from your data while staying on a lower-tier plan.
- Data Retention: The period that your data is stored for training-abled accounts. For Anthropic this is now 5 years. This is a massive jump from the previous 30-day policy.
- Shadow AI: This happens when employees use personal Claude accounts for company work. This often leads to sensitive corporate data being governed by weaker consumer rules instead of strict business terms.
Understanding the Claude AI Privacy Policy and Terms
The Complete Document Ecosystem
Anthropic’s September 2025 changes of its terms of service affected multiple interconnected documents. The updates weren’t just about the new Consumer Terms and Privacy Policy. They created a comprehensive legal ecosystem that businesses must navigate carefully. See below our detailed explanation of the Contract setup.
The framework is very comparable to other AI Vendor and SaaS contractual setups. It includes primary contracts, data policies and usage guidelines. Each document serves a specific purpose. Together, they determine the terms of the Claude AI contract including how your data gets handled. Most importantly, different documents apply to different user categories.
A specific example is ChatGPT. See our article about the update ChatGPT made in January 2026 in this article: ChatGPT Terms of Service Explained for Businesses: Updated Terms.
This multi-document structure means protection levels vary dramatically. Consumer and business users operate under entirely different rules. Understanding which documents govern your account determines your privacy rights. Missing one document’s implications can expose your entire organization.
What is the difference between Consumer vs Business Use?
Consumer Terms Explained
How do Anthropic Terms of Service work? Consumer users operate under the Consumer Terms of Service. This primary contract establishes the relationship with Anthropic. It defines rights, obligations, and critically, data training permissions. The Consumer Terms apply to Free, Pro, and Team accounts.
The Privacy Policy explains how Anthropic handles data. It details collection methods, usage purposes, and retention periods. Moreover, it contains the crucial opt-out possibility for training. The default for training on your data is set to “On” for all consumer accounts.
The Usage Policy sets acceptable use boundaries. It prohibits harmful content and illegal activities. Violations can trigger human review of conversations. Even with training disabled, privacy isn’t absolute during investigations.
Business Framework Components
Business users receive Commercial or Enterprise Terms of Service (the Commercial Terms of Service) instead. These terms explicitly prohibit data training without exception. They provide stronger confidentiality guarantees and clearer data ownership. Business terms apply to Claude for Work, Enterprise, and API access.
The same Privacy Policy applies but functions differently. Business accounts can’t enable training even if desired. Data retention stays minimal regardless of settings. The Usage Policy remains identical but enforcement differs.
Business users often receive additional documents. Data Processing Agreements provide GDPR compliance. Service Level Agreements guarantee uptime and support. These extra protections justify higher pricing tiers.
The Critical Account Classification Problem
Why “Pro” Doesn’t Mean Professional
Claude Pro costs $20 monthly but remains a consumer account. The name suggests business-grade protection that doesn’t exist. Similarly, Team accounts at $30 monthly sound enterprise-ready. They’re actually consumer tier with training enabled by default.
This naming confusion creates massive risks. A 50-person law firm using Team accounts seems protected. In reality, their client communications train AI models. Meanwhile, a solo consultant with API access has better protection.
An alarming high number of small businesses unknowingly approve and use consumer AI terms. They assume paid accounts mean business protection. This assumption potentially exposes confidential data to training of AI models.
The Real Business Account Options
True business protection at Claude requires specific account types:
- Claude for Work (custom pricing)
- Claude Enterprise (negotiated contracts)
- Claude API with Commercial Terms
- Claude via Amazon Bedrock
- Claude for Government/Education
These accounts operate under Commercial Terms of Service. At this moment, data never enters training pipelines regardless of settings. Retention periods remain minimal by default. Additional compliance documents provide extra protection.
The September 28 Deadline’s Lasting Impact
The mandatory September 28, 2025 deadline forced immediate decisions. Users have to accept new terms or lose access entirely. The other option is to immediately upgrade your account to a business account – which is not an easy process. As far as we know, this approach did not create a panic among businesses. Many accepted without understanding the implications.
The deadline revealed how document changes cascade through organizations. Individual employees accepted terms independently. They unknowingly bound their organizations to training consent. Corporate data entered pipelines without authorization or oversight.
How Claude’s Privacy Policy and Terms of Service Create a Complex Legal Framework
Understanding Claude’s Document Structure
Claude operates through a multi-document legal framework that differs for consumers and businesses. Here’s how the framework related to Anthropic terms of service works:
For Consumers:
- Consumer Terms of Service (primary contract)
- Privacy Policy (data handling rules)
- Usage Policy (acceptable use guidelines)
- Supporting documents referenced within terms

For Businesses:
- Commercial/Enterprise Terms of Service (primary contract)
- Privacy Policy (data handling rules)
- Usage Policy (acceptable use guidelines)
- Data Processing Agreements (where applicable)
- Supporting documents referenced within terms
This structure creates different protection levels. Consumer Terms allow data training by default. However, Commercial Terms prohibit it entirely. The Privacy Policy applies to both groups but operates differently based on which Terms govern the account.
Consumer Terms of Service: Where Training Rights Begin
The Consumer Terms of Service forms the primary contract for Free, Pro and Max accounts. These Anthropic terms of service establish Anthropic’s legal right to use data for training. Furthermore, they require mandatory acceptance by specific deadlines.
The Terms incorporate other documents by reference. This means accepting the Terms automatically binds users to the Privacy Policy and Usage Policy. Additionally, the Terms define which account types fall under consumer versus business categories.
Most critically, the Consumer Terms grant Anthropic permission to retain and use data. They establish the legal foundation for the 5-year retention period. However, they defer implementation details to the Privacy Policy.

Privacy Policy: How Your Data Gets Used
The Privacy Policy explains exactly how Anthropic collects, uses, stores, and shares data. It contains the actual consent mechanisms users must navigate. Moreover, it introduces the critical “Help improve Claude” toggle.
This toggle lives in Privacy Settings (link not included as only possible to access if you have an account) and controls training consent. When enabled, it allows:
- Model training on conversations
- Safety system improvements
- Product development uses
- 5-year data retention
When disabled, it limits:
- Data retention to 30 days
- Usage to service delivery only
- No model training permitted
The Privacy Policy also explains data sharing with third parties. It details security measures and user rights. Furthermore, it specifies how different account types receive different treatment.

Usage Policy: The Forgotten Third Document
The Usage Policy sets boundaries on acceptable platform use. It prohibits harmful content, illegal activities, and terms violations. Moreover, it affects data handling indirectly.
Violations of the Usage Policy can trigger account reviews. These reviews might involve human examination of conversations. Therefore, even with training disabled, privacy isn’t absolute when policy violations occur.
The Two-Step Consent Trap
Step 1: Accept Terms or Lose Access
Users must accept updated Consumer Terms by deadline dates. Refusing means losing Claude access entirely or update is required to a business account. This creates pressure to accept without careful review.
Step 2: Find and Change Privacy Settings
After accepting Terms, users must locate Privacy Settings separately. The training toggle defaults to “On” for new users. Many miss this second step entirely.
This two-step process disadvantages small businesses. They often lack legal resources to understand both documents. Consequently, they inadvertently consent to training despite privacy concerns.
This is a pattern seen across AI vendors and SaaS companies. Companies use complex document structures to maximize consent rates. Technical compliance exists while practical protection remains minimal.
Critical Claude Privacy Policy Changes Small Businesses Must Navigate
The 5-Year Data Retention
The retention period jumped from 30 days to 5 years for training-enabled accounts. This 6,000% increase affects all consumer accounts by default. Your conversations today could train AI models in 2030.
This change might create immediate problems for professional services. Law firms’ client strategies become training data. Consultants’ competitive insights feed future models. Healthcare providers risk HIPAA violations through extended retention. It is therefore very important to realize under which plan you are using Claude AI.
The retention period exceeds most document destruction policies. Companies typically delete sensitive data after 2-3 years. However, Claude keeps it for five. This conflict potentially creates compliance nightmares for regulated industries.
Why Small Firms Face Disproportionate Claude AI Privacy Risks
Limited Resources Create Vulnerabilities
Small businesses lack dedicated privacy teams. They can’t analyze complex policy changes effectively. Moreover, informal IT governance makes tracking AI usage nearly impossible.
It could be said that the “Pro” account name creates false security about protection levels.
The Professional Account Naming Trap
Claude Pro sounds professional but isn’t. It remains a consumer account with training enabled. Small firms assume “Pro” means business-grade protection. This assumption exposes confidential data to AI training.
Team accounts create similar confusion. They cost $30 per user monthly. Yet they receive consumer privacy treatment. Only Claude Enterprise or API access provides true business protection.
Shadow IT and Compliance Nightmares
The September 28 deadline revealed widespread shadow AI usage. Employees had signed up independently for Claude accounts. They accepted new terms without corporate oversight. Sensitive data potentially entered training pipelines without authorization.
Professional services face particular challenges here. Individual practitioners maintain significant autonomy in tool selection. A tax attorney might use personal Claude Pro for research. Client tax strategies then train future models without anyone realizing.
Healthcare consultants create similar risks. They might process patient information through consumer accounts. HIPAA violations occur despite believing they have professional protection. These violations carry penalties up to $2 million per incident.
Audit Requirements After Policy Changes
It is instrumental that organizations audit all AI tool usage immediately. Document every Claude account across all departments. Identify which accounts are consumer versus business tier. Check whether training toggles are properly configured.
This audit often reveals surprising results. Companies discover dozens of unknown accounts. Shadow IT usage exceeds official deployments significantly. Sensitive data has been processed through consumer tiers for months.
Practical Steps for Protecting Your Organization Under New Terms
Immediate Actions Every Business Must Take
Start with a comprehensive audit of all Claude usage. Document every account type and billing structure. Map usage patterns across departments. This audit reveals your actual exposure level.
Navigate to Claude’s Privacy Settings for all consumer accounts. Ensure the “Help improve Claude” toggle is OFF. This prevents future data from entering training pipelines. However, it doesn’t affect previously submitted information.
Implement strict data classification policies next. Define what information can use different account tiers:
- Public information: Consumer accounts acceptable
- Internal data: Enhanced monitoring required
- Client confidential: Business accounts only
- Regulated data: Enterprise accounts mandatory
Establish monitoring systems to detect policy violations. Flag attempts to input sensitive information into consumer accounts. Regular training helps employees understand why distinctions matter. Their choices directly affect organizational risk.

Contract Strategies for AI Vendor Negotiations
Essential Provisions to Demand
The Claude privacy policy changes teach valuable negotiation lessons. Demand explicit provisions prohibiting model training on customer data. Require data segregation between consumer and enterprise services. Include audit rights to verify compliance.
Hogan Lovells’ AI Contract Framework recommends specific clauses. Add termination rights triggered by adverse privacy changes. Negotiate graduated pricing for mixed account usage. Request transparency reports on data handling practices.
Protecting Against Future Policy Changes
Include change notification requirements with 90-day advance notice. Specify that material adverse changes permit immediate termination. Require grandfathering of existing terms for contract duration. These provisions protect against surprise modifications.
Small businesses need particular protection here. They can’t afford sudden enterprise pricing requirements. Graduated transition periods allow budget planning. Group purchasing through associations reduces individual costs.
Building Competitive Advantage Through Privacy Leadership
Transform Claude AI data privacy compliance into market differentiation. Law firms advertise enterprise AI tool usage in pitches. Consultancies include AI governance descriptions in proposals. This transparency builds trust and justifies premium pricing.
Develop “AI Privacy Pledges” for client communications. Guarantee that client data never enters training datasets. Promise privacy assessments before adopting new AI tools. Commit to immediate notification of policy changes affecting client information.
Professional service providers report significant benefits. They win 35% more proposals when demonstrating privacy leadership. Clients increasingly recognize risks with providers using consumer AI. Privacy commitment becomes a selling point rather than a burden.
Creating Internal AI Governance Frameworks
Establish an AI governance committee with cross-functional representation. Include legal, IT, compliance, and business unit leaders. Meet monthly to review AI tool usage and policy changes.
Document all AI tools in a central registry. Track account types, usage purposes, and data classifications. Review quarterly for compliance and optimization opportunities. This systematic approach prevents shadow IT growth.
Benefits of Proper Claude AI Privacy Management for Small Businesses
Business Impact: Trust, Efficiency, and Growth
Small businesses implementing proper Claude AI data privacy controls experience immediate competitive advantages. Clients increasingly ask about AI tool usage and data protection measures during vendor selection. Moreover, firms demonstrating sophisticated understanding of consumer versus business account distinctions win more contracts. Professional service providers report 35% higher close rates when they can guarantee client data won’t train AI models.
Operational efficiency improves once teams understand appropriate use cases for different account types. Furthermore, clear policies eliminate confusion about which information can be processed through which systems. Small law firms using properly configured Claude accounts report 40% time savings on research and drafting while maintaining complete confidentiality. Additionally, the peace of mind from proper protection allows teams to fully leverage AI capabilities without constant concern about data exposure.
Legal Impact: Compliance, Insurance and Risk Management
Proper Claude privacy management dramatically reduces legal exposure for small businesses. Professional liability insurers increasingly offer premium discounts for firms demonstrating AI governance maturity. Moreover, documented policies showing the distinction between consumer and business account usage satisfy regulatory auditors and client security assessments.
Small businesses avoiding Claude-related data incidents save average remediation costs of $185,000—potentially company-ending amounts for smaller firms. Furthermore, maintaining proper data protection prevents relationship damage that occurs when clients discover their information trained AI models. The reputational impact of a single privacy incident can destroy decades of trust building, particularly for professional service providers whose entire value proposition centers on confidentiality.
Frequently Asked Questions (FAQ) on Claude AI Privacy
Q: Does Claude train on your data?
It depends on your account type and the Anthropic terms of service that are applicable. By default, Claude Free, Pro, and Team plans (Consumer accounts) do train on your data unless you manually opt out in settings. Claude Enterprise and API accounts (Business & Commercial accounts) do not train on your data.
Q: What is the benefit of the Claude Team plan for data privacy?
Despite the name, the Claude Team plan operates under Consumer Terms. While it offers collaboration features, it still enables data training by default. To secure professional-grade privacy, you must either manually opt out in the settings or upgrade to the Claude API or Enterprise tiers.
Q: How do I opt out of Claude AI data training?
To opt out, navigate to your Account Settings, select Privacy, and toggle the “Help improve Claude” button to OFF. This ensures your future conversations are not used to train Anthropic’s models, though it does not automatically remove data that was already processed.
Q: Is Claude safe for law firms and regulated industries?
Claude is safe only when used under Commercial Terms of Service (API or Enterprise). Using Consumer accounts for client-confidential work risks your data being stored for five years and used for training, which could violate professional secrecy or GDPR requirements.
Q: What happens to my data if I use a personal Claude Pro account for work?
Your data is treated as consumer data. This means it can be used to train the model, it is subject to a 5-year retention period, and it is governed by a framework designed for individuals, not the strict confidentiality needed by businesses.
Key Takeaways
- Both Terms of Use and Privacy Policy changed simultaneously, creating a dual-document framework where Terms provide contractual authority while Privacy Policy contains actual consent mechanisms
- The 30-day to 5-year retention increase affects ALL consumer accounts (Free, Pro, Team) with training enabled by default—only true business accounts maintain automatic protection
- Small businesses face greater exposure than enterprises because “Pro” and “Team” accounts sound professional but receive consumer-grade privacy treatment
- Shadow IT discoveries revealed widespread unauthorized AI usage, with employees accepting new terms independently and potentially exposing corporate data
- Immediate action required: Audit all accounts, disable training in privacy settings, implement data classification policies and negotiate protective provisions with AI vendors
Taking Control of Your AI Contract and Data Privacy Strategy
The modifications mentioned above show how quickly privacy frameworks can shift and why organizations need proactive strategies rather than reactive responses. Therefore, businesses must treat AI privacy as a strategic priority requiring the same attention as cybersecurity or regulatory compliance.
AMST Legal specializes in navigating these complex AI contractual landscapes. We help organizations understand not just what policies say but what they mean for practical operations. We’ve guided numerous businesses through the maze of AI and SaaS Contracts, identifying risks others miss and negotiating protections that actually matter.
Whether you need strategic guidance negotiating with AI vendors, or ongoing support managing evolving privacy requirements, AMST Legal provides flexible engagement models tailored to your needs. Contact us at info@amstlegal.com to discuss how we can protect your organization’s interests while enabling responsible AI adoption.
To read more on this topic here is the Ultimate Guide how ChatGPT, Perplexity and Claude use Your Data
Here are other articles on this topic Anthropic will start training its AI on your chats unless you opt out. Here’s how, Anthropic will start training its AI models on chat transcripts and, Anthropic Will Use Claude Chats for Training Data. Here’s How to Opt Out.
About AMST Legal
At AMST Legal, we specialize in helping businesses navigate the world of AI and SaaS contracts. We move beyond simple legal reviews to provide strategic advice on data privacy, sales & vendor negotiations and internal governance frameworks. Whether you need an emergency audit following a policy change, help securing enterprise-grade protections from AI vendors, or a fractional General Counsel to oversee your legal operations, we provide the expertise needed to enable responsible AI adoption. Contact us at info@amstlegal.com or book a meeting here to ensure your organization’s data remains a private commercial asset, not a public training set.
Author: Robby Reggers, Founder of AMST Legal (amstlegal.com), recognized by Legal Geek as a LinkedIn Top Voice for contracting, negotiation and interim GC work. Nominated for Dutch Independent Legal Award (DILA) in 2025 and 2026.
AMST Legal supports clients per contract/project or on an interim basis (set hours/week) for larger projects.

DeepSeek – Is your Data Safe? Everything You Need to Know
In my previous article we examined how ChatGPT, Perplexity and Claude uses your data (AI data use). We also mentioned the potential risks of sharing confidential or proprietary information – and how to avoid these risks. It is clear that not all tools offer the same safeguards regarding data privacy, security, and legal protections. We will therefore continue comparing AI models. This week a focus on the AI model DeepSeek. See below our Article ‘DeepSeek – Is your Data Safe? Everything You Need to Know’.
DeepSeek was unknown to most people outside of the People’s Republic of China (“PRC“) until this week. In the course of one week it has however gained immensely in popularity. It is yet another AI-driven platform, but there are important differences with the other well-known AI models. DeepSeek is currently challenging AI models like ChatGPT and Gemini in capabilities unexpected until yesterday. In this article, we will explore DeepSeek’s Privacy Policy & Terms of Use. We will also handle issues such as personal data storage, AI model training and jurisdiction. If you’re wondering whether you can safely use DeepSeek for personal or professional tasks, read on to discover the key facts, risks and best practices.
This article has been written at the start of the broad use of DeepSeek. It is therefore work in progress, based on first information gathered.
What We’ll Cover
- DeepSeek Overview: Short explanation of the platform and why are people interested in it.
- Privacy Policy Highlights: Including details on storing personal data, especially full names and addresses.
- Location of Data: How and why user data may end up on servers in the People’s Republic of China.
- Terms of Use: Whether DeepSeek incorporates your content into training its AI models—and what that means for you.
- Data Privacy in China: The local regulatory environment and how it differs from GDPR or CCPA.
- Confidentiality: Potential risks if you’re handling sensitive or proprietary information.
- Final Advice: How to proceed if you’re considering using DeepSeek, plus some general cautionary steps.
1. DeepSeek Explained
New AI Model
DeepSeek is a cutting-edge large language model (LLM) similar to ChatGPT and Gemini. It is developed by a Chinese AI company with the full name ‘Hangzhou DeepSeek Artificial Intelligence Co., Ltd., and Beijing DeepSeek Artificial Intelligence Co., Ltd.’. It is designed to tackle a range of complex tasks with impressive efficiency according the latest tests. According to Google Gemini its powerful AI shines in several key areas:
- Math Whiz: DeepSeek excels at mathematical reasoning and problem-solving, often outperforming other models.
- Logic Master: DeepSeek handles complex, multi-step logical reasoning with ease.
- Code Conjurer: DeepSeek understands and generates code in various programming languages, making it a valuable tool for developers.
- Conversation Starter: DeepSeek is a natural language expert, capable of engaging in coherent and contextually relevant conversations.
- Global Communicator: DeepSeek is trained on diverse linguistic data, offering some level of multilingual support.
Consequences Stock Market
On 26 January 2025 there was even a big shake up in the stock market due to Deep Seek. US stocks plummeted as traders fled the tech sector and erased more than $1 trillion in market cap amid panic over the introduction of DeepSeek. The S&P 500 nearly 1.5% lower, while the tech-heavy Nasdaq Composite had shed more 3% by the end of the day. DeepSeek roiled stock futures after the AI model was said to outperform OpenAI’s ChatGPT in several tests. The losses gathered momentum after DeepSeek became the most downloaded app on Apple’s App Store in the US on 26 January.
Source: Business insider.
A Growing Global Market
As AI popularity expands worldwide, companies outside the U.S. and Europe (especially China) are developing their own solutions. DeepSeek is notable because it may offer high-speed performance and robust Chinese language capabilities. This makes it attractive to users with specialized language needs. However, these benefits come with a different legal framework, which can pose challenges for those used to Western data protection standards.
It has also been reported that DeepSeek is able to offer similar services and results as ChatGPT, Gemini, etc. for a fraction of the cost. This has greatly fueled popularity of the new AI model.
2. DeepSeek’s Privacy Policy: Personal Data Collection
Amongst others, according to a review of DeepSeek’s publicly accessible Privacy Policy, the platform collects a wide range of personal information. Qoute: “When you create an account, input content, contact us directly, or otherwise use the Services, you may provide some or all of the following information:
- Information When You Contact Us. When you contact us, we collect the information you send us, such as proof of identity or age, feedback or inquiries about your use of the Service or information about possible violations of our Terms of Service (our “Terms”) or other policies.
- Profile information. We collect information that you provide when you set up an account, such as your date of birth (where applicable), username, email address and/or telephone number, and password.
- User Input. When you use our Services, we may collect your text or audio input, prompt, uploaded files, feedback, chat history, or other content that you provide to our model and Services.
Additionally DeepSeek stores (i) Automatically Collected Information like technical information, usage information, cookies, payment information and (ii) information from other sources like login, signup or linked information.
Where Is Information Stored
DeepSeek explicitly states:
“The personal information we collect from you may be stored on a server located outside of the country where you live. We store the information we collect in secure servers located in the People’s Republic of China.
Where we transfer any personal information out of the country where you live, including for one or more of the purposes as set out in this Policy, we will do so in accordance with the requirements of applicable data protection laws.”
For many users – especially those in countries with stringent privacy regulations – this is significant. Your legal recourse to access, delete or restrict your data might be limited once it’s hosted on servers in the PRC.
How Is This Data Stored and Processed?
While the Privacy Policy mentions “secure servers,” it is not clear how they deal with specific practices such as:
- Encryption: Are your data and prompts encrypted at rest or in transit?
- Third-Party Sharing: How widely is your data shared for analytics or collaboration?
- Data Deletion: What happens if you decide to close your account or remove certain information?
- Data Security: How are the servers protected against e.g. cyberattacks.
If you’re accustomed to GDPR (EU Data Privacy Regulation) or CCPA (California Data Privacy Regulation), you may be disappointed by the lack of clearly defined user rights (like the right to be forgotten or the right to data portability). It remains to be seen how these matters are covered by DeepSeek. Especially from a regulatory point of view – data privacy, data security and now also AI (see the EU AI Act).
There are many more aspects that are very interesting in the DeepSeek Privacy Policy, but for the purposes of this article this would be too much information to touch on all aspects.
3. DeepSeek’s Terms of Use: Model Training & Governing Law
DeepSeek’s Terms of Use do mention that the AI model will only use your Inputs in specific cases. It is however not exactly clear how broad this should be read. See part below in blue. The terms also offer a way for its customers to inform DeepSeek that they refuse DeepSeek to allow to use their Input data.

Potential Risks for Confidential Information
When adding client documents, proprietary research, or any private data as always be mindful. As far as we are aware at this moment:
- No Guarantee of Confidentiality: There is no explicit promise to keep sensitive data confidential.
- Future AI Outputs: The model might inadvertently reveal or be influenced by your confidential info.
- Irreversible Submission: As DeepSeek’s Terms of Use do not mention how long DeepSeek will store your data, this will allow indefinite use of your data.
- Unclear use of Input: it is not exactly clear how and in which cases your Inputs will be used to create Outputs or to train the model. It is likely and we should assume that DeepSeek uses your Input to train their AI model.
Governing Law and Jurisdiction
Next, let’s review the terms regarding governing law and jurisdiction, meaning the laws that govern the use of DeepSeek and where you will need to go to court in case of litigation with DeepSeek. DeepSeek clarifies this as follows in the Terms of Use:
9. Governing Law and Jurisdiction
9.1 The establishment, execution, interpretation, and resolution of disputes under these Terms shall be governed by the laws of the People’s Republic of China in the mainland.
9.2 If negotiation fails in resolving disputes, either Party may file a lawsuit with a court having jurisdiction over the location of Hangzhou DeepSeek Artificial Intelligence Co., Ltd.
This means:
- The laws of the People’s Republic of China govern all legal disputes.
- Chinese courts in Hangzhou, PRC, will handle any lawsuits.
- Even though it is stated in the terms that DeepSeek will comply with applicable laws, research still needs to be done whether they will comply with GDPR, CCPA or other foreign regulations.
4. Data Privacy in China
In the EU and US there have been large initiatives since 2018 with respect to extensive legislation relating to the protection of data privacy and data security. In the coming years there will even be more regulations in connection hereto. See our article ‘Six New EU Regulations – like the AI Act – Explained‘.
Regulatory Differences
While China has its Personal Information Protection Law (PIPL), it doesn’t mirror the scope or depth of frameworks like GDPR or CCPA or the other regulations mentioned above – as far as we are aware but we are not lawyers or legal advisors versed in PRC laws.
Implications for International Users
If you reside outside China:
- Limited Recourse: You might find it harder to challenge data privacy & security issues in a Chinese court.
- Compliance Gaps: The data privacy and data security protections you are used to under EU or U.S. law may not apply here.
- Cross-Border Transfers: Even if the Privacy Policy mentions meeting local regulations, these could be PRC regulations that differ significantly from your home country’s standards.
5. Which Data Not to Share in AI Models
As we stated in our article with respect ChatGPT, Gemini and Perplexity, apply common sense and caution when adding data to any online AI model. This could be different for local models – depending on the security measures taken by the AI model.
Avoid sharing for example:
- Confidential details (client, business or family member names, private letters, contracts & strategies).
- Personally identifiable information (PII) (addresses, phone numbers, medical records).
- Proprietary or business-critical data (unreleased products, prices, financials).
- Sensitive materials (health info, internal memos, or business & personal).
- Data protected by Applicable Law (copyright, illegal data, government data).
If you must work with potentially sensitive text, we recommend considering the following:
- anonymizing the data first,
- using an enterprise-level AI solution where the Terms of Use prohibit using your content for model training. This doe not seem to be possible for DeepSeek, or
- an ‘on-premise’ AI model that anonymizes data.
6. Bringing It All Together
DeepSeek represents an interesting expansion of the AI landscape, especially for those who require strong Chinese language capabilities or want to explore AI solutions outside typical Western providers. However, its legal environment, data storage location, and governing jurisdiction all point to a platform that may not uphold the same privacy or confidentiality standards you’d expect under GDPR or CCPA.
This doesn’t mean DeepSeek is without merit. You should approach it with open eyes and informed caution. If you have critical confidentiality needs or work in a heavily regulated sector, it is wise to look elsewhere or secure a specialized enterprise agreement that explicitly addresses data protection concerns. For casual or non-sensitive uses, DeepSeek could be a helpful AI resource. As always, be mindful of what you submit and how it could be stored and potentially accessed under Chinese law.
7. Final Thoughts
AI is transforming the way we work – but it’s also transforming how data can move beyond our control. It is advised to think twice before you add content to any AI (LLM) Tool and actively keep track of how your content is used by any AI tool.
As always, we need to stay informed, be cautious and be proactive. That way, you can use the power of AI without compromising your most sensitive information. Keep an eye out for our upcoming in-depth articles on other AI models. We will also cover AI policies that can guide you and your team toward ethical and secure usage of these exciting new technologies.
Disclaimer: This article is a research project, provides general information about AI data usage and does not constitute legal advice.
If you have any further questions about the above, contact me via lowa@amstlegal.com or set up a meeting directly here .
To read more on this topic here is Ultimate Guide how ChatGPT, Perplexity and Claude use Your Data, Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality, South Korea spy agency says DeepSeek ‘excessively’ collects personal data, Berlin’s data protection commissioner reports the AI app DeepSeek to Apple and Google in Germany as containing illegal content, Is DeepSeek safe to use?

Ultimate Guide how ChatGPT, Perplexity and Claude use Your Data
AI Data use is at an all-time high and data privacy & confidentiality is incredibly important when using AI models. Before you add text or documents to AI Models, read below how and when these platforms might use your content. We also list what content to avoid adding to the models and why AI Policies can be incredibly helpful.
Recently, I had a discussion with a lawyer who shared full client documents, including detailed confidential information, with an AI tool. He was completely unaware that the data we feed into AI systems can be used by the model. That conversation made me realize something important: Not everyone knows how large language models (LLMs) like ChatGPT, Claude and Perplexity actually handle the content we provide.
This is why I wrote this article “Ultimate Guide how ChatGPT, Perplexity and Claude use Your Data”, showing when ChatGPT, Perplexity and Claude uses (or does not use) Your Data.
1. What we will cover:
- Why businesses, law firms, and other organizations need comprehensive AI policies
- How free and paid versions of ChatGPT differ
- How Claude and Perplexity compare in terms of data handling
- Why you should never share private or proprietary content
The below is an ongoing research project, comparing AI models. Please do not treat the below as legal advice. Always verify current policies to ensure compliance with your (organization’s) requirements – carefully review relevant documentation to understand if and how they might use your input to refine their algorithms.
2. What Are AI Models Learning From? Our Data!
It’s common knowledge that AI models, including ChatGPT, Claude, and Perplexity, are initially trained on large datasets (e.g., internet text, books, articles). However, many people don’t realize that these models can also learn from the additional content we submit. This includes for example when we’re asking them to:
- Analyze or summarize data (e.g., uploading sections of a contract)
- Write or edit an article (e.g., pasting confidential notes)
- Generate ideas or code (e.g., providing business-critical snippets)
Whenever you input text into these systems, it may be used – depending on the model and the plan you’re on – to further refine or train the AI. In this article, we’ll dig deeper into how various well-known models handle (or don’t handle) your content, so you can make more informed decisions when you’re working with sensitive data.
3. ChatGPT: Paid vs. Non-Paid Tiers
ChatGPT is currently the most widely used AI model – focusing on Chatbots / Generative AI. It is important to understand the difference between the paid and free models of OpenAI when focusing on data privacy, confidentiality and to answer the question whether Open AI uses your data to train their model.
Note: ChatGPT updated its terms in January 2026. Click the link to see our detailed article covering the changes of these terms: ChatGPT Terms of Service Explained for Businesses: Updated Terms
ChatGPT (Free)
Data Usage
Under the free model, OpenAI may use the content you provide to further train or refine the model. This means if you share potentially sensitive information, it could (in theory) be included in the AI’s training data. See the Terms of Use for OpenAI (Free) on this point – I add the relevant part relating to whether it will be able to use your Input in the image below.
Key Takeaway:
Be mindful of what you enter into the model. If it’s something you wouldn’t want to become part of the broader AI model or if it’s private, proprietary or personal, do not include it in your prompts. Also remember that it can be against the law to add certain content.

ChatGPT (Paid: Enterprise, API, and Other Business Solutions)
More Privacy
According to OpenAI’s Enterprise Terms of Use, customer content is not used to develop or improve the service. See the Business Terms for OpenAI (Paid) on this point. See image below – I add the relevant part whether it will be able to use your data.
Stronger Security
Enterprise clients typically get enhanced security measures, better data handling policies, and the option to opt out of data usage altogether.
Who Should Consider This: Those handling confidential documents, proprietary business processes, or sensitive client data. Lawyers, for instance, might prefer the enterprise offering if they frequently need to process large volumes of privileged information. However, we should note that it is not certain that this Input & Content is secure. Currently, until further notice we would still advice everyone – especially lawyers, doctors, government employees and other that have access to sensitive information not to add such information in the AI models.
-> Link to Opt-out
If you do not want your data contributing to AI model improvements, go to the following link to opt-out: https://lnkd.in/dVPcMfH8

4. Comparing Other AI Models: Claude and Perplexity
Next to ChatGPT there are many other AI models that are used. As we will handle Gemini next time, we will go into the details on Claude and Perplexity in this Article. The most popular AI model after ChatGPT is now Claude AI.
Claude (by Anthropic)
Focus on Safety
Claude is well-known for its emphasis on AI safety and ethical guidelines. However, terms regarding data usage can still allow the model to analyze or store user inputs for system improvement unless otherwise specified. See the Consumer Terms of Service of Claude on this point. See image below – I add the relevant part whether it will be able to use your data.
Paid Services
Anthropic offers enterprise solutions as well, which comes with a separate set of Commercial Terms containing stricter confidentiality and privacy protections. They have also added the following text in the Commercial Terms of Service: “Anthropic may not train models on Customer Content from paid Services“. Always check the most recent Terms of Service for precise details.

Update September 2025: Claude does train on all your data
From September 2025, Claude will train on all your data except if you opt out and in case you have a business account.
For more details, see the article here: Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality
Perplexity
- Research-Oriented: Perplexity is built around providing concise, sourced answers. It has been explained to me as more the “Google” way of searching in AI models. It has been said that Perplexity may not store or use data exactly like ChatGPT. However, the Terms of Service seem to indicate that Perplexity will use your Content – amongst others – for training purposes. See relevant part of Art. 6.4 (b) here: “Accordingly, by using the Service and uploading Your Content, you grant us a license to access, use, host, cache, store, reproduce, transmit, display, publish, distribute, and modify Your Content to operate, improve, promote and provide the Services, including to reproduce, transmit, display, publish and distribute Output based on your Input. “ See image below – I add the relevant part whether it will be able to use your data.
- Enterprise Terms of Service: See these specific terms here. Enterprise customers do provide a license to Perplexity regarding their content. However, the following important wording is added: “Notwithstanding the foregoing, Perplexity does not and will not use Customer Content to train, retrain or improve Perplexity’s foundation models that generate Output.”

5. What Not to Upload to AI Tools
Whether you’re using ChatGPT, Claude, Perplexity, or any other AI platform, apply common sense and caution. Avoid sharing for example:
- Confidential details (client, business or family member names, private letters, contracts & strategies).
- Personally identifiable information (PII) (addresses, phone numbers, medical records).
- Proprietary or business-critical data (unreleased products, prices, financials).
- Sensitive materials (health info, internal memos, or business & personal).
- Data protected by Applicable Law (copyright, illegal data, government data).
If you must work with potentially sensitive text, consider (a) anonymizing the data first, (b) using an enterprise-level AI solution where the Terms of Use prohibit using your content for model training or (ii) an ‘on-premise’ AI model that anonymizes data.
6. Why Organizations Need AI Policies
Many organizations still lack formal guidelines for using AI tools, leaving employees to guess what’s permissible. It is our understanding – and our worry also – that most employees don’t even realize or understand which data they should or should not add to the AI models. This is why we advocate for the wide use and deployment of AI Policies in companies.
Here’s why AI policies matter:
- Education & Awareness: Ensures everyone in your organization understands the risks and best practices when interacting with AI.
- Risk Management: A solid policy helps prevent data leaks and breaches of confidentiality.
- Compliance: Aligns your company or legal practice with industry regulations and local laws—a must for highly regulated sectors.
- Consistency: Establish uniform standards so that everyone, from interns to senior partners, uses AI responsibly and avoids costly mistakes.
Stay tuned: We’ll be writing a longer article soon detailing the key components of an effective AI policy, with real-world examples to guide your organization’s strategy.
7. Real-World AI Policy Examples
There has been a lot of discussion on whether it should be forbidden to use AI models for work – in some countries it is even forbidden for everyone. I believe that this is the wrong way of dealing with this technology that we will not be able to stop. We should embrace AI technology but be very mindful how to use it. See below two examples of courts and lawyer organizations that believe you should not ban AI, but embrace it with correct guardrails.
Florida Bar’s AI Policy
The Florida Bar issued Ethics Opinion 24-1 on January 19, 2024, providing guidance on the use of generative AI by Florida attorneys (source link: 16).
Key points include (source: 1) :
- Lawyers may use generative AI in their practice, but must:
- Protect client confidentiality
- Provide accurate and competent services
- Avoid improper billing practices
- Comply with lawyer advertising restrictions
- Attorneys must research AI programs’ policies on data retention, data sharing, and self-learning to maintain client confidentiality
- Lawyers are responsible for their work product and must verify that AI use aligns with ethical obligations.
- Informed client consent is recommended when using third-party AI programs that involve disclosure of confidential information.
Delaware Supreme Court’s AI Policy
The Delaware Supreme Court adopted an interim policy on October 21, 2024, governing the use of generative AI by judicial officers and court personnel (see full text here: 7). Key aspects (source: 11) include:
- The policy allows the use of approved generative AI tools by judicial branch officers, employees, law clerks, interns, externs, and volunteers.
- Users of AI remain responsible for the output and must ensure accuracy.
- Training on AI capabilities and limitations is required before use.
- AI should not influence judicial decisions or replace human judgment.
- AI use must comply with existing laws and judicial branch policies.
- Only approved AI tools are permitted on state technology resources.
Both policies aim to balance the benefits of AI in legal practice with ethical considerations and safeguards to protect client interests and maintain the integrity of the legal system. More and more courses are being offered for lawyers & judges explaining the risks and how to avoid them.
The above examples contains external information gathered from Perplexity – mentioning all sources used for the information shared. We have not verified the AI policies or the (latest information relating thereto) in detail at this moment.
8. Embrace AI—Responsibly
From streamlining legal tasks, writing articles to supporting creative brainstorming, generative AI offers enormous advantages. However, that conversation with the lawyer who unknowingly uploaded full client files to a free-tier AI tool reminds us that we to improve our understanding how AI models handle our data.
- Read the Terms: Familiarize yourself with the data usage policies of each AI platform you use.
- Choose Wisely: If data sensitivity is high, consider enterprise solutions or what we at this time advise most legal professionals: do not share sensitive data.
- Use Caution: Always think twice before uploading potentially sensitive content.
- Establish AI Policies: Protect your organization, employees, and clients by setting clear, enforceable guidelines.
Final Thoughts
AI is transforming the way we work – but it’s also transforming how data can move beyond our control. Actively keep track of how your content is used, whether you’re on the free or paid version of ChatGPT, Claude or Perplexity
The lesson? We need to stay informed, be cautious and be proactive. That way, you can use the power of AI without compromising your most sensitive information. Keep an eye out for our upcoming in-depth article on building AI policies that can guide you and your team toward ethical and secure usage of these exciting new technologies.
Disclaimer: This article is a research project, provides general information about AI data usage and does not constitute legal advice.
I wrote this article further to my post on LinkedIn on this subject. If you have any further questions about the above, contact me via lowa@amstlegal.com or set up a meeting directly here .
Here are more articles on this topic Anthropic’s Claude AI Updates – Impact on Privacy & Confidentiality, Italy fines OpenAI over ChatGPT privacy rules breach, User Privacy and Large Language Models: An Analysis of Frontier Developers’ Privacy Policies, Perplexity API Platform Privacy & Security
About AMST Legal
At AMST Legal, we specialize in helping businesses navigate the world of AI and SaaS contracts. We move beyond simple legal reviews to provide strategic advice on data privacy, sales & vendor negotiations and internal governance frameworks. Whether you need an emergency audit following a policy change, help securing enterprise-grade protections from AI vendors, or a fractional General Counsel to oversee your legal operations, we provide the expertise needed to enable responsible AI adoption. Contact us at info@amstlegal.com or book a meeting here to ensure your organization’s data remains a private commercial asset, not a public training set.
Author: Robby Reggers, Founder of AMST Legal (amstlegal.com), recognized by Legal Geek as a LinkedIn Top Voice for contracting, negotiation and interim GC work. Nominated for Dutch Independent Legal Award (DILA) in 2025 and 2026.
AMST Legal supports clients per contract/project or on an interim basis (set hours/week) for larger projects.
