ChatGPT, Claude, Gemini, Mistral, Perplexity:
The fact-based comparison

Five AI assistants, five different privacy models. Which ones use your chat history for training? How far back does their training data go? And what does that mean for the reliability of current information? A factual comparison based on official sources.

Important to know — Knowledge Cutoff Dates

How far back does the training data go?

AI models are trained on data up to a specific date — the so-called Knowledge Cutoff DateKnowledge CutoffThe date up to which an AI model was trained. Events or information after this date are unknown to the model from its own knowledge — it then needs a web search to provide current information.. Events or information after this date are not part of the model's training. This means: for current topics, a web search is often necessary, even if the AI has access to one.

⚠️
Important: Even if a model "talks" about a current topic, that does not mean the information comes from its training. Most assistants use web search or other tools to retrieve up-to-date information. Without those tools, the model only knows facts up to its cutoff date.

Knowledge cutoff dates for all 5 assistants

Assistant Current model Knowledge cutoff Release date
ChatGPT GPT-5.2 Instant 31 August 2025 11 Dec 2025
Claude Claude 4.6 Sonnet August 2025 (reliable)
January 2026 (training data)
17 Feb 2026
Gemini Gemini 3.0 January 2025 18 Nov 2025
Mistral Mistral Large 2 ~July 2024 July 2024
Perplexity Hybrid (uses other models) Variable (uses GPT, Claude, Mistral) Continuously updated

What does this mean in practice?

Bottom line: If you need current information (news, recent events, new products), enable web search or use Perplexity, which searches by default. Do not rely on the model's built-in knowledge alone.
The 5 assistants in detail

Who are they — and what do they offer?

ChatGPT
OpenAI (USA) — Microsoft partner
Market leader with 800 million weekly users. Strong coding capabilities, but privacy requires manual configuration.
Free: Yes (GPT-4o mini, limited)
Plus: $20/month (GPT-5.2, more requests)
Cutoff: 31 August 2025
Training: ON by default (opt-outOpt-outYou have to actively object to prevent your data from being used for training. Without objecting, your chats are used. The opposite of opt-in. possible)
Strengths: Coding, maths, wide availability
Claude
Anthropic (USA) — "Ethical AI"
Founded by ex-OpenAI staff with a focus on safety. No training without consent.
Free: Yes (Sonnet 4.6, limited)
Pro: $20/month (Opus 4.6, more requests)
Cutoff: August 2025 (reliable), Jan 2026 (training)
Training: OFF by default (opt-inOpt-inYour data is only used for training if you actively consent. Without your consent, nothing happens. More privacy-friendly than opt-out. required)
Strengths: Privacy, coding, long contexts
Gemini
Google (USA)
Integration with Google Workspace (Gmail, Drive, Calendar). Data is linked across Google services.
Free: Yes (Gemini 2.0 Flash, limited)
Advanced: $20/month (Gemini 3.0, 2TB Drive)
Cutoff: January 2025
Training: ON by default (complicated opt-out)
Strengths: Google integration, 100+ languages
Mistral
Mistral AI (France/EU)
European player with GDPRGDPRGeneral Data Protection Regulation — the EU's data protection law since 2018. It governs how companies may process personal data. Violations can be fined up to 4% of global annual turnover. compliance. Open-sourceOpen SourceThe AI model's source code is publicly visible and can be downloaded by anyone. Enables complete data control because the model can run on your own servers — no data leaves your network. models available.
Free: Yes (Large 2, limited)
Pro: €15/month
Cutoff: ~July 2024
Training: ON by default (simple opt-out)
Strengths: EU privacy, multilingual, open source
Perplexity
Perplexity AI (USA)
"Answer EngineAnswer EngineRather than returning links like Google, Perplexity searches the internet and formulates a complete answer with source citations. A hybrid of search engine and AI assistant." with automatic web search. Always with source citations.
Free: Yes (5 searches/day, limited)
Pro: $20/month (unlimited, GPT-5 + Claude)
Cutoff: Variable (uses external models)
Training: ON by default (simple opt-out, incognito mode)
Strengths: Research, source citations, current information
Who uses your data for training?

How providers use your chat history

Most AI assistants use chat histories to improve their models — if you do not object (opt-outOpt-out vs. Opt-inOpt-out = you must actively object, otherwise your data is used. Opt-in = your data is only used if you actively consent. Opt-in is more privacy-friendly.). The details vary considerably:

Provider Default setting Opt-out possible? Retention after opt-out Enterprise/API
ChatGPT Training ON Yes, manually 30 days (safety/abuse) No training
Claude Training OFF Opt-in required for training 30 days (safety)
5 years if opted in
No training
Gemini Training ON Complicated (Activity Control) 18 months default (adjustable: 3/36 months) No training
Mistral Training ON Yes, simple opt-out No exact figures given No training (API/Enterprise)
Perplexity Training ON Yes (AI Data Usage)
Incognito mode stores nothing
No exact figures given (incognito: 0 days) No training (API)

The details compared

ChatGPT (OpenAI):

Claude (Anthropic):

Gemini (Google):

Mistral:

Perplexity:

🔒
Important: Even with training turned off, all providers store conversations for 30 days or more for "safety and abuse detection". Furthermore, opting out only prevents future training — data already collected remains in the training dataset.

Privacy ranking

  1. Claude: No training without opt-in, transparent 30-day retention.
  2. Mistral: EU-based, GDPR-compliant, simple opt-out, open-source option.
  3. Perplexity: Simple opt-out, incognito mode, but data passes through OpenAI/Anthropic APIs.
  4. ChatGPT: Opt-out possible, but training is on by default. Temporary Chat available.
  5. Gemini: Complicated opt-out, 18-month retention, Google ecosystem linkage.
The facts problem

Hallucinations: AI confidently makes things up

What are hallucinations? AI models "hallucinateHallucinationWhen an AI confidently invents false information — such as non-existent studies or court rulings. The model does not know it is lying, because it has no concept of truth." when they confidently present false information — invented court rulings, non-existent studies, fabricated facts. The problem: they sound completely convincing while doing so.

Case 1: Mata v. Avianca (May 2023)

New York lawyer Steven Schwartz filed a brief in which 6 of the 9 cited court rulings were fabricated — all generated by ChatGPT. When the judge asked him to produce the rulings, Schwartz asked ChatGPT: "Are these cases real?" ChatGPT: "Yes, these are real cases."

Schwartz submitted the "full texts" — all invented by ChatGPT, including fabricated quotes and internal references. The judge called it "unprecedented" and imposed a $5,000 fine.

979 documented cases worldwide

French researcher Damien Charlotin maintains a database of lawyers worldwide who have been caught out by AI hallucinations. From 2 cases per week (spring 2025) to 2–3 cases per day (end of 2025). The numbers are rising exponentially.

Stanford study: 17–33% hallucination rate

Stanford University RegLab (May 2024) tested specialised legal AI tools (Lexis+ AI, Westlaw AI) — tools marketed as "hallucination-free". The result:

Why do AI models hallucinate?

AI language models work probabilisticallyProbabilisticProbability-based: the model calculates which next word is statistically most likely — based on patterns in billions of pages of text. It does not understand what it says. — they calculate which word is most likely to come next, based on patterns in training dataTraining dataBillions of texts from the internet, books, Wikipedia and other sources on which an AI model is trained. The quality and scope of this data determines what the AI can do and knows.. They have no understanding of "true" or "false". When you ask a difficult question for which there are no good answers in the training data, the AI invents a plausible-sounding answer.

Bottom line: The harder your question, the more likely the model is to hallucinate — because it wants to please you and generates a plausible answer even when no facts are available.

What does this mean in practice?

27 February 2026 — What happened yesterday

Pentagon vs. Anthropic: ethics or pragmatism?

On 27 February 2026, a conflict between the US Department of Defense and Anthropic (makers of Claude) escalated. The Pentagon gave Anthropic a deadline of 5:01 pm: remove the restrictions on mass surveillance and autonomous weapons, or lose the $200 million contract.

Anthropic CEO Dario Amodei refused: "We cannot in good conscience agree." The company insisted that Claude must not be used for mass surveillance of US citizens or fully autonomous weapons systemsAutonomous weaponsWeapons systems that can independently select and engage targets without human decision-making. Highly controversial internationally, as they remove human control over decisions to kill..

President Trump ordered: "All federal agencies must immediately stop using Anthropic." Defence Secretary Pete Hegseth declared Anthropic a "Supply Chain RiskSupply Chain RiskAn official US classification normally reserved for hostile foreign companies (e.g. Huawei). Applying it to a US company is unprecedented." — a designation normally reserved for foreign adversaries.

Hours later: OpenAI CEO Sam Altman announced that his company had signed a deal with the Pentagon — reportedly with the same "red lines" (no mass surveillance, no autonomous weapons).

What is the difference?

Unclear. OpenAI claims the restrictions are identical to Anthropic's demands. Anthropic says: the Pentagon reserved the right to override the rules "when necessary". Anthropic wanted binding guarantees. OpenAI accepted the Pentagon's promise to adhere to "existing laws and policies".

Context: OpenAI & the military

In January 2024, OpenAI removed the ban on military use from its terms of service. Since then, OpenAI has been actively working with the Pentagon. Yesterday (27 Feb 2026) that arrangement was formalised.

What does this mean?

This development reveals different priorities:

For privacy-conscious users the question is: do I trust a company that puts money above principles — or one that sacrifices money for principles?

💡 Conclusion — Which assistant for which purpose?

Privacy priority: Claude (no training without opt-in, Pentagon expulsion demonstrates principled stance) or Mistral (EU-based, GDPR-compliant).

Research & current information: Perplexity (automatic web search, source citations, most up-to-date data).

Coding & maths: ChatGPT (GPT-5.2) or Claude (Opus 4.6) — both strong, but different privacy models.

Google ecosystem: Gemini (Gmail, Drive, Calendar integration) — but data gets linked and stored for 18 months.

Market leader (despite concerns): ChatGPT (largest user base, but training on by default, Pentagon deal as of yesterday).

Most important rule for all: Never trust blindly. Check sources, verify facts. For current information, enable web search or use Perplexity.

Sources & references
[1]
ALLMO: Knowledge Cutoff Dates LLMs (Feb 2026)
Claude 4.6 Sonnet: Aug 2025 (reliable), Jan 2026 (training). ChatGPT 5.2: Aug 2025. Gemini 3: Jan 2025.
allmo.ai/articles/list-of-large-language-model-cut-off-dates
[2]
Medium: AI Chatbot Data Privacy ChatGPT/Claude/Gemini (Feb 2026)
Detailed comparison of training policies. ChatGPT: 30-day retention. Claude: opt-in/opt-out hybrid since Sept 2025. Gemini: 18 months.
medium.com/@aftab001x/the-truth-about-ai-chatbot-data-privacy
[3]
ObscureIQ: Privacy-Respecting LLMs (Dec 2025)
Ranking: Claude 3.5/5, ChatGPT 3.5/5, Perplexity 3.1/5. Details on opt-out mechanisms, retention and enterprise guarantees.
obscureiq.com/which-generative-ai-is-most-privacy-respecting/
[4]
Tom's Guide: Privacy Comparison ChatGPT/Claude/Gemini (Dec 2025)
Claude = best choice (no training without opt-in). ChatGPT = worst (training by default). Perplexity = granular control.
tomsguide.com/ai/compared-privacy-chatgpt-gemini-claude-and-perplexity
[5]
Legal Dive: ChatGPT Fake Cases Scandal (May 2023)
Mata v. Avianca: 6 of 9 cited court rulings invented. ChatGPT confirmed the fakes as real. $5,000 fine imposed.
legaldive.com/news/chatgpt-fake-legal-cases-generative-ai-hallucinations/651557/
[6]
Damien Charlotin: AI Hallucination Cases Database (2025)
979 documented cases worldwide. From 2 cases/week (spring 2025) to 2–3 cases/day (end of 2025).
damiencharlotin.com/hallucinations/
[7]
Stanford RegLab: Legal AI Hallucination Study (May 2024)
Lexis+ AI: 17% hallucination rate. Westlaw AI: 33%. Despite "hallucination-free" marketing.
hai.stanford.edu/news/ai-trial-legal-models-hallucinate-1-out-6-or-more
[8]
NPR: Pentagon drops Anthropic, OpenAI takes over (28 Feb 2026)
Trump orders halt to Anthropic. OpenAI signs Pentagon contract hours later with allegedly identical "red lines".
npr.org/2026/02/27/nx-s1-5729118/anthropic-pentagon-openai-ai-weapons
[9]
CNN: OpenAI Pentagon Deal Details (28 Feb 2026)
Different terms: Anthropic wanted binding guarantees, OpenAI accepted Pentagon promises.
cnn.com/2026/02/27/tech/openai-pentagon-deal-ai-systems
[10]
Fello AI: Stop AI from Training on Your Data (Dec 2025)
Step-by-step opt-out guide for ChatGPT, Claude and Gemini. Policy changes at end of 2025 towards "opt-out" models.
felloai.com/how-to-stop-ai-from-training-on-your-data/
Link copied ✓