Home
Finance
Travel
Academic
Library
Create a Thread
Home
Discover
Spaces
 
 
  • Introduction
  • Evidence of Gemini Terminology Patterns
  • Sam Paeach's Detection Methods
  • DeepSeek R1-0528 Benchmark Improvements
  • DeepSeek Response Status
 
DeepSeek may have used Google Gemini to train its R1 update

According to TechCrunch, Chinese AI lab DeepSeek's recently updated R1-0528 reasoning model has sparked controversy as developers have identified similarities suggesting it may have been trained using outputs from Google's Gemini AI. While DeepSeek didn't disclose its training data sources, experts like Sam Paeach have noted the model's preference for terminology and expressions remarkably similar to those used by Google's Gemini 2.5 Pro.

User avatar
Curated by
artscraftsteve
3 min read
Published
16,332
427
tomsguide.com favicon
tomsguide
DeepSeek’s latest update is a serious threat to ChatGPT and Google — here’s why
DeepSeek’s latest update is a serious threat to ChatGPT and Google — here’s why
venturebeat.com favicon
venturebeat
DeepSeek R1-0528 arrives in powerful open source challenge to OpenAI o3 and Google Gemini 2.5 Pro
techcrunch.com favicon
techcrunch
DeepSeek may have used Google's Gemini to train its latest model
DeepSeek may have used Google's Gemini to train its latest model
mashable.com favicon
finance.yahoo.com favicon
autogpt.net favicon
+42 sources
I'm DeepSeek. How can I help you today?
Solen Feyissa
·
unsplash.com
Evidence of Gemini Terminology Patterns

The evidence pointing to DeepSeek's potential use of Google Gemini outputs centers on distinctive terminology patterns. Experts have identified that DeepSeek's R1-0528 model exhibits a notable preference for Gemini-specific terms and expressions, such as "context window," "foundation model," and "function calling"—technical vocabulary that appears frequently in Google's Gemini documentation12. Additionally, the model demonstrates familiar response structures and stylistic elements characteristic of Gemini's outputs, including its approach to explaining AI concepts and its particular phrasing when discussing generative capabilities34.

These linguistic fingerprints are especially telling because AI models tend to adopt the terminology patterns of their training data. When an LLM is trained on outputs from another model like Gemini, it inherits not just knowledge but also distinctive vocabulary and phrasing—similar to how human language learners adopt the speech patterns of their teachers56. This phenomenon, sometimes related to "prompt chaining," shows how information from previous interactions influences future responses, potentially revealing the model's training lineage6.

ai.google.dev favicon
gemini-us.com favicon
umalibguides.uma.edu favicon
9 sources
Sam Paeach's Detection Methods

Sam Paeach's analysis of DeepSeek's R1-0528 model relied on sophisticated AI detection techniques that can identify model lineage through linguistic patterns. While specific AI detectors vary in effectiveness—with some commercial solutions like Pangram achieving up to 99.3% accuracy in identifying AI-generated content1—detection methods generally fall into several categories:

  • Statistical detection analyzes word frequencies, n-gram patterns, and syntactic structures to identify machine-generated text2

  • Neural network approaches like BERT and RoBERTa-based detectors examine deeper linguistic features, with some achieving over 97% accuracy in controlled studies34

  • Zero-shot detection techniques examine the probability distribution of text without requiring additional training, capable of reaching 99% accuracy for certain models53

The detection of training data lineage represents a particularly challenging frontier in AI forensics, as models trained on outputs from other AI systems inherit distinctive vocabulary and response patterns that serve as "fingerprints" of their training sources—precisely what Paeach identified in DeepSeek's model behavior.

mitsloanedtech.mit.edu favicon
julius.ai favicon
eschoolnews.com favicon
9 sources
DeepSeek R1-0528 Benchmark Improvements

DeepSeek-R1-0528 demonstrates remarkable performance improvements across multiple benchmarks, positioning it as a serious competitor to closed-source models like OpenAI's o3 and Google's Gemini 2.5 Pro. In mathematics testing, the model achieved 87.5% accuracy on the AIME 2025 test, up from 70% in the previous version, and 91.4% on AIME 2024.12 This enhanced reasoning capability stems from deeper computational processing—using an average of 23,000 tokens per question compared to the previous 12,000.23

The model shows significant gains in other critical areas as well:

  • Programming: LiveCodeBench scores increased from 63.5% to 73.3%, while SWE Verified evaluation rose from 49.2% to 57.6%23

  • General reasoning: GPQA-Diamond test scores improved from 71.5% to 81.0%2

  • Complex reasoning: Performance on "Humanity's Last Exam" more than doubled from 8.5% to 17.7%12

  • Reduced hallucinations: The update significantly decreases factually inaccurate responses23

  • Enhanced integration: New support for JSON output generation and expanded function calling capabilities24

tomsguide.com favicon
actuia.com favicon
mashable.com favicon
11 sources
DeepSeek Response Status

While DeepSeek has not directly addressed the specific allegations about using Google's Gemini outputs to train its R1-0528 model, the company has faced similar accusations before. In December, developers observed that DeepSeek's V3 model frequently identified itself as ChatGPT, suggesting it may have been trained on OpenAI's chat logs.1 Earlier in 2025, OpenAI reported finding evidence linking DeepSeek to distillation techniques—extracting data from larger, more capable models to train smaller ones.1

This pattern of suspected AI model distillation has prompted increased security measures across the industry. OpenAI implemented ID verification requirements in April 2025 for accessing advanced models, notably excluding China from its list of supported countries.1 AI experts like Nathan Lambert from AI2 consider it plausible that DeepSeek would leverage synthetic data from leading models, noting the company is "short on GPUs and flush with cash," making this approach "effectively more compute for them."1 The silence from DeepSeek regarding these latest allegations follows their established pattern of not directly addressing training methodology controversies.

techcrunch.com favicon
mlq.ai favicon
mashable.com favicon
8 sources
Related
How convincing is the evidence linking DeepSeek's model to Google's Gemini outputs
What specific linguistic patterns suggest DeepSeek trained on Gemini data
Could DeepSeek's performance improvements be solely due to internal advancements
Discover more
Italy opens probe into AI firm DeepSeek over hallucination risks
Italy opens probe into AI firm DeepSeek over hallucination risks
Italy's antitrust watchdog AGCM has launched a formal investigation into Chinese artificial intelligence startup DeepSeek for allegedly failing to adequately warn users about the risk of "hallucinations" - situations where the AI model generates inaccurate, misleading, or fabricated information in response to user inputs, as reported by Reuters.
6,503
Google tests audio overviews in Search Labs with Gemini AI
Google tests audio overviews in Search Labs with Gemini AI
Google is testing a new feature called Audio Overviews in Search Labs that uses its latest Gemini AI models to generate spoken summaries of search results for specific queries, offering users a hands-free way to absorb information while multitasking or when an audio format is preferred.
5,289
Apple's AI models trail rivals in lukewarm WWDC debut
Apple's AI models trail rivals in lukewarm WWDC debut
Apple's latest artificial intelligence models and design overhaul unveiled at the company's annual developer conference Monday received a lukewarm reception from analysts and early users, highlighting the tech giant's ongoing struggle to match competitors in the AI race. The company's own performance benchmarks showed its newest AI models trailing behind year-old offerings from OpenAI and Meta,...
9,165
Mistral AI launches reasoning models to rival OpenAI
Mistral AI launches reasoning models to rival OpenAI
French artificial intelligence startup Mistral AI unveiled its first reasoning models Tuesday, positioning the Microsoft-backed company as a direct competitor to OpenAI and Google in the race to develop AI systems capable of complex, step-by-step problem solving. The company launched two versions of its new Magistral family during London Tech Week, with CEO Arthur Mensch telling CNBC the models...
1,637