Begin typing your search above and press return to search.
proflie-avatar
Login
exit_to_app
Sharjeel Imam
access_time 30 Jan 2025 10:15 AM IST
Kerala amidst shocking murder rampage
access_time 29 Jan 2025 10:44 AM IST
Trump 2.0 quick in action hitting immigration
access_time 28 Jan 2025 12:12 PM IST
Trump has begun his work
access_time 28 Jan 2025 9:30 AM IST
Lets uphold this Constitution
access_time 26 Jan 2025 10:53 AM IST
DEEP READ
Munambam Waqf issue decoded
access_time 16 Nov 2024 10:48 PM IST
Ukraine
access_time 16 Aug 2023 11:16 AM IST
The Russian plan: Invade Japan and South Korea
access_time 16 Jan 2025 3:32 PM IST
Putin
access_time 2 Jan 2025 1:36 PM IST
What is Christmas?
access_time 26 Dec 2024 11:19 AM IST
exit_to_app
Homechevron_rightTechnologychevron_rightOpenAI suspects...

OpenAI suspects DeepSeek used its AI models to build DeepSeek-R1

text_fields
bookmark_border
OpenAI suspects DeepSeek used its AI models to build DeepSeek-R1
cancel

OpenAI has reportedly raised concerns that its proprietary AI models may have been used to develop DeepSeek-R1, a newly released AI model by Chinese firm DeepSeek.

According to a report by the Financial Times, OpenAI claims to have found evidence of distillation — a process in which AI models learn from the outputs of larger models - potentially involving its technology.

The San Francisco-based AI firm has stated that certain users may have been leveraging its application programming interface (API) to extract and distill information from its models. Suspecting that these activities were linked to DeepSeek, OpenAI, along with its cloud partner Microsoft, conducted an investigation and subsequently blocked the accounts involved.

In a statement to the Financial Times, OpenAI remarked, "We know [China]-based companies — and others — are constantly trying to distill the models of leading US AI companies." The company also noted that it is collaborating with the US government to safeguard its advanced AI models from competitors and potential adversaries.

DeepSeek recently released DeepSeek-R1, an open-source, reasoning-focused AI model, on GitHub and Hugging Face. The model has reportedly outperformed OpenAI's o1 AI models in multiple benchmarks, raising speculation about how it achieved such advancements in a short time.

AI model distillation is a well-known technique where knowledge from a large AI model is transferred to a smaller, more efficient model, improving performance while reducing computational demands. OpenAI’s GPT-4, for example, is believed to have around 1.8 trillion parameters, whereas DeepSeek-R1 operates on 1.5 billion parameters, making it significantly smaller.

Typically, companies apply distillation to their own proprietary models to refine them, as seen with Meta’s Llama 3, which was used to develop coding-specialised models. However, if OpenAI’s claims are accurate, DeepSeek may have extracted OpenAI’s model outputs by using prompt injections on its API, generating large datasets of text that were then converted into training data for DeepSeek-R1.

Despite these allegations, OpenAI has not yet released a public statement on the matter. Interestingly, Sam Altman, OpenAI’s CEO, recently praised DeepSeek for its advancements, acknowledging that its success fosters competition in the AI industry.

Show Full Article
TAGS:Artificial IntelligenceDeepSeek ModelSam Altman on DeepSeek
Next Story