DeepSeek’s success learning from bigger AI models raises questions about the billions being spent on the most advanced technology.
Whether it's ChatGPT since the past couple of years or DeepSeek more recently, the field of artificial intelligence (AI) has seen rapid advancements, with models becoming increasingly large and complex.
One possible answer being floated in tech circles is distillation, an AI training method that uses bigger "teacher" models to train smaller but faster-operating "student" models.
Top White House advisers this week expressed alarm that China's DeepSeek may have benefited from a method that allegedly piggybacks off the advances of US rivals called "distillation."
Microsoft and OpenAI are investigating whether DeepSeek, a Chinese artificial intelligence startup, illegally copying proprietary American technology, sources told Bloomberg
OpenAI believes DeepSeek used a process called “distillation,” which helps make smaller AI models perform better by learning from larger ones.
After DeepSeek AI shocked the world and tanked the market, OpenAI says it has evidence that ChatGPT distillation was used to train the model.
If there are elements that we want a smaller AI model to have, and the larger models contain it, a kind of transference can be undertaken, formally known as knowledge distillation since you ...
OpenAI is reviewing indications that DeepSeek may have trained its AI by mimicking responses from OpenAI’s models.
China's AI advancements, led by DeepSeek, are reshaping global power. Learn why Dario Amodei warns of the urgent need for ethical oversight
The tables turned for OpenAI this week, as it accused Chinese AI company DeepSeek of a “potential breach of intellectual property”. It claims that DeepSeek used its AI models to train its own open-source competitor using a technique known as ‘distillation’.