DeepSeek’s success learning from bigger AI models raises questions about the billions being spent on the most advanced technology.
Since Chinese artificial intelligence (AI) start-up DeepSeek rattled Silicon Valley and Wall Street with its cost-effective models, the company has been accused of data theft through a practice that is common across the industry.
Whether it's ChatGPT since the past couple of years or DeepSeek more recently, the field of artificial intelligence (AI) has seen rapid advancements, with models becoming increasingly large and complex.
One possible answer being floated in tech circles is distillation, an AI training method that uses bigger "teacher" models to train smaller but faster-operating "student" models.
AI-driven knowledge distillation is gaining attention. LLMs are teaching SLMs. Expect this trend to increase. Here's the insider scoop.
Microsoft and OpenAI are investigating whether DeepSeek, a Chinese artificial intelligence startup, illegally copying proprietary American technology, sources told Bloomberg
Top White House advisers this week expressed alarm that China's DeepSeek may have benefited from a method that allegedly piggybacks off the advances of U.S. rivals called "distillation."
After DeepSeek AI shocked the world and tanked the market, OpenAI says it has evidence that ChatGPT distillation was used to train the model.
Top White House advisers this week expressed alarm that China’s DeepSeek may have benefited from a method that allegedly piggybacks off the advances of US rivals called
CNBC's Deirdre Bosa joins 'The Exchange' to discuss what DeepSeek's arrival means for the AI race. White House says Trump funding freeze remains in effect despite rescinding OMB memo
There is a big lesson here for South Korea’s policymakers and tech companies. The country, which lags behind in AI solutions, may stumble into a fresh opportunity to catch up if the government and conglomerates realise it’s not too late to invest in making inroads into the AI battlefield.