Hosted on MSN
AI distillation could shrink models and cut costs
The AI industry is witnessing a transformative trend: the use of distillation to make AI models smaller and cheaper. This shift, spearheaded by companies like DeepSeek and OpenAI, is reshaping the AI ...
Model distillation is one of the technology trends that has reached a level of maturity identified in Gartner’s 2025 Hype Cycle for artificial intelligence (AI) as “the slope of enlightenment”.
The original version of this story appeared in Quanta Magazine. The Chinese AI company DeepSeek released a chatbot earlier this year called R1, which drew a huge amount of attention. Most of it ...
The AI industry stands at an inflection point. While the previous era pursued larger models—GPT-3's 175 billion parameters to PaLM's 540 billion—focus has shifted toward efficiency and economic ...
DeepSeek's R1 model attracted global attention in January Article in Nature reveals R1's compute training costs for the first time DeepSeek also addresses claims it distilled OpenAI's models in ...
China’s DeepSeek shook the tech world. Its developer just revealed the cost of training the AI model
Chinese artificial intelligence developer DeepSeek spent just $294,000 on training its R1 model, much less than reported for US rivals, it said in a paper that is likely to reignite debate over ...
Results that may be inaccessible to you are currently showing.
Hide inaccessible results