In the paper accompanying the launch of R1, DeepSeek explained how it took advantage of techniques such as synthetic data generation, distillation, and machine-driven reinforcement learning to produce a model that exceeded the current state-of-the-art. Each of these approaches can be explained another way as harnessing the capabilities of an existing AI model to assist in the training of a more advanced version.
DeepSeek is far from alone in using these AI techniques to advance AI. Mark Zuckerberg predicts that the mid-level engineers at https://fortune.com/company/facebook/” class=””>Meta may soon be replaced by AI counterparts, and that Llama 3 (his company’s LLM) “helps us experiment and iterate faster, building capabilities we want to refine and expand in Llama 4.” https://fortune.com/company/nvidia/” class=””>Nvidia CEO Jensen Huang has spoken at length about creating virtual environments in which AI systems supervise the training of robotic systems: “We can create multiple different multiverses, allowing robots to learn in parallel, possibly learning in 100,000 different ways at the same time.”
This isn’t quite yet the singularity, when intelligent machines autonomously self-replicate, but it is something new and potentially profound. Even amidst such dizzying progress in AI models, though, it’s not uncommon to hear some observers talk about the potential slowing of what’s called the “scaling laws”—the observed principles that AI models increase in performance in direct relationship to the quantity of data, power, and compute applied to them. The release from DeepSeek, and several subsequent announcements from other companies, suggests that arguments of the scaling laws’ demise may be greatly exaggerated. In fact, innovations in AI development are leading to entirely new vectors for scaling—all enabled by AI itself. Progress isn’t slowing down, it’s speeding up—thanks to AI.