DeepSeek-R1 last week, an open-source large language model (LLM) reportedly rivaling OpenAI’s top offerings, sending shockwaves through the industry and generating much excitement in the tech world. It apparently started as a side project at a Chinese hedge fund before being spun out.
Budget with ET
India, get ready for the modern warfare of tomorrow
Women, youth, farmers and poor can continue to be Budget 2025's ‘roti, kapda aur makan’
Modi govt has a key task in Budget 2025: Unlocking the PLI goldmine
Its efficacy, combined with claims of being built at a fraction of the cost and hardware requirements, has seriously
challenged BigAI’s notion that “foundation models” demand astronomical investments. I have personally been playing around with R1 and have found it to be excellent at writing code.
Speaking of foundation models, one rarely hears that term anymore; unsurprising, given that foundation is now commodity. Building a foundation-level LLM was once touted as the cornerstone of AI sovereignty, but that rhetoric has also waned. Muchhas changed regarding the idea of AI sovereignty.
When OpenAI’s and MidJourney’s generative AI models came out, they changed not just the mainstream psyche, but the techno-geopolitical landscape. Soon thereafter, Meta’s and StabilityAI’s open-weight models broke the floodgates to decentralised development and iteration in the tech community. More than that, the number of AI breakthroughs that have been coming out of the global open-source realm has been nothing short of astounding. AI