Sarvam AI, a Bengaluru-based artificial intelligence startup, has recently launched its flagship large language model (LLM), Sarvam-M, designed with a focus on Indian languages and reasoning capabilities. This open-source, 24-billion-parameter model is built upon Mistral Small and represents a significant stride towards building a sovereign AI ecosystem in India. Sarvam-M is engineered to power a diverse range of applications, including conversational agents, machine translation, and educational tools, specifically tailored to the Indian context.
Sarvam-M distinguishes itself through its unique training process, which involves three key steps: Supervised Fine-Tuning (SFT), Reinforcement Learning with Verifiable Rewards (RLVR), and inference optimization. During SFT, the model is trained using carefully crafted prompts to enhance its capabilities in general dialogue and complex reasoning. RLVR further refines its instruction-following and mathematical skills through custom reward engineering and curated datasets. Finally, inference is optimized using FP8 post-training quantization and techniques like lookahead decoding, ensuring efficient and accurate responses, especially in real-time applications.
The model has demonstrated strong performance in multilingual and reasoning benchmarks. It achieved an impressive 86% gain on a romanized Indian language version of the GSM-8K math dataset. Moreover, it showcased average performance boosts of 20% on Indian language benchmarks, 21.6% on math tasks, and 17.6% on programming tasks. When compared to other models, Sarvam-M outperforms Llama-4 Scout and is comparable to larger models like Llama-3.3 70B and Gemma 3 27B. However, it slightly lags in English benchmarks such as MMLU, indicating a trade-off for its enhanced multilingual and reasoning strengths.
Sarvam-M's architecture is designed for versatility, supporting a wide array of applications. Its accessibility is facilitated through Sarvam's API, a dedicated playground, and its availability for download on Hugging Face, enabling developers and researchers to experiment and integrate the model into various projects. The model supports 10 Indian languages, including Hindi, Bengali, Gujarati, Kannada, and Malayalam.
The launch of Sarvam-M is part of Sarvam AI's broader vision to create a sovereign AI ecosystem in India. This initiative is aligned with the Indian government's IndiaAI Mission, which aims to strengthen the country's domestic AI capabilities. Sarvam AI was selected by the Indian government to build a sovereign LLM under this mission.
Despite the model's capabilities, its initial reception has been mixed. While it has been praised for its focus on Indian languages, mathematics, and programming tasks, some critics have pointed out that it is not "good enough" compared to more established models. Some experts argue that there are cheaper and better models available from Google and other companies. The debate extends to whether India should focus on building AI for local needs or benchmark against Silicon Valley.
In response to the criticism, Sarvam AI has emphasized that Sarvam-M is a research model and a stepping stone towards building a comprehensive sovereign AI. The company plans to release models regularly and share detailed technical findings to foster collaboration and innovation.