Though funding remains limited, efforts are underway to launch India's own multilingual low-cost generative AI
Illustration: Chaitanya Dinesh Surpur; Photo imaging: freepik.com
The emergence of China’s DeepSeek, which is said to have been built at a small fraction of the cost compared to OpenAI’s GPT large language models (LLMs), has sparked excitement among India’s artificial intelligence (AI) scientists and engineers to build a home-grown LLM from scratch. India is entirely reliant on models such as OpenAI’s proprietary GPT or Meta’s Llama, an open-source model. It is OpenAI's second-largest market by number of users, said the company’s CEO Sam Altman during his visit to Delhi earlier in February. India also accounts for about 15.6 percent of DeepSeek’s AI chatbot downloads, according to AppFigures.
DeepSeek’s makers, Hangzhou DeepSeek Artificial Intelligence Basic Technology Research Co, say they spent only $5.576 million on training the model on H800 processors from Nvidia; these processors were designed to meet the restrictions placed on the chipmaker by the US government on exports to China. Some experts have pointed out that based on DeepSeek’s workload, the chips were equivalent to Nvidia’s H100 GPUs (graphical processing units).
In comparison, OpenAI’s GPT-4 is said to have cost $80-100 million to train, apart from the cost of developing the actual model itself and developing each new iteration. OpenAI’s 2024 losses were projected at $5 billion, CNBC reported last September. Such prohibitive costs had deterred any serious attempts to develop an AI model in India, but now DeepSeek seems to have changed that.
“The government is now asking us to build models for them. We're well-positioned to deliver, as we've already developed a model comparable to GPT,” says Vishnu Vardhan, founder and CEO of Hanooman AI, which is developing multilingual, multimodal LLMs, including one named Everest 1.0.