Tag: Sarvam AI

  • Launched by Sarvam AI, Sarvam 1 LLM is Trained in English and Ten Indic Languages

    On October 24, Sarvam AI, an artificial intelligence (AI) firm supported by Lightspeed, unveiled Sarvam 1, a Large Language Model (LLM). According to a tweet on X (previously Twitter), the business says it is India’s first indigenous multilingual LLM, trained from scratch on domestic AI infrastructure in ten Indian languages and English.

    Bengali, Gujarati, Hindi, Kannada, Malayalam, Marathi, Oriya, Punjabi, Tamil, and Telugu are among the ten major Indian languages that Sarvam 1 supports in addition to English. The LLM uses a two-billion-parameter language model and is trained on Nvidia’s H100 Graphics Processing Unit (GPU).

    Sarvam AI uses Nvidia services and AI4Bharat’s open-source technology

    In order to optimise and implement conversational AI agents with sub-second latency, Sarvam AI also makes use of a variety of Nvidia services and products, including its microservice, conversational AI, LLM software, and inference server.

    In addition to Nvidia, the LLM made use of AI4Bharat’s open-source technology and language resources, as well as Yotta’s data centres for computational infrastructure. According to a blog post by the AI startup, Sarvam-1’s strong performance and computational efficiency make it especially well-suited for real-world uses, such as deployment on edge devices.

    In specifics, Sarvam 1 clearly beats Gemma-2-2B and Llama-3.2-3B on a number of common benchmarks, such as MMLU, Arc-Challenge, and IndicGenBench, while attaining comparable results to Llama 3.1 8B, the company stated.

    Functioning of Various LLM Models Launched by the Company

    India’s first Hindi LLM, Open Hathi, was introduced by the AI firm in December 2023. The Llama2-7B architecture from Meta AI, which has 48,000 token extensions, served as the foundation for the model. However, a training corpus of two trillion tokens is used to develop Sarvam.

    Because of its effective tokeniser and unique data pipeline, which can produce diversified and high-quality text while preserving factual correctness, the LLM has two trillion tokens of synthetic Indic data. In addition to being four to six times faster during inference, Sarvam claimed that the most recent model from their stable meets or surpasses much larger models like Llama 3.1 8B.

    The process by which a trained model predicts or deduces from fresh data using the patterns it discovered during training is known as inference in artificial intelligence. Compared to current Indic datasets, the companies’ pretraining corpus, Sarvam-2T, supports eight times as much scientific material, three times as high quality, and two times as long documents. The total number of Indic tokens stored by Sarvam-2T is around 2 trillion. Apart from Hindi, which makes up over 20% of the data, the data is distributed nearly evenly among the ten supported languages.


    AI Firm Sarvam Unveils Blend of Open Source and Enterprise Products
    AI firm Sarvam unveils a new GenAI platform featuring a mix of open source and enterprise products, with support for 10 Indian languages.


  • AI firm Sarvam Unveiled a Blend of Open Source and Enterprise Products

    As part of its full-stack Generative AI (GenAI) platform, Sarvam AI, an AI startup based in Bengaluru, released a suite of products recently. These solutions cater to both enterprise usage and open source communities.

    With the backing of investors like Lightspeed and Peak XV (formerly Sequoia India), the company said that its upcoming products will support 10 Indian languages—Hindi, Tamil, Telugu, Malayalam, Punjabi, Odia, Gujarati, Marathi, Kannada, and Bengali—and be voice-enabled to run a variety of jobs.

    During an interview with a media outlet, Sarvam cofounder Vivek Raghavan said that the full-stack GenAI platform has been developed and deployed in collaboration with prominent industrial and technological partners. The product mix unveiled includes A1, Sarvam Models, Sarvam Agents, Shuka1.0, Sarvam 2B, and Sarvam Models.

    Products and their usage

    With their multilingual speech capabilities, the initial product, Sarvam Agents, will provide clients with the ability to communicate with agents through phone calls, WhatsApp, or in-app chat. Additionally, they will have the capability to act and make judgements in response to customer inputs. Businesses in industries such as banking, law, consumer products, telecommunications, media, and technology will be able to take advantage of the voice agents for as little as one rupee a minute.

    Another offering is an open source large-language model (LLM) named Sarvam 2B. According to Sarvam, the LLM can efficiently carry out targeted tasks in ten Indian languages thanks to its training on an internal dataset consisting of four trillion tokens.

    Meta, a digital giant, has an open source Llama 8B language model; their third product, Shuka1.0, will be an audio extension that supports Indian language usage. This product will also be available as open source.

    Additionally, a product titled “Sarvam Models” will be made accessible, which contains the Indic language models utilised in the development of Sarvam Agents. Application programming interfaces (APIs) will now be made available for these models. As part of Sarvam’s developer API platform, developers will have access to models for document parsing, speech synthesis, translation, and speech recognition.

    With tools like regulatory chat, document creation, redaction, and data extraction, the fifth product, ‘A1’, is a generative AI workbench made for solicitors to improve their skills.

    Enhancing the growth with partnership

    Yotta, Nvidia, Exotel, Microsoft Azure, and Google Cloud Platform (GCP) are some of the companies that Sarvam will team up with to power these offerings. During an event, Vishal Dhupar, Nvidia’s managing director for South Asia, said that the Sarvam stack will be powered by Nvidia’s DGX infrastructure.

    The Beckn Foundation and the financial technology company Pine Labs will be Sarvam’s partners. Open Network for Digital Commerce (ONDC), a government-backed e-commerce network, is powered by the Beckn Protocol.

    A rival to Sarvam, Krutrim AI, which has backing from Matrix, has recently launched a number of offerings, including GPU-as-a-service, cloud hosting for large language models, and access to other open source models. In May, Krutrim—which is part of the Ola group headed by Bhavish Aggarwal—also released consumer-facing smartphone applications for the Krutrim AI assistant.


    AI Firms Say Analysing Use Cases and Robust Data Strategies Key for AI
    Explore the AI landscape in India: From emerging startups to challenges companies face. Uncover the trends, myths, and the crucial need for AI training.