Trustworthy and Explainable LLMs for Indian Languages

Funded by the ANRF PM Early Career Research Grant, this project develops trustworthy and explainable LLMs for Indian languages

Indian language AI systems currently suffer from high hallucination rates and poor explainability, largely due to limited training data and complex linguistic structures. This project develops trustworthy and explainable LLMs for Indian languages — building on the BHRAM-IL benchmark to create AI systems capable of generating factually accurate, culturally appropriate, and explainable content.

Objectives

  • Design and implement domain-specific hallucination mitigation techniques using RAG and knowledge graph integration
  • Create annotated datasets and evaluation metrics tailored for Indian language trustworthiness assessment
  • Develop explainability frameworks combining grammar-informed linguistic rules and KG-based justifications
  • Build Indian-language-suitable evaluation metrics and conduct systematic validation across diverse languages and domains
  • Develop prototype applications including grammar correction assistants and knowledge-guided chatbots

Expected Outcomes

  • Open-source datasets for hallucination detection and explainability evaluation across major Indian languages
  • Specialized hallucination mitigation and explainability frameworks optimized for Indian linguistic structures
  • Comprehensive benchmark suite for trustworthiness in Indian-language LLMs
  • Research publications and trained personnel in trustworthy AI methodologies

Funding agency: Anusandhan National Research Foundation (ANRF), Government of India

Scheme: PM Early Career Research Grant (PM ECRG)

Amount: ₹60 Lakhs + overheads

Start Year: 2026

Duration: 3 years

PI: Hrishikesh Terdalkar