International Journal of Leading Research Publication
E-ISSN: 2582-8010
•
Impact Factor: 9.56
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Monthly Scholarly International Journal
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 7 Issue 4
April 2026
Indexing Partners
Reducing Hallucination and Improving Relevance in Telugu LLMs Through Prompt Design
| Author(s) | Dr. SriSudha Garugu, Ms. Harshitha Dudala, Mr. Jayanth Reddy Mandadi, Mr. Sai Kiran Ladineni |
|---|---|
| Country | India |
| Abstract | Telugu, a classical Dravidian language spoken by over 83 million people, remains significantly underrepresented in the training data of modern large language models. This imbalance causes two interrelated problems: the models frequently generate plausible-sounding but factually wrong outputs — a phenomenon called hallucination — and their responses often miss the point of what a Telugu user is actually asking. This paper tackles both problems head-on through a combination of LoRA-based fine-tuning and structured prompt engineering, without requiring expensive full-model retraining. We built a purpose-made bilingual instruction dataset covering eight distinct cross-lingual task variants, fine-tuned Mistral-7B-v0.3 on it, and then applied four complementary prompt design strategies: role declaration, confidence gating, chain-of-thought reasoning, and dialect-script cues. The results speak for themselves - accuracy jumped from 56.25% to 87.5% and precision from 54.55% to 90.0%, confirming that thoughtful prompt design is one of the most practical tools a developer has for improving Telugu NLP quality right now. |
| Keywords | Semantic Search, FAISS, OpenAI Embeddings, Vector Database, Natural Language Processing, Similarity Search, Information Retrieval, Artificial Intelligence, Machine Learning, Intelligent Search System |
| Field | Computer > Artificial Intelligence / Simulation / Virtual Reality |
| Published In | Volume 7, Issue 4, April 2026 |
| Published On | 2026-04-17 |
Share this

CrossRef DOI is assigned to each research paper published in our journal.
IJLRP DOI prefix is
10.70528/IJLRP
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.