Activity
6K followers
Experience
Education
-
Indian Institute of Technology (Indian School of Mines), Dhanbad
-
-
Activities and Societies: Dramatics Club
Licenses & Certifications
-
Six sigma Green Belt
Computer Society of India & Kinduz Business Consulting
Issued -
Honored Listee
Marquis Who's Who
Issued ExpiresCredential ID d179848f9682457aa29e209f5d32391f32f27d2948424a8395f489505405feac
Publications
-
CiteFix: Enhancing RAG Accuracy Through Post-Processing Citation Correction
The 63rd Annual Meeting of the Association for Computational Linguistics
Retrieval Augmented Generation (RAG) has emerged as a powerful application of Large Language Models (LLMs), revolutionizing information search and consumption. RAG systems combine traditional search capabilities with LLMs to generate comprehensive answers to user queries, ideally with accurate citations. However, in our experience of developing a RAG product, LLMs often struggle with source attribution, aligning with other industry studies reporting citation accuracy rates of only about 74% for…
Retrieval Augmented Generation (RAG) has emerged as a powerful application of Large Language Models (LLMs), revolutionizing information search and consumption. RAG systems combine traditional search capabilities with LLMs to generate comprehensive answers to user queries, ideally with accurate citations. However, in our experience of developing a RAG product, LLMs often struggle with source attribution, aligning with other industry studies reporting citation accuracy rates of only about 74% for popular generative search engines. To address this, we present efficient post-processing algorithms to improve citation accuracy in LLM-generated responses, with minimal impact on latency and cost. Our approaches cross-check generated citations against retrieved articles using methods including keyword + semantic matching, fine tuned model with BERTScore, and a lightweight LLM-based technique. Our experimental results demonstrate a relative improvement of 15.46% in the overall accuracy metrics of our RAG system. This significant enhancement potentially enables a shift from our current larger language model to a relatively smaller model that is approximately 12x more cost-effective and 3x faster in inference time, while maintaining comparable performance. This research contributes to enhancing the reliability and trustworthiness of AI-generated content in information retrieval and summarization tasks which is critical to gain customer trust especially in commercial products
Other authorsSee publication -
Contradiction Detection in RAG Systems: Evaluating LLMs as Context Validators for Improved Information Consistency
Retrieval Augmented Generation (RAG) systems have emerged as a powerful method for enhancing large language models (LLMs) with up-to-date information. However, the retrieval step in RAG can sometimes surface documents containing contradictory information, particularly in rapidly evolving domains such as news. These contradictions can significantly impact the performance of LLMs, leading to inconsistent or erroneous outputs. This study addresses this critical challenge in two ways. First, we…
Retrieval Augmented Generation (RAG) systems have emerged as a powerful method for enhancing large language models (LLMs) with up-to-date information. However, the retrieval step in RAG can sometimes surface documents containing contradictory information, particularly in rapidly evolving domains such as news. These contradictions can significantly impact the performance of LLMs, leading to inconsistent or erroneous outputs. This study addresses this critical challenge in two ways. First, we present a novel data generation framework to simulate different types of contradictions that may occur in the retrieval stage of a RAG system. Second, we evaluate the robustness of different LLMs in performing as context validators, assessing their ability to detect contradictory information within retrieved document sets. Our experimental results reveal that context validation remains a challenging task even for state-of-the-art LLMs, with performance varying significantly across different types of contradictions. While larger models generally perform better at contradiction detection, the effectiveness of different prompting strategies varies across tasks and model architectures. We find that chain-of-thought prompting shows notable improvements for some models but may hinder performance in others, highlighting the complexity of the task and the need for more robust approaches to context validation in RAG systems.
Other authorsSee publication -
Leveraging analytics to ace new product launches
Mu Sigma Analytics Blog
More than 20,000 new products are launched every month in the CPG market alone. However failure rates for these products are notoriously high, and this issue is prevalent across industries.
So how can marketers leverage data analytics to improve the odds of success for their product launch? Here are some key tenets:Other authorsSee publication
Languages
-
English
-
-
Hindi
-
-
Tamil
-
-
Telugu
-
Explore top content on LinkedIn
Find curated posts and insights for relevant topics all in one place.
View top content