Advancements in Retrieval-Augmented Generation and Large Language Models

The recent developments in the research area of Retrieval-Augmented Generation (RAG) and Large Language Models (LLMs) highlight a significant shift towards enhancing the efficiency, accuracy, and adaptability of these systems in various applications. A common theme across the studies is the focus on improving the interaction between LLMs and external knowledge sources, such as knowledge graphs and document databases, to mitigate issues like hallucinations and knowledge deficiency. Innovations in data augmentation, synthetic data generation, and the integration of advanced reasoning capabilities are prominent, aiming to refine the models' performance in low-resource scenarios and complex tasks. Additionally, there's a notable emphasis on developing frameworks and tools that facilitate the creation of high-quality, domain-specific datasets and benchmarks, thereby enabling more precise evaluation and optimization of RAG systems. The exploration of novel loss functions, reward models, and alignment methods further underscores the field's commitment to enhancing the interpretability, reliability, and user engagement of LLM-based applications.

Noteworthy Papers

  • Conversational Text Extraction with Large Language Models Using Retrieval-Augmented Systems: Introduces a system that leverages LLMs for text extraction from PDFs, offering a conversational interface for intuitive knowledge extraction.
  • Improving Automated Feedback Systems for Tutor Training in Low-Resource Scenarios through Data Augmentation: Explores the use of GPT models for generating synthetic datasets to enhance tutor training feedback systems.
  • Dialogue Benchmark Generation from Knowledge Graphs with Cost-Effective Retrieval-Augmented LLMs: Presents Chatty-Gen, a platform for generating high-quality dialogue benchmarks from knowledge graphs, reducing reliance on costly LLMs.
  • Passage Segmentation of Documents for Extractive Question Answering: Highlights the importance of chunking in RAG pipelines and introduces LGMGC, a novel framework for document segmentation.
  • FRAG: A Flexible Modular Framework for Retrieval-Augmented Generation based on Knowledge Graphs: Proposes FRAG, a framework that synergizes flexibility and retrieval quality in KG-RAG approaches.
  • AirRAG: Activating Intrinsic Reasoning for Retrieval Augmented Generation via Tree-based Search: Introduces AirRAG, a novel thinking pattern in RAG that integrates system analysis with efficient reasoning actions.
  • Towards Preventing Overreliance on Task-Oriented Conversational AI Through Accountability Modeling: Proposes an accountability model for LLM-based dialogue agents to address user overreliance.
  • BoK: Introducing Bag-of-Keywords Loss for Interpretable Dialogue Response Generation: Introduces BoK loss, a novel auxiliary loss function for enhancing dialogue generation interpretability.
  • Leveraging Chain of Thought towards Empathetic Spoken Dialogue without Corresponding Question-Answering Data: Presents LPE, a method for empathetic dialogue generation without the need for question-answering data.
  • Generative Retrieval for Book search: Introduces GBS, a generative retrieval framework for book search that features data augmentation and outline-oriented book encoding.
  • A Collection of Question Answering Datasets for Norwegian: Introduces a suite of question answering datasets for Norwegian, covering a wide range of skills and knowledge domains.
  • Question-to-Question Retrieval for Hallucination-Free Knowledge Access: An Approach for Wikipedia and Wikidata Question Answering: Introduces an approach to question answering over knowledge bases by performing question-to-question matching and retrieval.
  • Conversation Routines: A Prompt Engineering Framework for Task-Oriented Dialog Systems: Introduces Conversation Routines, a structured prompt engineering framework for developing task-oriented dialog systems.
  • Poison-RAG: Adversarial Data Poisoning Attacks on Retrieval-Augmented Generation in Recommender Systems: Presents Poison-RAG, a framework for adversarial data poisoning attacks targeting RAG-based recommender systems.
  • Systematic Abductive Reasoning via Diverse Relation Representations in Vector-symbolic Architecture: Proposes Rel-SAR, a model for systematic abductive reasoning with diverse relation representations.
  • ALoFTRAG: Automatic Local Fine Tuning for Retrieval Augmented Generation: Introduces ALoFTRAG, a framework for improving RAG accuracy on new data domains without manually labeled data.
  • Leveraging Graph Structures and Large Language Models for End-to-End Synthetic Task-Oriented Dialogues: Introduces GraphTOD, an end-to-end framework for generating task-oriented dialogues.
  • Reference-free Evaluation Metrics for Text Generation: A Survey: Surveys reference-free metrics for natural language generation systems.
  • Evaluating Efficiency and Engagement in Scripted and LLM-Enhanced Human-Robot Interactions: Compares scripted and LLM-enhanced responses in human-robot interactions.
  • Automatic Labelling with Open-source LLMs using Dynamic Label Schema Integration: Explores leveraging open-source models for automatic labelling.
  • Training Dialogue Systems by AI Feedback for Improving Overall Dialogue Impression: Focuses on improving dialogue impressions using reinforcement learning from AI feedback.
  • Generating Diverse Q&A Benchmarks for RAG Evaluation with DataMorgana: Introduces DataMorgana, a tool for generating diverse synthetic Q&A benchmarks for RAG evaluation.
  • Adaptive Retrieval Without Self-Knowledge? Bringing Uncertainty Back Home: Analyzes adaptive retrieval methods and uncertainty estimation techniques in RAG systems.
  • RAG-Reward: Optimizing RAG with Reward Modeling and RLHF: Introduces RAG-Reward, a dataset for optimizing RAG with reward modeling and reinforcement learning with human feedback.
  • K-COMP: Retrieval-Augmented Medical Domain Question Answering With Knowledge-Injected Compressor: Proposes K-COMP, a knowledge-injected compressor for medical domain question answering.
  • Improving Contextual Faithfulness of Large Language Models via Retrieval Heads-Induced Optimization: Introduces RHIO, a framework for improving contextual faithfulness in LLMs.
  • RPO: Retrieval Preference Optimization for Robust Retrieval-Augmented Generation: Introduces RPO, a method for robust RAG by adaptively leveraging multi-source knowledge.
  • A RAG-Based Institutional Assistant: Designs and evaluates a RAG-based virtual assistant for the University of São Paulo.

Sources

Conversational Text Extraction with Large Language Models Using Retrieval-Augmented Systems

Improving Automated Feedback Systems for Tutor Training in Low-Resource Scenarios through Data Augmentation

An LLM-Guided Tutoring System for Social Skills Training

Dialogue Benchmark Generation from Knowledge Graphs with Cost-Effective Retrieval-Augmented LLMs

Passage Segmentation of Documents for Extractive Question Answering

FRAG: A Flexible Modular Framework for Retrieval-Augmented Generation based on Knowledge Graphs

AirRAG: Activating Intrinsic Reasoning for Retrieval Augmented Generation via Tree-based Search

Towards Preventing Overreliance on Task-Oriented Conversational AI Through Accountability Modeling

BoK: Introducing Bag-of-Keywords Loss for Interpretable Dialogue Response Generation

Leveraging Chain of Thought towards Empathetic Spoken Dialogue without Corresponding Question-Answering Data

Generative Retrieval for Book search

A Collection of Question Answering Datasets for Norwegian

Question-to-Question Retrieval for Hallucination-Free Knowledge Access: An Approach for Wikipedia and Wikidata Question Answering

Conversation Routines: A Prompt Engineering Framework for Task-Oriented Dialog Systems

Poison-RAG: Adversarial Data Poisoning Attacks on Retrieval-Augmented Generation in Recommender Systems

Systematic Abductive Reasoning via Diverse Relation Representations in Vector-symbolic Architecture

ALoFTRAG: Automatic Local Fine Tuning for Retrieval Augmented Generation

Leveraging Graph Structures and Large Language Models for End-to-End Synthetic Task-Oriented Dialogues

Reference-free Evaluation Metrics for Text Generation: A Survey

Evaluating Efficiency and Engagement in Scripted and LLM-Enhanced Human-Robot Interactions

Automatic Labelling with Open-source LLMs using Dynamic Label Schema Integration

Training Dialogue Systems by AI Feedback for Improving Overall Dialogue Impression

Generating Diverse Q&A Benchmarks for RAG Evaluation with DataMorgana

Adaptive Retrieval Without Self-Knowledge? Bringing Uncertainty Back Home

RAG-Reward: Optimizing RAG with Reward Modeling and RLHF

K-COMP: Retrieval-Augmented Medical Domain Question Answering With Knowledge-Injected Compressor

Improving Contextual Faithfulness of Large Language Models via Retrieval Heads-Induced Optimization

RPO: Retrieval Preference Optimization for Robust Retrieval-Augmented Generation

A RAG-Based Institutional Assistant

Built with on top of