Enhancing Reasoning and Robustness in Large Language Models

The recent developments in the research area of large language models (LLMs) and their applications have shown a significant shift towards enhancing the models' reasoning capabilities and robustness. The field is moving towards more sophisticated methods of integrating external knowledge, such as through retrieval-augmented generation (RAG) frameworks, which aim to improve the accuracy and relevance of responses in various domains. Innovations in prompt compression and context-driven retrieval are also advancing the efficiency and effectiveness of LLMs, particularly in specialized fields like biomedicine and climate science. Additionally, there is a growing focus on evaluating and improving the consistency and logical integrity of LLM responses, as well as addressing the challenges of hallucination and out-of-distribution generalization. Noteworthy advancements include the development of frameworks that leverage formal logic for syllogistic reasoning in biomedical contexts and the creation of synthetic datasets for training LLMs in critical question generation. These developments collectively push the boundaries of what LLMs can achieve in terms of accuracy, reliability, and domain-specific applicability.

Sources

Style-Compress: An LLM-Based Prompt Compression Framework Considering Task-Specific Styles

Optimizing Retrieval-Augmented Generation with Elasticsearch for Enhanced Question-Answering Systems

Critical Questions Generation: Motivation and Challenges

SylloBio-NLI: Evaluating Large Language Models on Biomedical Syllogistic Reasoning

RAG-ConfusionQA: A Benchmark for Evaluating LLMs on Confusing Questions

From Test-Taking to Test-Making: Examining LLM Authoring of Commonsense Assessment Items

Are LLMs Good Zero-Shot Fallacy Classifiers?

Toward Robust RALMs: Revealing the Impact of Imperfect Retrieval on Retrieval-Augmented Language Models

BRIEF: Bridging Retrieval and Inference for Multi-hop Reasoning via Compression

ConTReGen: Context-driven Tree-structured Retrieval for Open-domain Long-form Text Generation

MedLogic-AQA: Enhancing Medical Question Answering with Abstractive Models Focusing on Logical Structures

Evaluating Consistencies in LLM responses through a Semantic Clustering of Question Answering

Do RAG Systems Cover What Matters? Evaluating and Optimizing Responses with Sub-Question Coverage

Reverse Question Answering: Can an LLM Write a Question so Hard (or Bad) that it Can't Answer?

"What is the value of {templates}?" Rethinking Document Information Extraction Datasets for LLMs

Leveraging Retrieval-Augmented Generation for Culturally Inclusive Hakka Chatbots: Design Insights and User Perceptions

RAG4ITOps: A Supervised Fine-Tunable and Comprehensive RAG Framework for IT Operations and Maintenance

Rulebreakers Challenge: Revealing a Blind Spot in Large Language Models' Reasoning with Formal Logic

No more hard prompts: SoftSRV prompting for synthetic data generation

ClimaQA: An Automated Evaluation Framework for Climate Foundation Models

Forewarned is Forearmed: Leveraging LLMs for Data Synthesis through Failure-Inducing Exploration

CLR-Bench: Evaluating Large Language Models in College-level Reasoning

An Adaptive Framework for Generating Systematic Explanatory Answer in Online Q&A Platforms

Leveraging the Domain Adaptation of Retrieval Augmented Generation Models for Question Answering and Reducing Hallucination

SimRAG: Self-Improving Retrieval-Augmented Generation for Adapting Large Language Models to Specialized Domains

LongRAG: A Dual-Perspective Retrieval-Augmented Generation Paradigm for Long-Context Question Answering

Aggregated Knowledge Model: Enhancing Domain-Specific QA with Fine-Tuned and Retrieval-Augmented Generation Models

From Blind Solvers to Logical Thinkers: Benchmarking LLMs' Logical Integrity on Faulty Mathematical Problems

Built with on top of