Advances in Large Language Models

The field of large language models (LLMs) is rapidly evolving, with a focus on improving their reasoning capabilities and performance in complex tasks. Recent developments have seen the introduction of novel methodologies and frameworks that enhance the abilities of LLMs, such as the use of external reformulation and retrieval-augmented generation. These advancements have led to significant improvements in the performance of LLMs, allowing them to achieve state-of-the-art results in various benchmarks and datasets. Notably, the integration of LLMs with other techniques, such as Monte Carlo Tree Search and reinforcement learning, has shown great promise in enhancing their decision-making and strategic planning abilities. Overall, the field is moving towards more sophisticated and powerful LLMs that can effectively tackle complex tasks and provide more accurate and informative responses. Noteworthy papers in this area include:

  • Open Deep Search, which introduces a novel framework for augmenting LLMs with search and reasoning capabilities, achieving state-of-the-art performance on several benchmarks.
  • MCTS-RAG, which enhances the reasoning capabilities of small language models using retrieval-augmented generation and Monte Carlo Tree Search, enabling them to achieve performance comparable to frontier LLMs.

Sources

Haunted House: A text-based game for comparing the flexibility of mental models in humans and LLMs

Solving Situation Puzzles with Large Language Model and External Reformulation

Open Deep Search: Democratizing Search with Open-source Reasoning Agents

MCTS-RAG: Enhancing Retrieval-Augmented Generation with Monte Carlo Tree Search

LLM-Gomoku: A Large Language Model-Based System for Strategic Gomoku with Self-Play and Reinforcement Learning

ReaRAG: Knowledge-guided Reasoning Enhances Factuality of Large Reasoning Models with Iterative Retrieval Augmented Generation

Built with on top of