Retrieval-augmented reasoning with lean language models
Abstract
A retrieval-augmented conversational agent using a lightweight model achieves high accuracy and consistency in domain-specific queries through fine-tuning and document compression.
This technical report details a novel approach to combining reasoning and retrieval augmented generation (RAG) within a single, lean language model architecture. While existing RAG systems typically rely on large-scale models and external APIs, our work addresses the increasing demand for performant and privacy-preserving solutions deployable in resource-constrained or secure environments. Building on recent developments in test-time scaling and small-scale reasoning models, we develop a retrieval augmented conversational agent capable of interpreting complex, domain-specific queries using a lightweight backbone model. Our system integrates a dense retriever with fine-tuned Qwen2.5-Instruct models, using synthetic query generation and reasoning traces derived from frontier models (e.g., DeepSeek-R1) over a curated corpus, in this case, the NHS A-to-Z condition pages. We explore the impact of summarisation-based document compression, synthetic data design, and reasoning-aware fine-tuning on model performance. Evaluation against both non-reasoning and general-purpose lean models demonstrates that our domain-specific fine-tuning approach yields substantial gains in answer accuracy and consistency, approaching frontier-level performance while remaining feasible for local deployment. All implementation details and code are publicly released to support reproducibility and adaptation across domains.
Community
This report explores combining reasoning and retrieval augmented generation (RAG) into a single, compact language model. The proposed system uses a lightweight backbone model to achieve strong performance on complex, domain-specific questions, making it suitable for deployment in environments with limited resources or strict privacy requirements.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Question Decomposition for Retrieval-Augmented Generation (2025)
- Inference Scaled GraphRAG: Improving Multi Hop Question Answering on Knowledge Graphs (2025)
- Leveraging LLM-Assisted Query Understanding for Live Retrieval-Augmented Generation (2025)
- SemRAG: Semantic Knowledge-Augmented RAG for Improved Question-Answering (2025)
- DIVER: A Multi-Stage Approach for Reasoning-intensive Information Retrieval (2025)
- LLM-Assisted Question-Answering on Technical Documents Using Structured Data-Aware Retrieval Augmented Generation (2025)
- KeyKnowledgeRAG (K^2RAG): An Enhanced RAG method for improved LLM question-answering capabilities (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper