ExplorerArtificial IntelligenceAI
Research PaperResearchia:202602.13078

AttentionRetriever: Attention Layers are Secretly Long Document Retrievers

David Jiahao Fu

Abstract

Retrieval augmented generation (RAG) has been widely adopted to help Large Language Models (LLMs) to process tasks involving long documents. However, existing retrieval models are not designed for long document retrieval and fail to address several key challenges of long document retrieval, including context-awareness, causal dependence, and scope of retrieval. In this paper, we proposed AttentionRetriever, a novel long document retrieval model that leverages attention mechanism and entity-based...

Submitted: February 13, 2026Subjects: AI; Artificial Intelligence

Description / Details

Retrieval augmented generation (RAG) has been widely adopted to help Large Language Models (LLMs) to process tasks involving long documents. However, existing retrieval models are not designed for long document retrieval and fail to address several key challenges of long document retrieval, including context-awareness, causal dependence, and scope of retrieval. In this paper, we proposed AttentionRetriever, a novel long document retrieval model that leverages attention mechanism and entity-based retrieval to build context-aware embeddings for long document and determine the scope of retrieval. With extensive experiments, we found AttentionRetriever is able to outperform existing retrieval models on long document retrieval datasets by a large margin while remaining as efficient as dense retrieval models.


Source: arXiv:2602.12278v1 - http://arxiv.org/abs/2602.12278v1 PDF: https://arxiv.org/pdf/2602.12278v1 Original Link: http://arxiv.org/abs/2602.12278v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Feb 13, 2026
Topic:
Artificial Intelligence
Area:
AI
Comments:
0
Bookmark