ACL 2026 Main Conference

Retrieval as Generation: A Unified Framework with Self-Triggered Information Planning

Bo Li1, Mingda Wang2, Gexiang Fang1, Shikun Zhang1, Wei Ye1
1National Engineering Research Center for Software Engineering, Peking University, 2Hebei University of Technology
GRIP rethinks RAG by making retrieval control part of generation itself. Instead of relying on external controllers or heuristic triggers, the model emits token-level actions to decide when to retrieve, how to reformulate follow-up queries, and when to stop. It further learns multi-step retrieval behavior through one-step decision optimization, avoiding heavy long-horizon search-policy optimization while preserving adaptive depth and controllable stopping.
[RETRIEVE] [INTERMEDIARY] [ANSWER] [SOLVED]

Two Representative GRIP Cases

Explore two GRIP trajectories: relation chaining and semantic reinterpretation with query reformulation.

English Case · Relation Chaining
Question
Who founded the company that Robinsons-May is a subsidiary of?

Current State

Only a related company name is recalled.

Current Action

Round 1 of a 3-round retrieval budget.

External Knowledge

Robinsons-May is a subsidiary of The May Department Stores Company.

What changed in this round?

The parent company is identified, but the founder is still missing.

Final Answer
Waiting for final answer

Why GRIP

Most RAG pipelines still treat retrieval as an external intervention: a separate controller decides whether to retrieve, the retriever runs, and the generator consumes fixed evidence. GRIP instead moves retrieval control into the autoregressive trajectory itself, so information planning, query reformulation, and termination become explicit, trainable actions represented in the same decoding space as language generation.

01

Retrieval should be part of decoding

GRIP internalizes retrieval timing into generation rather than delegating it to an external controller.

02

Control tokens make behavior explicit

When to retrieve, what to ask next, and when to stop are all represented as language-native actions.

03

Structured supervision teaches planning

Four answerability types expose the model to direct answering, retrieval triggering, multi-hop planning, and answer completion.

04

One-step decision optimization

GRIP learns multi-step retrieval behavior through one-step decision optimization rather than long-horizon search-policy optimization, making it simpler and more stable while preserving adaptive depth and controllable stopping.

Method

GRIP combines a unified token-control interface with Self-Triggered Information Planning. The full model figure below shows the complete training-and-inference pipeline: structured SFT and RL in the upper half, and a token-driven retrieval planning loop in the lower half, all within a single autoregressive trajectory.

GRIP model figure
Control Tokens

Token-level retrieval interface

[RETRIEVE], [INTERMEDIARY], [ANSWER], and [SOLVED] form a minimal but expressive action space for retrieval planning.

Planning Loop

Self-triggered planning

The model decides whether current knowledge is sufficient, emits a retrieval action if not, integrates evidence, and repeats until resolved.

Structured Data

Four answerability types

The supervision signal is scenario-typed, so the same model learns direct answering, retrieval triggering, multi-hop planning, and answer completion.

Optimization

SFT + rule-based RL

SFT learns token-controlled behaviors, and DAPO-based RL further improves answer fidelity and behavior control while reducing redundant retrievals.

Four Structured Supervision Types

GRIP uses four structured data types aligned with distinct token trajectories, allowing the model to learn when to answer, when to retrieve, and when to continue planning. Instead of a single generic supervision pattern, each type teaches a different decision boundary in the token-controlled retrieval loop.

GRIP four supervision types
α

Type-α · Direct Answer

Queries that can be answered directly from internal knowledge. The model emits [ANSWER] and ends with [SOLVED].

β

Type-β · Retrieval Needed

The model has partial knowledge but lacks enough information to finalize the answer. It emits [INTERMEDIARY] and then [RETRIEVE].

γ

Type-γ · Multi-hop Planning

Complex cases requiring iterative evidence gathering. The model produces a refined partial state and then issues a new follow-up query.

θ

Type-θ · Answer Completion

Retrieved passages contain the needed evidence, but the model must still synthesize and finalize the answer before stopping.

Main Results

The table below reproduces the paper’s main result table with the full metric set: EM, ROUGE, F1, and Avg.Score. GRIP is the strongest open-source system in the main comparison and remains competitive with GPT-4o using a much smaller backbone.

Table 1 in the paper reports EM, ROUGE (average of ROUGE-1/2/L), F1, and Avg.Score on five QA benchmarks.
Method HotpotQA PopQA NQ WebQ TriviaQA Avg.
Score
EMROUGEF1 EMROUGEF1 EMROUGEF1 EMROUGEF1 EMROUGEF1
Training-free Method
Instruct17.221.625.917.419.623.214.423.820.314.825.129.446.245.355.126.6
GPT-3.5 Turbo26.232.438.229.131.135.320.734.330.015.927.431.955.657.869.735.7
GPT-4o33.240.247.030.632.339.926.542.728.323.532.337.065.764.378.241.4
Single RAG26.131.637.222.826.631.119.328.424.814.022.726.646.447.056.830.8
FLARE23.227.932.814.316.018.414.722.421.324.230.334.748.648.556.428.9
DRAGIN27.932.638.715.516.819.823.932.828.525.231.535.755.353.664.633.5
ETC32.537.744.230.532.537.520.926.730.718.926.630.452.952.163.035.8
Training-based Method
SFT-RAG20.324.128.629.425.230.420.817.921.318.918.323.150.124.857.227.4
Self-RAG19.623.826.718.122.322.815.722.424.016.426.527.450.247.357.528.0
INFO-RAG19.923.726.918.322.623.017.222.924.918.127.728.950.847.858.128.7
RobustRAG27.631.837.529.727.732.426.425.129.221.525.029.148.847.757.933.2
GainRAG31.435.641.830.133.338.122.927.932.216.524.528.950.349.159.234.8
R1-Searcher26.029.134.941.635.241.325.824.928.721.826.130.656.053.364.936.0
RetRobust29.634.940.934.135.140.424.229.033.821.827.431.753.650.961.936.6
InstructRAG31.236.842.333.135.740.329.529.533.619.526.831.451.352.262.537.0
GRIP (ours)33.037.644.138.637.538.432.135.832.031.439.334.657.955.967.441.0
w/o RL31.636.643.038.137.137.632.636.132.732.039.935.157.055.266.840.7
GRIP reaches an Avg.Score of 41.0, outperforming all open-source baselines in the main table and approaching GPT-4o while using a substantially smaller backbone.

Behavior Analysis

Before turning to deeper control analyses, the paper first shows that GRIP learns task-aware retrieval depth and can improve retrieval quality by generating follow-up queries conditioned on evolving intermediate states.

Adaptive retrieval across tasks

GRIP retrieves more on HotpotQA and PopQA, fewer on NQ, and much less than R1-Searcher on average. RL reduces redundant retrieval while preserving the same task-aware pattern.

Table 4 in the paper reports mean retrieval count per dataset: GRIP averages 1.24 calls, versus 5.12 for R1-Searcher and 1.60 for GRIP without RL.

Improving retrieval quality with new queries

GRIP-generated follow-up queries substantially improve gold-answer coverage in top-1 and top-3 retrieved passages on both NQ and WebQ.

The larger gains on top-1 than top-3 suggest that GRIP improves evidence ranking, not just recall.

Controllable Retrieval Budget and Depth Extrapolation

This is one of the key analysis sections in the paper. GRIP does not blindly consume the available retrieval budget: even when the maximum budget B increases, realized retrieval count stays far below the cap, while performance improves modestly. The distribution plots below show selective extrapolation beyond the training depth of three retrieval steps.

Retrieval-count distributions under budget control

Retrieval-count distributions under budget control

Figure 14 shows that as the inference-time budget increases from B=3 to B=10, GRIP shifts some mass beyond r>3 while keeping most probability concentrated on small retrieval counts. This indicates selective and task-aware extrapolation rather than mechanical budget consumption.

Effect of maximum retrieval budget B

Max BHotpotPopQANQWebQTriviaAvg.CountAvg.Score
31.441.580.761.151.251.2441.0
51.561.731.011.451.311.4141.2
71.661.791.121.651.351.5141.5
101.741.861.221.901.391.6241.8
As B increases from 3 to 10, Avg.Score rises from 41.0 to 41.8 while Avg.Count grows only from 1.24 to 1.62.

Behavior shift after RL

Appendix G retrieval count shift after RL

Appendix G shows how RL shifts the retrieval-count distribution on NQ and WebQA. Compared with GRIP without RL, the final model places more mass on Retrieve = 0 for cases that can be answered directly, while also increasing the proportion of Retrieve = 1 where one retrieval step is sufficient. This supports the paper’s claim that RL improves behavior control rather than merely changing answer style.

General Capability Preservation

Another critical analysis in the paper asks whether retrieval-planning fine-tuning damages non-RAG abilities. GRIP shows only minimal degradation on closed-form benchmarks and remains competitive with the base instruct model on open-ended summarization judged by GPT-4o.

Closed-form non-RAG evaluation

ModelMMLU (Acc)MBPP (Pass@1)
Instruct66.5654.6
SFT-RAG62.9347.8
GRIP65.7353.8
Compared with the base instruct model, GRIP drops only 0.83 points on MMLU and 0.8 on MBPP, much less than the matched SFT-RAG baseline.

Pairwise GPT-4o evaluation on CNN/DailyMail

Pairwise comparison (GRIP vs.)WinEqualLoss
Instruct51.037.511.5
SFT-RAG90.58.51.0
R1-Searcher59.532.58.0
GRIP remains competitive with the base instruct model and is preferred much more often than SFT-RAG or R1-Searcher, suggesting that retrieval control does not come at the cost of general generation quality.

Resources

Reproducibility at a glance

T

Structured supervision types

GRIP training is organized around Type-α: direct answer, Type-β: retrieval needed, Type-γ: multi-hop planning, and Type-θ: answer completion.

D

Data construction

data_generation/first.sh, make_first_steps.py, use_gpt_for_data.py, merge_dataset.py, and index.py support data construction and BM25 indexing.

I

Inference

inference/agent.py provides the multi-step GRIP inference logic, and inference/inference.sh is the example launch script.

P

Preprocess & training

train/examples/data_preprocess/grip/sft.py, rl.py, run_sft_llama.sh, and dapo_4w_continue_rl_ep3_llama.sh cover the two-stage optimization pipeline.

E

Evaluation

eval/eval.py and eval/utils.py compute EM, F1, ROUGE, and related outputs under a unified protocol.

W

Released data and weights

The repository also links released structured datasets and trained checkpoints, making it possible to reproduce GRIP training and evaluate the final model directly.

Wikipedia / QA data → Type-α / β / γ / θ structured supervision → SFT parquet + RL parquet → supervised fine-tuning → DAPO-based RL → merged checkpoint → multi-step inference → benchmark evaluation
Citation

BibTeX

If you find this work useful, please cite the paper below.

@misc{li2026retrievalgenerationunifiedframework,
      title={Retrieval as Generation: A Unified Framework with Self-Triggered Information Planning}, 
      author={Bo Li and Mingda Wang and Gexiang Fang and Shikun Zhang and Wei Ye},
      year={2026},
      eprint={2604.11407},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2604.11407}, 
}