(논문 요약) Thinking to Recall: How Reasoning Unlocks Parametric Knowledge in LLMs (Paper)

핵심 내용

  • Reasoning Unlocks Latent Knowledge (reasoning on 성능 > reasoning off 성능)
  • Less Capable Models Benefit More from Reasoning.
  • Question Complexity is a Poor Predictor of Reasoning Effectiveness.
  • How Reasoning Improves Parametric Recall?
    • Reasoning Tokens as a Computational Buffer: generating extra tokens during reasoning allows models to perform additional latent operations and to bypass the depth limits of a single forward pass on the input.
    • Factual Priming
  • Hallucinating intermediate facts during reasoning increases the likelihood of hallucinations in the final answer.
  • Directly improve model accuracy by prioritizing reasoning trajectories that contain hallucination-free factual statements.