Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add new paper: #48

Open
wyzh0912 opened this issue Feb 23, 2025 · 0 comments
Open

Add new paper: #48

wyzh0912 opened this issue Feb 23, 2025 · 0 comments

Comments

@wyzh0912
Copy link
Contributor

Title

On Mechanistic Circuits for Extractive Question-Answering

Published Date

2025-02-12

Source

arXiv

Head Name

Attribution Head

Summary

  • Innovation: The paper introduces a method to extract mechanistic circuits from language models for extractive QA tasks using causal mediation analysis. It demonstrates how these circuits can elucidate the interplay between parametric memory and context usage, leading to practical applications like data attribution and model steering.

  • Tasks: The study involves designing a probe dataset and applying causal mediation analysis to identify circuits responsible for context and memory faithfulness in language models. The insights from these circuits are then used to develop ATTNATTRIB, a fast data attribution algorithm, and to steer models towards improved context faithfulness in QA tasks.

  • Significant Result: The research finds that a small set of attention heads within the identified circuits performs reliable data attribution by default, allowing for state-of-the-art attribution results in extractive QA benchmarks. Additionally, using these attribution insights, the model can be steered to improve its reliance on context over parametric memory, enhancing context faithfulness by up to 9% in QA datasets.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant