dataset_info:
features:
- name: relation
dtype: string
- name: relation_prefix
dtype: string
- name: relation_suffix
dtype: string
- name: prompt
dtype: string
- name: relation_id
dtype: string
- name: target_false_id
dtype: string
- name: target_true_id
dtype: string
- name: target_true
dtype: string
- name: target_false
dtype: string
- name: subject
dtype: string
splits:
- name: train
num_bytes: 3400668
num_examples: 21919
download_size: 1109314
dataset_size: 3400668
Dataset Card for "counterfact-tracing"
This is adapted from the counterfact dataset from the excellent ROME paper from David Bau and Kevin Meng.
This is a dataset of 21919 factual relations, formatted as data["prompt"]==f"{data['relation_prefix']}{data['subject']}{data['relation_suffix']}"
. Each has two responses data["target_true"]
and data["target_false"]
which is intended to go immediately after the prompt.
The dataset was originally designed for memory editing in models. I made this for a research project doing mechanistic interpretability of how models recall factual knowledge, building on their causal tracing technique, and so stripped their data down to the information relevant to causal tracing. I also prepended spaces where relevant so that the subject and targets can be properly tokenized as is (spaces are always prepended to targets, and are prepended to subjects unless the subject is at the start of a sentence).
Each fact has both a true and false target. I recommend measuring the logit difference between the true and false target (at least, if it's a single token target!), so as to control for eg the parts of the model which identify that it's supposed to be giving a fact of this type at all. (Idea inspired by the excellent Interpretability In the Wild paper).