TopK Transcoder Based on Llama 3.1 8B Instruct

This repository provides the TopK transcoder checkpoints used in the paper “Verifying Chain-of-Thought Reasoning via Its Computational Graph”.
The model is based on Llama 3.1 8B Instruct and trained with the TopK transcoder method described in the paper.

Installation

To run the model, you need the Circuit Tracer library.
It can be installed from the project page:

https://github.com/zsquaredz/circuit-tracer

Note that this is a fork of the original library as they don't yet support TopK transcoder.

After installing the library, you can load and run the transcoder as shown below.

Minimal Usage Example

from circuit_tracer import ReplacementModel
import torch

# Load transcoders into a ReplacementModel
model = ReplacementModel.from_pretrained("meta-llama/Llama-3.1-8B-Instruct", "facebook/crv-8b-instruct-transcoders", dtype=torch.bfloat16)

Once you have loaded the model, you can perform attribution or intervention as shown in this demo.

Citation

If you use this model, please cite our paper:

@article{zhao2025verifying,
      title={Verifying Chain-of-Thought Reasoning via Its Computational Graph},
      author={Zheng Zhao and Yeskendir Koishekenov and Xianjun Yang and Naila Murray and Nicola Cancedda},
      year={2025},
      eprint={2510.09312},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2510.09312},
}
Downloads last month
10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for facebook/crv-8b-instruct-transcoders

Finetuned
(2010)
this model

Dataset used to train facebook/crv-8b-instruct-transcoders