Skip to content
This repository has been archived by the owner on Nov 29, 2024. It is now read-only.

A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and fully reproducible.

Notifications You must be signed in to change notification settings

PaulPauls/llama3_interpretability_sae

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 

Repository files navigation

Llama 3 Interpretability with Sparse Autoencoders

This project is currently taken down. My apologies.

About

A complete end-to-end pipeline for LLM interpretability with sparse autoencoders (SAEs) using Llama 3.2, written in pure PyTorch and fully reproducible.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published