Transformers
English
Inference Endpoints
norabelrose commited on
Commit
e6bcdbf
1 Parent(s): b07ca63

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -0
README.md ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - togethercomputer/RedPajama-Data-V2
5
+ language:
6
+ - en
7
+ library_name: transformers
8
+ ---
9
+
10
+ This is a set of sparse autoencoders (SAEs) trained on [Llama 3.1 8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) using the 10B sample of the [RedPajama v2 corpus](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-V2), which comes out to roughly 8.5B tokens using the Llama 3 tokenizer. The SAEs are organized by hookpoint, and can be loaded using the EleutherAI [`sae` library](https://github.com/EleutherAI/sae).
11
+
12
+ Unlike [EleutherAI/sae-llama-3.1-8b-32x](https://huggingface.co/EleutherAI/sae-llama-3.1-8b-32x), these SAEs were trained with the MultiTopK loss, which allows them to be used at varying sparsity levels at inference time. For more information, see OpenAI's description of the loss in [this paper](https://cdn.openai.com/papers/sparse-autoencoders.pdf).
13
+
14
+ With the `sae` library installed, you can access an SAE like this:
15
+ ```python
16
+ from sae import Sae
17
+
18
+ sae = Sae.load_from_hub("EleutherAI/sae-llama-3.1-8b-64x", hookpoint="layers.23.mlp")
19
+ ```