Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | 
         @@ -15,7 +15,20 @@ See our [landing page](https://huggingface.co/google/gemma-scope) for details on 
     | 
|
| 15 | 
         
             
            - `2b-pt-`: These SAEs were trained on Gemma v2 2B base model.
         
     | 
| 16 | 
         
             
            - `att`: These SAEs were trained on the attention layer outputs, before the final linear projection.
         
     | 
| 17 | 
         | 
| 18 | 
         
            -
             
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 19 | 
         | 
| 20 | 
         
             
            Point of contact: Arthur Conmy
         
     | 
| 21 | 
         | 
| 
         @@ -26,4 +39,8 @@ Contact by email: 
     | 
|
| 26 | 
         
             
            ```
         
     | 
| 27 | 
         | 
| 28 | 
         
             
            HuggingFace account:
         
     | 
| 29 | 
         
            -
            https://huggingface.co/ArthurConmyGDM
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 15 | 
         
             
            - `2b-pt-`: These SAEs were trained on Gemma v2 2B base model.
         
     | 
| 16 | 
         
             
            - `att`: These SAEs were trained on the attention layer outputs, before the final linear projection.
         
     | 
| 17 | 
         | 
| 18 | 
         
            +
            # 3. How can I use these SAEs straight away?
         
     | 
| 19 | 
         
            +
             
     | 
| 20 | 
         
            +
            ```python
         
     | 
| 21 | 
         
            +
            from sae_lens import SAE  # pip install sae-lens
         
     | 
| 22 | 
         
            +
             
     | 
| 23 | 
         
            +
            sae, cfg_dict, sparsity = SAE.from_pretrained(
         
     | 
| 24 | 
         
            +
                release = "gemma-scope-2b-pt-att-canonical",
         
     | 
| 25 | 
         
            +
                sae_id = "layer_0/width_16k/canonical",
         
     | 
| 26 | 
         
            +
            )
         
     | 
| 27 | 
         
            +
            ```
         
     | 
| 28 | 
         
            +
             
     | 
| 29 | 
         
            +
            See https://github.com/jbloomAus/SAELens for details on this library.
         
     | 
| 30 | 
         
            +
             
     | 
| 31 | 
         
            +
            # 4. Point of Contact
         
     | 
| 32 | 
         | 
| 33 | 
         
             
            Point of contact: Arthur Conmy
         
     | 
| 34 | 
         | 
| 
         | 
|
| 39 | 
         
             
            ```
         
     | 
| 40 | 
         | 
| 41 | 
         
             
            HuggingFace account:
         
     | 
| 42 | 
         
            +
            https://huggingface.co/ArthurConmyGDM
         
     | 
| 43 | 
         
            +
             
     | 
| 44 | 
         
            +
            # 5. Citation
         
     | 
| 45 | 
         
            +
             
     | 
| 46 | 
         
            +
            Paper: https://arxiv.org/abs/2408.05147
         
     |