Add MoE tag to Mixtral (#29)
Browse files- Add MoE tag to Mixtral (560530cd80ab72072ea739d0d6093d847ce21f54)
Co-authored-by: Clémentine Fourrier <clefourrier@users.noreply.huggingface.co>
    	
        README.md
    CHANGED
    
    | 
         @@ -6,6 +6,8 @@ language: 
     | 
|
| 6 | 
         
             
            - de
         
     | 
| 7 | 
         
             
            - es
         
     | 
| 8 | 
         
             
            - en
         
     | 
| 
         | 
|
| 
         | 
|
| 9 | 
         
             
            ---
         
     | 
| 10 | 
         
             
            # Model Card for Mixtral-8x7B
         
     | 
| 11 | 
         
             
            The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mistral-8x7B outperforms Llama 2 70B on most benchmarks we tested.
         
     | 
| 
         | 
|
| 6 | 
         
             
            - de
         
     | 
| 7 | 
         
             
            - es
         
     | 
| 8 | 
         
             
            - en
         
     | 
| 9 | 
         
            +
            tags:
         
     | 
| 10 | 
         
            +
            - moe
         
     | 
| 11 | 
         
             
            ---
         
     | 
| 12 | 
         
             
            # Model Card for Mixtral-8x7B
         
     | 
| 13 | 
         
             
            The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mistral-8x7B outperforms Llama 2 70B on most benchmarks we tested.
         
     |