Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | 
         @@ -199,9 +199,10 @@ We recommend fine-tuning LFM2 models on your use cases to maximize performance. 
     | 
|
| 199 | 
         | 
| 200 | 
         
             
            | Notebook | Description | Link |
         
     | 
| 201 | 
         
             
            |-------|------|------|
         
     | 
| 202 | 
         
            -
            | SFT ( 
     | 
| 203 | 
         
            -
            | SFT ( 
     | 
| 204 | 
         
            -
            |  
     | 
| 
         | 
|
| 205 | 
         | 
| 206 | 
         
             
            ## 📈 Performance
         
     | 
| 207 | 
         | 
| 
         | 
|
| 199 | 
         | 
| 200 | 
         
             
            | Notebook | Description | Link |
         
     | 
| 201 | 
         
             
            |-------|------|------|
         
     | 
| 202 | 
         
            +
            | SFT (Unsloth) | Supervised Fine-Tuning (SFT) notebook with a LoRA adapter using Unsloth. | <a href="https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Liquid_LFM2_(1.2B)-Conversational.ipynb"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
         
     | 
| 203 | 
         
            +
            | SFT (Axolotl) | Supervised Fine-Tuning (SFT) notebook with a LoRA adapter using Axolotl. | <a href="https://colab.research.google.com/drive/155lr5-uYsOJmZfO6_QZPjbs8hA_v8S7t?usp=sharing"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
         
     | 
| 204 | 
         
            +
            | SFT (TRL) | Supervised Fine-Tuning (SFT) notebook with a LoRA adapter using TRL. | <a href="https://colab.research.google.com/drive/1j5Hk_SyBb2soUsuhU0eIEA9GwLNRnElF?usp=sharing"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
         
     | 
| 205 | 
         
            +
            | DPO (TRL) | Preference alignment with Direct Preference Optimization (DPO) using TRL. | <a href="https://colab.research.google.com/drive/1MQdsPxFHeZweGsNx4RH7Ia8lG8PiGE1t?usp=sharing"><img src="https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/vlOyMEjwHa_b_LXysEu2E.png" width="110" alt="Colab link"></a> |
         
     | 
| 206 | 
         | 
| 207 | 
         
             
            ## 📈 Performance
         
     | 
| 208 | 
         |