End of training
Browse files
    	
        README.md
    CHANGED
    
    | @@ -1,10 +1,11 @@ | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            base_model: meta-llama/Meta-Llama-3.1-8B
         | 
| 3 | 
             
            datasets:
         | 
| 4 | 
            -
            -  | 
| 5 | 
             
            library_name: peft
         | 
| 6 | 
             
            license: llama3.1
         | 
| 7 | 
             
            tags:
         | 
|  | |
| 8 | 
             
            - trl
         | 
| 9 | 
             
            - sft
         | 
| 10 | 
             
            - generated_from_trainer
         | 
| @@ -18,7 +19,7 @@ should probably proofread and complete it, then remove this comment. --> | |
| 18 |  | 
| 19 | 
             
            # llama3.1-8b-classification-gpt4o-100k
         | 
| 20 |  | 
| 21 | 
            -
            This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the  | 
| 22 | 
             
            It achieves the following results on the evaluation set:
         | 
| 23 | 
             
            - Loss: 3.0330
         | 
| 24 |  | 
|  | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            base_model: meta-llama/Meta-Llama-3.1-8B
         | 
| 3 | 
             
            datasets:
         | 
| 4 | 
            +
            - llama-duo/synth_classification_dataset_dedup
         | 
| 5 | 
             
            library_name: peft
         | 
| 6 | 
             
            license: llama3.1
         | 
| 7 | 
             
            tags:
         | 
| 8 | 
            +
            - alignment-handbook
         | 
| 9 | 
             
            - trl
         | 
| 10 | 
             
            - sft
         | 
| 11 | 
             
            - generated_from_trainer
         | 
|  | |
| 19 |  | 
| 20 | 
             
            # llama3.1-8b-classification-gpt4o-100k
         | 
| 21 |  | 
| 22 | 
            +
            This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the llama-duo/synth_classification_dataset_dedup dataset.
         | 
| 23 | 
             
            It achieves the following results on the evaluation set:
         | 
| 24 | 
             
            - Loss: 3.0330
         | 
| 25 |  | 
    	
        all_results.json
    CHANGED
    
    | @@ -1,5 +1,10 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
             
                "epoch": 10.0,
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
| 3 | 
             
                "total_flos": 4.416382035459834e+18,
         | 
| 4 | 
             
                "train_loss": 0.922980490487975,
         | 
| 5 | 
             
                "train_runtime": 12382.7598,
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
             
                "epoch": 10.0,
         | 
| 3 | 
            +
                "eval_loss": 3.0329792499542236,
         | 
| 4 | 
            +
                "eval_runtime": 0.4341,
         | 
| 5 | 
            +
                "eval_samples": 16,
         | 
| 6 | 
            +
                "eval_samples_per_second": 4.607,
         | 
| 7 | 
            +
                "eval_steps_per_second": 2.303,
         | 
| 8 | 
             
                "total_flos": 4.416382035459834e+18,
         | 
| 9 | 
             
                "train_loss": 0.922980490487975,
         | 
| 10 | 
             
                "train_runtime": 12382.7598,
         | 
    	
        eval_results.json
    CHANGED
    
    | @@ -1,8 +1,8 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
            -
                "epoch":  | 
| 3 | 
            -
                "eval_loss":  | 
| 4 | 
            -
                "eval_runtime": 0. | 
| 5 | 
             
                "eval_samples": 16,
         | 
| 6 | 
            -
                "eval_samples_per_second":  | 
| 7 | 
            -
                "eval_steps_per_second": 2. | 
| 8 | 
             
            }
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
            +
                "epoch": 10.0,
         | 
| 3 | 
            +
                "eval_loss": 3.0329792499542236,
         | 
| 4 | 
            +
                "eval_runtime": 0.4341,
         | 
| 5 | 
             
                "eval_samples": 16,
         | 
| 6 | 
            +
                "eval_samples_per_second": 4.607,
         | 
| 7 | 
            +
                "eval_steps_per_second": 2.303
         | 
| 8 | 
             
            }
         | 
    	
        runs/Aug07_09-32-26_deep-diver-main-weak-ape-1-0-0/events.out.tfevents.1723050187.deep-diver-main-weak-ape-1-0-0.531.1
    ADDED
    
    | @@ -0,0 +1,3 @@ | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:ae9bb2c7ad7d9ed7fdee930ed30ad79a06eab8961667d1b6c96f2278219ff488
         | 
| 3 | 
            +
            size 359
         | 

