YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Update: This model was fine-tuned with LoRA on a 170k dataset. Compared to version 0.4, date information in the dataset was masked, and low-quality data was further filtered out.

Metrics:

  • rouge: ~0.48

Dataset:

  • 170k data

Training Configuration:

  • precision: bf16
  • max sequence length: 2048
  • effective batch size: 16 (per-device batch size 2 × gradient accumulation 8)
  • learning rate: 1 × 10⁻⁴
  • epochs: 3
  • scheduler: cosine
  • warmup steps: 760
  • optimizer: AdamW
  • max gradient norm: 1.0
  • save checkpoint every 5 000 steps
  • output directory: saves/Llama-3.1-8B-Instruct/lora/train_2025-06-09-13-48-28

LoRA Hyperparameters:

  • rank: 8
  • alpha: 32
  • dropout: 0.05
  • target modules: all
  • finetuning type: LoRA

Downloads last month
1
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support