273 1 year ago

Model finetuned starting with llama 3.1 8b using a full precision LoRA he20, rank 64, alpha 16

tools

Models

View all →

Readme

Model Specifications

Base Model

  • Architecture: Llama 3.1
  • Size: 8B parameters
  • Type: Instruct model
  • Precision: FP16

Finetuning Parameters

  • Method: Full Precision LoRA
  • Epoch: 20
  • Rank: 64
  • Alpha: 16
  • Training Dataset: train_05_prompted_v2

Performance Metrics

Metric Score Base Model Score
Agentic Similarity 86 84.67
CoT Contextual Accuracy 56 / 3 54 / 5
Medical GPT Score 65 51.75

Benchmark Results

This model demonstrates superior performance compared to previous Llama 3.1 8B implementations across all measured metrics.

Loss Function:

training_loss.png