Uploaded model

  • Developed by: student-abdullah
  • License: apache-2.0
  • Finetuned from model: Meta/Meta-Llama-3.1-8B

Acknowledgement


Model Description

This model is fine-tuned from the Meta/Meta-Llama-3.1-8B base model to enhance its capabilities in generating relevant and accurate responses related to generic medications under the PMBJP scheme. The fine-tuning process included the following hyperparameters:

  • Max Tokens: 512
  • LoRA Alpha: 12
  • LoRA Rank (r): 128
  • Gradient Accumulation Steps: 32
  • Batch Size: 2
  • Qunatization: 8 bits

Model Quantitative Performace

  • Training Quantitative Loss: 0.262 (at final 160th epoch)

Limitations

  • Token Limitations: With a max token limit of 512, the model might not handle very long queries or contexts effectively.
  • Training Data Limitations: The model’s performance is contingent on the quality and coverage of the fine-tuning dataset, which may affect its generalizability to different contexts or medications not covered in the dataset.
  • Potential Biases: As with any model fine-tuned on specific data, there may be biases based on the dataset used for training.
Downloads last month
10
GGUF
Model size
8.03B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train student-abdullah/llama3.1_medicine_fine-tuned_21-09_8bits_gguf