You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

PSCManual Pre Trained Model

This model is a CPT version of TinyLlama/TinyLlama-1.1B-Chat-v1.0 on the NHSN 2025 Patient Safety Component Manual.

Intended uses & limitations

This is a Continued Pre-Training (CPT) model designed to function primarily as an autocomplete system. It was developed as an experimental exercise to evaluate knowledge injection into a language model, with continued pre-training on the NHSN 2025 Patient Safety Component Manual. This model is not intended for production use. Its outputs may be suboptimal because it was not trained with enough data to meet Chinchilla scaling laws, which recommend approximately 20 tokens per parameter for optimal performance.

Training procedure

CPT (Continued Pre Training) for knowledge injection.

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0002
  • train_batch_size: 2
  • eval_batch_size: 2
  • seed: 42
  • distributed_type: multi-GPU
  • num_devices: 4
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 32
  • total_eval_batch_size: 8
  • optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 10
  • training_steps: 16

Framework versions

  • Transformers 4.50.0
  • Pytorch 2.5.0+cu121
  • Datasets 3.4.1
  • Tokenizers 0.21.1
Downloads last month
91
Safetensors
Model size
1.1B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for sharadsin/PSCManual_CPT_Model

Finetuned
(313)
this model