Model Card for quantum-circuit-qubo-3B

This model is the one discussed in the paper Fine-Tuning Large Language Models on Quantum Optimization Problems for Circuit Generation.

This model is a fine-tuned version of Qwen/Qwen2.5-3B-Instruct. It has been trained using TRL.

Quick start

See general documentation.

Framework versions

  • TRL: 0.15.2
  • Transformers: 4.48.3
  • Pytorch: 2.5.1
  • Datasets: 3.2.0
  • Tokenizers: 0.21.1

Citations

Cite this model as:

@misc{jern2025finetuninglargelanguagemodels,
      title={Fine-Tuning Large Language Models on Quantum Optimization Problems for Circuit Generation}, 
      author={Linus Jern and Valter Uotila and Cong Yu and Bo Zhao},
      year={2025},
      eprint={2504.11109},
      archivePrefix={arXiv},
      primaryClass={quant-ph},
      url={https://arxiv.org/abs/2504.11109}, 
}
Downloads last month
8
Safetensors
Model size
3.4B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for linuzj/quantum-circuit-qubo-3B

Base model

Qwen/Qwen2.5-3B
Finetuned
(469)
this model
Quantizations
1 model

Dataset used to train linuzj/quantum-circuit-qubo-3B