Mol-Llama-3.1-8B-Instruct-Full-Weights

[Project Page] [Paper] [GitHub]

This repo contains the weights of Mol-LLaMA including the LoRA weights and projectors, build with Llama: meta-llama/Llama-3.1-8B-Instruct. Llama 3.1 is licensed under the Llama 3.1 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved.

Architecture

image.png

  1. Molecular encoders: Pretrained 2D encoder (MoleculeSTM) and 3D encoder (Uni-Mol)
  2. Blending Module: Combining complementary information from 2D and 3D encoders via cross-attention
  3. Q-Former: Embed molecular representations into query tokens based on SciBERT
  4. LoRA: Adapters for fine-tuning LLMs

Training Dataset

Mol-LLaMA is trained on Mol-LLaMA-Instruct, to learn the fundamental characteristics of molecules with the reasoning ability and explanbility.

Citation

If you find our model useful, please consider citing our work.

@misc{kim2025molllama,
    title={Mol-LLaMA: Towards General Understanding of Molecules in Large Molecular Language Model},
    author={Dongki Kim and Wonbin Lee and Sung Ju Hwang},
    year={2025},
    eprint={2502.13449},
    archivePrefix={arXiv},
    primaryClass={cs.LG}
}

Acknowledgements

We appreciate LLaMA, 3D-MoLM, MoleculeSTM, Uni-Mol and SciBERT for their open-source contributions.

Downloads last month
1
Safetensors
Model size
8.25B params
Tensor type
I64
·
F32
·
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for DongkiKim/Mol-Llama-3.1-8B-Instruct-Full-Weights

Finetuned
(1735)
this model

Dataset used to train DongkiKim/Mol-Llama-3.1-8B-Instruct-Full-Weights