tmoe-exp-v1 / README.md
ehristoforu's picture
Update README.md
bf91ab9 verified
|
raw
history blame
333 Bytes
---
pipeline_tag: other
library_name: transformers
---
This is fully exp model!
```
base_model: beomi/EXAONE-3.5-2.4B-Instruct-Llamafied
gate_mode: random
architecture: mixtral
experts_per_token: 2
dtype: bfloat16
experts:
- source_model: beomi/EXAONE-3.5-2.4B-Instruct-Llamafied
- source_model: unsloth/Phi-3.5-mini-instruct
```