open_llama_3b / README.md
danielhanchen's picture
Update README.md
5269678
---
license: apache-2.0
language:
- en
---
Original repo: https://huggingface.co/openlm-research/open_llama_3b
This repo just allows the tokenizer to allow the use of `use_fast = True` to work, which can speed up batched tokenization dramatically.
This repo DOES NOT host OpenLLAMA's models. For those, use OpenLLAMA's repo.
For eg:
```
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("danielhanchen/open_llama_3b")
model = AutoModelForCausalLM.from_pretrained("openlm-research/open_llama_3b")
```