Model name change
Browse files- README.md +7 -7
- config.json +1 -1
README.md
CHANGED
@@ -7,7 +7,7 @@ license_link: >-
|
|
7 |
|
8 |
# Model Overview
|
9 |
|
10 |
-
|
11 |
|
12 |
Deriving the Minitron 8B and 4B models from the base 15B model using our approach requires up to **40x fewer training tokens** per model compared to training from scratch; this results in **compute cost savings of 1.8x** for training the full model family (15B, 8B, and 4B). Minitron models exhibit up to a 16% improvement in MMLU scores compared to training from scratch, perform comparably to other community models such as Mistral 7B, Gemma 7B and Llama-3 8B, and outperform state-of-the-art compression techniques from the literature. Please refer to our [arXiv paper](https://arxiv.org/abs/2407.14679) for more details.
|
13 |
|
@@ -15,15 +15,15 @@ This model is for research and development only.
|
|
15 |
|
16 |
**Model Developer:** NVIDIA
|
17 |
|
18 |
-
**Model Dates:** Minitron-
|
19 |
|
20 |
## License
|
21 |
|
22 |
-
|
23 |
|
24 |
## Model Architecture
|
25 |
|
26 |
-
|
27 |
It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
|
28 |
|
29 |
**Architecture Type:** Transformer Decoder (auto-regressive language model)
|
@@ -54,14 +54,14 @@ Support for Nemotron models will be added in the upcoming transformers library r
|
|
54 |
pip install git+https://github.com/huggingface/transformers
|
55 |
```
|
56 |
|
57 |
-
The following code provides an example of how to load the
|
58 |
|
59 |
```python
|
60 |
import torch
|
61 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
62 |
|
63 |
# Load the tokenizer and model
|
64 |
-
model_path = 'nvidia/
|
65 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
66 |
|
67 |
device = 'cuda'
|
@@ -86,7 +86,7 @@ print(output_text)
|
|
86 |
|
87 |
**Labeling Method:** Not Applicable
|
88 |
|
89 |
-
**Properties:** The training corpus for
|
90 |
|
91 |
**Data Freshness:** The pretraining data has a cutoff of June 2023.
|
92 |
|
|
|
7 |
|
8 |
# Model Overview
|
9 |
|
10 |
+
Minitron-4B-Base is a large language model (LLM) obtained by pruning Nemotron-4 15B; specifically, we prune model embedding size, number of attention heads, and MLP intermediate dimension. Following pruning, we perform continued training with distillation using 94 billion tokens to arrive at the final model; we use the continuous pre-training data corpus used in Nemotron-4 15B for this purpose.
|
11 |
|
12 |
Deriving the Minitron 8B and 4B models from the base 15B model using our approach requires up to **40x fewer training tokens** per model compared to training from scratch; this results in **compute cost savings of 1.8x** for training the full model family (15B, 8B, and 4B). Minitron models exhibit up to a 16% improvement in MMLU scores compared to training from scratch, perform comparably to other community models such as Mistral 7B, Gemma 7B and Llama-3 8B, and outperform state-of-the-art compression techniques from the literature. Please refer to our [arXiv paper](https://arxiv.org/abs/2407.14679) for more details.
|
13 |
|
|
|
15 |
|
16 |
**Model Developer:** NVIDIA
|
17 |
|
18 |
+
**Model Dates:** Minitron-4B-Base was trained between February 2024 and June 2024.
|
19 |
|
20 |
## License
|
21 |
|
22 |
+
Minitron-4B-Base is released under the [NVIDIA Open Model License Agreement](https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf).
|
23 |
|
24 |
## Model Architecture
|
25 |
|
26 |
+
Minitron-4B-Base uses a model embedding size of 3072, 32 attention heads, and an MLP intermediate dimension of 9216.
|
27 |
It also uses Grouped-Query Attention (GQA) and Rotary Position Embeddings (RoPE).
|
28 |
|
29 |
**Architecture Type:** Transformer Decoder (auto-regressive language model)
|
|
|
54 |
pip install git+https://github.com/huggingface/transformers
|
55 |
```
|
56 |
|
57 |
+
The following code provides an example of how to load the Minitron-4B-Base model and use it to perform text generation.
|
58 |
|
59 |
```python
|
60 |
import torch
|
61 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
62 |
|
63 |
# Load the tokenizer and model
|
64 |
+
model_path = 'nvidia/Minitron-4B-Base'
|
65 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
66 |
|
67 |
device = 'cuda'
|
|
|
86 |
|
87 |
**Labeling Method:** Not Applicable
|
88 |
|
89 |
+
**Properties:** The training corpus for Minitron-4B-Base consists of English and multilingual text, as well as code. Our sources cover a variety of document types such as: webpages, dialogue, articles, and other written materials. The corpus spans domains including legal, math, science, finance, and more. In our continued training set, we introduce a small portion of question-answering, and alignment style data to improve model performance.
|
90 |
|
91 |
**Data Freshness:** The pretraining data has a cutoff of June 2023.
|
92 |
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "nvidia/
|
3 |
"architectures": [
|
4 |
"NemotronForCausalLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "nvidia/Minitron-4B-Base",
|
3 |
"architectures": [
|
4 |
"NemotronForCausalLM"
|
5 |
],
|