anamikac2708
commited on
Commit
•
4203da5
1
Parent(s):
7381c70
Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,9 @@ base_model: meta-llama/Meta-Llama-3-8B
|
|
20 |
|
21 |
This llama model was trained with Huggingface's TRL library and NEFTune https://arxiv.org/abs/2310.05914 using open-sourced finance dataset https://huggingface.co/datasets/FinLang/investopedia-instruction-tuning-dataset developed for finance application by FinLang Team
|
22 |
|
23 |
-
|
|
|
|
|
24 |
|
25 |
## How to Get Started with the Model
|
26 |
|
|
|
20 |
|
21 |
This llama model was trained with Huggingface's TRL library and NEFTune https://arxiv.org/abs/2310.05914 using open-sourced finance dataset https://huggingface.co/datasets/FinLang/investopedia-instruction-tuning-dataset developed for finance application by FinLang Team
|
22 |
|
23 |
+
NEFTune paper propose to add random noise to the embedding vectors of the training data during the forward pass of fine-tuning as a result the model overfits less to the
|
24 |
+
specifics of the instruction-tuning dataset, such as formatting details, exact wording, and text length. Instead of collapsing to the exact instruction distribution, the model is more capable of providing
|
25 |
+
answers that incorporate knowledge and behaviors of the pretrained base model.
|
26 |
|
27 |
## How to Get Started with the Model
|
28 |
|