rkumar1999 commited on
Commit
ac14dea
·
verified ·
1 Parent(s): d7b5923

End of training

Browse files
Files changed (2) hide show
  1. README.md +3 -2
  2. config.json +1 -1
README.md CHANGED
@@ -1,15 +1,16 @@
1
  ---
2
  base_model: microsoft/Phi-tiny-MoE-instruct
3
  library_name: transformers
4
- model_name: phi-tiny-moe-lean-sft
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
10
  ---
11
 
12
- # Model Card for phi-tiny-moe-lean-sft
13
 
14
  This model is a fine-tuned version of [microsoft/Phi-tiny-MoE-instruct](https://huggingface.co/microsoft/Phi-tiny-MoE-instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
 
1
  ---
2
  base_model: microsoft/Phi-tiny-MoE-instruct
3
  library_name: transformers
4
+ model_name: rkumar1999/phi-tiny-moe-lean-sft
5
  tags:
6
  - generated_from_trainer
7
+ - alignment-handbook
8
  - trl
9
  - sft
10
  licence: license
11
  ---
12
 
13
+ # Model Card for rkumar1999/phi-tiny-moe-lean-sft
14
 
15
  This model is a fine-tuned version of [microsoft/Phi-tiny-MoE-instruct](https://huggingface.co/microsoft/Phi-tiny-MoE-instruct).
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
config.json CHANGED
@@ -37,6 +37,6 @@
37
  "sliding_window": 2047,
38
  "tie_word_embeddings": false,
39
  "transformers_version": "4.57.1",
40
- "use_cache": false,
41
  "vocab_size": 32064
42
  }
 
37
  "sliding_window": 2047,
38
  "tie_word_embeddings": false,
39
  "transformers_version": "4.57.1",
40
+ "use_cache": true,
41
  "vocab_size": 32064
42
  }