Phrasing and other minor changes to sync with other readme
Browse files
README.md
CHANGED
@@ -12,9 +12,9 @@ tags:
|
|
12 |
|
13 |
## Model Summary
|
14 |
|
15 |
-
Phi-3
|
16 |
|
17 |
-
The model underwent a
|
18 |
When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.
|
19 |
|
20 |
Resources and Technical Documentation:
|
@@ -33,7 +33,7 @@ The model is intended for commercial and research use in English. The model prov
|
|
33 |
|
34 |
1) Memory/compute constrained environments
|
35 |
2) Latency bound scenarios
|
36 |
-
3) Strong reasoning (especially math and logic)
|
37 |
|
38 |
Our model is designed to accelerate research on language and multimodal models, for use as a building block for generative AI powered features.
|
39 |
|
|
|
12 |
|
13 |
## Model Summary
|
14 |
|
15 |
+
The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered websites data with a focus on high-quality and reasoning dense properties. The model belongs to the Phi-3 family with the Mini version in two variants [4K](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) and [128K](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) which is the context length (in tokens) that it can support.
|
16 |
|
17 |
+
The model has underwent a post-training process that incorporates both supervised fine-tuning and direct preference optimization for the instruction following and safety measures.
|
18 |
When assessed against benchmarks testing common sense, language understanding, math, code, long context and logical reasoning, Phi-3 Mini-4K-Instruct showcased a robust and state-of-the-art performance among models with less than 13 billion parameters.
|
19 |
|
20 |
Resources and Technical Documentation:
|
|
|
33 |
|
34 |
1) Memory/compute constrained environments
|
35 |
2) Latency bound scenarios
|
36 |
+
3) Strong reasoning (especially code, math and logic)
|
37 |
|
38 |
Our model is designed to accelerate research on language and multimodal models, for use as a building block for generative AI powered features.
|
39 |
|