Triangle104 commited on
Commit
51ea55a
·
verified ·
1 Parent(s): 577ca36

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -0
README.md CHANGED
@@ -12,6 +12,28 @@ tags:
12
  This model was converted to GGUF format from [`deepcogito/cogito-v1-preview-llama-8B`](https://huggingface.co/deepcogito/cogito-v1-preview-llama-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/deepcogito/cogito-v1-preview-llama-8B) for more details on the model.
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ## Use with llama.cpp
16
  Install llama.cpp through brew (works on Mac and Linux)
17
 
 
12
  This model was converted to GGUF format from [`deepcogito/cogito-v1-preview-llama-8B`](https://huggingface.co/deepcogito/cogito-v1-preview-llama-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/deepcogito/cogito-v1-preview-llama-8B) for more details on the model.
14
 
15
+ ---
16
+ The Cogito LLMs are instruction tuned generative models (text in/text
17
+ out). All models are released under an open license for commercial use.
18
+
19
+ -Cogito models are hybrid reasoning models. Each model can answer
20
+ directly (standard LLM), or self-reflect before answering (like
21
+ reasoning models).
22
+
23
+ -The LLMs are trained using Iterated Distillation and Amplification (IDA) - an scalable and efficient alignment strategy for superintelligence using iterative self-improvement.
24
+
25
+ -The models have been optimized for coding, STEM, instruction
26
+ following and general helpfulness, and have significantly higher
27
+ multilingual, coding and tool calling capabilities than size equivalent
28
+ counterparts.
29
+
30
+ --In both standard and reasoning modes, Cogito v1-preview models
31
+ outperform their size equivalent counterparts on common industry
32
+ benchmarks.
33
+
34
+ Each model is trained in over 30 languages and supports a context length of 128k.
35
+
36
+ ---
37
  ## Use with llama.cpp
38
  Install llama.cpp through brew (works on Mac and Linux)
39