warshanks commited on
Commit
af9bb1c
·
verified ·
1 Parent(s): f9d2502

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +108 -0
README.md ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Menlo/Jan-nano-128k
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
9
+ ---
10
+
11
+ # Jan-Nano-128k: Empowering deeper research through extended context understanding.
12
+ <sub>*Note: Jan-Nano is a non-thinking model.*</sub>
13
+
14
+ [![GitHub](https://img.shields.io/badge/GitHub-Repository-blue?logo=github)](https://github.com/menloresearch/deep-research)
15
+ [![License](https://img.shields.io/badge/License-Apache%202.0-yellow)](https://opensource.org/licenses/Apache-2.0)
16
+
17
+ <div align="center">
18
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/NP7CvcjOtLX8mST0t7eAM.png" width="300" alt="Jan-Nano-128k">
19
+ </div>
20
+
21
+ **Authors:** [Alan Dao](https://scholar.google.com/citations?user=eGWws2UAAAAJ&hl=en), [Bach Vu Dinh](https://scholar.google.com/citations?user=7Lr6hdoAAAAJ&hl=vi)
22
+
23
+
24
+ ![image/gif](https://cdn-uploads.huggingface.co/production/uploads/62d7b2339b629105a5d6888a/aLL8fyMLE3ujV75qD4WKI.gif)
25
+
26
+
27
+ ## Overview
28
+
29
+ Jan-Nano-128k represents a significant advancement in compact language models for research applications. Building upon the success of [Jan-Nano](https://huggingface.co/Menlo/Jan-nano), this enhanced version features a **native 128k context window** that enables deeper, more comprehensive research capabilities without the performance degradation typically associated with context extension methods.
30
+
31
+ **Key Improvements:**
32
+ - **🔍 Research Deeper**: Extended context allows for processing entire research papers, lengthy documents, and complex multi-turn conversations
33
+ - **⚡ Native 128k Window**: Built from the ground up to handle long contexts efficiently, maintaining performance across the full context range
34
+ - **📈 Enhanced Performance**: Unlike traditional context extension methods, Jan-Nano-128k shows improved performance with longer contexts
35
+
36
+ This model maintains full compatibility with Model Context Protocol (MCP) servers while dramatically expanding the scope of research tasks it can handle in a single session.
37
+
38
+ ## Evaluation
39
+
40
+ Jan-Nano-128k has been rigorously evaluated on the SimpleQA benchmark using our MCP-based methodology, demonstrating superior performance compared to its predecessor:
41
+
42
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65713d70f56f9538679e5a56/Bc0ehij86l_NX52OfxeOj.png)
43
+
44
+ ## Why Jan-Nano-128k?
45
+
46
+ Traditional approaches to extending context length, such as YaRN (Yet another RoPE extensioN), often result in performance degradation as context length increases. Jan-Nano-128k breaks this paradigm:
47
+
48
+ This fundamental difference makes Jan-Nano-128k ideal for research applications requiring deep document analysis, multi-document synthesis, and complex reasoning over large information sets.
49
+
50
+ ## 🖥️ How to Run Locally
51
+
52
+ Jan desktop will eventually support this model (WIP). Otherwise you can check the deployment options below that we have tested.
53
+
54
+ For additional tutorials and community guidance, visit our [Discussion Forums](https://huggingface.co/Menlo/Jan-nano-128k/discussions).
55
+
56
+ ### Deployment
57
+
58
+ Deploy using VLLM:
59
+ ```bash
60
+ vllm serve Menlo/Jan-nano-128k \
61
+ --host 0.0.0.0 \
62
+ --port 1234 \
63
+ --enable-auto-tool-choice \
64
+ --tool-call-parser hermes \
65
+ --rope-scaling '{"rope_type":"yarn","factor":3.2,"original_max_position_embeddings":40960}' --max-model-len 131072
66
+ ```
67
+
68
+ Or `llama-server` from `llama.cpp`:
69
+ ```bash
70
+ llama-server ... --rope-scaling yarn --rope-scale 3.2 --yarn-orig-ctx 40960
71
+ ```
72
+ **Note:** The chat template is included in the tokenizer. For troubleshooting, download the [Non-think chat template](https://qwen.readthedocs.io/en/latest/_downloads/c101120b5bebcc2f12ec504fc93a965e/qwen3_nonthinking.jinja).
73
+
74
+ ### Recommended Sampling Parameters
75
+
76
+ ```yaml
77
+ Temperature: 0.7
78
+ Top-p: 0.8
79
+ Top-k: 20
80
+ Min-p: 0.0
81
+ ```
82
+
83
+ ## FAQ:
84
+ - I have Jinja template issue with LMStudio, how can i fix? [Here](https://huggingface.co/Menlo/Jan-nano-128k-gguf/discussions/1#6862fe2375cb85f79b28d69c)
85
+
86
+ ## 🤝 Community & Support
87
+
88
+ - **Discussions**: [HuggingFace Community](https://huggingface.co/Menlo/Jan-nano-128k/discussions)
89
+ - **Issues**: [GitHub Repository](https://github.com/menloresearch/jan/issues)
90
+ - **Documentation**: [Official Docs](https://menloresearch.github.io/deep-research/)
91
+
92
+ ## 📄 Citation
93
+
94
+ ```bibtex
95
+ @misc{dao2025jannanotechnicalreport,
96
+ title={Jan-nano Technical Report},
97
+ author={Alan Dao and Dinh Bach Vu},
98
+ year={2025},
99
+ eprint={2506.22760},
100
+ archivePrefix={arXiv},
101
+ primaryClass={cs.CL},
102
+ url={https://arxiv.org/abs/2506.22760},
103
+ }
104
+ ```
105
+
106
+ ---
107
+
108
+ *Jan-Nano-128k: Empowering deeper research through extended context understanding.*