File size: 1,322 Bytes
f16e6c8 a3f8143 f16e6c8 a3f8143 f16e6c8 a9d35db f16e6c8 532d1f2 9966a3e f16e6c8 a9d35db f16e6c8 a9d35db f16e6c8 07a47f8 382c57d 71cd0b3 382c57d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
quantized_by: Menlo
pipeline_tag: text-generation
base_model: Menlo/Jan-nano
base_model_relation: quantized
license: apache-2.0
---
# Jan Nano
Authors: [Alan Dao](https://scholar.google.com/citations?user=eGWws2UAAAAJ&hl=en), [Bach Vu Dinh](https://scholar.google.com/citations?user=7Lr6hdoAAAAJ&hl=vi), [Thinh Le](https://scholar.google.com/citations?user=8tcN7xMAAAAJ&hl=en)

## Overview
Jan Nano is a fine-tuned language model built on top of the Qwen3 architecture. Developed as part of the Jan ecosystem, it balances compact size and extended context length, making it ideal for efficient, high-quality text generation in local or embedded environments.
## Features
- **Tool Use**: Excellent function calling and tool integration
- **Research**: Enhanced research and information processing capabilities
- **Small Model**: VRAM efficient for local deployment
## Use it with Jan (UI)
1. Install **Jan** using [Quickstart](https://jan.ai/docs/quickstart)
Original weight: https://huggingface.co/Menlo/Jan-nano
## Recommended Sampling Parameters
- Temperature: 0.7
- Top-p: 0.8
- Top-k: 20
- Min-p: 0
## Documentation
[Setup, Usage & FAQ](https://menloresearch.github.io/deep-research/) |