Mitchins commited on
Commit
2a9bf4e
Β·
verified Β·
1 Parent(s): c87b146

Update README for flux1-schnell-bf16-blackwell

Browse files
Files changed (1) hide show
  1. README.md +102 -26
README.md CHANGED
@@ -1,52 +1,128 @@
1
- # FLUX1-SCHNELL-BF16-BLACKWELL
 
 
 
 
 
 
 
 
 
 
 
2
 
3
- TensorRT-RTX engines for FLUX1-SCHNELL optimized for Blackwell architecture with BF16 quantization.
4
 
5
- ## Specifications
6
 
7
- - **Model**: FLUX1-SCHNELL
8
- - **Base Model**: black-forest-labs/FLUX.1-schnell
9
- - **Architecture**: Blackwell
10
- - **Quantization**: BF16
11
- - **Batch Size**: 1 (optimized)
12
- - **Resolution**: 1024x1024 (optimized)
13
- - **Estimated Size**: 12.0 GB
14
 
15
- ## Performance Estimates
16
 
17
- Based on architecture and quantization:
18
- - **Memory Usage**: ~12.0GB VRAM
19
- - **Speed**: ~2.5s (H200)
 
 
20
 
21
- ## Files
22
 
23
- - `engines/`: TensorRT engine files (.plan)
24
- - `config.json`: Configuration metadata
25
- - `README.md`: This file
26
 
27
- ## Usage
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
  ```python
30
  from imageai_server.tensorrt.nvidia_sdxl_pipeline import NVIDIASDXLPipeline
31
 
32
  pipeline = NVIDIASDXLPipeline()
33
  pipeline.load_engines(
34
- engine_dir="./engines",
35
- framework_model_dir="./framework",
36
- onnx_dir="./onnx"
37
  )
38
  pipeline.activate_engines()
39
 
40
  images, time_ms = pipeline.infer(
41
- prompt="a beautiful landscape",
42
  height=1024,
43
  width=1024
44
  )
45
  ```
46
 
47
- ## Requirements
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
 
49
- - **GPU**: Blackwell architecture (Compute Capability 8.0+)
50
- - **VRAM**: 12.0GB+
51
  - **TensorRT-RTX**: 1.0.0.21+
52
  - **CUDA**: 12.0+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: tensorrt-rtx
3
+ license: apache-2.0
4
+ base_model: black-forest-labs/FLUX.1-schnell
5
+ tags:
6
+ - tensorrt-rtx
7
+ - flux1
8
+ - bf16
9
+ - schnell
10
+ - optimized
11
+ inference: false
12
+ ---
13
 
14
+ # FLUX1 TensorRT-RTX: SCHNELL-Bf16 πŸ”¨ Building
15
 
16
+ Optimized TensorRT-RTX engines for **FLUX1** on **Bf16** architecture with **SCHNELL** quantization.
17
 
18
+ ## 🎯 This Repository
 
 
 
 
 
 
19
 
20
+ **One variant, one download** - only get exactly what you need!
21
 
22
+ - **Model**: FLUX1
23
+ - **Architecture**: Bf16 (Compute Capability 8.0+)
24
+ - **Quantization**: SCHNELL
25
+ - **Memory**: TBD
26
+ - **Speed**: TBD for 1024x1024 generation
27
 
28
+ ## πŸš€ Quick Start
29
 
30
+ ### Automatic (Recommended)
 
 
31
 
32
+ ```bash
33
+ # ImageAI server downloads automatically
34
+ curl -X POST "http://localhost:8001/generate" \
35
+ -H "Content-Type: application/json" \
36
+ -d '{
37
+ "prompt": "a beautiful landscape",
38
+ "model": "flux1-tensorrt_rtx:schnell",
39
+ "width": 1024,
40
+ "height": 1024
41
+ }'
42
+ ```
43
+
44
+ ### Manual Download
45
+
46
+ ```python
47
+ from huggingface_hub import snapshot_download
48
+
49
+ # Download this specific variant only
50
+ engines_path = snapshot_download(
51
+ repo_id="imgailab/flux1-trtx-schnell-bf16-blackwell"
52
+ )
53
+
54
+ # Engines are in: engines_path/engines/*.plan
55
+ ```
56
+
57
+ ### Direct Integration
58
 
59
  ```python
60
  from imageai_server.tensorrt.nvidia_sdxl_pipeline import NVIDIASDXLPipeline
61
 
62
  pipeline = NVIDIASDXLPipeline()
63
  pipeline.load_engines(
64
+ engine_dir=f"{engines_path}/engines",
65
+ framework_model_dir=f"{engines_path}/framework",
66
+ onnx_dir=f"{engines_path}/onnx"
67
  )
68
  pipeline.activate_engines()
69
 
70
  images, time_ms = pipeline.infer(
71
+ prompt="a serene mountain landscape",
72
  height=1024,
73
  width=1024
74
  )
75
  ```
76
 
77
+ ## πŸ“Š Performance
78
+
79
+ | Metric | Value |
80
+ |--------|-------|
81
+ | **Memory Usage** | TBD |
82
+ | **Inference Speed** | TBD |
83
+ | **Resolution** | 1024x1024 (optimized) |
84
+ | **Batch Size** | 1 (optimized) |
85
+ | **Precision** | SCHNELL |
86
+
87
+ ## πŸ”§ Requirements
88
+
89
+ ### Hardware
90
+ - **GPU**: Bf16 architecture
91
+ - Ampere: RTX 3090, A100, etc.
92
+ - Ada Lovelace: RTX 4090, etc.
93
+ - Blackwell: H200, etc.
94
+ - **VRAM**: TBD minimum
95
+ - **Compute Capability**: 8.0+
96
 
97
+ ### Software
 
98
  - **TensorRT-RTX**: 1.0.0.21+
99
  - **CUDA**: 12.0+
100
+ - **Python**: 3.8+
101
+
102
+ ## πŸ“ Repository Structure
103
+
104
+ ```
105
+ flux1-trtx-schnell-bf16-blackwell/
106
+ β”œβ”€β”€ engines/ # TensorRT engine files
107
+ β”‚ β”œβ”€β”€ *.plan # Optimized engines
108
+ β”œβ”€β”€ config.json # Configuration metadata
109
+ └── README.md # This file
110
+ ```
111
+
112
+ ## 🌐 Related Repositories
113
+
114
+ Other variants for FLUX1:
115
+ - [Ampere BF16](https://huggingface.co/imgailab/flux1-trtx-bf16-ampere)\n- [Ada FP8](https://huggingface.co/imgailab/flux1-trtx-fp8-ada)\n- [Ada BF16](https://huggingface.co/imgailab/flux1-trtx-bf16-ada)\n- [Blackwell FP4](https://huggingface.co/imgailab/flux1-trtx-fp4-blackwell)\n- [Blackwell FP8](https://huggingface.co/imgailab/flux1-trtx-fp8-blackwell)\n- [Blackwell BF16](https://huggingface.co/imgailab/flux1-trtx-bf16-blackwell)\n
116
+
117
+ ## πŸ“ License
118
+
119
+ Inherits license from base model: [black-forest-labs/FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell)
120
+
121
+ ## πŸ”„ Updates
122
+
123
+ - **2025-08-12**: Initial release
124
+ - Optimized for single-variant downloads
125
+
126
+ ---
127
+
128
+ *Part of the ImageAI TensorRT-RTX engine collection*