Upload DeepHermes-3-Llama-3-8B-Preview-Linux_CUDA_NGL_KV_F32_NO_Output_Weights_Processed-F32.imatrix with huggingface_hub
Browse files
.gitattributes
CHANGED
@@ -137,3 +137,4 @@ DeepHermes-3-Mistral-24B-Preview_Linux_CUDA_NGL_KV_F32-F32.imatrix filter=lfs di
|
|
137 |
Llama-3.1-8B_Linux_CUDA_NGL_KV_F32-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
138 |
Mistral-Small-3.1-24B-Instruct-2503-Linux_CUDA_NGL_KV_F32-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
139 |
Mistral-Small-3.1-24B-Instruct-2503-Linux_CUDA_NGL_KV_F32_NO_Output_Weights_Processed-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
|
|
|
137 |
Llama-3.1-8B_Linux_CUDA_NGL_KV_F32-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
138 |
Mistral-Small-3.1-24B-Instruct-2503-Linux_CUDA_NGL_KV_F32-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
139 |
Mistral-Small-3.1-24B-Instruct-2503-Linux_CUDA_NGL_KV_F32_NO_Output_Weights_Processed-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
140 |
+
DeepHermes-3-Llama-3-8B-Preview-Linux_CUDA_NGL_KV_F32_NO_Output_Weights_Processed-F32.imatrix filter=lfs diff=lfs merge=lfs -text
|
DeepHermes-3-Llama-3-8B-Preview-Linux_CUDA_NGL_KV_F32_NO_Output_Weights_Processed-F32.imatrix
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79571e6b859fc54da451f9656d477c5d4b1ad7fc604d4c52017b555e81e5f66b
|
3 |
+
size 4988191
|