InfinityLake-2x7B

Q8_0 GGUF quant forInfinityLake-2x7B. imatrix file provided for other quant if you want.

Experimental model from Endevor/InfinityRP-v1-7B and senseable/WestLake-7B-v2 models. Merged to MoE model with 2x7B parameters.

Okay enough. similar to InfinityKuno-2x7B but more creative.

Switch: FP16 - GGUF

Downloads last month
1
GGUF
Model size
12.9B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support