Model Details

Model Description

Repository

Usage

Method 1: llama.cpp Backend Server + Chatbox

Step 1: Start .llama.cpp Server

./llama-server \
  -m /path/to/model.gguf \
  -c 2048 \          # Context length
  --host 0.0.0.0 \   # Allow remote connections
  --port 8080 \      # Server port
  --n-gpu-layers 35  # GPU acceleration (if available)

Step 2: Connect via Chatbox

  1. Download Chatbox
  2. Configure API endpoint:
    API URL: http://localhost:8080
    Model: (leave empty)
    API Type: llama.cpp
    
  3. Set generation parameters:
    {
      "temperature": 0.7,
      "max_tokens": 512,
      "top_p": 0.9
    }
    

Method 2: LM Studio

  1. Download LM Studio
  2. Load GGUF file:
    • Launch LM Studio
    • Search Slipstream-Max/Emollm-InternLM2.5-7B-chat-GGUF-fp16
  3. Configure settings:
    Context Length: 2048
    GPU Offload: Recommended (enable if available)
    Batch Size: 512
    
  4. Start chatting through the built-in UI

Precision Details

Filename Precision Size Characteristics
CPsyCounX.gguf FP16 [15.5GB] Full original model precision

Hardware Requirements

Minimum:

  • 24GB RAM (for 7B model)
  • CPU with AVX/AVX2 instruction set support

Recommended:

  • 32GB RAM
  • CUDA-capable GPU (for acceleration)
  • Fast SSD storage (due to large model size)

Key Notes

  1. Requires latest llama.cpp (v3+ recommended)
  2. Use --n-gpu-layers 35 for GPU acceleration (requires CUDA-enabled build)
  3. Initial loading takes longer (2-5 minutes)
  4. Requires more memory/storage than quantized versions
  5. Use --mlock to prevent swapping

Advantages

  • Preserves original model precision
  • Ideal for precision-sensitive applications
  • No quantization loss
  • Suitable for continued fine-tuning
Downloads last month
7
GGUF
Model size
7.74B params
Architecture
internlm2
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Slipstream-Max/CPsyCounX-InternLM2-Chat-7B-GGUF-fp16

Quantized
(27)
this model

Dataset used to train Slipstream-Max/CPsyCounX-InternLM2-Chat-7B-GGUF-fp16