GGUF
conversational

image/png

GGUF Quantizations for Caller

Caller (32B) is a robust model engineered for seamless integrations and optimized for managing complex tool-based interactions and API function calls. Its strength lies in precise execution, intelligent orchestration, and effective communication between systems, making it indispensable for sophisticated automation pipelines.

Model Details

  • Architecture Base: Qwen2.5-32B
  • Parameter Count: 32B
  • License: Apache-2.0

Use Cases:

  • Managing integrations between CRMs, ERPs, and other enterprise systems
  • Running multi-step workflows with intelligent condition handling
  • Orchestrating external tool interactions like calendar scheduling, email parsing, or data extraction
  • Real-time monitoring and diagnostics in IoT or SaaS environments

License

Caller (32B) is released under the Apache-2.0 License. You are free to use, modify, and distribute this model in both commercial and non-commercial applications, subject to the terms and conditions of the license.

If you have questions or would like to share your experiences using Caller (32B), please connect with us on social media. We’re excited to see what you build—and how this model helps you innovate!

Downloads last month
0
GGUF
Model size
32.8B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for arcee-ai/Caller-GGUF

Base model

Qwen/Qwen2.5-32B
Finetuned
arcee-ai/Caller
Quantized
(2)
this model