qwen-14b-hyperswitch-v1

qwen14

Model Details

  • Base Model: Qwen/Qwen3-4B
  • Training Data: Hyperswitch repository code
  • Final Training Loss: 0.4966
  • Training Progress: 92% complete (excellent results)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model = AutoModelForCausalLM.from_pretrained(
    "archit11/qwen-14b-hyperswitch-v1",
    torch_dtype=torch.float16,
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("archit11/qwen-14b-hyperswitch-v1")

# Generate code
prompt = "use crate::"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=150, temperature=0.3)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)

Training Details

This model was fine-tuned on the Hyperswitch codebase to improve code completion and generation for Rust payment processing systems.

Downloads last month
4
Safetensors
Model size
14.8B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for archit11/qwen-14b-hyperswitch-v1

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Finetuned
(273)
this model