bitskip1

This is a fine-tuned BitNet model with layer skipping capabilities.

Model Details

  • Model Type: BitNet with Layer Skipping
  • Base Model: Unknown
  • Architecture: Unknown

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("USERNAME/MODEL_NAME")
tokenizer = AutoTokenizer.from_pretrained("USERNAME/MODEL_NAME")

# Generate text
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training

This model was trained using the LayerSkip framework with BitNet architecture.

License

[Add your license information here]

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support