YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
English to Vietnamese Translation (Quantized Model)
This repository contains a quantized English-To-Vietnamese translation model fine-tuned on the "HelloWorld2307/eng_viet_translation" dataset and optimized using dynamic quantization for efficient CPU inference.
π§ Model Details
- Base model: Helsinki-NLP/opus-mt-en-vi
- Dataset: HelloWorld2307/eng_viet_translation
- Training platform: Kaggle (CUDA GPU)
- Fine-tuned: On English-Vietnamese pairs from the Hugging Face dataset
- Quantization: PyTorch Dynamic Quantization (
torch.quantization.quantize_dynamic
) - Tokenizer: Saved alongside the model
π Folder Structure
quantized_model/ βββ config.json βββ pytorch_model.bin βββ tokenizer_config.json βββ tokenizer.json βββ vocab.json / merges.txt
π Usage
πΉ 1. Load Quantized Model for Inference
import torch
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline
# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained("./quantized_model")
# Load quantized model
model = AutoModelForSeq2SeqLM.from_pretrained("./quantized_model")
model.eval()
# Run translation
translator = pipeline("translation_en_to_vi", model=model, tokenizer=tokenizer, device=-1)
text = "Hello, How are you?"
print("English:", translator(text)[0]['translation_text'])
Model Training Summary
Loaded dataset: HelloWorld2307/eng_viet_translation
Mapped translation data: {"en": ..., "vi": ...} before training
Training: 3 epochs using GPU
Disabled: wandb logging
Skipped: Evaluation phase
Saved: Trained + Quantized model and tokenizer
Quantization: torch.quantization.Quantize_dynamic is used for efficient CPU inference
- Downloads last month
- 6
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support