Core implementation of Jina XLM-RoBERTa

This implementation is adapted from XLM-Roberta. In contrast to the original implementation, this model uses Rotary positional encodings and supports flash-attention 2.

Models that use this implementation

Converting weights

Weights from an original XLMRoberta model can be converted using the convert_roberta_weights_to_flash.py script in the model repository.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for jinaai/xlm-roberta-flash-implementation

Quantizations
1 model