YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
Turing
Turing is a character-level AI language model based on the GCLM (Global Convolutional Language Model) architecture. It is designed to learn from text using a hybrid approach consisting of local 1-dimensional convolutions for short-range dependencies and FFT-based global 1D convolutions for long-range context.
Architecture
The model (GCLM) processes sequences using a stack of blocks that alternate between:
- LocalConv1D: Captures local context (small chunks of n tokens)
- GlobalConv1D: Uses the FFT (Fast Fourier Transform) to capture global context across the entire sequence length.
Usage
Training
To train the model on your own text data:
- Place
.txtfiles in thedata/directory. - Run the training script:
This will automatically detect available hardware (CUDA, MPS, or CPU) and start training, saving checkpoints topython train.pyTuring_<params>.pt.
Inference
To generate text, run:
python sample.py
Requirements
- Python 3 (install at https://python.org)
- PyTorch (run
pip install torch) - tqdm (
pip install tqdm)
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
