YAML Metadata Error: "tags" must be an array

T5-VAE-Wiki (flax)

A Transformer-VAE made using flax.

It has been trained to interpolate on sentences form wikipedia.

Done as part of Huggingface community training (see forum post).

Builds on T5, using an autoencoder to convert it into an MMD-VAE (more info).

How to use from the 🤗/transformers library

Add model repo as a submodule:

git submodule add https://github.com/Fraser-Greenlee/t5-vae-flax.git t5_vae_flax
from transformers import AutoTokenizer
from t5_vae_flax.src.t5_vae import FlaxT5VaeForAutoencoding

tokenizer = AutoTokenizer.from_pretrained("t5-base")

model = FlaxT5VaeForAutoencoding.from_pretrained("flax-community/t5-vae-wiki")

Setup

Run setup_tpu_vm_venv.sh to setup a virtual enviroment on a TPU VM for training.

Downloads last month
4
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model’s pipeline type.

Space using flax-community/t5-vae-wiki 1