Public Models
Collection
Some of them are finetuned and some of them are created by me!
β’
6 items
β’
Updated
Boringβs Experimental Transformer for Autoregression (Flax) β A 100M parameter autoregressive model built in Flax. Lightweight, chaotic, and surprisingly good (I mean ok). Because who needs sanity when youβve got tokens to predict?
Trained on determination, fueled by suffering, powered by free TPUs. π₯
WICKED4950/Raw-GPT-traindata
) Or a total of about 7.6B tokensThis model was trained using a custom JAX/Flax pipeline optimized for free Google TPUs.
Interested in the tech? Contact me for consulting/licensing.
Use pip install BwETAF
to install it.
** It does not include a Trainer**
import BwETAF
# You can use this function for quick testing of the model
prompt = "The meaning of life is"
output = BwETAF.SetUpAPI(prompt, "WICKED4950/BwETAF-IID-100M")
print(output) # Example: "The meaning of life is... (model's actual output)"
# Load from Hugging Face
model = BwETAF.load_hf("WICKED4950/BwETAF-IID-100M")
# Load from local directory
BwETA.load_model("path/to/model")
# Save locally
model.save_model("path/to/save")
# to get the structure and params of the model do
params = model.trainable_variables
structure = model.model_struct
Open an google collab notebook
As a 17-year-old solo developer, I built this to:
If you got anything to talk realted to this... Contact me at Instagram