view article Article Efficient LLM Pretraining: Packed Sequences and Masked Attention By sirluk โข Oct 7, 2024 โข 39
view article Article ๐ค PEFT: Parameter-Efficient Fine-Tuning of Billion-Scale Models on Low-Resource Hardware By smangrul and 1 other โข Feb 10, 2023 โข 85
view article Article How to train a new language model from scratch using Transformers and Tokenizers By julien-c โข Feb 14, 2020 โข 37
view article Article Fine-Tune W2V2-Bert for low-resource ASR with ๐ค Transformers By ylacombe โข Jan 19, 2024 โข 33