QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Models Paper • 2310.08041 • Published Oct 12, 2023 • 1
Mesa: A Memory-saving Training Framework for Transformers Paper • 2111.11124 • Published Nov 22, 2021 • 1
TFMQ-DM: Temporal Feature Maintenance Quantization for Diffusion Models Paper • 2311.16503 • Published Nov 27, 2023
EcoFormer: Energy-Saving Attention with Linear Complexity Paper • 2209.09004 • Published Sep 19, 2022 • 1
Less is More: Pay Less Attention in Vision Transformers Paper • 2105.14217 • Published May 29, 2021 • 1
Scalable Vision Transformers with Hierarchical Pooling Paper • 2103.10619 • Published Mar 19, 2021 • 1
EfficientDM: Efficient Quantization-Aware Fine-Tuning of Low-Bit Diffusion Models Paper • 2310.03270 • Published Oct 5, 2023
TeacherLM: Teaching to Fish Rather Than Giving the Fish, Language Modeling Likewise Paper • 2310.19019 • Published Oct 29, 2023 • 9