Pychop: Emulating Low-Precision Arithmetic in Numerical Methods and Neural Networks
Abstract
Pychop, a Python library, facilitates low-precision numerical computations and mixed-precision algorithms in deep learning, reducing memory and energy consumption with minimal impact on model performance.
Motivated by the growing demand for low-precision arithmetic in computational science, we exploit lower-precision emulation in Python -- widely regarded as the dominant programming language for numerical analysis and machine learning. Low-precision training has revolutionized deep learning by enabling more efficient computation and reduced memory and energy consumption while maintaining model fidelity. To better enable numerical experimentation with and exploration of low precision computation, we developed the Pychop library, which supports customizable floating-point formats and a comprehensive set of rounding modes in Python, allowing users to benefit from fast, low-precision emulation in numerous applications. Pychop also introduces interfaces for both PyTorch and JAX, enabling efficient low-precision emulation on GPUs for neural network training and inference with unparalleled flexibility. In this paper, we offer a comprehensive exposition of the design, implementation, validation, and practical application of Pychop, establishing it as a foundational tool for advancing efficient mixed-precision algorithms. Furthermore, we present empirical results on low-precision emulation for image classification and object detection using published datasets, illustrating the sensitivity of the use of low precision and offering valuable insights into its impact. Pychop enables in-depth investigations into the effects of numerical precision, facilitates the development of novel hardware accelerators, and integrates seamlessly into existing deep learning workflows. Software and experimental code are publicly available at https://github.com/inEXASCALE/pychop.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities (2025)
- Dual Precision Quantization for Efficient and Accurate Deep Neural Networks Inference (2025)
- Quartet: Native FP4 Training Can Be Optimal for Large Language Models (2025)
- FP4 All the Way: Fully Quantized Training of LLMs (2025)
- Efficient Precision-Scalable Hardware for Microscaling (MX) Processing in Robotics Learning (2025)
- Low-bit Model Quantization for Deep Neural Networks: A Survey (2025)
- Tilus: A Virtual Machine for Arbitrary Low-Precision GPGPU Computation in LLM Serving (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper