oneAPI Deep Neural Network Library (oneDNN)
-
Updated
May 27, 2024 - C++
oneAPI Deep Neural Network Library (oneDNN)
Half-precision floating point types f16 and bf16 for Rust.
A LLaMA2-7b chatbot with memory running on CPU, and optimized using smooth quantization, 4-bit quantization or Intel® Extension For PyTorch with bfloat16.
Customizable floating point types, with all standard floating point operations implemented from scratch.
Floating-Point Arithmetic Library for Z80
Comparison of PageRank algorithm using various datatypes.
Basic linear algebra routines implemented using the chop rounding function
Comparison of vector element sum using various data types.
IEEE 754-style floating-point converter
A JAX implementation of stochastic addition.
A Pytorch implementation of stochastic addition.
Round matrix elements to lower precision in MATLAB
Hybridized On-Premise and Cloud (HOPC) Deployment Experimentation with Bfloat16
Add a description, image, and links to the bfloat16 topic page so that developers can more easily learn about it.
To associate your repository with the bfloat16 topic, visit your repo's landing page and select "manage topics."