semiprecision
Semiprecision refers to a range of floating-point formats that offer more precision than standard single-precision (FP32) but less than double-precision (FP64). These formats are often used in machine learning and deep learning applications to reduce memory usage and computational cost without a significant loss of accuracy for many tasks.
Common semiprecision formats include half-precision (FP16) and bfloat16. FP16 uses 16 bits, with 1 sign bit,
The adoption of semiprecision formats has been driven by the increasing demand for faster training and inference