Publikace: Comparison of Floating-point Representations for the Efficient Implementation of Machine Learning Algorithms
Konferenční objektopen accesspeer-reviewedpostprint (accepted version)Načítá se...
Datum
Autoři
Mishra, Saras Mani
Tiwari, Ankita
Shekhawat, Hanumant Singh
Guha, Prithwijit
Trivedi, Gaurav
Pidanič, Jan
Němec, Zdeněk
Název časopisu
ISSN časopisu
Název svazku
Nakladatel
IEEE
Abstrakt
Smart systems are enabled by artificial intelligence (AI), which is realized using machine learning (ML) techniques. ML algorithms are implemented in the hardware using fixedpoint, integer, and floating-point representations. The performance of hardware implementation gets impacted due to very small or large values because of their limited word size. To overcome this limitation, various floating-point representations are employed, such as IEEE754, posit, bfloat16 etc. Moreover, for the efficient implementation of ML algorithms, one of the most intuitive solutions is to use a suitable number system. As we know, multiply and add (MAC), divider and square root units are the most common building blocks of various ML algorithms. Therefore, in this paper, we present a comparative study of hardware implementations of these units based on bfloat16 and posit number representations. It is observed that posit based implementations perform 1.50x better in terms of accuracy, but consume 1.51x more hardware resources as compared to bfloat16 based realizations. Thus, as per the trade-off between accuracy and resource utilization, it can be stated that the bfloat16 number representation may be preferred over other existing number representations in the hardware implementations of ML algorithms.
Popis
Klíčová slova
floating-point representations, deep learning, posit, training, reprezentace s plovoucí desetinnou čárkou, hluboké učení, pozice, trénink