Show simple item record

dc.contributor.advisorStine, James E.
dc.contributor.authorNguyen, Tuan Danh
dc.date.accessioned2019-07-19T14:33:29Z
dc.date.available2019-07-19T14:33:29Z
dc.date.issued2018-12-01
dc.identifier.urihttps://hdl.handle.net/11244/320978
dc.description.abstractMultiplication has long been an important part of any computer architecture. It has usually been a common case for most computer architecture decisions to include in any microarchitecture. However, the difficulty in creating hardware for multiplication because of its inherent shifting of the radix point has been a cogent reason for the need for floating-point hardware in scientific applications. The IEEE 754 floating-point standard was originally ratified in 1985 and later amended in 2008 to make floating-point multiplication easier for users to implement applications. Although floating-point arithmetic creates a mechanism to make things easier for using multiplication, it is complicated both algorithmically and practically for hardware implementations.
dc.description.abstractThis dissertation discusses possible architectural improvements in IEEE-compliant floating-point multiplication for Machine Learning/Deep Learning applications. First, a combined IEEE half and single precision floating-point multipliers is proposed to reduce power dissipation for Deep Learning applications. Second, a novel rounding scheme is proposed that is simpler but comparable with the state-of-the-art rounding schemes. Third, an optimized design is proposed that can handle both denormal and normal numbers. Finally, a hybrid precision design is proposed, aiming to improve the power consumption of Machine Learning/Deep Learning applications. Proposed designs are targeted to Machine Learning/Deep Learning applications-specific processors to improve the latency and power consumption. All designs are implemented in RTL-level Verilog, verified for correctness against open-source TestFloat generated test vectors, and synthesized using an ARM 32nm CMOS library for Global Foundries (GF) cmos32soi technology for estimated power, area and delay analysis.
dc.formatapplication/pdf
dc.languageen_US
dc.rightsCopyright is held by the author who has granted the Oklahoma State University Library the non-exclusive right to share this material in its institutional repository. Contact Digital Library Services at lib-dls@okstate.edu or 405-744-9161 for the permission policy on the use, reproduction or distribution of this material.
dc.titleArchitectural Improvements in IEEE-Compliant Floating-Point Multiplication
dc.contributor.committeeMemberTeague, Keith A.
dc.contributor.committeeMemberLatino, Carl D.
dc.contributor.committeeMemberDeYong, Camille F.
osu.filenameNguyen_okstate_0664D_16076.pdf
osu.accesstypeOpen Access
dc.description.departmentElectrical Engineering
dc.type.genreDissertation
dc.type.materialText


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record