Search Results - "Schelm, Kerstin"
-
1
A 7-nm Four-Core Mixed-Precision AI Chip With 26.2-TFLOPS Hybrid-FP8 Training, 104.9-TOPS INT4 Inference, and Workload-Aware Throttling
Published in IEEE journal of solid-state circuits (01-01-2022)“…Reduced precision computation is a key enabling factor for energy-efficient acceleration of deep learning (DL) applications. This article presents a 7-nm…”
Get full text
Journal Article -
2
RaPiD: AI Accelerator for Ultra-low Precision Training and Inference
Published in 2021 ACM/IEEE 48th Annual International Symposium on Computer Architecture (ISCA) (01-06-2021)“…The growing prevalence and computational demands of Artificial Intelligence (AI) workloads has led to widespread use of hardware accelerators in their…”
Get full text
Conference Proceeding -
3
SIMD Multi Format Floating-Point Unit on the IBM z15(TM)
Published in 2020 IEEE 27th Symposium on Computer Arithmetic (ARITH) (01-06-2020)“…The IBM z Systems(TM) is the backbone of the insurance, banking, and retail industry. Innovation in these markets is driving the demand for new and additional…”
Get full text
Conference Proceeding -
4
9.1 A 7nm 4-Core AI Chip with 25.6TFLOPS Hybrid FP8 Training, 102.4TOPS INT4 Inference and Workload-Aware Throttling
Published in 2021 IEEE International Solid- State Circuits Conference (ISSCC) (13-02-2021)“…Low-precision computation is the key enabling factor to achieve high compute densities (TOPS/W and TOPS/mm 2 ) in AI hardware accelerators across cloud and…”
Get full text
Conference Proceeding -
5
The POWER7 Binary Floating-Point Unit
Published in 2011 IEEE 20th Symposium on Computer Arithmetic (01-07-2011)“…The binary Floating-Point Unit (FPU) of the POWER7 processor is a 5.5 cycle Fused Multiply-Add (FMA) design, fully compliant with the IEEE 754-2008 standard…”
Get full text
Conference Proceeding