An Arabic Slot Grammar parser
Michael C. McCord, Violetta Cavalli-Sforza
ACL 2007
Advances in deep neural networks (DNNs) and the availability of massive real-world data have enabled superhuman levels of accuracy on many AI tasks and ushered the explosive growth of AI workloads across the spectrum of computing devices. However, their superior accuracy comes at a high computational cost, which necessitates approaches beyond traditional computing paradigms to improve their operational efficiency. Leveraging the application-level insight of error resilience, we demonstrate how approximate computing (AxC) can significantly boost the efficiency of AI platforms and play a pivotal role in the broader adoption of AI-based applications and services. To this end, we present RaPiD, a multi-tera operations per second (TOPS) AI hardware accelerator core (fabricated at 14-nm technology) that we built from the ground-up using AxC techniques across the stack including algorithms, architecture, programmability, and hardware. We highlight the workload-guided systematic explorations of AxC techniques for AI, including custom number representations, quantization/pruning methodologies, mixed-precision architecture design, instruction sets, and compiler technologies with quality programmability, employed in the RaPiD accelerator.
Michael C. McCord, Violetta Cavalli-Sforza
ACL 2007
Bowen Zhou, Bing Xiang, et al.
SSST 2008
Eric Price, David P. Woodruff
FOCS 2011
Kafai Lai, Alan E. Rosenbluth, et al.
SPIE Advanced Lithography 2007