May 22, 2020
Machine-learning (ML) hardware and software system demand is burgeoning. Driven by ML applications, the number of different ML inference systems has exploded. Over 100 organizations are building ML inference chips, and the systems that incorporate existing models span at least three orders of magnitude in power consumption and five orders of magnitude in performance; they range from embedded devices to data-center solutions. Fueling the hardware are a dozen or more software frameworks and libraries. The myriad combinations of ML hardware and ML software make assessing ML-system performance in an architecture-neutral, representative, and reproducible manner challenging. There is a clear need for industry-wide standard ML benchmarking and evaluation criteria. MLPerf Inference answers that call. In this paper, we present our benchmarking method for evaluating ML inference systems. Driven by more than 30 organizations as well as more than 200 ML engineers and practitioners, MLPerf prescribes a set of rules and best practices to ensure comparability across systems with wildly differing architectures. The first call for submissions garnered more than 600 reproducible inference-performance measurements from 14 organizations, representing over 30 systems that showcase a wide range of capabilities. The submissions attest to the benchmark’s flexibility and adaptability.
Written by
Vijay Janapa Reddi
Christine Cheng
David Kanter
Peter Mattson
Guenther Schmuelling
Brian Anderson
Maximilien Breughe
Mark Charlebois
William Chou
Ramesh Chukka
Cody Coleman
Sam Davis
Pan Deng
Greg Diamos
Jared Duke
Dave Fick
J. Scott Gardner
Itay Hubara
Sachin Idgunji
Thomas B. Jablin
Jeff Jiao
Tom St. John
Pankaj Kanwar
David Lee
Jeffery Liao
Anton Lokhmotov
Francisco Massa
Peng Meng
Paulius Micikevicius
Colin Osborne
Gennady Pekhimenko
Arun Tejusve Raghunath Rajan
Dilip Sequeira
Ashish Sirasao
Fei Sun
Hanlin Tang
Michael Thomson
Frank Wei
Ephrem Wu
Lingjie Xu
Koichi Yamada
Bing Yu
George Yuan
Aaron Zhong
Peizhao Zhang
Yuchen Zhou
Publisher
International Symposium on Computer Architecture (ISCA)
November 27, 2022
Nicolas Ballas, Bernhard Schölkopf, Chris Pal, Francesco Locatello, Li Erran, Martin Weiss, Nasim Rahaman, Yoshua Bengio
November 27, 2022
November 27, 2022
Andrea Tirinzoni, Aymen Al Marjani, Emilie Kaufmann
November 27, 2022
November 16, 2022
Kushal Tirumala, Aram H. Markosyan, Armen Aghajanyan, Luke Zettlemoyer
November 16, 2022
November 10, 2022
Unnat Jain, Abhinav Gupta, Himangi Mittal, Pedro Morgado
November 10, 2022
April 08, 2021
Caner Hazirbas, Joanna Bitton, Brian Dolhansky, Jacqueline Pan, Albert Gordo, Cristian Canton Ferrer
April 08, 2021
April 30, 2018
Tomer Galanti, Lior Wolf, Sagie Benaim
April 30, 2018
April 30, 2018
Yedid Hoshen, Lior Wolf
April 30, 2018
December 11, 2019
Eliya Nachmani, Lior Wolf
December 11, 2019
Foundational models
Latest news
Foundational models