EE Times writes the first release of MLPerf will focus on training jobs, while later versions will add inferencing benchmarks. There is still a need for a lot more performance. Below is a pretty interesting quote from Baidu, it reveals one AI model the company really wants to train currently requires two years of computing on all the GPUs the company currently has!
“To train one model we really want to run would take all GPUs we have for two years,” given the size of the model and its data sets, said Greg Diamos, a senior researcher in Baidu’s deep-learning group, giving an example of the issue for web giants.
“If systems become faster, we can unlock the potential of machine learning a lot quicker,” said Peter Mattson, a staff engineer on the Google Brain project who announced MLPerf at a May 2 event.