UPDATED 14:41 EST / APRIL 05 2017

EMERGING TECH

Google says its ultra-fast AI chips crush the competition

Google Inc. made headlines at its I/O developer event last May after revealing the existence of an internally developed chip (pictured) for running artificial intelligence workloads. Nearly a year later, the search giant is finally opening up about how the processor matches up against commercially available alternatives.

Members of Google’s hardware team released a paper today that claims the system beats central processing units and graphics processing unit in its weight class on several key fronts. One of them is power consumption, which is a major economic factor for a company that operates as much hardware as the search giant does. Its engineers highlight that the Tensor Processing Unit, as the chip is called, can provide 30 to 80 times more horsepower per watt than a comparable Intel Corp. Haswell CPU or Nvidia Inc.’s Tesla K80 GPU.

Google’s TSU leads in overall speed as well. Internal tests have shown that the chip can consistently provide 15 to 30 times better performance than commercial alternatives when handling AI workloads. One of the models that Google used during the trials, which the paper refers to only as CNN1, ran 70 times faster.

The company’s engineers have managed to pack all this horsepower into a chip that is smaller than Nvidia’s K80. It’s housed on a board configured to fit into the hard drive slots on the likewise custom-made server racks that Google employs in its data centers. According to the search giant, more than 100 internal teams are using TSUs to power support for Street View and the voice recognition features of other key services.

Google is one of several web-scale giants that use silicon optimized for their specific requirements. The group also includes Microsoft Corp. and Amazon.com Inc., which employ custom processors commissioned from Intel to power their public clouds. Moreover, the retail giant has a subsidiary called Annapurna Labs that sells specialized chips designed for use in networking equipment.

It’s worth mentioning that these chips are for actual operation of AI workloads, known as “inference.” As Patrick Moorhead, president and principal analyst at Moor Insights & Strategy, pointed out, the training of machine learning models is still carried out on vast server farms using CPUs and GPUs.

Image: Google

A message from John Furrier, co-founder of SiliconANGLE:

Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.

  • 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
  • 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About SiliconANGLE Media
SiliconANGLE Media is a recognized leader in digital media innovation, uniting breakthrough technology, strategic insights and real-time audience engagement. As the parent company of SiliconANGLE, theCUBE Network, theCUBE Research, CUBE365, theCUBE AI and theCUBE SuperStudios — with flagship locations in Silicon Valley and the New York Stock Exchange — SiliconANGLE Media operates at the intersection of media, technology and AI.

Founded by tech visionaries John Furrier and Dave Vellante, SiliconANGLE Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.