“We’re excited to announce that our second-generation Tensor Processing Units (TPUs) are coming to Google Cloud to accelerate a wide range of machine learning workloads, including both training and inference. We call them Cloud TPUs, and they will initially be available via Google Compute Engine. We’ve witnessed extraordinary advances in machine learning (ML) over the past few years. Neural networks have dramatically improved the quality of Google Translate, played a key role in ranking Google Search results and made it more convenient to find the photos you want with Google Photos. Machine learning allowed DeepMind’s AlphaGo program to defeat Lee Sedol, one of the world’s top Go players, and also made it possible for software to generate natural-looking sketches. These breakthroughs required enormous amounts of computation, both to train the underlying machine learning models and to run those models once they’re trained (this is called “inference”). We’ve designed, built and deployed a family of Tensor Processing Units, or TPUs, to allow us to support larger and larger amounts of machine learning computation, first internally and now externally. While our first TPU was designed to run machine learning models quickly and efficiently—to translate a set of sentences or choose the next move in Go—those models still had to be trained separately. Training a machine learning model is even more difficult than running it, and days or weeks of computation on the best available CPUs and GPUs are commonly required to reach state-of-the-art levels of accuracy. Research and engineering teams at Google and elsewhere have made great progress scaling machine learning training using readily-available hardware. However, this wasn’t enough to meet our machine learning needs, so we designed an entirely new machine learning system to eliminate bottlenecks and maximize overall performance. At the heart of this system is the second-generation TPU we’re announcing today, which can both train and run machine learning models.”
Related Content
Related Posts:
- ASUS & Google Team Up for ‘Tinker Board’ AI-Focused Credit-Card Sized Computers
- Google announces a new generation for its TPU machine learning hardware
- Intel Gaudi AI Accelerator Gains 2x Performance Leap on GPT-3 with FP8 Software
- Microchip Teams Up with Intelligent Hardware Korea (IHWK) to Develop an Analog Compute Platform to Accelerate Edge AI/ML Inferencing
- New MLCommons Results Highlight Impressive Competitive AI Gains for Intel
- Intel Labs Introduces AI Diffusion Model, Generates 360-Degree Images from Text Prompts
- Chip Manufacturing ‘Ideal Application’ for AI, NVIDIA CEO Says
- Raspberry Pi Receives Strategic Investment from Sony Semiconductor Solutions Corporation Establishing strategic partnership for the further enhancement of the edge AI solution development
- First AI-enhanced smart accelerometers from STMicroelectronics raise performance and efficiency for always-aware applications
- Renesas to Demonstrate First AI Implementations on the Arm Cortex-M85 Processor Featuring Helium Technology at Embedded World