Exploration of TPUs for AI Applications

Tensor Processing Units (TPUs) are specialized hardware accelerators for deep learning developed by Google. This paper explores the performance of TPU with a focus on AI and its implementation in edge computing. It first provides an overview of TPUs, specifically their design in relation to neural networks, their general architecture, compilation techniques and supporting frameworks. Furthermore, we provide a comparative analysis of Cloud and Edge TPU performance against other counterpart chip architectures. It is then discussed how TPUs can be used to speed up AI workloads. The results show that TPUs can provide significant performance improvements both in cloud and edge computing. Additionally, we address the need for further research for the deployment of more architectures in the Edge TPU, as well as the need for the development of more robust comparisons in edge computing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/27/2020

AI on the Edge: Rethinking AI-based IoT Applications Using Specialized Edge Architectures

Edge computing has emerged as a popular paradigm for supporting mobile a...
research
07/17/2020

Klessydra-T: Designing Vector Coprocessors for Multi-Threaded Edge-Computing Cores

Convolutional computation kernels are fundamental to today's edge comput...
research
12/08/2020

The Why, What and How of Artificial General Intelligence Chip Development

The AI chips increasingly focus on implementing neural computing at low ...
research
06/01/2020

Understanding Uncertainty of Edge Computing: New Principle and Design Approach

Due to the edge's position between the cloud and the users, and the rece...
research
02/10/2020

AI-oriented Medical Workload Allocation for Hierarchical Cloud/Edge/Device Computing

In a hierarchically-structured cloud/edge/device computing environment, ...
research
11/27/2021

Roadmap for Edge AI: A Dagstuhl Perspective

Based on the collective input of Dagstuhl Seminar (21342), this paper pr...
research
04/18/2023

Heterogeneous Integration of In-Memory Analog Computing Architectures with Tensor Processing Units

Tensor processing units (TPUs), specialized hardware accelerators for ma...

Please sign up or login with your details

Forgot password? Click here to reset