TPU (Tensor Processing Unit)

A Tensor Processing Unit (TPU) is a custom-built processor built to speed up machine learning workloads, especially those based on neural networks. It is an application-specific integrated circuit (ASIC), meaning it is designed for a specific purpose rather than general computing. TPUs are optimized for the large matrix calculations that are common in deep learning, making them very efficient for training and running AI models.

Google began using TPUs internally in 2015 and later made them available to others through Google Cloud. They support popular machine learning frameworks such as TensorFlow, JAX, and PyTorch, and are widely used for tasks like training large language models, recommendation systems, computer vision, and speech processing.

Compared to CPUs and GPUs, TPUs focus on doing fewer types of operations but doing them extremely well. CPUs are flexible and good for many tasks, GPUs excel at parallel processing and are widely used in AI, while TPUs are purpose-built to maximize speed and energy efficiency for matrix-heavy AI computations.

We use cookies

Our website uses cookies to ensure you get the best experience. By browsing the website you agree to our use of cookies. Please note, we don’t collect sensitive data and child data.

To learn more and adjust your preferences click Cookie Policy and Privacy Policy. Withdraw your consent or delete cookies whenever you want here.

Allow all cookies