Jump to content
  • Google takes on Nvidia Blackwell GPUs with the new Trillium TPUs


    Karlston

    • 1 comment
    • 317 views
    • 2 minutes
     Share


    • 1 comment
    • 317 views
    • 2 minutes

    Google started developing custom AI accelerators in the name of Tensor Processing Units (TPUs) a decade ago. Early this year, Google announced Trillium, its sixth-generation TPU, that delivers better performance with efficiency than its predecessors. Today, Google announced the general availability of Trillium TPUs for Google Cloud customers. Google today also revealed that it used Trillium TPUs to train the new Gemini 2.0.

     

    Nvidia's GPUs are incredibly popular among developers for AI workloads, not just because of their hardware capabilities but also because of their software support. To make Trillium TPUs popular among AI developers, Google has made several improvements to its software layer. It has optimized the XLA compiler and AI frameworks such as JAX, PyTorch, and TensorFlow so that developers can achieve price-performance across AI training, tuning, and serving.

     

    Compared to the previous generation TPU, Trillium offers the following improvements:

     

    • Over 4x improvement in training performance
    • Up to 3x increase in inference throughput
    • A 67% increase in energy efficiency
    • An impressive 4.7x increase in peak compute performance per chip
    • Double the High Bandwidth Memory (HBM) capacity
    • Double the Interchip Interconnect (ICI) bandwidth
    • 100K Trillium chips in a single Jupiter network fabric
    • Up to 2.5x improvement in training performance per dollar and up to 1.4x improvement in inference performance per dollar

     

    Google also claimed that Trillium TPUs can achieve 99% scaling efficiency with a deployment of 12 pods consisting of 3072 chips and can achieve 94% scaling efficiency across 24 pods with 6144 chips to pre-train gpt3-175b.

     

    Trillium is now available in North America (US East region), Europe (West region), and Asia (Northeast region). For evaluation, Trillium is available starting at $2.7000 per chip-hour. With 1-year and 3-year commitments, it is available starting at $1.8900 per chip-hour and $1.2200 per chip-hour, respectively.

     

    With its ability to scale to hundreds of thousands of chips and improved software support, Trillium represents a significant leap forward for Google in the cloud AI infrastructure market.

     

    Source


    Hope you enjoyed this news post.

    Thank you for appreciating my time and effort posting news every day for many years.

    2023: Over 5,800 news posts | 2024 (till end of November): 5,298 news posts

    RIP Matrix | Farewell my friend  :sadbye:


    User Feedback

    Recommended Comments



    Join the conversation

    You can post now and register later. If you have an account, sign in now to post with your account.
    Note: Your post will require moderator approval before it will be visible.

    Guest
    Add a comment...

    ×   Pasted as rich text.   Paste as plain text instead

      Only 75 emoji are allowed.

    ×   Your link has been automatically embedded.   Display as a link instead

    ×   Your previous content has been restored.   Clear editor

    ×   You cannot paste images directly. Upload or insert images from URL.


  • Recently Browsing   0 members

    • No registered users viewing this page.
×
×
  • Create New...