Category
1 post
Google's TPU 8 ships as two chips. The 8t handles training at 121 ExaFlops per pod; the 8i carries 288GB HBM for inference. The split is the real story.