Building data centre networks for AI workloads: key requirements and considerations for CSPs

11 October 2023 | Research

Joseph Attwood

Strategy report | PPTX and PDF (6 slides) | Cloud and AI Infrastructure


"For AI workloads, the performance of back-end networks in data centres will have a significant impact on GPU usage and consequently job completion time."

data_735x70_1075876368.jpg

As communications service providers (CSPs) train continuously larger AI models (especially generative AI models), models must be trained on an increasingly large number of graphics processing units (GPUs). CSPs that want to train these models in their own data centres will need to upgrade their back-end networks, that is, the networks which connect different GPUs within a data centre.

Information included in this report

  • Insights into the back-end data centre networking requirements of AI workloads
  • Analysis of the approaches to AI networking that CSPs can adopt in their data centres

 

USD1499

Log in

Log in to check if this content is included in your content subscription.

Author

Joseph Attwood

Analyst