Accelerating Generative AI – Options for Conquering the Dataflow Bottlenecks

Logo
Presented by

M. Baldi, AMD; R. Davis, NVIDIA; D. Eggleston, Microchip; D. McIntyre, Samsung; A. Rodriguez, Intel; Joe White, Dell

About this talk

Workloads using generative artificial intelligence trained on large language models are frequently throttled by insufficient resources (e.g., memory, storage, compute, or network dataflow bottlenecks). If not identified and addressed, these dataflow bottlenecks can constrain Gen AI application performance well below optimal levels. Given the compelling uses across natural language processing (NLP), video analytics, document resource development, image processing, image generation, and text generation, being able to run these workloads efficiently has become critical to many IT and industry segments. The resources that contribute to generative AI performance and efficiency include CPUs, DPUs, GPUs, FPGAs, plus memory and storage controllers. This webinar, with a broad cross-section of industry veterans, provides insight into the following: • Defining the Gen AI dataflow bottlenecks • Tools and methods for identifying acceleration options • Matchmaking the right xPU solution to the target Gen AI workload(s) • Optimizing the network to support acceleration options • Moving data closer to processing, or processing closer to data • The role of the software stack in determining Gen AI performance
Related topics:

More from this channel

Upcoming talks (2)
On-demand talks (123)
Subscribers (56100)
SNIA is a not-for-profit global organization made up of corporations, universities, startups, and individuals. The members collaborate to develop and promote vendor-neutral architectures, standards, and education for management, movement, and security for technologies related to handling and optimizing data. SNIA focuses on the transport, storage, acceleration, format, protection, and optimization of infrastructure for data.