Published September 10, 2025 | Version v1
Poster Open

Efficient Data Movement for Machine Learning Inference in Heterogeneous CMS Software

Authors/Creators

Description

Efficient data processing using machine learning relies on heterogeneous computing approaches, but optimizing input and output data movements remains a challenge. In GPU-based workflows data already resides on GPU memory, but machine learning models requires the input and output data to be provided in specific tensor format, often requiring unnecessary copying outside of the GPU device and conversion steps. To address this, we present an interface that allows seamless conversion of Structure of Arrays (SoA) data into lists of PyTorch tensors without explicit data movement. Our approach computes the necessary strides for various data types, including scalars and rows of vectors, matrices, allowing PyTorch tensors to directly access the data on the GPU memory. The introduced metadata structure provides a flexible mechanism for defining the columns to be used and specifying the order of the resulting tensor list. This user-friendly interface minimizes the amount of code required, allowing direct integration with machine learning models. Implemented within the CMS computing framework and using the Alpaka library for heterogeneous applications, this solution significantly improves GPU efficiency. By avoiding unnecessary CPU-GPU transfers, it accelerates model execution while maintaining flexibility and ease of use.

Files

ACAT__Efficient_Data_Movement.pdf

Files (1.6 MB)

Name Size Download all
md5:28648f95a350ab8279dd2a405b66f225
1.6 MB Preview Download

Additional details

Funding

Schmidt Family Foundation

Conference

Acronym
ACAT2025
Dates
8-12 September 2025
Place
Hamburg, Germany