JOURNAL ARTICLE

Accelerating Machine Learning Inference on GPUs with SYCL

Abstract

Recently, machine learning has established itself as a valuable tool for researchers to analyze their data and draw conclusions in various scientific fields, such as High Energy Physics (HEP). Commonly used machine learning libraries, such as Keras and PyTorch, might provide functionality for inference, but they only support their own models and are constrained by heavy dependencies, which render their deployment on embedded or bare-metal environments infeasible. SOFIE [3], which stands for System for Optimized Fast Inference code Emit, a part of the ROOT project developed at CERN, creates standalone C++ inference code from an input model in one of the popular machine learning formats. This code is directly invokable from other C++ projects and has minimal dependencies. In this work, we extend the functionality of SOFIE to generate SYCL code for machine learning model inference that can run on various GPU platforms and is only dependent on Intel MKL BLAS and portBLAS libraries.

Keywords:
Inference Computer science Code (set theory) Artificial intelligence Large Hadron Collider Software deployment Source code Machine learning Deep learning Programming language Operating system Particle physics

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
4
Refs
0.02
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

Particle Detector Development and Performance
Physical Sciences →  Physics and Astronomy →  Nuclear and High Energy Physics
Parallel Computing and Optimization Techniques
Physical Sciences →  Computer Science →  Hardware and Architecture
Advanced Data Storage Technologies
Physical Sciences →  Computer Science →  Computer Networks and Communications

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.