Machine Learning at the Edge of Scale and Speed: Nanosecond Inference at the CERN Large Hadron Collider

Summary

Disclaimer: This summary has been generated by AI. It is experimental, and feedback is welcomed. Please reach out to info@qconlondon.com with any comments or concerns.

Thea discusses the integration of machine learning with particle physics at the CERN Large Hadron Collider (LHC) and the challenges associated with processing massive data streams in real-time.

Introduction

The CERN LHC generates data on a massive scale, producing 10,000 exabytes annually from high-energy proton collisions. The data handling must be real-time due to constraints in storage and compute capacity. The High-Luminosity LHC poses additional challenges with increased data complexity and requires enhanced machine learning solutions for efficient data processing.

Data Processing and Machine Learning Challenges

  • The first trigger stage at the LHC handles data rates approximating 5% of global internet traffic, needing low-latency, high-throughput machine learning solutions
  • FPGA and custom silicon are used for sub-microsecond latency, processing data faster than traditional CPU/GPU solutions
  • Emerging techniques involve hardware-aware model design, quantization, sparsity, and hardware-software co-design to enhance nanosecond inference

Innovative Machine Learning Techniques

  • Use of variational autoencoders for data compression and efficient embedding transmission via optical links
  • Graph neural networks facilitate event reconstruction with tight latency constraints
  • Applying quantization aware training and FPGA-specific optimizations to manage resource constraints and latency

Impact and Future Directions

The implementation of machine learning at CERN is not only a necessity for current operations but also serves as a potential model for similar data-intensive projects in other scientific domains and industries. The collaboration extends to industry partners like Google, Siemens, and Volvo, highlighting the broad applicability and advancement of these technologies.

The presentation concludes with a call to expand the application of these machine learning techniques to various scientific and industrial fields, offering immense opportunities for future research and collaboration.

This is the end of the AI-generated content.


Abstract

The CERN Large Hadron Collider (LHC) produces O(10,000) exabytes of raw data annually from high-energy proton collisions. Handling this volume under strict compute and storage limits requires real-time event filtering capable of processing millions of collisions per second. A multi-tier architecture of ASICs, FPGAs, CPUs, and GPUs reconstructs and analyzes events, rejecting >98% of data within microseconds.

With the transition to the High-Luminosity Large Hadron Collider (HL-LHC), the first trigger stage, located in radiation-shielded caverns ~100 m underground, must handle data rates approaching ~5% of global internet traffic and significantly increased event complexity. Maintaining physics sensitivity therefore requires highly efficient machine-learning algorithms optimized for real-time inference with extreme throughput and ultra-low latency.

In this talk, we will discuss emerging techniques for low-power, low-latency inference, including hardware-aware model design, quantization, sparsity, and hardware-software co-design. Using examples from particle physics and other domains, we will show how real-time machine learning is both a practical necessity and a powerful tool for scientific discovery.


Speaker

Thea Klaeboe Aarrestad

Particle Physics and Real-Time ML @CERN @ETH Zürich

Thea Klæboe Aarrestad is a particle physicist and fellow at the Institute for Particle Physics and Astrophysics at ETH Zürich. She earned a PhD in particle physics from the University of Zurich and previously worked as a research fellow at CERN in Geneva. Her work centers on applying machine learning to particle physics, with an emphasis on real-time techniques for discovering new phenomena, including low-power nanosecond inference on FPGAs and ML-based anomaly detection for analyzing proton collision data from the Large Hadron Collider. She was awarded the Young Experimental Physicist Prize by the European Physical Society for her contributions to integrating machine learning into experimental particle physics.

Read more
Find Thea Klaeboe Aarrestad at:

Date

Wednesday Mar 18 / 10:35AM GMT ( 50 minutes )

Location

Windsor (5th Fl.)

Topics

AI/ML systems real-time fpga asics

Share

From the same track

Session compilers

Automatically Retrofitting JIT Compilers

Wednesday Mar 18 / 03:55PM GMT

We as a community have attempted, multiple times, to speed up languages such as Lua, Python, and Ruby by hand-writing JIT compilers. Sometimes we've had short-term success, but the size, and pace of change, of their standard implementations has proven difficult to keep up with over time.

Speaker image - Laurence Tratt

Laurence Tratt

Shopify / Royal Academy of Engineering Research Chair in Language Engineering @King's College London

Session architecture

Not Just I/O: Using Async/Await for Computational Scheduling

Wednesday Mar 18 / 01:35PM GMT

In the past two years I have developed a new query execution engine for Polars, which not only tries to execute as much of your query in parallel as possible, but in a streaming fashion as well, such that you can process data sets which do not fit in memory.

Speaker image - Orson Peters

Orson Peters

Senior Engineer of Query Execution @Polars, (Co-)Author of Stdlib Sort in Rust & Go

Session Data Management

Looking Under the Hood: Data Processing Systems Performance Tricks (and How to Apply Them to Your Code)

Wednesday Mar 18 / 02:45PM GMT

Modern data processing systems—databases, analytics engines, vector stores, and stream processors—hide an extraordinary amount of performance engineering beneath their abstractions.

Speaker image - Holger Pirk

Holger Pirk

Associate Professor for Data Management Systems at Imperial College London and Avid Runner — Minimizing Cache Misses, Thread Divergence and Aerobic Decoupling

Session Data Systems

Vector Search on Columnar Storage

Wednesday Mar 18 / 11:45AM GMT

Managing vector data entails storing, updating, and searching collections of large and multi-dimensional pieces of data. Some believe that this justifies the creation of a new class of data systems specialized for this.

Speaker image - Peter Boncz

Peter Boncz

Professor @CWI, Co-Creator of MonetDB, VectorWise and MotherDuck, Database Systems Researcher, and Entrepreneur