TY - JOUR
T1 - Symbolic Regression on FPGAs for Fast Machine Learning Inference
AU - Tsoi, Ho Fung
AU - Pol, Adrian Alan
AU - Loncar, Vladimir
AU - Govorkova, Ekaterina
AU - Cranmer, Miles
AU - Dasu, Sridhara
AU - Elmer, Peter
AU - Harris, Philip
AU - Ojalvo, Isobel
AU - Pierini, Maurizio
N1 - Publisher Copyright:
© The Authors, published by EDP Sciences, 2024.
PY - 2024/5/6
Y1 - 2024/5/6
N2 - The high-energy physics community is investigating the potential of deploying machine-learning-based solutions on Field-Programmable Gate Arrays (FPGAs) to enhance physics sensitivity while still meeting data processing time constraints. In this contribution, we introduce a novel end-to-end procedure that utilizes a machine learning technique called symbolic regression (SR). It searches the equation space to discover algebraic relations approximating a dataset. We use PySR (a software to uncover these expressions based on an evolutionary algorithm) and extend the functionality of hls4ml (a package for machine learning inference in FPGAs) to support PySR -generated expressions for resource-constrained production environments. Deep learning models often optimize the top metric by pinning the network size because the vast hyperparameter space prevents an extensive search for neural architecture. Conversely, SR selects a set of models on the Pareto front, which allows for optimizing the performance-resource trade-off directly. By embedding symbolic forms, our implementation can dramatically reduce the computational resources needed to perform critical tasks. We validate our method on a physics benchmark: the multiclass classification of jets produced in simulated proton-proton collisions at the CERN Large Hadron Collider. We show that our approach can approximate a 3-layer neural network using an inference model that achieves up to a 13-fold decrease in execution time, down to 5 ns, while still preserving more than 90% approximation accuracy.
AB - The high-energy physics community is investigating the potential of deploying machine-learning-based solutions on Field-Programmable Gate Arrays (FPGAs) to enhance physics sensitivity while still meeting data processing time constraints. In this contribution, we introduce a novel end-to-end procedure that utilizes a machine learning technique called symbolic regression (SR). It searches the equation space to discover algebraic relations approximating a dataset. We use PySR (a software to uncover these expressions based on an evolutionary algorithm) and extend the functionality of hls4ml (a package for machine learning inference in FPGAs) to support PySR -generated expressions for resource-constrained production environments. Deep learning models often optimize the top metric by pinning the network size because the vast hyperparameter space prevents an extensive search for neural architecture. Conversely, SR selects a set of models on the Pareto front, which allows for optimizing the performance-resource trade-off directly. By embedding symbolic forms, our implementation can dramatically reduce the computational resources needed to perform critical tasks. We validate our method on a physics benchmark: the multiclass classification of jets produced in simulated proton-proton collisions at the CERN Large Hadron Collider. We show that our approach can approximate a 3-layer neural network using an inference model that achieves up to a 13-fold decrease in execution time, down to 5 ns, while still preserving more than 90% approximation accuracy.
UR - http://www.scopus.com/inward/record.url?scp=85196279926&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85196279926&partnerID=8YFLogxK
U2 - 10.1051/epjconf/202429509036
DO - 10.1051/epjconf/202429509036
M3 - Conference article
AN - SCOPUS:85196279926
SN - 2101-6275
VL - 295
JO - EPJ Web of Conferences
JF - EPJ Web of Conferences
M1 - 09036
T2 - 26th International Conference on Computing in High Energy and Nuclear Physics, CHEP 2023
Y2 - 8 May 2023 through 12 May 2023
ER -