Neural Architecture Codesign for Fast Physics Applications

FOS: Computer and information sciences Computer Science - Machine Learning Condensed Matter - Materials Science High Energy Physics - Experiment (hep-ex) Physics - Instrumentation and Detectors Materials Science (cond-mat.mtrl-sci) FOS: Physical sciences Instrumentation and Detectors (physics.ins-det) High Energy Physics - Experiment Machine Learning (cs.LG)
DOI: 10.48550/arxiv.2501.05515 Publication Date: 2025-01-01
ABSTRACT
21 pages, 6 figures<br/>We develop a pipeline to streamline neural architecture codesign for physics applications to reduce the need for ML expertise when designing models for novel tasks. Our method employs neural architecture search and network compression in a two-stage approach to discover hardware efficient models. This approach consists of a global search stage that explores a wide range of architectures while considering hardware constraints, followed by a local search stage that fine-tunes and compresses the most promising candidates. We exceed performance on various tasks and show further speedup through model compression techniques such as quantization-aware-training and neural network pruning. We synthesize the optimal models to high level synthesis code for FPGA deployment with the hls4ml library. Additionally, our hierarchical search space provides greater flexibility in optimization, which can easily extend to other tasks and domains. We demonstrate this with two case studies: Bragg peak finding in materials science and jet classification in high energy physics, achieving models with improved accuracy, smaller latencies, or reduced resource utilization relative to the baseline models.<br/>
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....