Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
194 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Graph Neural Network-based Tracking as a Service (2402.09633v1)

Published 15 Feb 2024 in physics.comp-ph, hep-ex, and physics.data-an

Abstract: Recent studies have shown promising results for track finding in dense environments using Graph Neural Network (GNN)-based algorithms. However, GNN-based track finding is computationally slow on CPUs, necessitating the use of coprocessors to accelerate the inference time. Additionally, the large input graph size demands a large device memory for efficient computation, a requirement not met by all computing facilities used for particle physics experiments, particularly those lacking advanced GPUs. Furthermore, deploying the GNN-based track-finding algorithm in a production environment requires the installation of all dependent software packages, exclusively utilized by this algorithm. These computing challenges must be addressed for the successful implementation of GNN-based track-finding algorithm into production settings. In response, we introduce a ``GNN-based tracking as a service'' approach, incorporating a custom backend within the NVIDIA Triton inference server to facilitate GNN-based tracking. This paper presents the performance of this approach using the Perlmutter supercomputer at NERSC.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (8)
  1. ATLAS Collaboration, “ATLAS Inner Tracker Strip Detector: Technical Design Report.” ATLAS-TDR-025; CERN-LHCC-2017-005, 2017.
  2. ATLAS Collaboration, “ATLAS Inner Tracker Pixel Detector: Technical Design Report.” ATLAS-TDR-030; CERN-LHCC-2017-021, 2017.
  3. arXiv:2010.08556.
  4. LiXin Xue, “A fixed radius nearest neighbors search implemented on cuda.” https://github.com/lxxue/FRNN.
  5. NVIDIA Corporation, “Triton inference server: An optimized cloud and edge inferencing solution.” https://github.com/triton-inference-server/server.
  6. NVIDIA Corporation, “Triton client.” https://github.com/triton-inference-server/client.
  7. N. Corporation, “Nvidia triton inference server: Metrics summary.” https://github.com/triton-inference-server/server/blob/main/docs/user_guide/metrics.md.
  8. P. Authors, “Exposition formats.” https://prometheus.io/docs/instrumenting/exposition_formats/.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com