MCNNTUNES: Tuning Shower Monte Carlo generators with machine learning,☆☆

https://doi.org/10.1016/j.cpc.2021.107908Get rights and content

Abstract

The parameters tuning of event generators is a research topic characterized by complex choices: the generator response to parameter variations is difficult to obtain on a theoretical basis, and numerical methods are hardly tractable due to the long computational times required by generators. Event generator tuning has been tackled by parametrization-based techniques, with the most successful one being a polynomial parametrization. In this work, an implementation of tuning procedures based on artificial neural networks is proposed. The implementation was tested with closure testing and experimental measurements from the ATLAS experiment at the Large Hadron Collider.

Program summary

Program Title: MCNNTUNES

CPC Library link to program files: https://doi.org/10.17632/dmkydsxgd3.1

Developer’s repository link: https://github.com/N3PDF/mcnntunes

Licensing provisions: GPLv3

Programming language: Python

Nature of problem: Shower Monte Carlo generators introduce many parameters that must be tuned to reproduce the experimental measurements. The dependence of the generator output on these parameters is difficult to obtain on a theoretical basis.

Solution method: Implementation of a tuning method using supervised machine learning algorithms based on neural networks, which are universal approximators.

Introduction

Shower Monte Carlo (SMC) event generators are tools that apply shower algorithms to simulated collisions of particles at high energies. They introduce many parameters, mainly due to the usage of phenomenological models, like the hadronization model or the underlying event model, needed to describe the low-energy limit of quantum chromodynamics (QCD) which is not easily calculable from first principles. These parameters are difficult to obtain on a theoretical basis, so they must be carefully tuned in order to make the generators reproduce the experimental measurements. The procedure of estimating the best value for each parameter is called event generator tuning.

This tuning procedure is made more difficult by the high computational cost of running a generator, so it requires methods to study the dependence between a generator output and its parameters. Moreover, since the observables considered while analysing the generator output play a pivotal role in determining the tuning, one needs to model this dependence for different observables at the same time.

The current state-of-the-art tuning procedure is based on a polynomial parametrization of the generator response to parameter variations, followed by a numerical fit of the parametrized behaviour to experimental data. This is the procedure which is implemented in Professor [1], the primary tool for SMC event generator tuning at the Large Hadron Collider (LHC). However, the assumption that the dependence of the generator output on its parameters is polynomial is not always justified.

This paper investigates new tuning procedures based on artificial neural networks. Artificial neural networks are universal function approximators [2], [3], [4], providing accurate predictive models with a finite number of parameters, as shown in early attempts [5]. Two different tuning procedures are presented, called Per Bin and Inverse from now on. The former follows the same approach of Professor, but with a different parametrization model made of fully-connected neural networks and a different minimization algorithm: the evolutionary algorithm CMA-ES [6]. The latter takes a completely different approach: by using a fully-connected neural network, it learns to predict directly the parameters that the generator needs to output a given result. These two procedures were implemented in the Python package MCNNTUNES [7] and then tested with the event generator PYTHIA8 [8]. Two different datasets of Monte Carlo runs were generated, with three and four tunable parameters respectively. The procedures were tested with closure tests and with real experimental data taken from the ATLAS experiment [9], [10].

A description of the procedures and their technical implementation is presented in Section 2, while Section 3 contains the details of the testing phase. Finally, the conclusion of this work and future development directions are presented in Section 4.

Section snippets

Implementation

MCNNTUNES implements two different strategies for generator tuning, both based on feedforward neural networks. In this section these two strategies are presented in detail, along with a description of their technical implementation.

Results

This section presents the testing phase of MCNNTUNES. The choice of the generator, the parameters with their variation ranges, the process and the observables on which performing the tunes were chosen following the AZ tune [9] as reference. This should be considered only as a study of the efficiency and reliability of the MCNNTUNES approach for some specific observables and data and not as an attempt to devise a new exhaustive tune for the LHC.

The generation of some datasets of Monte Carlo runs

Outlook

A deep learning approach to event generator tuning was presented by introducing two different procedures, called Per Bin strategy and Inverse strategy respectively. The former is a variation of the Professor tuning procedure, that improves over it by relaxing the assumption of a polynomial dependence of the generator response to variations of the parameters. The latter is a novel and completely different approach. The procedures were tested with closure tests and real experimental data, though

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

This research used resources of the National Energy Research Scientific Computing Center (NERSC), a U.S. Department of Energy Office of Science User Facility operated under Contract No. DE-AC02-05CH11231. S.C. is supported by the European Research Council under the European Union’s Horizon 2020 research and innovation Programme (grant agreement number 740006). The work of S.A. is supported by the ERC Starting Grant REINVENT-714788. He also acknowledges funding from Fondazione Cariplo, Italy and

References (23)

  • BuckleyA. et al.

    Eur. Phys. J. C

    (2010)
  • HornikK.

    Neural Netw.

    (1991)
  • LeshnoM. et al.

    Neural Netw.

    (1993)
  • LuZ. et al.
  • AndreassenA. et al.

    Phys. Rev. D

    (2020)
  • HansenN.

    He CMA evolution strategy: A tutorial (2016)

    (2016)
  • CarrazzaS. et al.

    N3pdf/mcnntunes: mcnntunes 0.1.0

    (2020)
  • SjöstrandT. et al.

    Comput. Phys. Comm.

    (2015)
  • J. High Energy Phys.

    (2014)
  • Phys. Lett. B

    (2013)
  • CzarneckiW.M. et al.
  • Cited by (6)

    • Novel approaches in hadron spectroscopy

      2022, Progress in Particle and Nuclear Physics
      Citation Excerpt :

      The second factor is related to rapid and multifaceted development of architectures, algorithms and computational techniques like convolutional neural networks [86], rectified linear unit (ReLU) [87,88], improved stochastic gradient optimization [89] or batch normalization [90]. ML in nuclear and high-energy physics has already quite a long history: Applications to experimental studies include event selection [91–93], jet classification [94,95], track reconstruction [96,97], and event generation [98,99]. On the theory front, ML has been extensively used for fitting [70,71,100–102] and to provide model-independent parametrizations of structure and spectral functions [101–104], as well as of solutions of Schrödinger equations [105,106].

    The review of this paper was arranged by Prof. Z. Was.

    ☆☆

    This paper and its associated computer program are available via the Computer Physics Communication homepage on ScienceDirect (http://www.sciencedirect.com/science/journal/00104655)

    View full text