19–20 Jun 2024
Uni Mail - University of Geneva
Europe/Zurich timezone

Masked particle modelling

19 Jun 2024, 11:25
12m
MR060

MR060

Speaker

Samuel Byrne Klein (Universite de Geneve (CH))

Description

The Bert pretraining paradigm has proven to be highly effective in many domains including natural language processing, image processing and biology. To apply the Bert paradigm the data needs to be described as a set of tokens, and each token needs to be labelled. To date the Bert paradigm has not been explored in the context of HEP. The samples that form the data used in HEP can be described as a set of particles (tokens) where each particle is represented as a continuous vector. We explore different approaches for discretising/labelling particles such that the Bert pretraining can be performed and demonstrate the utility of the resulting pretrained models on common downstream HEP tasks.

Authors

Samuel Byrne Klein (Universite de Geneve (CH)) Tobias Golling (Universite de Geneve (CH))

Presentation materials