Transformers for jet tagging

19 Feb 2025, 13:50
20m
seminar hall (Building 3)

seminar hall

Building 3

KEK, Tsukuba campus

Speaker

Ahmed Hammad

Description

Attention-based transformer models have become increasingly prevalent in collider analysis, offering enhanced performance for tasks such as jet tagging. However, they are computationally intensive and require substantial data for training. we introduce a new jet classification network using an MLP mixer, where two subsequent MLP operations serve to transform particle and feature tokens over the jet constituents. The transformed particles are combined with subjet information using multi-head cross-attention so that the network is invariant under the permutation of the jet constituents. The network structure is closely related to the multiscale nature of HEP events.
The proposed network demonstrates comparable classification performance to state-of-the-art models while boosting computational efficiency drastically. The network structure can be applied to the various collider processes.

Presentation materials