Speaker
Description
Attention-based transformer models have become increasingly prevalent in collider analysis, offering enhanced performance for tasks such as jet tagging. However, they are computationally intensive and require substantial data for training. In this paper, we introduce a new jet classification network using an MLP mixer, where two subsequent MLP operations serve to transform particle and feature tokens over the jet constituents. The transformed particles are combined with subjet information using multi-head cross-attention so that the network is invariant under the permutation of the jet constituents. The network structure is closely related to the multiscale nature of HEP events.
The proposed network demonstrates comparable classification performance to state-of-the-art models while boosting computational efficiency drastically. The network structure can be applied to the various collider processes.