Anomalies, representations, and self-supervision

Barry Dillon, Luigi Favaro, Friedrich Feiden, Tanmoy Modak, Tilman Plehn

Research output: Contribution to journalArticlepeer-review

Abstract

We develop a self-supervised method for density-based anomaly detection using contrastive learning, and test it using event-level anomaly data from CMS ADC2021. The AnomalyCLR technique is data-driven and uses augmentations of the background data to mimic non-Standard-Model events in a model-agnostic way. It uses a permutation-invariant Transformer Encoder architecture to map the objects measured in a collider event to the representation space, where the data augmentations define a representation space which is sensitive to potential anomalous features. An AutoEncoder trained on background representations then computes anomaly scores for a variety of signals in the representation space. With AnomalyCLR we find significant improvements on performance metrics for all signals when compared to the raw data baseline.
Original languageEnglish
Article number056
Pages (from-to)1-19
Number of pages19
JournalSciPost physics core
Volume7
Issue number3
Early online date16 Aug 2024
DOIs
Publication statusPublished online - 16 Aug 2024

Bibliographical note

Publisher Copyright:
Copyright B. M. Dillon et al.

Keywords

  • machine-learning
  • physics
  • self-supervised
  • LHC
  • anomaly detection
  • autoencoder
  • transformer

Fingerprint

Dive into the research topics of 'Anomalies, representations, and self-supervision'. Together they form a unique fingerprint.

Cite this