OmniLearn: Facilitating All Jet Physics Tasks
1
Vinicius M. Mikuni
vmikuni@lbl.gov
vinicius-mikuni
Foundational Models
2
Option 1: Human language is the communication medium between the user and the machine
Foundational Models
3
Option 1: Human language is the communication medium between the user and the machine
Foundational Models
4
Option 1: Human language is the communication medium between the user and the machine
Foundational Models
5
Option 2: Data is the communication medium between the user and the machine
Foundational Models
6
Option 2: Data is the communication medium between the user and the machine
Foundational Models
7
Option 2: Data is the communication medium between the user and the machine
Jets
8
Jets are the most common signatures at the LHC
Jets
9
Measurements
Searches
Tagging
Jets
How to teach AI about jets?
10
Encoding jet information
Create a neural network model that aims to accomplish 2 tasks:
11
Diffusion 101
Diffusion models are the go to for data generation
12
Encoding jet information
Create a neural network model that aims to accomplish 2 tasks:
13
Encoding jet information
Point-Edge Transformer (PET)
14
+
Input Dropout
Not all datasets contain the same information:
15
More details at: https://arxiv.org/abs/2404.16091
f1, f2, f3, f4, f5, f6, f7, f8
f5, f6, f7, f8
0,0,0,0
p = 0.9
p = 0.1
Comparison Between Models
Language inspired models
OmniLearn
16
Training
JetClass dataset used for training
Use the pre-trained model as the starting point and fine-tune using different datasets
17
Huilin Qu, Congqiao Li, Sitian Qian, arXiv:2202.03772
Evaluation
2 different jet categories, AK8 jets simulated in pp collisions with Madgraph + Pythia8 with ATLAS Delphes detector simulation
18
Better than all non-fine-tuned models and similar to PartT performance
Evaluation datasets: 1
Evaluation
2 different jet categories, AK4 jets simulated in pp collisions with Madgraph + Pythia8 with CMS Delphes detector simulation
19
Better than all non-fine-tuned models and similar to PartT performance
Evaluation datasets: 2
Evaluation
20
Evaluation datasets: 2
Faster training and better convergence
Evaluation
2 different jet categories, AK5 jets simulated in pp collisions with Pythia6 with Geant4 Simulation + CMS Particle flow reconstruction
21
Evaluation datasets: 3
Evaluation
2 different jet categories, AK10 jets simulated in ep collisions with Rapgap with Geant3 Simulation + H1 Particle flow reconstruction
22
Evaluation datasets: 4
Jet Generation
23
Evaluation datasets: 6
Great generation quality across multiple metrics
Application Highlight
24
“
FastSim to FullSim
25
Evaluation datasets: 7
OmniLearn is trained on cheap Delphes simulations. Can we fine-tune to Run 2 ATLAS Full simulation + Reconstruction?
Unfolding
26
What we measure
What we want
OmniFold
27
Source: Andreassen et al. PRL 124, 182001 (2020)
2-step iterative process
ATLAS OmniFold analysis
28
OmniFold dataset consisting of Z(𝜈𝜈) + Jets events. Unfold the particles directly and then build the jet observables
Evaluation datasets: 8
Unfolding
29
Evaluation datasets: 8
Unbinned Unfolding using the OmniFold workflow. More precise than traditional unfolding and more efficient than previous ML models
Anomaly Detection
30
Evaluation datasets: 9
Bump-hunting using ML:
Anomaly Detection
31
Evaluation datasets: 9
Bump-hunting using ML:
LHCO dataset
32
LHCO R&D dataset
Evaluation datasets: 9
Anomaly Detection
33
Evaluation datasets: 9
SIC = Significance Improvement Curve (TPR/sqrt(FPR) vs TPR) “By how much can I improve the significance of a particular signal given an initial significance.”
Anomaly Detection
34
Evaluation datasets: 9
Previous results were limited by the amount of data in the SR: Only sensitive to NP when S/B > 3% ~ 4𝜎
OmniLearn founds the NP with S/B = 0.7% ~ 2𝜎
Conclusion
35
THANKS!
Any questions?
36
Backup
37
ATLAS Loss Curves
38
OmniLearn for reweighting
39
OmniLearn for Unfolding
40
PET
Train one model that learns to classify and generate jets
41
More details at: https://arxiv.org/abs/2404.16091
Diffusion Generative Models
42
Loss function
Straightforward loss function:
43
More details at: https://arxiv.org/abs/2404.16091