bachelor_thesis/talk/vortrag.org
2020-06-21 21:25:59 +02:00

12 KiB
Raw Blame History

What the heck should be in there. Let's draft up an outline. 20 minutes: bloody short, so just results

Intro   1_30m

Importance of MC Methods   SHORT

  • important tool in particle physics

    • not just numerical
  • also applications in stat. phys and lattice QCD
  • somewhat romantic: distilling information with entropy
  • interface with exp
  • precision predictions within, beyond sm
  • validation of new theories

    • some predictions are often more subtle than just the existense of new particles
    • backgrounds have to be substracted

Diphoton Process

  • feynman diags and reaction formula
  • higgs decay channel
  • dihiggs decay
  • pure QED

Calculation of the XS :TOO_LONG:   5m

Approach

  • formalism well separated from underlying theory

    • but can fool intuition (spin arguments)
    • in the course of semester: learned more about the theory :)
  • translating feynman diagrams to abstract matrix elements straight forward
  • first try: casimir's trick

    • error in calculation + one identity unknown
  • second try: evaluating the matrices directly

    • discovered a lot of tricks
    • error prone
  • back to studying the formalism: completeness relation for real photons

    • a matter of algebraic gymnastics
    • boils down to some trace and dirac matrix gymnastics
    • mixing terms cancel out, not zero in themselves
  • resulting expression for ME essentially t/u channel propagator (1/(t*u)) and spin correlation 1 + cos(x)^2
  • only angular dependencies, no kinematics, "nice" factors
  • symmetric in θ

Result + Sherpa

  • apply the golden rule for 2->2 processes
  • show plots and total xs
  • shape verified later -> we need sampling techniques first

Monte Carlo Methods   8m

  • one simple idea, can be exploited and refined
  • how to extract information from a totally unknown function

    • look at it -> random points are the most "symmetric" choice
    • statistics to the rescue
  • what does this have to do with minecraft
  • theory deals with truly random (uncorrelated) so that statistics apply, prng's cater to that: deterministic, efficient (we don't do crypto)

Integration

  • integration as mean value
  • convergence due to law of large numbers

    • independent of dimension
    • trivially parallelism
  • result normal distributed with σ due to central limit theorem
  • goal: speeding up convergence

    1. modify distribution
    2. integration variable
    3. subdivide integration volume
  • all those methods can be somewhat intertwined
  • focus on some simple methods

Naive Integration

  • why mix in that distribution: we choose it uniform
  • integral is mean
  • variance is variance of function: stddev linear in Volume!
  • include result
  • rediculous sample size
TODO compare to other numeric

Change of Variables

  • drastic improvement by transf. to η
  • only works by chance (more or less)

    • pseudo rapidity eats up angular divergence
  • can be shown: same effect as propability density
  • implementation is different

VEGAS

  • a simple ρ: step function on hypercubes, can be trivially generated
  • effectively subdividing the integration volume
  • optimal: same variance in every cube
  • easier to optimize: approximate optimal rho by step function
  • clarify: use rectangular grid and blank out unwated edges with θ function
  • nice feature: integrand does not have to be smooth :)
  • similar efficiency as the travo case

    • but a lot of room for parameter adjustment and tuning
TODO research the drawbacks that led to VEGAS
TODO nice visualization of vegas working
TODO look at original vegas
  • in 70s/80s memory a constraint

Sampling

  • why: generate events

    • same as exp. measurements
    • (includes statistical effects)
    • events can be "dressed" with more effects
  • usual case: we have access to uniformly distributed random values
  • task: convert this sample into a sample of another distribution
  • short: solve equation

Hit or Miss

  • we don't always know f, may have complicated (inexplicit) form
  • solve "by proxy": generate sample of g and accept with propability f/g
  • the closer g to f, the better the efficiency
  • simplest choice: flat upper bound
  • show results etc
  • one can optimize upper bound with VEGAS

Change of Variables

  • reduction of variance similar to integration
  • simplify or reduce variance
  • one removes the step of generating g-samples
  • show results etc
  • hard to automate, but intuition and 'general rules' may serve well

    • see later case with PDFs -> choose eta right away

Hit or Miss VEGAS

  • use scaled vegas distribution as g and to hit or miss
  • samples for g are trivial to generate
  • vegas again approximates optimal distribution
  • results etc
  • advantage: no function specific input
  • problem: isolated parts of the distribution can drag down efficiency

    • where the hypercube approx does not work well
    • especially at discontinuities
TODO add pic that i've sent Frank

Stratified Sampling

  • avoid global effects: subdivide integration interval and sample independently
  • first generate coarse samples and distribute them in the respective grid points
  • optimizing: make cubes with low efficiency small! -> VEGAS
  • this approach was used for the self-made event generator and improved the efficiency greatly (< 1% to 30%)
  • disadvantage: accuracies of upper bounds and grid weights has to be good

    • will come back to this

Observables

  • particle identities and kinematics determine final state
  • other observables can be calculated on a per-event base

    • as can be shown, this results in the correct distributions without knowledge of the Jacobian

Outlook

TODO Other modern Stuff

Toy Event Generator   3m

Basics   SHORT

  • just sampling the hard xs not realistic

    1. free quarks do not occur in nature
    2. hadron interaction more complicated in general
  • we address the first problem here
  • quarks in protons: no analytical bound state solution known so-far

Parton Density Functions

  • in leading order, high momentum limit: propability to encounter parton at some energy scale with some momentum fraction
  • can not be calcualated from first principles

  • scale has to be chosen appropriately: in deep inelastic scattering -> momentum transfer

    • p_T good choice
    • here s/2 (mean of t and u in this case)
  • xs formula
  • here LO fit and evolution of PDFs
TODO check s/2

Implementation

  • find xs in lab frame
  • impose more cuts

    • guarantee applicability of massless limit
    • satisfy experimental requirements
  • used vegas to integrate
  • cuts now more complicated because photons not back to back
  • apply stratified sampling variant along with VEGAS

    • 3 dimensions: x1, x2 (symmetric), η
    • use VEGAS to find grid, grid-weights and maxima
    • improve maxima by gradient ascend (usually very fast)
    • improve performance by cythonizing the xs and cut computation
    • sampling routines JIT compiled with numba, especially performant for loops and very easy
    • trivial parallelism through python multiprocessing
    • overestimating the maxima corrects for numerical maximization error
    • assumptions: mc found maximum and VEGAS weights are precise enough
  • most time consuming part: multidimensional implementation + debugging
  • along the way: validation of kinematics and PDF values through sherpa

Results

Integration with VEGAS

  • Python Tax: very slow, parallelism implemented, but omitted due to complications with the PDF library

    • also very inefficient memory management :P
  • result compatible with sherpa
  • that was the easy part

Sampling and Observables

  • observables:

    • usual: η and cosθ
  • p_t of one photon and invariant mass are more interesting
  • influence of PDF:

    • more weight to the central angles (see eta)
    • p_t cutoff due to cuts, very steep falloff due to pdf
    • same picture in inv mass
  • compatibilty problematic: just within acceptable limits

    • for p_t and inv mass: low statistic and very steep falloff
    • very sensitive to uncertainties of weights (can be improved by improving accuracy of VEGAS)
    • prompts a more rigorous study of uncertainties in the vegas step!

Pheno Stuff   2m

  • non LO effects completely neglected
  • sherpa generator allows to model some of them

    • always approximations

Short review of HO Effects

  • always introduce stage and effects along with the nice event picture

LO

  • same as toy generator

LO+PS

  • parton shower CSS (dipole) activated
  • radiation of gluons, and splitting into quarks -> shower like cascades QCD
  • as there are no QCD particles in FS: initial state radiation
  • due to 4-mom conservation: recoil momenta (and energies)

LO+PS+pT

  • beam remnants and primordial transverse momenta simulated
  • additinal radiation and parton showers
  • primordial p_T due to localization of quarks, modeled like gaussian distribution

    • mean, sigma: .8 GeV, standard values in sherpa
    • consistent with the notion of "fermi motion"

LO+PS+pT+Hadronization

  • AHADIC activated (cluster hadr)
  • jets of parton cluster into hadrons: non perturbative
  • models inspired by qcd but still just models
  • mainly affects isolation of photons (come back to that)
  • in sherpa, unstable are being decayed (using lookup tables) with correct kinematics

LO+PS+pT+Hadronization+MI

  • Multiple Interactions (AMISIC) turned on
  • no reason for just one single scattering in event
  • based on overlap of hadrons and the most important QCD scattering processes
  • in sherpa: shower corrections
  • generally more particles in FS, affects isolation

Presentation and Discussion of selected Histograms

pT of γγ system

  • Parton showers enhance at higher pT
  • intrinsic pT at lower pT (around 1GeV)
  • some isolation impact
  • but highest in phase space cuts

    • increase is almost one percent
    • pT recoils to the diphoton system usually substract pT from one photon -> harder to pass cuts -> amplified through big probability of low pT events!

pT of leading and sub-leading photon

  • shape similar to LO
  • first photon slight pT boost
  • second almost untouched

    • cut bias to select events that have little effect on sub-lead photon

Invariant Mass

  • events with lower m are allowed throgh cuts
  • events with very high recoil suppressed: colinear limit…

Angular Observables

  • mostly untouched
  • biggest difference: total xs and details
  • but LO gives good qualitative picture
  • reasonable, because LO should be dominating

Effects of Hadronization and MI

  • fiducial XS differs because of isolation and cuts in the phase space
  • we've seen: parton shower affect kinematics and thus the shape of observables and phase space cuts
  • isolation critera:

    • photon has to be isolated in detector
    • allow only certain amount of energy in cone around photon
    • force moinimum separation of photons to prevent cone overlap
  • Hadronization spreads out FS particles (decay kinematics) and produces particles like muons and neutrinos that aren't detectable or easily filtered out -> decrase in isolation toll
  • MI increases hadr activity in FS -> more events filtered out

Summary

  • LO gives qualitative picture
  • NLO affect observables shape, create new interesting observables
  • some NLO effects affect mainly the isolation
  • caveat: non-exhaustive, no QED radiation enabled

Wrap-Up

Summary

  • calculated XS
  • studied and applied simple MC methods
  • built a basic working event generator
  • looked at what lies beyond that simple generator

Lessons Learned (if any)

  • calculations have to be done verbose and explicit
  • spending time on tooling is OK
  • have to put more time into detailed diagnosis
  • event generators are marvelously complex
  • should have introduced the term importance sampling properly

Outlook

  • more effects
  • multi channel mc
  • better validation of vegas