This document summarizes a study of missing transverse energy (MET) resolution in minimum bias data collected by the ATLAS detector at the LHC. It finds that the MET resolution decreases with increasing numbers of proton-proton interactions (pileup) in an event. Specifically:
1) MET resolution is measured as a function of the total transverse energy in minimum bias events with 1 vertex.
2) MET resolution is also measured for events with 2, 3, and 4 vertices and found to decrease with increasing numbers of vertices.
3) Comparisons of skimmed and unskimmed data sets show inconsistencies, suggesting issues with the event selection cuts used to isolate pileup events. Further work is
Calculating transition amplitudes by variational quantum eigensolversQunaSys
This is our poster planned to be presented at APS March.
We proposed a method to calculate transition amplitudes between two orthogonal states on NISQ devices.
This work is a joint research between QunaSys and Mitsubishi Chemical Corporation.
Calculating transition amplitudes by variational quantum eigensolversTenninYan
This is our poster planned to presented at APS March.
We proposed a method to calculate transition amplitudes between two orthogonal states on NISQ devices.
Fermilab Muon g − 2 Experiment: Laser Based Gain Calibration SystemAtanu Nath
The Muon g-2 experiment at Fermilab (E989) is currently measuring the muon magnetic anomaly with a goal precision of 140 parts per billion, which will be a fourfold precision improvement over the current best measurement by the previous muon g-2 experiment at the Brookhaven Laboratory (BNL). The BNL-measured value of the muon magnetic anomaly and the corresponding Standard Model (SM) best estimate, differ by more than three standard deviation which inspired the current measurement as well as a theoretical drive for a significantly more precise calculation of the muon magnetic anomaly to rule out (or establish) statistical fluctuation as the origin of such a huge discrepancy. Stable central values along with 4-fold precision improvements in both theoretical (SM) and experimental fronts, would imply a ∼ 7σ discrepancy and that will be a clear hint of the physics beyond the Standard Model. Such unprecedented precision demands state-of-the-art technological improvements in all involved components to keep the systematic uncertainty below 70 ppb. This paper reports the current status of the E989 experiment after two years of data acquisition.
Calculating transition amplitudes by variational quantum eigensolversQunaSys
This is our poster planned to be presented at APS March.
We proposed a method to calculate transition amplitudes between two orthogonal states on NISQ devices.
This work is a joint research between QunaSys and Mitsubishi Chemical Corporation.
Calculating transition amplitudes by variational quantum eigensolversTenninYan
This is our poster planned to presented at APS March.
We proposed a method to calculate transition amplitudes between two orthogonal states on NISQ devices.
Fermilab Muon g − 2 Experiment: Laser Based Gain Calibration SystemAtanu Nath
The Muon g-2 experiment at Fermilab (E989) is currently measuring the muon magnetic anomaly with a goal precision of 140 parts per billion, which will be a fourfold precision improvement over the current best measurement by the previous muon g-2 experiment at the Brookhaven Laboratory (BNL). The BNL-measured value of the muon magnetic anomaly and the corresponding Standard Model (SM) best estimate, differ by more than three standard deviation which inspired the current measurement as well as a theoretical drive for a significantly more precise calculation of the muon magnetic anomaly to rule out (or establish) statistical fluctuation as the origin of such a huge discrepancy. Stable central values along with 4-fold precision improvements in both theoretical (SM) and experimental fronts, would imply a ∼ 7σ discrepancy and that will be a clear hint of the physics beyond the Standard Model. Such unprecedented precision demands state-of-the-art technological improvements in all involved components to keep the systematic uncertainty below 70 ppb. This paper reports the current status of the E989 experiment after two years of data acquisition.
Parity-Violating and Parity-Conserving Asymmetries in ep and eN Scattering in...Wouter Deconinck
Invited workshop presentation at the Amherst Center for Fundamental Interactions at UMass Amherst. This presentation includes the official Qweak results and discussion of unofficial beam normal single spin asymmetries.
This presentation is the introduction to Density Functional Theory, an essential computational approach used by Physicist and Quantum Chemist to study Solid State matter.
Measurement-induced long-distance entanglement of superconducting qubits usin...Ondrej Cernotik
Although superconducting systems provide a promising platform for quantum computing, their networking poses a challenge as they cannot be interfaced to light---the medium used to send quantum signals through channels at room temperature. We show that mechanical oscillators can mediated such coupling and light can be used to measure the joint state of two distant qubits. The measurement provides information on the total spin of the two qubits such that entangled qubit states can be postselected. Entanglement generation is possible without ground-state cooling of the mechanical oscillators for systems with optomechanical cooperativity moderately larger than unity; in addition, our setup tolerates a substantial transmission loss. The approach is scalable to generation of multipartite entanglement and represents a crucial step towards quantum networks with superconducting circuits.
Introduction to Quantum Computing & Quantum Information TheoryRahul Mee
Note:This is just presentation created for study purpose.
This comprehensive introduction to the field offers a thorough exposition of quantum computing and the underlying concepts of quantum physics.
In this work Predestination of Particles Wavering Search (PPS) algorithm has been applied to solve optimal reactive power problem. PPS algorithm has been modeled based on the motion of the particles in the exploration space. Normally the movement of the particle is based on gradient and swarming motion. Particles are permitted to progress in steady velocity in gradient-based progress, but when the outcome is poor when compared to previous upshot, immediately particle rapidity will be upturned with semi of the magnitude and it will help to reach local optimal solution and it is expressed as wavering movement. In standard IEEE 14, 30, 57,118,300 bus systems Proposed Predestination of Particles Wavering Search (PPS) algorithm is evaluated and simulation results show the PPS reduced the power loss efficiently.
What is Quantum Computing
What is Quantum bits (Qubit)
What is Reversible Logic gates and Logic Circuits
What is Quantum Neuron (Quron)
What are the methods of implementing ANN using Quantum computing
UCSD NANO 266 Quantum Mechanical Modelling of Materials and Nanostructures is a graduate class that provides students with a highly practical introduction to the application of first principles quantum mechanical simulations to model, understand and predict the properties of materials and nano-structures. The syllabus includes: a brief introduction to quantum mechanics and the Hartree-Fock and density functional theory (DFT) formulations; practical simulation considerations such as convergence, selection of the appropriate functional and parameters; interpretation of the results from simulations, including the limits of accuracy of each method. Several lab sessions provide students with hands-on experience in the conduct of simulations. A key aspect of the course is in the use of programming to facilitate calculations and analysis.
Parity-Violating and Parity-Conserving Asymmetries in ep and eN Scattering in...Wouter Deconinck
Invited workshop presentation at the Amherst Center for Fundamental Interactions at UMass Amherst. This presentation includes the official Qweak results and discussion of unofficial beam normal single spin asymmetries.
This presentation is the introduction to Density Functional Theory, an essential computational approach used by Physicist and Quantum Chemist to study Solid State matter.
Measurement-induced long-distance entanglement of superconducting qubits usin...Ondrej Cernotik
Although superconducting systems provide a promising platform for quantum computing, their networking poses a challenge as they cannot be interfaced to light---the medium used to send quantum signals through channels at room temperature. We show that mechanical oscillators can mediated such coupling and light can be used to measure the joint state of two distant qubits. The measurement provides information on the total spin of the two qubits such that entangled qubit states can be postselected. Entanglement generation is possible without ground-state cooling of the mechanical oscillators for systems with optomechanical cooperativity moderately larger than unity; in addition, our setup tolerates a substantial transmission loss. The approach is scalable to generation of multipartite entanglement and represents a crucial step towards quantum networks with superconducting circuits.
Introduction to Quantum Computing & Quantum Information TheoryRahul Mee
Note:This is just presentation created for study purpose.
This comprehensive introduction to the field offers a thorough exposition of quantum computing and the underlying concepts of quantum physics.
In this work Predestination of Particles Wavering Search (PPS) algorithm has been applied to solve optimal reactive power problem. PPS algorithm has been modeled based on the motion of the particles in the exploration space. Normally the movement of the particle is based on gradient and swarming motion. Particles are permitted to progress in steady velocity in gradient-based progress, but when the outcome is poor when compared to previous upshot, immediately particle rapidity will be upturned with semi of the magnitude and it will help to reach local optimal solution and it is expressed as wavering movement. In standard IEEE 14, 30, 57,118,300 bus systems Proposed Predestination of Particles Wavering Search (PPS) algorithm is evaluated and simulation results show the PPS reduced the power loss efficiently.
What is Quantum Computing
What is Quantum bits (Qubit)
What is Reversible Logic gates and Logic Circuits
What is Quantum Neuron (Quron)
What are the methods of implementing ANN using Quantum computing
UCSD NANO 266 Quantum Mechanical Modelling of Materials and Nanostructures is a graduate class that provides students with a highly practical introduction to the application of first principles quantum mechanical simulations to model, understand and predict the properties of materials and nano-structures. The syllabus includes: a brief introduction to quantum mechanics and the Hartree-Fock and density functional theory (DFT) formulations; practical simulation considerations such as convergence, selection of the appropriate functional and parameters; interpretation of the results from simulations, including the limits of accuracy of each method. Several lab sessions provide students with hands-on experience in the conduct of simulations. A key aspect of the course is in the use of programming to facilitate calculations and analysis.
Artificial intelligence (AI) is everywhere, promising self-driving cars, medical breakthroughs, and new ways of working. But how do you separate hype from reality? How can your company apply AI to solve real business problems?
Here’s what AI learnings your business should keep in mind for 2017.
Study: The Future of VR, AR and Self-Driving CarsLinkedIn
We asked LinkedIn members worldwide about their levels of interest in the latest wave of technology: whether they’re using wearables, and whether they intend to buy self-driving cars and VR headsets as they become available. We asked them too about their attitudes to technology and to the growing role of Artificial Intelligence (AI) in the devices that they use. The answers were fascinating – and in many cases, surprising.
This SlideShare explores the full results of this study, including detailed market-by-market breakdowns of intention levels for each technology – and how attitudes change with age, location and seniority level. If you’re marketing a tech brand – or planning to use VR and wearables to reach a professional audience – then these are insights you won’t want to miss.
Investigation of repeated blasts at Aitik mine using waveform cross correlationIvan Kitov
We present results of signal detection from repeated events at the Aitik and Kiruna mines in Sweden as based on waveform cross correlation. Several advanced methods based on tensor Singular Value Decomposition is applied to waveforms measured at seismic array ARCES, which consists of three-component sensors.
Electrocardiogram Denoised Signal by Discrete Wavelet Transform and Continuou...CSCJournals
One of commonest problems in electrocardiogram (ECG) signal processing is denoising. In this paper a denoising technique based on discrete wavelet transform (DWT) has been developed. To evaluate proposed technique, we compare it to continuous wavelet transform (CWT). Performance evaluation uses parameters like mean square error (MSE) and signal to noise ratio (SNR) computations show that the proposed technique out performs the CWT.
FPGA Design & Simulation Modeling of Baseband Data Transmission SystemIOSR Journals
Abstract: This paper describes a study on a baseband data transmission system developed for undergraduate
students studying communication engineering. Theoretical material, developed in the lectures, is briefly
covered. A practical system is presented with pre-detection filtering being employed to improve the bit error
rate. A simulation of the complete system is carried out on a Sun work station using the MATLAB simulation
package. Simulation and theoretical results are compared.
Diagnosis of Faulty Sensors in Antenna Array using Hybrid Differential Evolut...IJECEIAES
In this work, differential evolution based compressive sensing technique for detection of faulty sensors in linear arrays has been presented. This algorithm starts from taking the linear measurements of the power pattern generated by the array under test. The difference between the collected compressive measurements and measured healthy array field pattern is minimized using a hybrid differential evolution (DE). In the proposed method, the slow convergence of DE based compressed sensing technique is accelerated with the help of parallel coordinate decent algorithm (PCD). The combination of DE with PCD makes the minimization faster and precise. Simulation results validate the performance to detect faulty sensors from a small number of measurements.
MET Energy Resolution in Pileup Minimum Bias Events using 7 TeV LHC Data
1. Missing ET Resolution in Minimum Bias data at 7 TeV
K.Wang & R. Keeler
University of Victoria
*http://www.phys.ufl.edu/~matchev/LHCJC/cern-lhc.jpg
s
2. Outline
1.1 Introductions - LHC
1.2 Introduction - ATLAS
2. Missing Energy
3. Minimum Bias
3.1 Data
3.2 Run 153565
3.3 Rates
3.4 1 Vertex MET Resolution
4. Pile up
4.1 Definitions
4.2 Rates
4.3 In Data
5. MET Resolution
6. Comparing Datasets
7. Further Work
8. Conclusions
3. 1.1 Introduction - LHC
CERN – European Organization for Nuclear Research, formed in 1954
Important source of Particle physics research ever since
Constructed in the tunnel originally used for LEP the LHC is the most powerful particle
accelerator built to date. Center of mass energies in hadron collisions to reach 14 TeV center
of mass energy, with peak luminosity 1034
cm-2
s-1
The success of many hypothetical extensions to the Standard Model rests on the success of
the LHC
As of March 2010 7 TeV collisions have occurred
3.5 pb-1
of real data is now available for analysis
Parameter Performance Goals Run Analyzed Current August
CM energy 14 TeV 7 TeV 7 TeV
Peak Luminosity 1034
cm-2
s-1
0.014*1030
cm-2
s-1
6.7*1030
cm-2
s-1
Bunch Pairs 2808 2 48
Bunch Spacing 25 ns 45 μs Trains
Beam Current 584 mA ~0.05 mA ~8.5 mA
4. 1.2 Introduction - ATLAS
General purpose spectrometer of proton-proton collisions
The detector is designed to fulfill many objectives – it is a high precision test of qcd, flavor
physics and electroweak theory. In addition it is of course, searching for evidence of new
physics, Higgs, SUSY, Heavy gauge bosons...
The parameters of the LHC puts stringent requirements on its detectors
Fast electronics, high
granularity
High hermiticity (angular
coverage)
Good particle momentum
resolution and reconstruction
efficiency
Efficient triggering on low
momentum objects while
rejecting background
Good particle identification
*http://www.slipperybrick.com/wp-
content/uploads/2006/12/atlas-detector.jpg
8. 2 Missing Energy
An interaction occurs in the ATLAS detector, the final state particles diverge outwards from the
interaction point
The particles are seen by the calorimetry and the triggering system records it – this is now an
event
Sum of final state vector momenta from an interaction
Given that for highly energetic events
The resultant vector tells you about what your detector missed and about its orientation
If our detector was 100% efficient and had perfect resolution MET should be 0.
A good measurement of missing transverse energy is crucial to many studies. In particular
searches for new physics (SUSY, extra dimensions) all depend a great deal on very good MET
measurements for a given interaction.
True missing energy is of course the smoking gun of new physics
E2
=p2
c2
m2
c4
≈p2
c2
9. 2 Missing Energy
In minimum bias data MET is reconstructed using the calorimeters with corrections for
muons
Ex
miss
=−i=1
N cell
Ei sin icos i
Ey
miss
=−i=1
N cell
Ei sin icos i
Emiss
=Ex
miss
2
Ey
miss
2
Due to granularity (calorimeter has ~187 000 cells), it is vital to suppress noise contributions.
The sum over Ncell
is restricted to topological clusters which have received hits, there is an
algorithm that determines what goes into Ncell
based on noise levels
Cells flagged by ATLAS as noisy are also not included.
10. 2 Missing Energy
The ntuple variables we used to
analyze MET are “refined final” -
i.e. MET_RefFinal_et
These are the recommended
analysis variables
The MET variables in ATLAS are
reconstructed in a step by step
process.
First the muon contribution is
obtained from the muon
spectrometer reconstruction
Then there is a correction for
energy losses in the cryostat
Finally the contributions from each
reconstructed particle in the
collision are added on to create the
final term
The contributions of MET_RefFinal,
which we used are shown on the
right
MET_Ex(y)MissRefEle Contribution from cells in
electrons
MET_Ex(y)MissRefGamma Contribution from cells in
gamma
MET_Ex(y)MissRefTau Contribution from cells in tau-
jets
MET_Ex(y)MissRefJet Contribution from cells inside
jets
MET_Ex(y)MissRefMuon Contribution from cells in
calorimeters crossed by
muons
MET_Ex(y)MissCellOut Contribution from cells in
Topo outside objects
MET_Ex(y)MissMuonBoy Contribution from the Staco
muon collection
MET_Ex(y)MissCryo Cryostat correction from
ConeTopoJets
= =
MET_RefFinal Linear sum of above
*Atlas TWiki
11. We can parametrize the MET resolution in terms of the total energy deposited in the
detector.
This is the sum of the scalars of the energy.
The d3pd variable is MET_RefFinal_sumet
Its construction is similar to that for MET_RefFinal_et.
Following the same procedure used for
MET_RefFinal components above
Specifically then in terms of the physical
accelerator the scalar sum of the missing
transverse energy is
ET=i=1
Ncell
Ei sini
2 Missing Energy
12. 3 Minimum Bias Data
Minimum bias data is data taken with an inclusive trigger
It is called minimum bias data because the goal is to trigger on and record any kind of p-p
interaction with no bias
In ATLAS the minimum bias events can be triggered by the dedicated Minimum Bias Trigger
Scintillators (MBTS) in coincidence with the Beam Pickup Timing Device (BPTX).
The BPTX are mounted 175 m away from the bunch crossing point in both directions in the
ATLAS detector.
The BPTX allows the observer to see the bunches as they pass towards the collision point,
and thereby allows monitoring of the beam.
More importantly for us, the BPTX provides the L1 trigger with accurate timing references
whenever a bunch passes it. This is used as a triggering condition for minimum bias events.
The MBTS are mounted on the Liquid Argon Endcap cryostats, with a coverage of 2.1<|η|
<3.8
This consists of two sets of 16 scintillator counters
The MBTS will only be functional for a short time at low luminosity as it is expected radiation
damage will render them useless within several months of beam collisions
14. TOT=ESDDDNDCD
3 Minimum Bias Data
MB=NDDDSD
The usual experiment definition of minimum bias cross section is Non-Single diffractive
events (that is Non-diffractive inelastic and double diffractive events) but ATLAS is sensitive
also to single diffractive events.
The expected cross section for NSD is ~65 mb if we include single diffractive events this is
~90 mb.
The number of events as seen by the minimum triggers for the run we looked at is
21,282,442 and for the operating time this gives a rate of 148.417 Hz.
We should note there is always some bias in the data taking. In practice the detector will
probably not see glancing blows, or particles that just go down the beam pipe.
Cross Section (mb) – (Pythia 6.205)
Elastic 22
Single Diffractive 28
Double Diffractive 10
Non-Diffractive Inelastic 55
Central Diffractive ~
15. 3.1 Data
Run 153565 with ~21 million events
The original raw data is reconstructed, 21 282 442 events this is the number seen by the minimum bias
trigger, into
group10.perf-
jets.data10_7TeV.00153565.physics_MinBias.recon.ESD.f251_JetEtMissDPDModifier000016.v4_E
XT0
A skim was made on the RAW for pileup events, and then reconstructed with 89334 events -
group10.perf-
jets.data10_7TeV.00153565.physics_MinBias_extractedBS.merge.ESD.o5_r1297_p161_JetEtMiss
DPDModifier000025.v1_EXT0
Monte carlo of minbias+pileup events do not yet exist, but has been requested
The unskimmed raw data was processed according to the following criteria for events with multiple p-p
interactions – primary vertices >1* -
- a primary vertex (PV) must be compatible with the beam spot
- each PV must have more than 7 tracks with pT > 150 MeV;
- a PV must not have more than 25% of outlier tracks (to reject overlapping
vertices).
This skimmed data consists almost exclusively of pileup events, in practice there are 65 1 vertex events
that appear in the skimmed data
*https://twiki.cern.ch/twiki/bin/view/Atlas/InterestingSkims
17. 3.3 Minimum Bias – Rates
In a vacuum, the rate at the LHC is
The rate seen by the minimum bias trigger is
Also we know the number of bunches in
the detector. The bunch crossing rate is
We can then calculate the events/
bunch crossing rate from the minbias
rate
This is now properly – the rate of minimum bias events per bunch crossing as seen by
the minbias trigger.
If we take the predicted minimum bias cross section as ~90 mb then the total number
of minimum bias events predicted with a total run luminosity of 0.78 nb-1
is
Why the discrepancy with what the trigger sees? We know the total luminosity 0.78
nb-1
from machine monitors, so if we worked backwards to obtain the cross section we
find
A question of efficiency?
R=pp L
Events
Time
=148.417 Hz
N=∫ Ldt=9∗10−26
×7.8∗1032
=7.02∗107
Circumference
velocity of bunches
×bunch pairs≈
27 km2
c
=22.22kHz
minbiasrate
bunchrate
=0.0067
events
BC
minbias≈29mb
18. 3.4 Minimum Bias Data
With the minimum bias data we can plot the MET Resolution plotted against the absolute
value of the sum of the transverse energy deposited into the detector (~essentially the
strength of the interaction)
19.
20.
21.
22. 4 Pileup - Definitions
We distinguish between in-time pileup and out-of-time pileup
We imagine as proton bunches circle around in the accelerator in any given collision we may
have more then one pair of protons colliding, this is in time pileup
Alternatively if multiple proton bunches are traveling in the accelerator and the bunch spacing
is small, the first collision will occur within the detection chamber. As the product particles
stream outwards the next collision (proton bunches trailing behind) occurs before the initial
post-collision particles can fully evacuate the chamber. This is out-of-time pileup.
In this analysis we are strictly concerned with in time pile up. There should be no out-of-time
pileup due to the spacing between the bunches.
What constitutes a pileup event is determined by the number of vertices. We consider any
two events with more then two vertices as being pileup (excluding decay)
To distinguish between vertices due to additional collisions and vertices due to decay, there
is a D3PD variable called vxp_type. Pileup vertices have vxp_type=3.
In addition there is a dummy vertex in every event, these we ignore.
23.
24. 4.1 Pileup
We expect the number of vertices in a collision to go as a poisson distribution
Examining the skimmed data, which contains all pile up events that passed the appropriate
cuts we can fit the number of 2, 3, 4 vertices events to a Poisson distribution
If we extrapolate back to 1 vertex events, there would presumably be (2.78±0.03)*106
- that
passed those same cuts
P=A
n
e−
n!
A=4.71±0.05×107
=6.29±0.03×10−2
Vertices Events
2 87399
3 1841
4 29
28. 5 MET Resolution
We can quantitatively analyze the difference between the MET resolutions with respect to the
number of vertices by approximating the statistics with an acceptable function.
This does become problematic if one does not have enough statistics
MET _ RefFinal _ et=A[ MET _ RefFinal _ sumet ]
Vertex A α (10-1
)
1 18.00±0.04 5.030±0.004
2 17.1±0.8 5.06±0.04
3 6±3 5.9±0.4
4 (2.8±0.1)*103
1±2
29.
30. 6 Comparing the Datasets
In order to understand the effects of multi-interaction events with respect to single
interaction events we have to make sure the skimmed and unskimmed data sets are
actually consistent with one another
Any comparison that we make then to the original data set (unskimmed) wouldn't make
sense unless we applied the same cuts.
For example – if we use the Poisson distribution to extrapolate back to however many 1
vertex events are expected – that number would only correspond to the number of 1 vertex
events that passed the cuts above and not the complete number of 1 events in the
unskimmed data.
We have to make sure the two data sets (skimmed and unskimmed) are consistent.
31. 6 Comparing the Datasets
How do we do this?
The skimmed data should be completely contained within the original unskimmed data
Let's work with a few variables (in principle any number of variables present in both d3pds
should do)
1. Number of Tracks per Vertex (vxp_nTracks)
2. Sum of transverse momentum per Vertex (vxp_sumPt)
3. Sum of energy per vertex (vxp_E)
We want to know if the vertices we find in the skimmed data are consistent with what is in the
unskimmed data.
For the skimmed data create 3 plots which consist of 1, 2 and 3 of the above but per event
rather then per vertex (sum up the vertices for every event).
For the unskimmed data create plots of 1, 2 and 3 as they are.
Generate random numbers according to the unskimmed distributions above in accordance to
the vertex information in the skimmed data ,creating generated plots of 1, 2 and 3 per event
If everything was ideal the generated distributions of 1 2 and 3 per event should match the real
distributions created from the skimmed data
32.
33.
34.
35. 6.1 Problems
Obviously this doesn't work perfectly well...yet
What is wrong? This tells us our cuts are not correct.
Also the above consistency test was not made on the full unskimmed data set but only a
portion of them (so the first thing to try is to run over all the data!)
The easiest way to solve this is to find the person who did the reconstruction and ask
him/her!
So far no luck in that department, but also have not tried very hard
But since the data was skimmed from the RAW it's possible the reconstructed d3pds lack the
necessary tools for skimming
The stupid way of doing this would be to skim the original ~21 million d3pd dataset for pileup
events, then try to match variables in the JetETMiss skim to our skim, presumably one would
see signs of cuts
For example in the JetETMiss d3pd one may find there are no vertices with less 7 tracks,
indicating that such a cut has to be made the unskimmed data.
Rinse and repeat this until our skim is reduced to ~89k pileup events. Then reskim the ~21
million events, this time apply the pileup cuts to 1 vertex events
In this manner we could then obtain all 1 vertex events cut according to the pileup criteria
36. 7 Life is a work in progress
Obtain all 1 vertex events that pass the skimmed dataset cuts (Make correct cleaning and
selection cuts as prescribed by JetEtMiss Group)
Monte Carlo of minbias+pileup does not yet exist. Once these are made we can compare
real data results to simulations
If we can correctly apply the cuts to the original data, a Kolomogorov-Smirnov test is the
quantitative description of the probability that two distributions originate from the one and
same underlying distribution. This applies to the consistency tests in the previous couple of
slides.
Look at more data?
37. 8 Conclusions
We have looked at the MET resolution in 7 TeV minimum bias data coming from ATLAS with
respect to the effects of pileup
We validated the consistency of the data by examining various physical parameters and
checking they made sense
The form of the MET resolution as parametrized by total transverse energy output into the
detector does not seem to change with increasing pileup -
Although at greater then 3 vertices we do not have enough statistics to say for certain
As a function of total energy deposited in the calorimeter, the MET resolution for 1, 2 and 3
vertex seem to go roughly as the square root of the former.
We predicted the average number of interactions from the skimmed minimum bias data to
be-
Predictions of minimum bias cross section from MC do not seem to match the cross section
seen by the trigger system. There is probably a question of efficiency here, this needs to be
looked at more carefully.
Monte Carlo data has yet to be examined but would prove a useful check on these results
For 1 vertex events our data analysis methodology is not yet fully formulate and needs more
corrections for selection and cleaning.
6.29±0.03×10−2
39. References
ATLAS Collaboration. "Data-Quality Requirements and Event Cleaning for Jets and Missing
Transverse Energy Reconstruction with the ATLAS Detector in Proton-Proton Collisions at a
Center-of-Mass Energy of Sqrt(s)=7 TeV." ATLAS NOTES (2010). Print.
ATLAS Collaboration. "Measurement of Missing Transverse Energy." ATLAS NOTES (2009).
Print.
ATLAS Collaboration. "Performance of the Missing Transverse Energy Reconstruction in
Minimum Bias Collision at Center-of-mass Energy of Sqrt(s) 7 TeV with the ATLAS
Detector." ATLAS NOTES (2010). Print.
Dawson, I., and K. Prokofiev. "Determining the Inelastic Proton-proton Cross Section at the
Large Hadron Collider Using Minimum Bias Events." Print.
Moraes, A., C. Butter, and I. Dawson. "Prediction for Minimum Bias and the Underlying Event
at LHC Energies." The European Journal of Physics C 50 (2007): 435-66. Print.