This document summarizes Juan Rojo's presentation on a new method for combining PDF sets called compressed Monte Carlo PDFs (CMC-PDFs). The method involves combining Monte Carlo replicas from different PDF sets, then compressing the large combined set into a smaller set that still accurately reproduces properties like uncertainties. Validation shows CMC-PDFs with only 25 replicas can reproduce uncertainties for a variety of LHC processes, providing a computationally efficient implementation of the PDF4LHC recommendation.
Parton Distributions at a 100 TeV Hadron Colliderjuanrojochacon
This document discusses parton distributions for a potential future 100 TeV hadron collider. It begins with an outline of topics to be covered, including the increased kinematic coverage from 14 TeV to 100 TeV, PDF luminosities over this energy range, and tools needed for PDF studies at 100 TeV. The document then examines the kinematic coverage at 100 TeV in detail, showing how it probes PDFs at smaller x values and allows access to higher mass scales. Plots demonstrate that PDF luminosities increase much more rapidly from 14 to 100 TeV for higher mass states. The discussion covers tools needed for FCC PDF studies, including behavior of PDFs at large-x, inclusion of a top quark PDF,
Non-Gaussian Methods for Learning Linear Structural Equation Models: Part IShiga University, RIKEN
This document provides an overview of a tutorial on non-Gaussian methods for learning linear structural equation models (SEMs). The tutorial will cover how linear SEMs can be used to model data generating processes and review a new approach that utilizes non-Gaussianity of data for model identification. The tutorial is divided into two parts, with the first part providing an overview of linear SEMs and the identifiability problems of conventional methods. The second part will discuss recent advances in applying these non-Gaussian methods to time series data and models with latent confounders.
1. The document discusses estimating causal direction between two variables in the presence of hidden common causes.
2. A key challenge is that the hidden common causes introduce dependence between the error terms, making regression coefficients an unreliable guide to causal direction.
3. The author proposes a non-Gaussian structural equation model that can estimate causal direction without specifying the number of hidden common causes by exploiting the fact that different causal directions imply different distributions over the data, even when the error terms are dependent.
A non-Gaussian model for causal discovery in the presence of hidden common ca...Shiga University, RIKEN
1) The document proposes a Bayesian linear non-Gaussian structural equation model (SEM) approach for estimating causal direction between observed variables in the presence of hidden common causes.
2) Rather than explicitly model the hidden common causes, the approach transforms the model to one without hidden causes by including observation-specific intercepts representing the sums of hidden causes.
3) The approach compares marginal likelihoods of the transformed models under different causal directions to select the most likely direction, without needing to specify the number or distributions of hidden causes. It was shown to successfully estimate causal directions on a sociology data set.
Image compression using dpcm with lms algorithm ranbeerRanbeer Tyagi
The document discusses image compression using differential pulse code modulation (DPCM) with a least mean square (LMS) algorithm. It begins with an introduction to image compression and DPCM. It then provides details on the LMS algorithm and how it can be used with DPCM to provide almost 2 bits per pixel reduction in transmission rate compared to standard DPCM while maintaining similar distortion levels. The document presents simulation results showing the performance of DPCM and DPCM with LMS on an image in terms of distortion, histogram, and prediction mean square error. It concludes that LMS has lower computational complexity than DPCM and can achieve better compression.
That presentation explains everything you always wanted to know about JPEG 2000.
All the benefits, how it works, what applications, how to implement JPEG2000.
Particularly for Digital Cinema, Broadcast Contribution, Video Archiving, Post-production.
Image compression uses four stages to reduce file sizes: 1) transforming RGB pixels to the YCbCr color space, 2) applying discrete cosine transformation to concentrate pixel data into a few matrix elements, 3) quantizing pixel values to reduce their range, and 4) using Huffman coding to assign shorter bit codes to more common pixel values. This allows JPEG images to be compressed to smaller file sizes for storage and transmission while still maintaining good visual quality.
Parton Distributions at a 100 TeV Hadron Colliderjuanrojochacon
This document discusses parton distributions for a potential future 100 TeV hadron collider. It begins with an outline of topics to be covered, including the increased kinematic coverage from 14 TeV to 100 TeV, PDF luminosities over this energy range, and tools needed for PDF studies at 100 TeV. The document then examines the kinematic coverage at 100 TeV in detail, showing how it probes PDFs at smaller x values and allows access to higher mass scales. Plots demonstrate that PDF luminosities increase much more rapidly from 14 to 100 TeV for higher mass states. The discussion covers tools needed for FCC PDF studies, including behavior of PDFs at large-x, inclusion of a top quark PDF,
Non-Gaussian Methods for Learning Linear Structural Equation Models: Part IShiga University, RIKEN
This document provides an overview of a tutorial on non-Gaussian methods for learning linear structural equation models (SEMs). The tutorial will cover how linear SEMs can be used to model data generating processes and review a new approach that utilizes non-Gaussianity of data for model identification. The tutorial is divided into two parts, with the first part providing an overview of linear SEMs and the identifiability problems of conventional methods. The second part will discuss recent advances in applying these non-Gaussian methods to time series data and models with latent confounders.
1. The document discusses estimating causal direction between two variables in the presence of hidden common causes.
2. A key challenge is that the hidden common causes introduce dependence between the error terms, making regression coefficients an unreliable guide to causal direction.
3. The author proposes a non-Gaussian structural equation model that can estimate causal direction without specifying the number of hidden common causes by exploiting the fact that different causal directions imply different distributions over the data, even when the error terms are dependent.
A non-Gaussian model for causal discovery in the presence of hidden common ca...Shiga University, RIKEN
1) The document proposes a Bayesian linear non-Gaussian structural equation model (SEM) approach for estimating causal direction between observed variables in the presence of hidden common causes.
2) Rather than explicitly model the hidden common causes, the approach transforms the model to one without hidden causes by including observation-specific intercepts representing the sums of hidden causes.
3) The approach compares marginal likelihoods of the transformed models under different causal directions to select the most likely direction, without needing to specify the number or distributions of hidden causes. It was shown to successfully estimate causal directions on a sociology data set.
Image compression using dpcm with lms algorithm ranbeerRanbeer Tyagi
The document discusses image compression using differential pulse code modulation (DPCM) with a least mean square (LMS) algorithm. It begins with an introduction to image compression and DPCM. It then provides details on the LMS algorithm and how it can be used with DPCM to provide almost 2 bits per pixel reduction in transmission rate compared to standard DPCM while maintaining similar distortion levels. The document presents simulation results showing the performance of DPCM and DPCM with LMS on an image in terms of distortion, histogram, and prediction mean square error. It concludes that LMS has lower computational complexity than DPCM and can achieve better compression.
That presentation explains everything you always wanted to know about JPEG 2000.
All the benefits, how it works, what applications, how to implement JPEG2000.
Particularly for Digital Cinema, Broadcast Contribution, Video Archiving, Post-production.
Image compression uses four stages to reduce file sizes: 1) transforming RGB pixels to the YCbCr color space, 2) applying discrete cosine transformation to concentrate pixel data into a few matrix elements, 3) quantizing pixel values to reduce their range, and 4) using Huffman coding to assign shorter bit codes to more common pixel values. This allows JPEG images to be compressed to smaller file sizes for storage and transmission while still maintaining good visual quality.
A description about image Compression. What are types of redundancies, which are there in images. Two classes compression techniques. Four different lossless image compression techiques with proper diagrams(Huffman, Lempel Ziv, Run Length coding, Arithmetic coding).
Comparison between JPEG(DCT) and JPEG 2000(DWT) compression standardsRishab2612
This topic comes under the Image Processing.In this comparison between JPEG and JPEG 2000 compression standard techniques is made.The PPT comprises of results, analysis and conclusion along with the relevant outputs
JPEG is a lossy image compression algorithm, not a file format. It uses a 4-step process to compress images: 1) transforming RGB to YCbCr color space, 2) applying a discrete cosine transformation to identify redundant data, 3) quantizing the remaining data, and 4) encoding the result to minimize storage requirements. Typical compression ratios are 10:1 to 20:1 without visible loss and up to 100:1 compression for low quality applications.
The document discusses the JPEG image compression standard. It describes the basic JPEG compression pipeline which involves encoding, decoding, colour space transform, discrete cosine transform (DCT), quantization, zigzag scan, differential pulse code modulation (DPCM) on the DC component, run length encoding (RLE) on the AC components, and entropy coding using Huffman or arithmetic coding. It provides details on quantization methods, quantization tables, zigzag scan, DPCM, RLE, and Huffman coding used in JPEG to achieve maximal compression of images.
This document discusses image compression techniques. It begins by defining image compression as reducing the data required to represent a digital image. It then discusses why image compression is needed for storage, transmission and other applications. The document outlines different types of redundancies that can be exploited in compression, including spatial, temporal and psychovisual redundancies. It categorizes compression techniques as lossless or lossy and describes several algorithms for each type, including Huffman coding, LZW coding, DPCM, DCT and others. Key aspects like prediction, quantization, fidelity criteria and compression models are also summarized.
The document provides an overview of JPEG image compression. It discusses that JPEG is a commonly used lossy compression method that allows adjusting the degree of compression for a tradeoff between file size and image quality. The JPEG compression process involves splitting the image into 8x8 blocks, converting color space, applying discrete cosine transform (DCT), quantization, zigzag scanning, differential pulse-code modulation (DPCM) on DC coefficients, run length encoding on AC coefficients, and Huffman coding for entropy encoding. Quantization is the main lossy step that discards high frequency data imperceptible to human vision to achieve higher compression ratios.
NNPDF3.0: Next Generation Parton Distributions for the LHC Run IIjuanrojochacon
NNPDF3.0 is a new PDF release from the NNPDF collaboration that incorporates recent experimental data from HERA and the LHC, improved theory calculations, and methodological advances. Key aspects of NNPDF3.0 include the inclusion of new data like HERA structure functions, LHC jets and electroweak data, and top quark production data. It also utilizes approximate NNLO calculations for jets and NLO electroweak corrections for Drell-Yan production. The fitting methodology has been improved with a C++ rewrite of the code and validation on closure tests. Preliminary results show good agreement with NNPDF2.3 and reduced uncertainties for some PDFs from the new data and methodology
The impact of new collider data into the NNPDF global analysisJuan Rojo
The document summarizes Juan Rojo's presentation on the impact of new collider data in the NNPDF global analysis. It discusses updates and improvements to the NNPDF methodology, including adopting the public code APFEL, adding new LHC datasets like LHCb and top quark pair differential distributions, and analyzing the impact on parton distributions from including precise Tevatron and LHC Z boson data. Preliminary results from NNPDF3.1 indicate good stability compared to the previous NNPDF3.0 analysis, with reduced uncertainties and improved flavor separation from new experimental inputs.
Constraints on the gluon PDF from top quark differential distributions at NNLOjuanrojochacon
- The document discusses constraints on the gluon PDF from top quark production at hadron colliders.
- It describes using the inclusive top quark pair production cross section to reduce uncertainties in the gluon PDF, especially in the large-x region between 0.1 and 0.5.
- Cross section ratios between different beam energies, such as 8 TeV/7 TeV, are highlighted as powerful precision tests that can discriminate between PDFs and probe BSM physics.
News from NNPDF: new data and fits with intrinsic charmjuanrojochacon
Juan Rojo presented recent work by the NNPDF collaboration including: 1) inclusion of the final HERA legacy dataset which provides a moderate reduction in PDF uncertainties, 2) inclusion of new LHC data which constrains the large-x gluon PDF, and 3) ongoing work to perform fits with intrinsic charm and investigate implications for LHC phenomenology.
The document discusses the NNPDF3.1 global analysis of parton distribution functions (PDFs). It provides an update to NNPDF3.0 motivated by new high-precision collider data and progress in NNLO calculations. NNPDF3.1 fits both a perturbative and fitted charm PDF and finds slightly better fit quality for the fitted charm case. Comparisons to NNPDF3.0 show agreement within uncertainties and reduced PDF errors in NNPDF3.1 due to the new LHC data.
NNLO PDF fits with top-quark pair differential distributionsJuan Rojo
Juan Rojo presented a study on including top-quark pair differential distributions in NNLO global PDF fits. The distributions provide stringent constraints on the large-x gluon, comparable to inclusive jet data. Fitting normalized distributions and including one distribution from ATLAS and CMS improves the description of data and reduces PDF uncertainties, particularly at high masses important for BSM searches. Some tension is seen between ATLAS and CMS measurements that can be reduced by fitting the experiments separately. Differential top data will be essential for future global PDF fits.
Parton Distributions at a 100 TeV Hadron Colliderjuanrojochacon
Usage of modern PDF sets with LHAPDF6 v6.1.5 is suitable for FCC studies and simulations. At a 100 TeV hadron collider, PDFs would need to be evaluated in more extreme regions of small-x, large-x, and large invariant masses than at the LHC. Photon-initiated processes could contribute significantly at the FCC due to large uncertainties in the photon PDF. Heavy quark PDFs, including for the top quark, should be included in matched calculations for FCC simulations rather than using a purely massless scheme.
NNPDF3.0: parton distributions for the LHC Run IIjuanrojochacon
NNPDF3.0 is a new PDF determination that includes updated data and theory improvements compared to NNPDF2.3. It includes all HERA-II data and new LHC measurements. The fitting code was rewritten in C++ and validated using closure tests. NNPDF3.0 shows reasonable agreement with NNPDF2.3 while improving descriptions of data and reducing uncertainties in some regions. It provides PDFs for use at the LHC Run II.
Statistical issues in global fits: Lessons from PDF determinationsJuan Rojo
The document discusses statistical issues that arise in global fits to determine parton distribution functions (PDFs) from experimental data. It notes that PDF fits must combine data from different collision types and experiments, which can have inconsistent measurements. Traditional PDF fitting methods make restrictive assumptions that introduce bias, while the NNPDF approach uses neural networks and Monte Carlo replicas to avoid biases and faithfully represent uncertainties, including in regions with limited data. Inconsistent data poses challenges and requires delicate handling in global fits to obtain statistically sound PDF results.
News from NNPDF: QED, small-x, and alphas(MZ) fitsjuanrojochacon
Juan Rojo presented recent work from the NNPDF collaboration on three spin-off fits from their NNPDF3.1 global analysis: NNPDF3.1QED, fits including small-x resummation, and a determination of the strong coupling constant αS(mZ). For NNPDF3.1QED, they are imposing the LUXqed formalism to constrain the photon PDF rather than extracting it from data. For small-x resummation fits, they find that including NNLO+NLLx theory stabilizes the small-x gluon and improves description of HERA data. Their preliminary αS(mZ) value is consistent with other determinations.
The document summarizes the NNPDF3.1 global analysis which provides an updated determination of parton distribution functions (PDFs) from experimental data. Key points include:
1) NNPDF3.1 includes new high-precision measurements from the LHC as well as NNLO QCD calculations, allowing more data to be included. It also fits the charm PDF rather than assuming it is purely perturbative.
2) The new data provides stronger constraints on PDFs, particularly the gluon and down quark, significantly reducing their uncertainties. It also shows good agreement with the previous NNPDF3.0 analysis.
3) For the first time, NNPDF3.1 includes LHC
aMCfast: Automation of Fast NLO Computations for PDF fitsjuanrojochacon
MadGraph5_aMCatNLO provides NLO calculations for arbitrary processes and their matching to parton showers, but existing fast interfaces are limited. A new tool called aMCfast provides a fast interface to MadGraph5_aMCatNLO, allowing its predictions to be used in global PDF fits. It precomputes matrix elements and interpolates them using grids, then reconstructs distributions for any PDFs or scales. This will increase the number and accuracy of processes in PDF fits, including electroweak corrections and photon-initiated effects, improving determination of PDFs from LHC data.
PDFs at the LHC: Lessons from Run I and preparation for Run IIjuanrojochacon
This document summarizes parton distribution functions (PDFs) at the start of LHC Run II. It discusses the status of recent PDF sets from NNPDF, MMHT, CT, ABM, and HERAPDF. It notes some differences between these sets and the importance of PDF uncertainties for LHC measurements. The document also discusses benchmarks between PDF sets, comparisons using the APFEL online tool, and the prospects for including more ATLAS data in global PDF fits to further constrain PDFs.
NNPDF3.0: Next generation parton distributions for the LHC Run IIjuanrojochacon
The document provides an overview of the forthcoming NNPDF3.0 PDF release from the NNPDF Collaboration. Key points include:
1) NNPDF3.0 includes over 1000 new data points from HERA and LHC experiments like ATLAS and CMS, improving constraints on PDFs.
2) Improved theory calculations are used, including approximate NNLO corrections for jet data and full NLO electroweak corrections.
3) The NNPDF methodology has been upgraded with a C++ code rewrite, validation on closure tests, and improvements to the fitting strategy and basis choices.
A description about image Compression. What are types of redundancies, which are there in images. Two classes compression techniques. Four different lossless image compression techiques with proper diagrams(Huffman, Lempel Ziv, Run Length coding, Arithmetic coding).
Comparison between JPEG(DCT) and JPEG 2000(DWT) compression standardsRishab2612
This topic comes under the Image Processing.In this comparison between JPEG and JPEG 2000 compression standard techniques is made.The PPT comprises of results, analysis and conclusion along with the relevant outputs
JPEG is a lossy image compression algorithm, not a file format. It uses a 4-step process to compress images: 1) transforming RGB to YCbCr color space, 2) applying a discrete cosine transformation to identify redundant data, 3) quantizing the remaining data, and 4) encoding the result to minimize storage requirements. Typical compression ratios are 10:1 to 20:1 without visible loss and up to 100:1 compression for low quality applications.
The document discusses the JPEG image compression standard. It describes the basic JPEG compression pipeline which involves encoding, decoding, colour space transform, discrete cosine transform (DCT), quantization, zigzag scan, differential pulse code modulation (DPCM) on the DC component, run length encoding (RLE) on the AC components, and entropy coding using Huffman or arithmetic coding. It provides details on quantization methods, quantization tables, zigzag scan, DPCM, RLE, and Huffman coding used in JPEG to achieve maximal compression of images.
This document discusses image compression techniques. It begins by defining image compression as reducing the data required to represent a digital image. It then discusses why image compression is needed for storage, transmission and other applications. The document outlines different types of redundancies that can be exploited in compression, including spatial, temporal and psychovisual redundancies. It categorizes compression techniques as lossless or lossy and describes several algorithms for each type, including Huffman coding, LZW coding, DPCM, DCT and others. Key aspects like prediction, quantization, fidelity criteria and compression models are also summarized.
The document provides an overview of JPEG image compression. It discusses that JPEG is a commonly used lossy compression method that allows adjusting the degree of compression for a tradeoff between file size and image quality. The JPEG compression process involves splitting the image into 8x8 blocks, converting color space, applying discrete cosine transform (DCT), quantization, zigzag scanning, differential pulse-code modulation (DPCM) on DC coefficients, run length encoding on AC coefficients, and Huffman coding for entropy encoding. Quantization is the main lossy step that discards high frequency data imperceptible to human vision to achieve higher compression ratios.
NNPDF3.0: Next Generation Parton Distributions for the LHC Run IIjuanrojochacon
NNPDF3.0 is a new PDF release from the NNPDF collaboration that incorporates recent experimental data from HERA and the LHC, improved theory calculations, and methodological advances. Key aspects of NNPDF3.0 include the inclusion of new data like HERA structure functions, LHC jets and electroweak data, and top quark production data. It also utilizes approximate NNLO calculations for jets and NLO electroweak corrections for Drell-Yan production. The fitting methodology has been improved with a C++ rewrite of the code and validation on closure tests. Preliminary results show good agreement with NNPDF2.3 and reduced uncertainties for some PDFs from the new data and methodology
The impact of new collider data into the NNPDF global analysisJuan Rojo
The document summarizes Juan Rojo's presentation on the impact of new collider data in the NNPDF global analysis. It discusses updates and improvements to the NNPDF methodology, including adopting the public code APFEL, adding new LHC datasets like LHCb and top quark pair differential distributions, and analyzing the impact on parton distributions from including precise Tevatron and LHC Z boson data. Preliminary results from NNPDF3.1 indicate good stability compared to the previous NNPDF3.0 analysis, with reduced uncertainties and improved flavor separation from new experimental inputs.
Constraints on the gluon PDF from top quark differential distributions at NNLOjuanrojochacon
- The document discusses constraints on the gluon PDF from top quark production at hadron colliders.
- It describes using the inclusive top quark pair production cross section to reduce uncertainties in the gluon PDF, especially in the large-x region between 0.1 and 0.5.
- Cross section ratios between different beam energies, such as 8 TeV/7 TeV, are highlighted as powerful precision tests that can discriminate between PDFs and probe BSM physics.
News from NNPDF: new data and fits with intrinsic charmjuanrojochacon
Juan Rojo presented recent work by the NNPDF collaboration including: 1) inclusion of the final HERA legacy dataset which provides a moderate reduction in PDF uncertainties, 2) inclusion of new LHC data which constrains the large-x gluon PDF, and 3) ongoing work to perform fits with intrinsic charm and investigate implications for LHC phenomenology.
The document discusses the NNPDF3.1 global analysis of parton distribution functions (PDFs). It provides an update to NNPDF3.0 motivated by new high-precision collider data and progress in NNLO calculations. NNPDF3.1 fits both a perturbative and fitted charm PDF and finds slightly better fit quality for the fitted charm case. Comparisons to NNPDF3.0 show agreement within uncertainties and reduced PDF errors in NNPDF3.1 due to the new LHC data.
NNLO PDF fits with top-quark pair differential distributionsJuan Rojo
Juan Rojo presented a study on including top-quark pair differential distributions in NNLO global PDF fits. The distributions provide stringent constraints on the large-x gluon, comparable to inclusive jet data. Fitting normalized distributions and including one distribution from ATLAS and CMS improves the description of data and reduces PDF uncertainties, particularly at high masses important for BSM searches. Some tension is seen between ATLAS and CMS measurements that can be reduced by fitting the experiments separately. Differential top data will be essential for future global PDF fits.
Parton Distributions at a 100 TeV Hadron Colliderjuanrojochacon
Usage of modern PDF sets with LHAPDF6 v6.1.5 is suitable for FCC studies and simulations. At a 100 TeV hadron collider, PDFs would need to be evaluated in more extreme regions of small-x, large-x, and large invariant masses than at the LHC. Photon-initiated processes could contribute significantly at the FCC due to large uncertainties in the photon PDF. Heavy quark PDFs, including for the top quark, should be included in matched calculations for FCC simulations rather than using a purely massless scheme.
NNPDF3.0: parton distributions for the LHC Run IIjuanrojochacon
NNPDF3.0 is a new PDF determination that includes updated data and theory improvements compared to NNPDF2.3. It includes all HERA-II data and new LHC measurements. The fitting code was rewritten in C++ and validated using closure tests. NNPDF3.0 shows reasonable agreement with NNPDF2.3 while improving descriptions of data and reducing uncertainties in some regions. It provides PDFs for use at the LHC Run II.
Statistical issues in global fits: Lessons from PDF determinationsJuan Rojo
The document discusses statistical issues that arise in global fits to determine parton distribution functions (PDFs) from experimental data. It notes that PDF fits must combine data from different collision types and experiments, which can have inconsistent measurements. Traditional PDF fitting methods make restrictive assumptions that introduce bias, while the NNPDF approach uses neural networks and Monte Carlo replicas to avoid biases and faithfully represent uncertainties, including in regions with limited data. Inconsistent data poses challenges and requires delicate handling in global fits to obtain statistically sound PDF results.
News from NNPDF: QED, small-x, and alphas(MZ) fitsjuanrojochacon
Juan Rojo presented recent work from the NNPDF collaboration on three spin-off fits from their NNPDF3.1 global analysis: NNPDF3.1QED, fits including small-x resummation, and a determination of the strong coupling constant αS(mZ). For NNPDF3.1QED, they are imposing the LUXqed formalism to constrain the photon PDF rather than extracting it from data. For small-x resummation fits, they find that including NNLO+NLLx theory stabilizes the small-x gluon and improves description of HERA data. Their preliminary αS(mZ) value is consistent with other determinations.
The document summarizes the NNPDF3.1 global analysis which provides an updated determination of parton distribution functions (PDFs) from experimental data. Key points include:
1) NNPDF3.1 includes new high-precision measurements from the LHC as well as NNLO QCD calculations, allowing more data to be included. It also fits the charm PDF rather than assuming it is purely perturbative.
2) The new data provides stronger constraints on PDFs, particularly the gluon and down quark, significantly reducing their uncertainties. It also shows good agreement with the previous NNPDF3.0 analysis.
3) For the first time, NNPDF3.1 includes LHC
aMCfast: Automation of Fast NLO Computations for PDF fitsjuanrojochacon
MadGraph5_aMCatNLO provides NLO calculations for arbitrary processes and their matching to parton showers, but existing fast interfaces are limited. A new tool called aMCfast provides a fast interface to MadGraph5_aMCatNLO, allowing its predictions to be used in global PDF fits. It precomputes matrix elements and interpolates them using grids, then reconstructs distributions for any PDFs or scales. This will increase the number and accuracy of processes in PDF fits, including electroweak corrections and photon-initiated effects, improving determination of PDFs from LHC data.
PDFs at the LHC: Lessons from Run I and preparation for Run IIjuanrojochacon
This document summarizes parton distribution functions (PDFs) at the start of LHC Run II. It discusses the status of recent PDF sets from NNPDF, MMHT, CT, ABM, and HERAPDF. It notes some differences between these sets and the importance of PDF uncertainties for LHC measurements. The document also discusses benchmarks between PDF sets, comparisons using the APFEL online tool, and the prospects for including more ATLAS data in global PDF fits to further constrain PDFs.
NNPDF3.0: Next generation parton distributions for the LHC Run IIjuanrojochacon
The document provides an overview of the forthcoming NNPDF3.0 PDF release from the NNPDF Collaboration. Key points include:
1) NNPDF3.0 includes over 1000 new data points from HERA and LHC experiments like ATLAS and CMS, improving constraints on PDFs.
2) Improved theory calculations are used, including approximate NNLO corrections for jet data and full NLO electroweak corrections.
3) The NNPDF methodology has been upgraded with a C++ code rewrite, validation on closure tests, and improvements to the fitting strategy and basis choices.
Parton Distributions: future needs, and the role of the High-Luminosity LHCjuanrojochacon
1) Improved PDFs are needed to match the accuracy of higher-order calculations of cross sections and characterize properties of the Higgs boson and search for new physics.
2) The HL-LHC could significantly reduce PDF uncertainties through high-statistics measurements, especially in processes sensitive to large-x gluons and quarks like top quark pair production and Drell-Yan.
3) Preliminary studies generating HL-LHC pseudo-data show PDF uncertainties on the gluon-gluon and quark-antiquark luminosities could be reduced by 10-30% from top pair and Drell-Yan measurements respectively.
This document discusses parton distribution functions (PDFs) with intrinsic charm. It presents the motivation for fitting a charm PDF in global analyses, including stabilizing charm mass dependence, quantifying the non-perturbative charm component, and exploring implications for LHC phenomenology. The document summarizes previous fits allowing for intrinsic charm and issues they faced. It then presents new NNPDF fits that allow the charm PDF to be fitted, finding these fits describe EMC charm data and have stable gluon and charm PDFs under charm mass variations, with implications for precision LHC calculations.
The document summarizes progress in the NNPDF global analysis of parton distribution functions. It discusses past NNPDF analyses from 2012-2015 and plans for future analyses, including NNPDF3.1 which will include new LHC data and improve the determination of intrinsic charm. It also presents results from fits allowing the charm content of the proton to be determined from data rather than assumed from perturbation theory, finding the data support a small intrinsic charm component within uncertainties.
Parton Distributions and Standard Model Physics at the LHCjuanrojochacon
This document discusses parton distribution functions (PDFs) and recent developments. It notes that NNLO calculations are essential to reduce uncertainties in PDF analysis. Several key processes like inclusive jet production and top quark production are now available at NNLO. The document also discusses the inclusion of LHC data like W+charm, top quark, and jet data in global PDF fits. It highlights updates to various PDF fitting groups and the upcoming NNPDF3.0 release.
New dynamics in parton distributions at a 100 TeV hadron colliderjuanrojochacon
This document discusses parton distribution functions (PDFs) relevant for a future 100 TeV hadron collider. It summarizes that such a collider would probe PDFs at smaller x values than currently possible, down to x=10-9. It would also probe higher mass scales, up to masses of 10 TeV. Current PDF determinations have little constraint for x<10-4 or masses above 1 TeV. The increased energy would require considering effects like top quark PDFs, electroweak corrections, and high-energy resummation not important at lower energies. Polarized collisions could also provide insights into PDFs.
Parton distributions with QED corrections and LHC phenomenologyjuanrojochacon
The document discusses parton distribution functions (PDFs) that include quantum electrodynamics (QED) corrections. It summarizes the NNPDF2.3QED PDF set, which is the first to include next-to-next-to-leading order (NNLO) QCD and leading order (LO) QED effects. The photon PDF is directly constrained by LHC data for the first time. The PDF set improves constraints on the photon PDF from both DIS and LHC Drell-Yan data. It also discusses implications for LHC phenomenology from photon-initiated contributions.
Similar to PDF uncertainties the LHC made easy: a compression algorithm for the combination of PDF sets (20)
Higgs Pair Production at the LHC and future collidersjuanrojochacon
Juan Rojo gave a seminar at Boston University Physics on November 26th, 2018 about Higgs pair production at the LHC and future colliders. He discussed how measuring Higgs pair production can provide insights into the Higgs sector and electroweak symmetry breaking. However, observing Higgs pair production is challenging due to small cross-sections in the standard model. Future colliders with higher energies, like the FCC, may enable more precise studies of Higgs pair production and help address open questions in particle physics.
Parton distributions in the LHC precision erajuanrojochacon
This document summarizes Juan Rojo's presentation on parton distribution functions (PDFs) at the Zurich Phenomenology Workshop 2018. The key points are:
1) PDF fits require combining perturbative cross-sections calculated using the Standard Model Lagrangian with non-perturbative PDFs extracted from a global analysis of experimental data.
2) More precise PDFs are needed to reduce uncertainties on calculations of processes like Higgs production and measurements of its couplings at the LHC.
3) Recent PDF analyses have included new data like differential top quark production from LHC and NNLO calculations, improving determinations of the gluon PDF over a wide range of x values.
4) Small-x data
Parton distributions from high-precision collider datajuanrojochacon
Juan Rojo gave a seminar at the Technical University of Munich on July 13, 2017 about parton distributions from high-precision collider data. He discussed how parton distribution functions are essential for calculating cross sections at hadron colliders like the LHC, since they describe the probability of finding quarks and gluons within protons. Rojo explained that global analyses fit PDFs to diverse experimental data using statistical techniques like neural networks, and the PDFs can then provide predictions for new processes. He highlighted recent updates from the NNPDF collaboration in version 3.1 to include more precise LHC data and the option to fit the charm quark distribution.
Recent progress in proton and nuclear PDFs at small-xjuanrojochacon
1) The document discusses recent progress in proton and nuclear parton distribution functions (PDFs) at small values of x. PDFs describe the momentum distribution of quarks and gluons inside protons and nuclei.
2) Global analyses of experimental data from various processes are used to determine PDFs at hadronic scales, which are then evolved perturbatively to higher scales relevant for LHC predictions. Recent analyses include data from the LHC.
3) Probing PDFs at small x requires processes dominated by gluons at leading order, produced in the forward region with low invariant masses. Examples discussed are direct photon and charm production. LHCb and future forward calorimeter data provide constraints on the small-
Higgs pair production in vector-boson fusion at the LHC and beyondjuanrojochacon
The document discusses Higgs pair production, which can help uncover the mechanism of electroweak symmetry breaking. Double Higgs production allows reconstructing the full Higgs potential and probing the Higgs self-interaction. While production rates are small in the Standard Model, deviations from SM couplings could significantly increase rates. The document focuses on vector boson fusion channel, where requiring two forward jets and vetoing central jets can reduce backgrounds, and sensitivity to non-SM couplings improves at high di-Higgs invariant masses. Precision studies of Higgs pair production may reveal insights into electroweak symmetry breaking and new physics.
Precision determination of the small-x gluon from charm production at LHCbjuanrojochacon
This document discusses using LHCb data on charm production to constrain the small-x gluon and improve predictions for neutrino fluxes. LHCb data at 7 TeV, 5 TeV and 13 TeV provides stringent constraints on the small-x gluon beyond HERA. This improved gluon allows more accurate predictions for signals and backgrounds at neutrino telescopes. At a 100 TeV collider, inclusive cross sections depend directly on small-x PDFs, but using LHCb data leads to stabilized predictions with reduced uncertainties.
This document discusses potential areas where lattice QCD calculations could provide input to help constrain parton distribution functions (PDFs) which are currently not well known. It identifies "benchmarks" where PDFs are already well determined, as well as "opportunities" where lattice calculations could have more impact. These include PDFs at large values of Bjorken x, the strange quark content of the proton, and charm content. The document also discusses how lattice data could be included in PDF global fits to help reduce PDF uncertainties.
Discovery through precision: perturbative QCD at the dawn of the LHC Run IIjuanrojochacon
This document summarizes a seminar given by Juan Rojo on perturbative QCD at the dawn of LHC Run II. In the seminar, Rojo discussed how recent advances in perturbative QCD, such as higher precision parton distribution functions and higher-order calculations, have improved the prospects for new physics searches at the LHC. Rojo also presented recent work on topics like PDFs with threshold resummation, Higgs pair production, and charm production that further push the boundaries of perturbative QCD and its applications at the LHC.
Parton Distributions and the search for New Physics at the LHCjuanrojochacon
Juan Rojo gave a seminar at King's College London on September 23, 2015 about parton distribution functions (PDFs) and their importance for precision physics at the Large Hadron Collider (LHC). PDFs describe the momentum distributions of quarks and gluons within protons and are crucial for determining cross sections and uncertainties for many LHC processes. The accurate determination of PDFs requires global analyses of experimental data using flexible parametrizations like neural networks to avoid biases. PDF uncertainties now limit characterization of the Higgs boson and searches for new physics at the LHC.
CERN, Particle Physics and the Large Hadron Colliderjuanrojochacon
The document discusses particle physics research done at CERN using the Large Hadron Collider (LHC). It describes the LHC as the most powerful particle accelerator ever built, with a 27km long tunnel housing four detectors. The LHC collides protons together at high energies to study their constituent particles like quarks and search for new particles like the Higgs boson. It also allows researchers to recreate conditions shortly after the Big Bang and potentially observe mini black holes or extra dimensions at very small scales. The future includes planning for an even larger successor to the LHC to continue advancing understanding of fundamental physics.
The structure of the proton in the Higgs boson erajuanrojochacon
Juan Rojo gave a seminar at SLAC on July 4th, 2015 about the structure of the proton in the Higgs boson era and the role of the Large Hadron Collider. He discussed how the discovery of the Higgs boson completed the Standard Model but also opened new questions. Determining the parton distribution functions of protons with precision is important for phenomenology at the LHC, such as characterizing Higgs couplings and searching for new physics. Over the next 20 years, the LHC will play a key role in exploring these open questions.
Particle Physics, CERN and the Large Hadron Colliderjuanrojochacon
The document discusses particle physics research done at CERN's Large Hadron Collider (LHC). It describes the LHC as the most powerful particle accelerator ever built, with a 27 km long tunnel housing detectors that observe proton collisions at very high energies. One of the LHC's major discoveries was the Higgs boson particle in 2012. The document outlines how the LHC allows scientists to study the fundamental building blocks of matter at the smallest observable scales.
PDF uncertainties and the W mass: Report from the Workshop “Parton Distributi...juanrojochacon
The document summarizes discussions from a workshop about reducing uncertainties in parton distribution functions (PDFs) that affect measurements of the W boson mass. Key topics discussed included issues modeling the Z boson's transverse momentum, inconsistencies between PDF sets, and the need for additional LHC measurements like of W/Z production to reduce PDF uncertainties in future W mass measurements to below 10 MeV. Participants agreed more work is needed to understand differences between PDF sets and include relevant LHC data in global fits.
Characterizing New Physics with Polarized Beams in Hadron Collisionsjuanrojochacon
This document discusses how polarized proton beams could help characterize new physics discovered at the LHC or FCC. Polarized and unpolarized parton distribution functions have different behaviors, so cross sections with polarized beams would provide insights into couplings to different quark flavors. Single and double spin asymmetries could distinguish between new physics models that predict similar unpolarized signatures. While more study is needed, polarized beams could significantly aid understanding of the structure and couplings of any discovered heavy new particles.
The Standard Model and the LHC in the Higgs Boson Erajuanrojochacon
The document discusses the Standard Model of particle physics and the role of the Large Hadron Collider (LHC) following the discovery of the Higgs boson. It provides background on the development of the Standard Model and discovery of its key particles like quarks, gluons, and weak bosons. It describes the LHC as the most powerful particle collider built to explore physics at the highest energies and probe unanswered questions left by the Standard Model. Four main detectors at the LHC, including ATLAS and CMS, precisely measure collision products to explore fundamental particles and forces.
The Structure of the Proton in the Higgs Boson Erajuanrojochacon
Juan Rojo gave a seminar at NIKHEF in Amsterdam on January 22, 2015 about the structure of the proton in the Higgs boson era. He discussed how the discovery of the Higgs boson completed the Standard Model but also opened new questions. He explained that the Large Hadron Collider will play a key role in exploring these questions over the next 20 years. Accurately determining the parton distribution functions of the proton is vital for phenomenology at the LHC.
Boosting Strong Higgs Pair Production at the LHCjuanrojochacon
This document summarizes Juan Rojo's presentation on boosting strong Higgs pair production at the LHC. The key points are:
1) Higgs pair production allows for stringent tests of the understanding of electroweak symmetry breaking, but has low rates in the Standard Model.
2) Higgs pair production in vector boson fusion is small in the Standard Model but provides unique information on the hhVV coupling. It can be substantially enhanced in composite Higgs models.
3) Preliminary results show the hhVV coupling can be measured with 25-30% precision at the 14 TeV LHC with 300 fb-1, and 10-15% precision with 3000 fb-1, while the FCC could achieve
Pinning down the hhVV coupling from Higgs boson pair production in vector-b...juanrojochacon
The document discusses using Higgs boson pair production in vector-boson fusion (VBF) to probe the hhVV coupling. VBF Higgs pair production provides unique sensitivity to this coupling. While the standard model rate for VBF Higgs pair production is small, it can be substantially enhanced in composite Higgs models with new strong dynamics. Preliminary results indicate the hhVV coupling can be measured with 25-30% precision at the LHC with 300 fb-1, improving to 10-15% precision with 3000 fb-1. The FCC could determine the coupling with few-percent accuracy. Analyzing the boosted 4b and 2b2l final states allows probing deviations in the hhVV coupling growing with di-Higgs mass.
Pinning down the hhVV coupling from Higgs boson pair production in vector-b...
PDF uncertainties the LHC made easy: a compression algorithm for the combination of PDF sets
1. !
PDF uncertainties at the LHC made easy:
A compression algorithm for the combination of PDF sets
Juan Rojo!
STFC Rutherford Fellow!
Rudolf Peierls Center for Theoretical Physics!
University of Oxford!
!
in collaboration with S. Carrazza, J. I. Latorre and G. Watt!
!
PDF4LHC Meeting!
CERN, 21/01/2015
Juan Rojo PDF4LHC Meeting, 21/01/2015
2. 2
Motivation
!
To provide a practical implementation of the PDF4LHC recommendation, easy to use by the
experiments and computationally less intensive that the original prescription!
Based on the Monte Carlo statistical combination of different PDF sets, followed by a compression
algorithm to end up with a reduced number of replicas!
Similar in spirit to the Meta-PDF approach (Gao and Nadolsky 14) but important conceptual and
practical differences!
Having a single combined PDF sets (even with large number of eigenvector/sets) would already
simplify the life of many people since widely-used tools like MadGraph5_aMC@NLO, POWHEG or
FEWZ provide the PDF uncertainties without any additional cost!
But this is not true for all theory tools used at the LHC, so there is still a strong motivation to be able
to use a combined PDF set with a small number of eigenvectors/replicas!
Here all results obtained for fixed alphas(MZ)=0.118, adding the combined PDF+alphas uncertainty
in quadrature (updated PDF4LHC recommendation) trivial in our approach!
In addition, the compression algorithm can also be used in native MC sets, like NNPDF, starting
from a 1000 replica sample and reducing it to a smaller sample while reproducing all statistical
estimators
Juan Rojo PDF4LHC Meeting, 21/01/2015
3. 3
Basic strategy (I)!
Select the PDF sets that enter the combination. Results here based on NNPDF3.0, CT10 and
MMHT14, but any other choice possible!
Transform the Hessian PDF sets into their Monte Carlo representation (Watt and Thorne 12)
!
Now combine the same number of replicas from each of the three sets (assume equal weight in the
combination). First proposed by Forte 12 !
The resulting Monte Carlo ensemble has a robust statistical interpretation, and in many cases leads
to similar results, with somewhat smaller uncertainties, compared to the original PDF4LHC envelope.
Forte and Watt 13Juan Rojo PDF4LHC Meeting, 21/01/2015
4. 4
The combined PDF set!
Since there is reasonable agreement between CT10, MMHT14 and NNPDF3.0, the resulting
combined distribution is in general Gaussian, but there are also important cases where the non-
gaussianity of the combined PDFs is substantial
x
5
10 4
10
3
10 2
10 1
10
1
0
1
2
3
4
5
6
Q = 1.4142 GeV
NNPDF30_nnlo_as_0118
MMHT2014nnlo68cl_rand1002
CT10nnlo_rand1004
MCcompPDFnnlo
Q = 1.4142 GeV
x * PDF
0.76 0.78 0.8 0.82 0.84 0.86 0.88 0.9 0.92 0.94 0.96 0.98
Probabilityperbin
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
Gluon PDF, x=0.1, Q=100 GeV
histo1
Entries 100
Mean 0.8887
RMS 0.0151
NNPDF3.0
CT10
MMHT14
MCcompPDFs
Gluon PDF, x=0.1, Q=100 GeV
!
It is possible to have smoother distributions by increasing the number of replicas for each set, but
this does not seem to be required by phenomenology!
For typical applications using Nrep=100 for each of the three PDF sets is enough!
Note that in general, the combination of Gaussian distributions is not a Gaussian itself
Juan Rojo PDF4LHC Meeting, 21/01/2015
5. 5
Combined MC set vs PDF4LHC envelope!
As already noted in the Forte-Watt study, the MC combination leads to somewhat smaller uncertainties
than the PDF4LHC envelope (same here and in the Meta-PDFs)!
This can be understood because now each PDF set receives the same weight, while the PDF4LHC
envelope effectively gives more weight to the outliers
Juan Rojo PDF4LHC pre-Meeting, 16/01/2014
(pb)
40
40.5
41
41.5
42
42.5
43
43.5
44
=0.118sggH, ggHiggs NNLO, LHC 13 TeV,
NNPDF3.0
MMHT14
CT10
CMCPDF
=0.118sggH, ggHiggs NNLO, LHC 13 TeV,
(pb)
750
760
770
780
790
800
810
820
830
840
850
=0.118sttbar, top++ NNLO, LHC 13 TeV,
NNPDF3.0
MMHT14
CT10
CMCPDF
=0.118sttbar, top++ NNLO, LHC 13 TeV,
(nb)
3300
3350
3400
3450
3500
3550
3600
3650
=0.118sW+, VRAP NNLO, LHC 13 TeV,
NNPDF3.0
MMHT14
CT10
CMCPDF
=0.118sW+, VRAP NNLO, LHC 13 TeV,
(nb)
460
470
480
490
500
510
=0.118sZ0, VRAP NNLO, LHC 13 TeV,
NNPDF3.0
MMHT14
CT10
CMCPDF
=0.118sZ0, VRAP NNLO, LHC 13 TeV,
PDF4LHC envelope
PDF4LHC envelope
PDF4LHC envelope
PDF4LHC envelope
6. 6
Compression as a mathematical problem!
The goal now is to compressed the combined set of Nrep=300 replicas to a smaller subset, in a way
that this subset reproduces the statistical properties of the original distribution
!
Mathematically, this is a well-defined problem:
compression is finding the subset that minimises
the distance between two probability
distributions
!
Many equally good minimisations possible,
so choice of minimisation algorithm not crucial
(similar to the travelling salesman problem)!
Mathematically well-posed problem, with a
number of robust solutions!
1. Kolmogorov distance!
2. Kullback-Leibler entropy!
3. …..!
Optimal choice determined by the requirements
of the problem at hand, in this case LHC
phenomenology
Juan Rojo PDF4LHC Meeting, 21/01/2015
7. 7
Basic strategy (II)!
Now we have a single combined set, but number of MC replicas still too large!
Compress the original probability distribution to one with a smaller number of replicas, in a way
that all the relevant estimators (mean, variances, correlations etc) for the PDFs are reproduced
!
The compression is applied at Q = 2 GeV,
though the results are robust wrt other choices!
Various options about how the error
function to be minimised can be defined, ie.,
to reproduce central values add a term
!
The algorithm also minimises the Kolmogorov distance
between the original and compressed distributions
!
Same for variances, correlations and higher
moments!
At the end, optimal choice decided by the
resulting phenomenology
8. 8
Results of the compression!
To gauge improvements due to compression, compare various contributions to the error function in
the best compression and in randoms selection with the same number of replicas
!
Substantial improvements as
compared to random
compressions, typically by one
order or magnitude or more!
Compression is also able to
successfully reproduce higher
moments like skewness or
kurtosis!
Similar improvements for the
correlations and the Kolmogorov
distances
Horizontal dashed line: !
lower limit of 68%CL range for random
compressions with Nrep=100
Juan Rojo PDF4LHC Meeting, 21/01/2015
9. 9
Results of the compression
Juan Rojo PDF4LHC Meeting, 03/11/2014
!
For example, for Nrep=40 replicas the compressed and the original PDFs are virtually identical
x
5
10 4
10
3
10 2
10 1
10
Gluon,ratiotoprior
0.75
0.8
0.85
0.9
0.95
1
1.05
1.1
1.15
1.2
1.25
Q = 100 GeV
Prior, 300 MC replicas
Compressed set, 40 MC replicas
Q = 100 GeV
x
5
10 4
10
3
10 2
10 1
10
Up,ratiotoprior
0.75
0.8
0.85
0.9
0.95
1
1.05
1.1
1.15
1.2
1.25
Q = 100 GeV
Prior, 300 MC replicas
Compressed set, 40 MC replicas
Q = 100 GeV
x
5
10 4
10
3
10 2
10 1
10
Up,ratiotoprior
0.75
0.8
0.85
0.9
0.95
1
1.05
1.1
1.15
1.2
1.25
Q = 100 GeV
Prior, 300 MC replicas
Compressed set, 10 MC replicas
Q = 100 GeV
x
5
10 4
10
3
10 2
10 1
10
Gluon,ratiotoprior
0.75
0.8
0.85
0.9
0.95
1
1.05
1.1
1.15
1.2
1.25
Q = 100 GeV
Prior, 300 MC replicas
Compressed set, 10 MC replicas
Q = 100 GeV
!
As expected, for a very small number of replicas (10 in this case) agreement is much worse
10. 10
The compressed PDF set!
Since there is reasonable agreement between CT10, MMHT14 and NNPDF3.0, the resulting
combined distribution looks typically Gaussian
!
On average, the same number of replicas from each of the three sets is selected in the compressed
set, a further demonstration that the algorithm is unbiased
Replicas
0 50 100 150 200 250 300
Entries
0
1
CMC-PDF NLO - 25 replica distribution
NNPDF3.0 CT10 MMHT14
9 replicas 8 replicas 8 replicas
Juan Rojo PDF4LHC Meeting, 21/01/2015
11. 11
Gaussian vs non-Gaussian!
Even if the original PDF sets in the combination are approximately Gaussian, their combination in general will
be non-Gaussian, and linear propagation might not be adequate!
Working in Gaussian approximation might not be reliable: i.e. skewness is not reproduced in the compression
(despite central values and variances are) unless we explicitly include it in the minimised figure of merit
Juan Rojo PDF4LHC pre-Meeting, 16/01/2014
Skewness included
Skewness excluded
12. 12
Compressing native MC sets!
The compression algorithm can of course be also used in native MC sets, like NNPDF. We have
shown that starting from NNPDF3.0 with Nrep=1000 replicas we can compress down to 40-50 replicas
maintaining all relevant statistical properties
!
Central values and variances well reproduced, but also, non-trivially, also higher moments and
correlations!
Sets with Nrep=1000 replicas are still useful for other applications, like Bayesian reweighting
All plots done with the APFEL Web plotter
Juan Rojo PDF4LHC Meeting, 21/01/2015
13. 13
Phenomenology!
The ultimate validation is of course to check that the compressed set reproduces the original PDF
combination for a wide variety of LHC observables!
We have tested a very large number of processes, both at the inclusive and differential level and
always found that Nrep=20-30 replicas are enough for phenomenology
!
NNLO cross-sections:!
gg->H with ggHiggs!
tt with top++!
W,Z with Vrap
Juan Rojo PDF4LHC Meeting, 21/01/2015
14. 14
Phenomenology!
The ultimate validation is of course to check that the compressed set reproduces the original
combination for a wide variety of observables!
We have tested a very large number of processes, both at the inclusive and differential level and
always found that Nrep=20-30 replicas are enough for phenomenology
!
NLO inclusive cross sections
with MCFM!
H VBF!
WW!
WH
Juan Rojo PDF4LHC Meeting, 21/01/2015
15. 15
Phenomenology!
The ultimate validation is of course to check that the compressed set reproduces the original
combination for a wide variety of observables!
We have tested a very large number of processes, both at the inclusive and differential level and
always found that Nrep=20-30 replicas are enough for phenomenology
Compression also works for
fully differential distributions!
Tested on a large number of
processes: jets, Drell-Yan, WW,
W+charm, Z+jets, ….!
Calculations use fast NLO
interfaces:!
1. aMCfast/applgrid for
MadGraph5_aMC@NLO!
2. applgrid for MCFM/
NLOjet++!
Very flexible to redo validation for
any other compressed set
Juan Rojo PDF4LHC Meeting, 21/01/2015
16. 16
Correlations!
The compression algorithm also manages to reproduce the correlations between physical
observables, even for a small number of replicas
Correlationcoefficient
1
0.5
0
0.5
1
Correlation Coefficient for ttbar
ggH tt W+ W- Z
= 40repN
Reference
Compressed
Correlation Coefficient for ttbar
!
Direct consequence of the fact that correlations between PDFs are reproduced
Juan Rojo PDF4LHC Meeting, 21/01/2015
17. 17
Correlations
!
Not an accident: selecting replicas at random fails to reproduce the correlation accurately enough
Correlationcoefficient
1
0.5
0
0.5
1
Correlation Coefficient for ttbar
= 40repN
Reference
Compressed
Random (68% CL)
Correlation Coefficient for ttbar
Juan Rojo PDF4LHC Meeting, 21/01/2015
18. 18
Comparison with Meta-PDFs!
To compare with the available Meta-PDFs in LHAPDF6, we have produced compressed sets based
on MSTW08, CT10 and NNPDF2.3!
Reasonable agreement found for central values and variances, except perhaps small- and at large-x!
Need to redo the comparison when the two approaches use NNPDF3.0, MMHT14 and CT14
Juan Rojo PDF4LHC Meeting, 21/01/2015
19. 19
Summary and discussion!
Our investigations suggest that the Compressed MC PDFs (CMC-PDFs) provide an efficient and easy to use
implementation of the PDF4LHC combination!
A single set with only 25 replicas seems to be enough to reproduce central values and variances of all the LHC
processes, inclusive and differential, that we have explored. Correlations both between PDFs and cross-sections
are also well reproduced!
Our plan is to release publicly the combination and compression codes, so that users can create their own
compressed sets in LHAPDF6 format with the only input the names of the sets to be used!
Some of the advantages of our approach wrt the MetaPDFs are!
1. !Improvement in CPU time (baseline Meta-PDF has 100 eigenvector members)!
2. Streamlined construction of the combined sets with no need of refitting (and thus of testing the
accuracy of the refit, which strongly depends on the PDF sets used) nor to redo the PDF evolution and
reconstruct PDF interpolation grids: for CMC-PDFs only the LHAPDF6 files needed as input!
3. No need of Gaussian/linear approximations: the compression algorithm reproduces the full
probability distribution of the combined set (which in general is non-Gaussian)!
4. No strong need for process-dependent PDF combined sets (like a Higgs-specific Meta-PDFs)!
However we believe that the two approaches can nicely complement each other: The agreement between the
compressed PDFs and the Meta-PDFs, for the same input tests, provide a non-trivial validation of the
combination. It is conceivable that the two of them could be used in LHC analysis.!
To make further progress, we need that:!
1. !PDF4LHC decides which PDF sets (and how) should be used in the combination!
2. This input is used to construct Meta-PDFs and CMC-PDFs!
3. Agree on benchmark tests for LHC processes and test the performance of both methods!