SlideShare a Scribd company logo
1 of 24
Download to read offline
THE OPEN SCIENCE GRID (OSG)
And how it relates to PRAGMA
by Igor Sfiligoi, UCSD/SDSC
(isfiligoi@sdsc.edu)
PRAGMA 37 - SEPT 13TH, 2019
Premise – Not the Grid you were used to
• It is my understanding Grid has a negative connotation in PRAGMA
• While OSG has “Grid” in its name,
I believe it is far from the “Grid” you experienced 15 years ago
• Please give me a chance to convince you of that
PRAGMA 37 - SEPT 13TH, 2019 2
OSG evolution
• 10 years ago, OSG provided an integrated software stack that organizations
had to deploy in order for their researchers to benefit from OSG.
•We learned the hard way that this is too high a bar
to jump across for most researchers and institutions.
• Today, we offer to operate services to the groups we serve in order
to minimize the threshold to entry for everybody.
• Leveraging work performed by the likes of PRP/NRP-P/TNRP (e.g. Kubernetes)
PRAGMA 37 - SEPT 13TH, 2019 3
OSG Aspirational Goal
A campus Research IT Organization
should not have to learn anything ”non-standard”
in order to have their researchers benefit from OSG,
or have their resources be available via OSG
PRAGMA 37 - SEPT 13TH, 2019 4
Now …. Not there yet, but getting closer by the day…
OSG Mission
Advancing Open Science
through distributed High
Throughput Computing
PRAGMA 37 - SEPT 13TH, 2019 5
Open Science
•All of open science irrespective of discipline
•Include the maximum possible dynamic range
of science, groups, and institutions
• From individual undergraduates
to international collaborations with thousands of members.
• From small colleges, museums, zoos,
to national scale centers of open science.
PRAGMA 37 - SEPT 13TH, 2019 6
What Sciences use OSG?
• Order here is from most active to least active.
• Category is based on self-identification … i.e. people say what they do.
PRAGMA 37 - SEPT 13TH, 2019 7
Not all users are the same
• The individual researchers and small groups
• The campus Research Support Organizations
• Multi-institutional Science Teams
• XENON, GlueX, SPT, Simons, and many many more
• Collaborations between multiple campuses
• The 4 “big science” projects:
• US-ATLAS, US-CMS, LIGO, IceCube
Used services
change between
different categories
PRAGMA 37 - SEPT 13TH, 2019 8
OSG Mission
Advancing Open Science
through distributed High
Throughput Computing
PRAGMA 37 - SEPT 13TH, 2019 9
distributed High Throughput Computing
DRIVING PRINCIPLES:
• Maximize total compute, do not
optimize for any single application
• Allow for resources coming from
anywhere in the world
• Embrace heterogeneity
• Federated policy control
PRAGMA 37 - SEPT 13TH, 2019 10
(as opposed to e.g. HPC)
distributed High Throughput Computing
DRIVING PRINCIPLES:
• Maximize total compute, do not
optimize for any single application
• Allow for resources coming from
anywhere in the world
• Embrace heterogeneity
• Federated policy control
IMPLICATIONS:
• Expect high network latencies
• Expect varying walltime
between identical runs
• Job restarts are the norm
PRAGMA 37 - SEPT 13TH, 2019 11
(Although throughput
should be high)
Thriving in dHTC
• The path to successful dHTC involves two aspects:
• Separate a big computing problem in
many individually schedulable small problems.
• Minimize your requirements in order to
maximize the raw capacity that you can effectively use.
• The infrastructure will transparently take care of job restarts
Ingenious
Parallelism
PRAGMA 37 - SEPT 13TH, 2019 12
dHTC is special
•dHTC scales by definition perfectly.
• When a researcher understands how to partition their workflow into
many individually schedulable compute problems they can scale out
with ease to seemingly arbitrary scales of computing.
• When we integrate all IT resources at Universities, National Labs,
and the commercial cloud, we arrive at a near infinite resource pool.
PRAGMA 37 - SEPT 13TH, 2019 13
Stronger together
• For OSG VO, more than half of the compute comes from smaller sites
(and this does not include any of the big sciences, that have their own pools)
15k cores
PRAGMA 37 - SEPT 13TH, 2019 14
OSG IN DETAIL
PRAGMA 37 - SEPT 13TH, 2019 15
OSG Compute Services
• OSG provides the means for operating a compute federation
• Mostly in the form of Metascheduling Service
• But also operates one access point itself for the smaller players
PRAGMA 37 - SEPT 13TH, 2019 16
Federation i.e. Distributed control
• OSG works on three simple principles:
• Resource Owners determine policy of use
• This means that all policy of use is set locally by the clusters that join the federation.
• Resource Consumers specify the types of resources they are willing to use.
• How much RAM? How many cores per node? …
• OSG submits HTCondor batch system as (pilot) payload into all local batch
systems that match requirements, creating an overlay dHTC system
• Jobs are submitted locally, queue centrally, and execute anywhere that
matches requirements after resource becomes available.
PRAGMA 37 - SEPT 13TH, 2019 17
OSG Data Federation
• OSG operates a content delivery network (CDN)
• Based on XRootD caching technology
• Data origins operated by Science communities
• Data transparently published in our federated namespace.
• Caches hide data access latencies and reduce network traffic.
PRAGMA 37 - SEPT 13TH, 2019 18
OSG Software Distribution
•OSG can host and distribute community software on CVMFS
• A CDN at heart, but based on Apache Squid technology
• CVMFS provides a Read-Only POSIX interface
•OSG encourages the use of containerization
• Encourages participating sites to provide Singularity
• Which is optimized for use by unprivileged users (unlike Docker)
• Containers can be distributed through CVMFS (for performance)
PRAGMA 37 - SEPT 13TH, 2019 19
Facilitation Services
• Proactive, personalized guidance
and support for:
• Institution-specific submit points
• Sharing institutional resources via OSG
• Data federation across OSG sites
• Individual researchers using the
OSG submit infrastructure
• We also offer
• Local workshops
• OSG-hosted education and training
• Learning from the OSG Facilitation
Community
PRAGMA 37 - SEPT 13TH, 2019 20
OSG AND PRAGMA
PRAGMA 37 - SEPT 13TH, 2019 21
OSG AND PRAGMA
• OSG would be happy to facilitate
any Open Science endeavor
that has at least some of its
members in the US
• We are a NSF-funded project
• All OSG services are at your disposal
• Compute federation
• Data federation
• Software distribution
• Facilitation
PRAGMA 37 - SEPT 13TH, 2019 22
OSG and International Setups
•All our major users are international
• CMS, ATLAS, LIGO, IceCube, XENON, GlueX
•But, alas, we have been mostly collaborating with Europe
•PRAGMA could help us engage more
Science communities in the Pacific Rim
PRAGMA 37 - SEPT 13TH, 2019 23
Summary and Conclusion
• OSG’s objective is to “Advance Open Science
through distributed High Throughput Computing”
• We are committed to help any Open Science community
that has at least some members in the USA – Nice overlap with PRAGMA
• OSG can offer a wide variety of services
• Both in terms of technology and facilitation
• We could use some help getting more involved in the Pacific Rim
• Contact us at: help@opensciencegrid.org
PRAGMA 37 - SEPT 13TH, 2019 24

More Related Content

Similar to The Open Science Grid and how it relates to PRAGMA

GRP 19 - Nautilus, IceCube and LIGO
GRP 19 - Nautilus, IceCube and LIGOGRP 19 - Nautilus, IceCube and LIGO
GRP 19 - Nautilus, IceCube and LIGOIgor Sfiligoi
 
GridIIT Open Science Grid
GridIIT Open Science GridGridIIT Open Science Grid
GridIIT Open Science GridSciCompIIT
 
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computingISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computingAlan Sill
 
Bringing OSG users to the PRP Kubernetes Cluster
Bringing OSG users to the PRP Kubernetes ClusterBringing OSG users to the PRP Kubernetes Cluster
Bringing OSG users to the PRP Kubernetes ClusterIgor Sfiligoi
 
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...Haripds Shrestha
 
The Open Science Grid
The Open Science GridThe Open Science Grid
The Open Science GridRob Gardner
 
Cloud Busting: Understanding Cloud-based Digital Forensics
Cloud Busting: Understanding Cloud-based Digital ForensicsCloud Busting: Understanding Cloud-based Digital Forensics
Cloud Busting: Understanding Cloud-based Digital ForensicsKerry Hazelton
 
Edge Computing and OGC SensorThings API
Edge Computing and OGC SensorThings APIEdge Computing and OGC SensorThings API
Edge Computing and OGC SensorThings APISensorUp
 
OGF Introductory Overview - OGF 44 at EGI Conference 2015
OGF Introductory Overview - OGF 44 at EGI Conference 2015OGF Introductory Overview - OGF 44 at EGI Conference 2015
OGF Introductory Overview - OGF 44 at EGI Conference 2015Alan Sill
 
Calgary-Splunk-User-Group-March-2023.pdf
Calgary-Splunk-User-Group-March-2023.pdfCalgary-Splunk-User-Group-March-2023.pdf
Calgary-Splunk-User-Group-March-2023.pdfPremDomingo
 
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...IRJET Journal
 
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...EUDAT
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardLarry Smarr
 
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...Ed Dodds
 
Internet2 - US Ignite Application Summit 2013
Internet2 - US Ignite Application Summit 2013Internet2 - US Ignite Application Summit 2013
Internet2 - US Ignite Application Summit 2013US-Ignite
 
Science Gateways: one portal, many e-Infrastructures and related services
Science Gateways: one portal, many e-Infrastructures and related servicesScience Gateways: one portal, many e-Infrastructures and related services
Science Gateways: one portal, many e-Infrastructures and related servicesriround
 
Big Data on Cloud Native Platform
Big Data on Cloud Native PlatformBig Data on Cloud Native Platform
Big Data on Cloud Native PlatformSunil Govindan
 

Similar to The Open Science Grid and how it relates to PRAGMA (20)

GRP 19 - Nautilus, IceCube and LIGO
GRP 19 - Nautilus, IceCube and LIGOGRP 19 - Nautilus, IceCube and LIGO
GRP 19 - Nautilus, IceCube and LIGO
 
GridIIT Open Science Grid
GridIIT Open Science GridGridIIT Open Science Grid
GridIIT Open Science Grid
 
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computingISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
ISC Cloud13 Sill - Crossing organizational boundaries in cloud computing
 
Bringing OSG users to the PRP Kubernetes Cluster
Bringing OSG users to the PRP Kubernetes ClusterBringing OSG users to the PRP Kubernetes Cluster
Bringing OSG users to the PRP Kubernetes Cluster
 
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
Designing and Implementing a cloud-hosted SaaS for data movement and Sharing ...
 
The Open Science Grid
The Open Science GridThe Open Science Grid
The Open Science Grid
 
QuSandbox+NVIDIA Rapids
QuSandbox+NVIDIA RapidsQuSandbox+NVIDIA Rapids
QuSandbox+NVIDIA Rapids
 
Cloud Busting: Understanding Cloud-based Digital Forensics
Cloud Busting: Understanding Cloud-based Digital ForensicsCloud Busting: Understanding Cloud-based Digital Forensics
Cloud Busting: Understanding Cloud-based Digital Forensics
 
Edge Computing and OGC SensorThings API
Edge Computing and OGC SensorThings APIEdge Computing and OGC SensorThings API
Edge Computing and OGC SensorThings API
 
THE OPEN SCIENCE GRID Ruth Pordes
THE OPEN SCIENCE GRID Ruth PordesTHE OPEN SCIENCE GRID Ruth Pordes
THE OPEN SCIENCE GRID Ruth Pordes
 
OGF Introductory Overview - OGF 44 at EGI Conference 2015
OGF Introductory Overview - OGF 44 at EGI Conference 2015OGF Introductory Overview - OGF 44 at EGI Conference 2015
OGF Introductory Overview - OGF 44 at EGI Conference 2015
 
Calgary-Splunk-User-Group-March-2023.pdf
Calgary-Splunk-User-Group-March-2023.pdfCalgary-Splunk-User-Group-March-2023.pdf
Calgary-Splunk-User-Group-March-2023.pdf
 
Orchestration, Automation and Virtualisation (OAV) in GÉANT
Orchestration, Automation and Virtualisation (OAV) in GÉANTOrchestration, Automation and Virtualisation (OAV) in GÉANT
Orchestration, Automation and Virtualisation (OAV) in GÉANT
 
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
IRJET-Open Curltm Cloud Computing Test Structure:Confederate Data Centers for...
 
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
Using the EGI Fed-Cloud for Data Analysis - EUDAT Summer School (Giuseppe La ...
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forward
 
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
Creating a Climate for Innovation on Internet2 - Eric Boyd Senior Director, S...
 
Internet2 - US Ignite Application Summit 2013
Internet2 - US Ignite Application Summit 2013Internet2 - US Ignite Application Summit 2013
Internet2 - US Ignite Application Summit 2013
 
Science Gateways: one portal, many e-Infrastructures and related services
Science Gateways: one portal, many e-Infrastructures and related servicesScience Gateways: one portal, many e-Infrastructures and related services
Science Gateways: one portal, many e-Infrastructures and related services
 
Big Data on Cloud Native Platform
Big Data on Cloud Native PlatformBig Data on Cloud Native Platform
Big Data on Cloud Native Platform
 

More from Igor Sfiligoi

Preparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROPreparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROIgor Sfiligoi
 
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...Igor Sfiligoi
 
Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Igor Sfiligoi
 
The anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingThe anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingIgor Sfiligoi
 
Auto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesAuto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesIgor Sfiligoi
 
Speeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateSpeeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateIgor Sfiligoi
 
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsPerformance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsIgor Sfiligoi
 
Comparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeComparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeIgor Sfiligoi
 
Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Igor Sfiligoi
 
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessAccelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessIgor Sfiligoi
 
Using A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputUsing A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputIgor Sfiligoi
 
Using commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsUsing commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsIgor Sfiligoi
 
Modest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROModest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROIgor Sfiligoi
 
Data-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstData-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstIgor Sfiligoi
 
Scheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyScheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyIgor Sfiligoi
 
Accelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCAccelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCIgor Sfiligoi
 
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Igor Sfiligoi
 
Porting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsPorting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsIgor Sfiligoi
 
Demonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsDemonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsIgor Sfiligoi
 
TransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksTransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksIgor Sfiligoi
 

More from Igor Sfiligoi (20)

Preparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROPreparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYRO
 
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
 
Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...
 
The anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingThe anachronism of whole-GPU accounting
The anachronism of whole-GPU accounting
 
Auto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesAuto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resources
 
Speeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateSpeeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rate
 
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsPerformance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
 
Comparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeComparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance compute
 
Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...
 
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessAccelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
 
Using A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputUsing A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific Output
 
Using commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsUsing commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobs
 
Modest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROModest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYRO
 
Data-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstData-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud Burst
 
Scheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyScheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with Admiralty
 
Accelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCAccelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACC
 
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
 
Porting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsPorting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUs
 
Demonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsDemonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public Clouds
 
TransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksTransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud links
 

Recently uploaded

Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenHervé Boutemy
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embeddingZilliz
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionDilum Bandara
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfAddepto
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024Stephanie Beckett
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 

Recently uploaded (20)

Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
DevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache MavenDevoxxFR 2024 Reproducible Builds with Apache Maven
DevoxxFR 2024 Reproducible Builds with Apache Maven
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Training state-of-the-art general text embedding
Training state-of-the-art general text embeddingTraining state-of-the-art general text embedding
Training state-of-the-art general text embedding
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
Advanced Computer Architecture – An Introduction
Advanced Computer Architecture – An IntroductionAdvanced Computer Architecture – An Introduction
Advanced Computer Architecture – An Introduction
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
Gen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdfGen AI in Business - Global Trends Report 2024.pdf
Gen AI in Business - Global Trends Report 2024.pdf
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024What's New in Teams Calling, Meetings and Devices March 2024
What's New in Teams Calling, Meetings and Devices March 2024
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 

The Open Science Grid and how it relates to PRAGMA

  • 1. THE OPEN SCIENCE GRID (OSG) And how it relates to PRAGMA by Igor Sfiligoi, UCSD/SDSC (isfiligoi@sdsc.edu) PRAGMA 37 - SEPT 13TH, 2019
  • 2. Premise – Not the Grid you were used to • It is my understanding Grid has a negative connotation in PRAGMA • While OSG has “Grid” in its name, I believe it is far from the “Grid” you experienced 15 years ago • Please give me a chance to convince you of that PRAGMA 37 - SEPT 13TH, 2019 2
  • 3. OSG evolution • 10 years ago, OSG provided an integrated software stack that organizations had to deploy in order for their researchers to benefit from OSG. •We learned the hard way that this is too high a bar to jump across for most researchers and institutions. • Today, we offer to operate services to the groups we serve in order to minimize the threshold to entry for everybody. • Leveraging work performed by the likes of PRP/NRP-P/TNRP (e.g. Kubernetes) PRAGMA 37 - SEPT 13TH, 2019 3
  • 4. OSG Aspirational Goal A campus Research IT Organization should not have to learn anything ”non-standard” in order to have their researchers benefit from OSG, or have their resources be available via OSG PRAGMA 37 - SEPT 13TH, 2019 4 Now …. Not there yet, but getting closer by the day…
  • 5. OSG Mission Advancing Open Science through distributed High Throughput Computing PRAGMA 37 - SEPT 13TH, 2019 5
  • 6. Open Science •All of open science irrespective of discipline •Include the maximum possible dynamic range of science, groups, and institutions • From individual undergraduates to international collaborations with thousands of members. • From small colleges, museums, zoos, to national scale centers of open science. PRAGMA 37 - SEPT 13TH, 2019 6
  • 7. What Sciences use OSG? • Order here is from most active to least active. • Category is based on self-identification … i.e. people say what they do. PRAGMA 37 - SEPT 13TH, 2019 7
  • 8. Not all users are the same • The individual researchers and small groups • The campus Research Support Organizations • Multi-institutional Science Teams • XENON, GlueX, SPT, Simons, and many many more • Collaborations between multiple campuses • The 4 “big science” projects: • US-ATLAS, US-CMS, LIGO, IceCube Used services change between different categories PRAGMA 37 - SEPT 13TH, 2019 8
  • 9. OSG Mission Advancing Open Science through distributed High Throughput Computing PRAGMA 37 - SEPT 13TH, 2019 9
  • 10. distributed High Throughput Computing DRIVING PRINCIPLES: • Maximize total compute, do not optimize for any single application • Allow for resources coming from anywhere in the world • Embrace heterogeneity • Federated policy control PRAGMA 37 - SEPT 13TH, 2019 10 (as opposed to e.g. HPC)
  • 11. distributed High Throughput Computing DRIVING PRINCIPLES: • Maximize total compute, do not optimize for any single application • Allow for resources coming from anywhere in the world • Embrace heterogeneity • Federated policy control IMPLICATIONS: • Expect high network latencies • Expect varying walltime between identical runs • Job restarts are the norm PRAGMA 37 - SEPT 13TH, 2019 11 (Although throughput should be high)
  • 12. Thriving in dHTC • The path to successful dHTC involves two aspects: • Separate a big computing problem in many individually schedulable small problems. • Minimize your requirements in order to maximize the raw capacity that you can effectively use. • The infrastructure will transparently take care of job restarts Ingenious Parallelism PRAGMA 37 - SEPT 13TH, 2019 12
  • 13. dHTC is special •dHTC scales by definition perfectly. • When a researcher understands how to partition their workflow into many individually schedulable compute problems they can scale out with ease to seemingly arbitrary scales of computing. • When we integrate all IT resources at Universities, National Labs, and the commercial cloud, we arrive at a near infinite resource pool. PRAGMA 37 - SEPT 13TH, 2019 13
  • 14. Stronger together • For OSG VO, more than half of the compute comes from smaller sites (and this does not include any of the big sciences, that have their own pools) 15k cores PRAGMA 37 - SEPT 13TH, 2019 14
  • 15. OSG IN DETAIL PRAGMA 37 - SEPT 13TH, 2019 15
  • 16. OSG Compute Services • OSG provides the means for operating a compute federation • Mostly in the form of Metascheduling Service • But also operates one access point itself for the smaller players PRAGMA 37 - SEPT 13TH, 2019 16
  • 17. Federation i.e. Distributed control • OSG works on three simple principles: • Resource Owners determine policy of use • This means that all policy of use is set locally by the clusters that join the federation. • Resource Consumers specify the types of resources they are willing to use. • How much RAM? How many cores per node? … • OSG submits HTCondor batch system as (pilot) payload into all local batch systems that match requirements, creating an overlay dHTC system • Jobs are submitted locally, queue centrally, and execute anywhere that matches requirements after resource becomes available. PRAGMA 37 - SEPT 13TH, 2019 17
  • 18. OSG Data Federation • OSG operates a content delivery network (CDN) • Based on XRootD caching technology • Data origins operated by Science communities • Data transparently published in our federated namespace. • Caches hide data access latencies and reduce network traffic. PRAGMA 37 - SEPT 13TH, 2019 18
  • 19. OSG Software Distribution •OSG can host and distribute community software on CVMFS • A CDN at heart, but based on Apache Squid technology • CVMFS provides a Read-Only POSIX interface •OSG encourages the use of containerization • Encourages participating sites to provide Singularity • Which is optimized for use by unprivileged users (unlike Docker) • Containers can be distributed through CVMFS (for performance) PRAGMA 37 - SEPT 13TH, 2019 19
  • 20. Facilitation Services • Proactive, personalized guidance and support for: • Institution-specific submit points • Sharing institutional resources via OSG • Data federation across OSG sites • Individual researchers using the OSG submit infrastructure • We also offer • Local workshops • OSG-hosted education and training • Learning from the OSG Facilitation Community PRAGMA 37 - SEPT 13TH, 2019 20
  • 21. OSG AND PRAGMA PRAGMA 37 - SEPT 13TH, 2019 21
  • 22. OSG AND PRAGMA • OSG would be happy to facilitate any Open Science endeavor that has at least some of its members in the US • We are a NSF-funded project • All OSG services are at your disposal • Compute federation • Data federation • Software distribution • Facilitation PRAGMA 37 - SEPT 13TH, 2019 22
  • 23. OSG and International Setups •All our major users are international • CMS, ATLAS, LIGO, IceCube, XENON, GlueX •But, alas, we have been mostly collaborating with Europe •PRAGMA could help us engage more Science communities in the Pacific Rim PRAGMA 37 - SEPT 13TH, 2019 23
  • 24. Summary and Conclusion • OSG’s objective is to “Advance Open Science through distributed High Throughput Computing” • We are committed to help any Open Science community that has at least some members in the USA – Nice overlap with PRAGMA • OSG can offer a wide variety of services • Both in terms of technology and facilitation • We could use some help getting more involved in the Pacific Rim • Contact us at: help@opensciencegrid.org PRAGMA 37 - SEPT 13TH, 2019 24