SlideShare a Scribd company logo
1 of 25
Download to read offline
Bringing OSG users to
the PRP Kubernetes Cluster
Presented by Igor Sfiligoi, UCSD
May 13, 2019 1
Outline
• PRP and Kubernetes
• OSG Overview
• OSG and Opportunistic use
• IceCube and LIGO science
• CVMFS and Unprivileged containers
• Lack of nested containerization
• Dealing with multiple users
• Some stats
May 13, 2019 2
PRP and Kubernetes
• Pacific Research Platform (PRP) originally created
as a regional networking project
• Establishing end-to-end links between 10Gbps and 100Gbps
• It has recently become also a major resource provider
• About 3.5k CPU cores and 330 GPUs
• About 2PB of storage space
• Kubernetes chosen for resource management
• Industry standard – Large and active development and support community
• Container based – More freedom for users
• Flexible scheduling – Allows for easy mixing of service and user workloads
May 13, 2019 3
The Open Science Grid (OSG)
• The Open Science Grid (OSG) is a NSF funded effort
• Open to all open science irrespective of discipline,
but does manage differently:
• The 4 “big science” projects
• Multi-institutional Science Teams,
• Campus Research Support Organizations,
• Individual researchers
• Focused on supporting dHTC workflows
• Due to almost perfect scalability
• Owns no compute resources
• Acts as a glue between and among
resource providers and users
Advancing
Open Science
through
distributed High Throughput Computing
(dHTC)
May 13, 2019 4
A few words about dHTC
The challenge in successful dHTC
is two-fold:
• Separate a big computing
problem in many
individually schedulable
small problems.
• Minimize your requirements
in order to maximize the raw
capacity that you can
effectively use.
As a computing paradigm, dHTC is special because
it scales by definition perfectly.
• When a researcher understands how to partition
their workflow into many individually schedulable
compute problems they can scale out with ease to
seemingly arbitrary scales of computing.
• When one integrates all IT resources at Universities,
National Labs, and the commercial cloud, you arrive
at a near infinite resource pool.
Ingenious ParallelismAttribution: From Frank Wuerthwein’s OSG talk May 13, 2019 5
dHTC and Opportunistic Use
• By creating a global virtual resource pool, OSG can both
• Help distributed organizations to spread their usage among many sites
• Give unused resources to external users – Opportunistic use
• Opportunistic use must be minimally invasive
• Resources owners should get their resources back
when they need them again (ideally, within seconds)
• dHTC ideally suited for this operational model
• All tasks are short and independent
• An opportunistic task can be killed without much loss;
the system will automatically reschedule it somewhere else
Never let
a cycle
go unused!
May 13, 2019 6
OSG use of PRP
• PRP currently has very little resource contention – GPU and CPU cycles were going unused
• OSG has several user communities who could use more resources
In this talk, OSG integration with
PRP is by means of opportunistic
use
• Low priority containers are only scheduled if there is no contention by higher priority ones
• When a higher priority container needs the resource, the lower priority container is
instantly killed
Kubernetes naturally allows for
opportunistic use by means of
priorities
• But not part of this talk
• Other future uses possible, too
OSG is using PRP also for running
some of its internal services
May 13, 2019 7
OSG users on PRP
PRP was the first Kubernetes-
based resource provider for OSG
• Lots of unknowns
• So we decided to only target a
few, well behaved user
communities as a starting point
We started with the two smaller
of the four big science projects
• IceCube
• LIGO
May 13, 2019 8
IceCube
• The IceCube Neutrino Observatory is designed
to observe the cosmos from deep within the
South Pole ice. Encompassing a cubic kilometer
of ice, IceCube searches for nearly massless
subatomic particles called neutrinos.
• One of the most compute intensive activities is
simulating the properties of ice and its photon
propagation in the presence of neutrinos.
Direct Photon Propagation
May 13, 2019 9
LIGO • LIGO’s mission is to open the field of gravitational-
wave astrophysics through the direct detection of
gravitational waves. LIGO detectors use laser
interferometry to measure the distortions in space-
time occurring between stationary, hanging masses
(mirrors) caused by passing gravitational waves.
• The main compute activities of LIGO are event
template timeseries searches and parameter fitting.
May 13, 2019 10
Integration with OSG
OSG natively does
not know how to talk
to Kubernetes
• We needed a batch
system interface
Instantiated a HTCondor pool as a
Kubernetes/Containerized deployment
• Pretty straightforward, just needed to create
images with HTCondor binaries in them
• Configuration wise, not much different than a
bare metal setup
• HTCondor deals gracefully with dynamic host
names
The OSG gateway
(known as a CE)
was also
containerized
• Here I needed some
elevated privileges
• Must use host IP and
DNS due to the use of
GSI/X.509
May 13, 2019 11
The drawbacks of containerization
OSG normally ask resource providers to
provide two services on all execute nodes
• CVMFS – A FUSE-mounted global filesystem
• Singularity – So users can launch their own containers
Both need elevated privileges
• Could not just put them in
the HTCondor execute Image
and run as a regular
Kubernetes Container
May 13, 2019 12
CVMFS and Kubernetes CSI
•Turns out it is a common enough problem (think Box and Netflix)
•From the technical point of view, it is implemented as admin-deployed, privileged
side containers
Kubernetes answer is CSI
(Container Storage Interface)
•CERN had developed against a beta version of the API (now deprecated)
•Dima Mishin did the re-factoring – Contributed back the changes
•Also switched from CERN-internal version to OSG-provided version of the RPMs
Had to fix CERN-
provided version
•Normally, CVMFS relies on autofs
•But autofs does not work in side-containers, so explicit mounting needed
Some minor
problems still remain
May 13, 2019 13
CVMFS and Kubernetes CSI
•Turns out it is a common enough problem (think Box and Netflix)
•From the technical point of view, it is implemented as admin-deployed, privileged
side containers
Kubernetes answer is CSI
(Container Storage Interface)
•CERN had developed against a beta version of the API (now deprecated)
•Dima Mishin did the re-factoring – Contributed back the changes
•Also switched from CERN-internal version to OSG-provided version of the RPMs
Had to fix CERN-
provided version
•Normally, CVMFS relies on autofs
•But autofs does not work in side-containers, so explicit mounting needed
Some minor
problems still remain
Working well enough nowMay 13, 2019 14
No Singularity
•OSG pilots can only invoke Singularity directly
•But nested containerization is not supported
Currently no real solution for
Singularity
•With only a couple supported users it is doable
•But still time consuming
•And will not scale
Have been creating
user specific
HTCondor execute Images
•HTCondor adding native support for Kubernetes
•OSG pilots could then provide the container Image to launch
•Exploring dynamic side-container options in Kubernetes
Looking at longer term
alternatives
May 13, 2019 15
Dockerfile blues
• Usually expect to be root when running
• Optimized for Singularity, not Kubernetes/Docker
• Differences in GPU driver integration particularly nasty
Users have images
they use on OSG, but
• Users not root anymore
• May have conflicts in required libraries
• Environment differences
Need to inject OSG
environment and HTCondor
• It is usually just a set of yum installs and the like
• Adding the PRP-specific additions then gets easy
• But a few back-and-forts usually still needed
Easier with access to
their Dockerfile
May 13, 2019 16
Contention in Opportunistic use
We started with
a single OSG user
•Using all opportunistic
resources was easy
•Just keep enough low-priority
pods in the system
Adding a second user
was easy, too
•First was a GPU user, the
second was a CPU user
•No contention between them
First user wants the same
resources as second user
•They use different container Images
• Contention!
May 13, 2019 17
Learning to deal with contention
• Was not a problem in the absence of opportunistic contention
• Should be solvable, similar to how pilots operate in OSG
OSG pods currently stay alive
even if no work available
• Not with the default scheduler
• Basically priority-FIFO
Kubernetes not good at
contention management
• A couple of times a week was good enough so far
• But will need a better solution longer term
Currently manually adjusting
pressure
May 13, 2019 18
OSG a special case in Nautilus
• OSG “user” is different than other Nautilus users
• Not a regular user – Preemption-tolerant, low priority
• Not a service – Heavy users of GPUs and CPUs
• Nautilus admins had to create special rules for us
• E.g. it is OK to “waste” GPUs
• But the same rules would likely apply to any opportunistic user
May 13, 2019 19
Some stats – IceCube GPU
• First use case
• over 2 months now
• Periods of demand
• Followed by only small bursts of requests
May 13, 2019 20
Some Stats – CPU usage
• Started with public LIGO
• They had no GPU needs
• But progress was slow
• Recently added IceCube
• Output of CPU jobs needed to run GPU jobs
• Should result in higher GPU demand
May 13, 2019 21
Some stats – LIGO GPU
• And LIGO experiment now needs GPUs, too
• Although demand still low
May 13, 2019 22
Preemption in action
• Kubernetes will automatically regulate number of slots
• Kill containers when higher-priority users need them
• Re-start the OSG containers when nobody else is requesting them
May 13, 2019 23
Questions
welcome
May 13, 2019 24
Acknowledgements
• PRP/TNRP is supported by US National
Science Foundation (NSF) awards CNS-
1456638, CNS-1730158, ACI-1540112,
ACI-1541349, OAC-1826967, OAC
1450871 and OAC-1659169.
• OSG Multi-Messenger Astrophysics
activities are supported by US
National Science Foundation (NSF)
award OAC-1841530.
May 13, 2019 25

More Related Content

Similar to Bringing OSG users to the PRP Kubernetes Cluster

The Open Science Grid and how it relates to PRAGMA
The Open Science Grid and how it relates to PRAGMAThe Open Science Grid and how it relates to PRAGMA
The Open Science Grid and how it relates to PRAGMAIgor Sfiligoi
 
Latest (storage IO) patterns for cloud-native applications
Latest (storage IO) patterns for cloud-native applications Latest (storage IO) patterns for cloud-native applications
Latest (storage IO) patterns for cloud-native applications OpenEBS
 
Exploring a simpler, more portable, less overhead solution to deploy Elastics...
Exploring a simpler, more portable, less overhead solution to deploy Elastics...Exploring a simpler, more portable, less overhead solution to deploy Elastics...
Exploring a simpler, more portable, less overhead solution to deploy Elastics...LetsConnect
 
CloudLab Overview
CloudLab OverviewCloudLab Overview
CloudLab OverviewEd Dodds
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Adventures in Research
Adventures in ResearchAdventures in Research
Adventures in ResearchNETWAYS
 
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick OpenNebulaConf 2013 -Adventures in Research by Joel Merrick
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick OpenNebula Project
 
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Region
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain RegionPEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Region
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Regionpruprecht
 
GridIIT Open Science Grid
GridIIT Open Science GridGridIIT Open Science Grid
GridIIT Open Science GridSciCompIIT
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardLarry Smarr
 
Kubernetes The New Research Platform
Kubernetes The New Research PlatformKubernetes The New Research Platform
Kubernetes The New Research PlatformBob Killen
 
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...chiportal
 
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scale
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production ScaleGPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scale
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scalesparktc
 
Future of Data Platform in Cloud Native world
Future of Data Platform in Cloud Native worldFuture of Data Platform in Cloud Native world
Future of Data Platform in Cloud Native worldSrivatsan Srinivasan
 
2014 11-05 hpcac-kniep_christian_dockermpi
2014 11-05 hpcac-kniep_christian_dockermpi2014 11-05 hpcac-kniep_christian_dockermpi
2014 11-05 hpcac-kniep_christian_dockermpiQNIB Solutions
 
Consul connect
Consul connectConsul connect
Consul connectjabizz
 
Demonstrating 100 Gbps in and out of the Clouds
Demonstrating 100 Gbps in and out of the CloudsDemonstrating 100 Gbps in and out of the Clouds
Demonstrating 100 Gbps in and out of the CloudsIgor Sfiligoi
 
Technical standards & the RDTF Vision: some considerations
Technical standards & the RDTF Vision: some considerationsTechnical standards & the RDTF Vision: some considerations
Technical standards & the RDTF Vision: some considerationsPaul Walk
 

Similar to Bringing OSG users to the PRP Kubernetes Cluster (20)

The Open Science Grid and how it relates to PRAGMA
The Open Science Grid and how it relates to PRAGMAThe Open Science Grid and how it relates to PRAGMA
The Open Science Grid and how it relates to PRAGMA
 
Latest (storage IO) patterns for cloud-native applications
Latest (storage IO) patterns for cloud-native applications Latest (storage IO) patterns for cloud-native applications
Latest (storage IO) patterns for cloud-native applications
 
Exploring a simpler, more portable, less overhead solution to deploy Elastics...
Exploring a simpler, more portable, less overhead solution to deploy Elastics...Exploring a simpler, more portable, less overhead solution to deploy Elastics...
Exploring a simpler, more portable, less overhead solution to deploy Elastics...
 
CloudLab Overview
CloudLab OverviewCloudLab Overview
CloudLab Overview
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Adventures in Research
Adventures in ResearchAdventures in Research
Adventures in Research
 
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick OpenNebulaConf 2013 -Adventures in Research by Joel Merrick
OpenNebulaConf 2013 -Adventures in Research by Joel Merrick
 
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Region
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain RegionPEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Region
PEARC17: Deploying RMACC Summit: An HPC Resource for the Rocky Mountain Region
 
GridIIT Open Science Grid
GridIIT Open Science GridGridIIT Open Science Grid
GridIIT Open Science Grid
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forward
 
Kubernetes The New Research Platform
Kubernetes The New Research PlatformKubernetes The New Research Platform
Kubernetes The New Research Platform
 
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...
TRACK F: OpenCL for ALTERA FPGAs, Accelerating performance and design product...
 
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scale
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production ScaleGPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scale
GPU Support in Spark and GPU/CPU Mixed Resource Scheduling at Production Scale
 
Future of Data Platform in Cloud Native world
Future of Data Platform in Cloud Native worldFuture of Data Platform in Cloud Native world
Future of Data Platform in Cloud Native world
 
2014 11-05 hpcac-kniep_christian_dockermpi
2014 11-05 hpcac-kniep_christian_dockermpi2014 11-05 hpcac-kniep_christian_dockermpi
2014 11-05 hpcac-kniep_christian_dockermpi
 
Consul connect
Consul connectConsul connect
Consul connect
 
Demonstrating 100 Gbps in and out of the Clouds
Demonstrating 100 Gbps in and out of the CloudsDemonstrating 100 Gbps in and out of the Clouds
Demonstrating 100 Gbps in and out of the Clouds
 
Technical standards & the RDTF Vision: some considerations
Technical standards & the RDTF Vision: some considerationsTechnical standards & the RDTF Vision: some considerations
Technical standards & the RDTF Vision: some considerations
 

More from Igor Sfiligoi

Preparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROPreparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROIgor Sfiligoi
 
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...Igor Sfiligoi
 
Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Igor Sfiligoi
 
The anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingThe anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingIgor Sfiligoi
 
Auto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesAuto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesIgor Sfiligoi
 
Speeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateSpeeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateIgor Sfiligoi
 
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsPerformance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsIgor Sfiligoi
 
Comparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeComparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeIgor Sfiligoi
 
Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Igor Sfiligoi
 
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessAccelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessIgor Sfiligoi
 
Using A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputUsing A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputIgor Sfiligoi
 
Using commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsUsing commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsIgor Sfiligoi
 
Modest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROModest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROIgor Sfiligoi
 
Data-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstData-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstIgor Sfiligoi
 
Scheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyScheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyIgor Sfiligoi
 
Accelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCAccelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCIgor Sfiligoi
 
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Igor Sfiligoi
 
Porting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsPorting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsIgor Sfiligoi
 
Demonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsDemonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsIgor Sfiligoi
 
TransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksTransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksIgor Sfiligoi
 

More from Igor Sfiligoi (20)

Preparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYROPreparing Fusion codes for Perlmutter - CGYRO
Preparing Fusion codes for Perlmutter - CGYRO
 
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
O&C Meeting - Evaluation of ARM CPUs for IceCube available through Google Kub...
 
Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...Comparing single-node and multi-node performance of an important fusion HPC c...
Comparing single-node and multi-node performance of an important fusion HPC c...
 
The anachronism of whole-GPU accounting
The anachronism of whole-GPU accountingThe anachronism of whole-GPU accounting
The anachronism of whole-GPU accounting
 
Auto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resourcesAuto-scaling HTCondor pools using Kubernetes compute resources
Auto-scaling HTCondor pools using Kubernetes compute resources
 
Speeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rateSpeeding up bowtie2 by improving cache-hit rate
Speeding up bowtie2 by improving cache-hit rate
 
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence SimulationsPerformance Optimization of CGYRO for Multiscale Turbulence Simulations
Performance Optimization of CGYRO for Multiscale Turbulence Simulations
 
Comparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance computeComparing GPU effectiveness for Unifrac distance compute
Comparing GPU effectiveness for Unifrac distance compute
 
Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...Managing Cloud networking costs for data-intensive applications by provisioni...
Managing Cloud networking costs for data-intensive applications by provisioni...
 
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory AccessAccelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
Accelerating Key Bioinformatics Tasks 100-fold by Improving Memory Access
 
Using A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific OutputUsing A100 MIG to Scale Astronomy Scientific Output
Using A100 MIG to Scale Astronomy Scientific Output
 
Using commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobsUsing commercial Clouds to process IceCube jobs
Using commercial Clouds to process IceCube jobs
 
Modest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYROModest scale HPC on Azure using CGYRO
Modest scale HPC on Azure using CGYRO
 
Data-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud BurstData-intensive IceCube Cloud Burst
Data-intensive IceCube Cloud Burst
 
Scheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with AdmiraltyScheduling a Kubernetes Federation with Admiralty
Scheduling a Kubernetes Federation with Admiralty
 
Accelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACCAccelerating microbiome research with OpenACC
Accelerating microbiome research with OpenACC
 
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
Demonstrating a Pre-Exascale, Cost-Effective Multi-Cloud Environment for Scie...
 
Porting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUsPorting and optimizing UniFrac for GPUs
Porting and optimizing UniFrac for GPUs
 
Demonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public CloudsDemonstrating 100 Gbps in and out of the public Clouds
Demonstrating 100 Gbps in and out of the public Clouds
 
TransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud linksTransAtlantic Networking using Cloud links
TransAtlantic Networking using Cloud links
 

Recently uploaded

The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxKatpro Technologies
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksSoftradix Technologies
 

Recently uploaded (20)

The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptxFactors to Consider When Choosing Accounts Payable Services Providers.pptx
Factors to Consider When Choosing Accounts Payable Services Providers.pptx
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other Frameworks
 

Bringing OSG users to the PRP Kubernetes Cluster

  • 1. Bringing OSG users to the PRP Kubernetes Cluster Presented by Igor Sfiligoi, UCSD May 13, 2019 1
  • 2. Outline • PRP and Kubernetes • OSG Overview • OSG and Opportunistic use • IceCube and LIGO science • CVMFS and Unprivileged containers • Lack of nested containerization • Dealing with multiple users • Some stats May 13, 2019 2
  • 3. PRP and Kubernetes • Pacific Research Platform (PRP) originally created as a regional networking project • Establishing end-to-end links between 10Gbps and 100Gbps • It has recently become also a major resource provider • About 3.5k CPU cores and 330 GPUs • About 2PB of storage space • Kubernetes chosen for resource management • Industry standard – Large and active development and support community • Container based – More freedom for users • Flexible scheduling – Allows for easy mixing of service and user workloads May 13, 2019 3
  • 4. The Open Science Grid (OSG) • The Open Science Grid (OSG) is a NSF funded effort • Open to all open science irrespective of discipline, but does manage differently: • The 4 “big science” projects • Multi-institutional Science Teams, • Campus Research Support Organizations, • Individual researchers • Focused on supporting dHTC workflows • Due to almost perfect scalability • Owns no compute resources • Acts as a glue between and among resource providers and users Advancing Open Science through distributed High Throughput Computing (dHTC) May 13, 2019 4
  • 5. A few words about dHTC The challenge in successful dHTC is two-fold: • Separate a big computing problem in many individually schedulable small problems. • Minimize your requirements in order to maximize the raw capacity that you can effectively use. As a computing paradigm, dHTC is special because it scales by definition perfectly. • When a researcher understands how to partition their workflow into many individually schedulable compute problems they can scale out with ease to seemingly arbitrary scales of computing. • When one integrates all IT resources at Universities, National Labs, and the commercial cloud, you arrive at a near infinite resource pool. Ingenious ParallelismAttribution: From Frank Wuerthwein’s OSG talk May 13, 2019 5
  • 6. dHTC and Opportunistic Use • By creating a global virtual resource pool, OSG can both • Help distributed organizations to spread their usage among many sites • Give unused resources to external users – Opportunistic use • Opportunistic use must be minimally invasive • Resources owners should get their resources back when they need them again (ideally, within seconds) • dHTC ideally suited for this operational model • All tasks are short and independent • An opportunistic task can be killed without much loss; the system will automatically reschedule it somewhere else Never let a cycle go unused! May 13, 2019 6
  • 7. OSG use of PRP • PRP currently has very little resource contention – GPU and CPU cycles were going unused • OSG has several user communities who could use more resources In this talk, OSG integration with PRP is by means of opportunistic use • Low priority containers are only scheduled if there is no contention by higher priority ones • When a higher priority container needs the resource, the lower priority container is instantly killed Kubernetes naturally allows for opportunistic use by means of priorities • But not part of this talk • Other future uses possible, too OSG is using PRP also for running some of its internal services May 13, 2019 7
  • 8. OSG users on PRP PRP was the first Kubernetes- based resource provider for OSG • Lots of unknowns • So we decided to only target a few, well behaved user communities as a starting point We started with the two smaller of the four big science projects • IceCube • LIGO May 13, 2019 8
  • 9. IceCube • The IceCube Neutrino Observatory is designed to observe the cosmos from deep within the South Pole ice. Encompassing a cubic kilometer of ice, IceCube searches for nearly massless subatomic particles called neutrinos. • One of the most compute intensive activities is simulating the properties of ice and its photon propagation in the presence of neutrinos. Direct Photon Propagation May 13, 2019 9
  • 10. LIGO • LIGO’s mission is to open the field of gravitational- wave astrophysics through the direct detection of gravitational waves. LIGO detectors use laser interferometry to measure the distortions in space- time occurring between stationary, hanging masses (mirrors) caused by passing gravitational waves. • The main compute activities of LIGO are event template timeseries searches and parameter fitting. May 13, 2019 10
  • 11. Integration with OSG OSG natively does not know how to talk to Kubernetes • We needed a batch system interface Instantiated a HTCondor pool as a Kubernetes/Containerized deployment • Pretty straightforward, just needed to create images with HTCondor binaries in them • Configuration wise, not much different than a bare metal setup • HTCondor deals gracefully with dynamic host names The OSG gateway (known as a CE) was also containerized • Here I needed some elevated privileges • Must use host IP and DNS due to the use of GSI/X.509 May 13, 2019 11
  • 12. The drawbacks of containerization OSG normally ask resource providers to provide two services on all execute nodes • CVMFS – A FUSE-mounted global filesystem • Singularity – So users can launch their own containers Both need elevated privileges • Could not just put them in the HTCondor execute Image and run as a regular Kubernetes Container May 13, 2019 12
  • 13. CVMFS and Kubernetes CSI •Turns out it is a common enough problem (think Box and Netflix) •From the technical point of view, it is implemented as admin-deployed, privileged side containers Kubernetes answer is CSI (Container Storage Interface) •CERN had developed against a beta version of the API (now deprecated) •Dima Mishin did the re-factoring – Contributed back the changes •Also switched from CERN-internal version to OSG-provided version of the RPMs Had to fix CERN- provided version •Normally, CVMFS relies on autofs •But autofs does not work in side-containers, so explicit mounting needed Some minor problems still remain May 13, 2019 13
  • 14. CVMFS and Kubernetes CSI •Turns out it is a common enough problem (think Box and Netflix) •From the technical point of view, it is implemented as admin-deployed, privileged side containers Kubernetes answer is CSI (Container Storage Interface) •CERN had developed against a beta version of the API (now deprecated) •Dima Mishin did the re-factoring – Contributed back the changes •Also switched from CERN-internal version to OSG-provided version of the RPMs Had to fix CERN- provided version •Normally, CVMFS relies on autofs •But autofs does not work in side-containers, so explicit mounting needed Some minor problems still remain Working well enough nowMay 13, 2019 14
  • 15. No Singularity •OSG pilots can only invoke Singularity directly •But nested containerization is not supported Currently no real solution for Singularity •With only a couple supported users it is doable •But still time consuming •And will not scale Have been creating user specific HTCondor execute Images •HTCondor adding native support for Kubernetes •OSG pilots could then provide the container Image to launch •Exploring dynamic side-container options in Kubernetes Looking at longer term alternatives May 13, 2019 15
  • 16. Dockerfile blues • Usually expect to be root when running • Optimized for Singularity, not Kubernetes/Docker • Differences in GPU driver integration particularly nasty Users have images they use on OSG, but • Users not root anymore • May have conflicts in required libraries • Environment differences Need to inject OSG environment and HTCondor • It is usually just a set of yum installs and the like • Adding the PRP-specific additions then gets easy • But a few back-and-forts usually still needed Easier with access to their Dockerfile May 13, 2019 16
  • 17. Contention in Opportunistic use We started with a single OSG user •Using all opportunistic resources was easy •Just keep enough low-priority pods in the system Adding a second user was easy, too •First was a GPU user, the second was a CPU user •No contention between them First user wants the same resources as second user •They use different container Images • Contention! May 13, 2019 17
  • 18. Learning to deal with contention • Was not a problem in the absence of opportunistic contention • Should be solvable, similar to how pilots operate in OSG OSG pods currently stay alive even if no work available • Not with the default scheduler • Basically priority-FIFO Kubernetes not good at contention management • A couple of times a week was good enough so far • But will need a better solution longer term Currently manually adjusting pressure May 13, 2019 18
  • 19. OSG a special case in Nautilus • OSG “user” is different than other Nautilus users • Not a regular user – Preemption-tolerant, low priority • Not a service – Heavy users of GPUs and CPUs • Nautilus admins had to create special rules for us • E.g. it is OK to “waste” GPUs • But the same rules would likely apply to any opportunistic user May 13, 2019 19
  • 20. Some stats – IceCube GPU • First use case • over 2 months now • Periods of demand • Followed by only small bursts of requests May 13, 2019 20
  • 21. Some Stats – CPU usage • Started with public LIGO • They had no GPU needs • But progress was slow • Recently added IceCube • Output of CPU jobs needed to run GPU jobs • Should result in higher GPU demand May 13, 2019 21
  • 22. Some stats – LIGO GPU • And LIGO experiment now needs GPUs, too • Although demand still low May 13, 2019 22
  • 23. Preemption in action • Kubernetes will automatically regulate number of slots • Kill containers when higher-priority users need them • Re-start the OSG containers when nobody else is requesting them May 13, 2019 23
  • 25. Acknowledgements • PRP/TNRP is supported by US National Science Foundation (NSF) awards CNS- 1456638, CNS-1730158, ACI-1540112, ACI-1541349, OAC-1826967, OAC 1450871 and OAC-1659169. • OSG Multi-Messenger Astrophysics activities are supported by US National Science Foundation (NSF) award OAC-1841530. May 13, 2019 25