SlideShare a Scribd company logo
1 of 57
1
Computing Just What You Need:
Online Data Analysis and Reduction
at Extreme Scales
Ian Foster
Argonne National Lab & University of Chicago
December 21, 2017
HiPC, Jaipur, India
https://www.researchgate.net/publication/317703782
foster@anl.gov
2
Earth to be paradise; distance to lose enchantment
“If, as it is said to be not unlikely in
the near future, the principle of
sight is applied to the telephone
as well as that of sound, earth will
be in truth a paradise, and
distance will lose its enchantment
by being abolished altogether.”
— Arthur Mee, 1898
3
Cooperating at a distance
1994 1999, 2003 2017
4
Automating the research data lifecycle.
Reducing barriers to cooperation at a distance
globus.org
5
Automating research data lifecycle
5
major
services
13
national labs
use Globus
340PB
transferred
10,000
active endpoints
50 Bn
files processed
75,000
registered users
99.5%
uptime
65+
institutional
subscribers
1 PB
largest single
transfer to date
3 months
longest continuously
managed transfer
300+
federated
campus
identities
12,000
active users/year
6
Transferring 1PB in a day
Argonne → NCSA
• Cosmology simulation
on Mira @ Argonne
produces 1 PB in 24
hours
• Data streamed to Blue
Waters for analytics
• Application reveals
feasibility of real-time
streaming at scale
Without checksums
With checksums
7
Time to discovery
Simula on me
Ultra
scale
Where does HPC fit in the research lifecycle?
8
The challenges of managing data and
computation at the 1018 scale
US Department of Energy
9
US Exascale Computing Program: Using codesign
and integration to achieve capable exascale
US Department of Energy
10
A. C. Bauer et al., EuroVis 2016
Computation
125 PB/s
Node memory
4.5 PB/s
Interconnect (largest cross-sectional b/w)
24 TB/s
Storage
1.4 TB/s
Interconnect
24 TB/s
Node memory
4.5 PB/s
Changing storage geography a major challenge for exascale
11
Disk didn’t use to be the problem
~1980-2000
Patterson,
CACM, 2004
12
Disks are getting larger, but not faster
~1980-2000
Patterson,
CACM,
2004
https://www.backblaze.com/blog/hard-drive-cost-per-gigabyte/
13
10 180
x18
0.3 1
x3
14
15
Exascale climate goal: Ensembles of 1km models at
15 simulated years/24 hours
Full state once per model day  260 TB every 16 seconds
 1.4 EB/day
16
Time to discovery
Simula on me
Ultra
scale
17
Time to discovery
Simula on me
Ultra
scale
Data Space tools: Popula on, naviga on, manipula on, dissemina on
Leadership
class facility
Smaller systems
Leadership
class facility
Smaller systems
Ultra
scale
18
Time to discovery
Simula on me
Ultra
scale
Data Space tools: Popula on, naviga on, manipula on, dissemina on
Leadership
class facility
Smaller systems
Leadership
class facility
Smaller systems
Ultra
scale
19
Time to discovery
Simula on me
Ultra
scale
Data Space tools: Popula on, naviga on, manipula on, dissemina on
Leadership
class facility
Smaller systems
Leadership
class facility
Smaller systems
Ultra
scale
20
Time to discovery
Simula on me
Ultra
scale
Data Space tools: Popula on, naviga on, manipula on, dissemina on
Leadership
class facility
Smaller systems
Leadership
class facility
Smaller systems
Ultra
scale
21
The need for online data analysis and reduction
Traditional approach:
Simulate, output, analyze
Write simulation output to secondary
storage; read back for analysis
Decimate in time when simulation
output rate exceeds output rate of
computer
Online: y = F(x)
Offline: a = A(y), b= B(y), …
22
The need for online data analysis and reduction
Traditional approach:
Simulate, output, analyze
Write simulation output to secondary
storage; read back for analysis
Decimate in time when simulation
output rate exceeds output rate of
computer
Online: y = F(x)
Offline: a = A(y), b= B(y), …
New approach:
Online data analysis & reduction
Co-optimize simulation, analysis,
reduction for performance and
information output
Substitute CPU cycles for I/O, via online
data (de)compression and/or analysis
a) Online: a = A(F(x)), b = B(F(x)), …
b) Online: r = R(F(x))
Offline: a = A’(r), b = B’(r), or
a = A(U(r)), b = B(U(r))
[R = reduce, U = un-reduce]
23
Exascale computing at Argonne by 2021
Precision medicine
Data from
sensors and
scientific
instruments
Simulation and
modeling of
materials and
physical
systems
Support for three types of computing:
Traditional: HPC simulation and modeling
Learning: Machine learning, deep learning, AI
Data: Data analytics, data science, big data
[Artists impression]
25
Real-time analysis and experimental steering
• Current protocols process
and validate data only after
experiment, which can lead
to undetected errors and
prevents online steering
• Process data streamed
from beamline to
supercomputer; control
feedback loop makes
decisions during
experiment
• Tests in TXM beamline (32-
ID@APS) in cement wetting
experiment (2 experiments,
each with 8 hours of data
acquisition time)
Sustained # Projections/seconds
CircularBufferSize
Reconstruction Frequency
Image Quality w.r.t. Streamed Projections
SimilarityScore
# Streamed Projections Reconstructed
Image Sequence
Tekin Bicer et al., eScience 2017
26
Deep learning for precision medicine
https://de.mathworks.com/company/newsletters/articles/cancer-diagnostics-with-deep-learning-and-photonic-time-stretch.html
27
Model
selection
Model
training
Inference
Incremental
training
Training
data
Q
A
Training
data
New
data
Human
expertise
model
architecture
trained
model
Deep learning on HPC systems
Evaluate 1M alternative models,
each with 100M parameters
 1014 parameter values
Using learning to optimize simulation studies
Simulation data Learning methods New capabilities
New simulations
Using learning to optimize simulation studies
Logan Ward and Ben Blaiszik
30
Synopsis: Applications are changing
Single
program
Multiple
program
Offline
analysis
Online
analysis
A few or many tasks:
• Loosely or tightly coupled
• Hierarchical or not
• Static or dynamic
• Fail-stop or recoverable
• Shared state
• Persistent and transient state
• Scheduled or data driven
Multiple
simulations
+ analyses
Simulation
+ analysis
Multiple
simulations
31
Many interesting codesign problems
Big simulation
Machine learning
Deep learning
Streaming
Online analysis
Online reduction
Heterogeneity
Prog. models
- Many task
- Streaming
Libraries
- Analysis, reduction
- Communications
System software
- Fault tolerance
- Resource mgmt
Complex nodes
- Many core
- Accelerators
- Heterogeneous
NVRAM
Networks
- Internal
- External
Node configuration
Internal networks
External networks
Memory hierarchy
Storage systems
Heterogeneity
Operating policies
32
Reduction comes with challenges
• Handling high entropy
• Performance – no benefit
otherwise
• Not only error in variable:
Ε ≡ 𝑓 − 𝑓
• Must also consider impact
on derived quantities:
Ε ≡ (𝑔𝑙
𝑡
(𝑓 𝑥, 𝑡 ) − 𝑔𝑙
𝑡
( 𝑓𝑙
𝑡
( 𝑥, 𝑡 )
S. Klasky
33
Key research challenge:
How to manage the impact
of errors on derived
quantities?
Where did it go???
S. Klasky
Reduction comes with challenges
34
CODAR: Codesign center for Online Data
Analysis and Reduction
• Infrastructure development and deployment
• Enable rapid composition of application and “data services” (data
reduction methods, data analysis methods, etc.)
• Support CODAR-developed and other data services
• Method development: new reduction & analysis routines
• Motif-specific: e.g., finite difference mesh vs. particles vs. finite elements
• Application-specific: e.g., reduced physics to understand deltas
• Application engagement
• Understand data analysis and reduction requirements
• Integrate, deploy, evaluate impact
https://codarcode.github.io codar-info@cels.anl.gov
35
Cross-cutting research questions
What are the best data analysis and reduction algorithms for
different application classes, in terms of speed, accuracy, and
resource needs? How can we implement those algorithms to
achieve scalability and performance portability?
What are tradeoffs in analysis accuracy, resource needs,
and overall application performance between using various
data reduction methods online prior to offline data reconstruction
and analysis vs. performing more online data analysis? How do
tradeoffs vary with hardware & software choices?
How do we effectively orchestrate online data analysis and
reduction to reduce associated overheads? How can hardware
and software help with orchestration?
36
Prototypical data analysis and reduction pipeline
CODAR runtime
Reduced output and reconstruction info
I/O
system
CODAR data API
Running simulation
Multivariate statistics
Feature analysis
Outlier detection
Application-aware
Transforms
Encodings
Error calculation
Refinement hints
CODARdataAPI
Offlinedataanalysis
Simulation knowledge: application, models, numerics, performance optimization, …
CODAR
data analysis
CODAR
data reduction
CODAR
data monitoring
37
Overarching data reduction challenges
• Understanding the science requires massive data reduction
• How do we reduce
• The time spent in reducing the data to knowledge?
• The amount of data moved on the HPC platform?
• The amount of data read from the storage system?
• The amount of data stored in memory, on storage system, moved over WAN?
• Without removing the knowledge.
• Requires deep dives into application post processing routines and simulations
• Goal is to create both (a) co-design infrastructure and (b)
reduction and analysis routines
• General: e.g., reduce Nbytes to Mbytes, N<<M
• Motif-specific: e.g., finite difference mesh vs. particles vs. finite elements
• Application-specific: e.g. reduced physics allows us to understand deltas
38
HPC floating point compression
• Current interest is with lossy algorithms, some use preprocessing
• Lossless may achieve up to ~3x reduction
• ISABELA
• SZ
• ZFP
• Linear auditing
• SVD
• Adaptive gradient methods
Compress each variable separately: Several variables simultaneously:
• PCA
• Tensor decomposition
• …
39
Lossy compression with SZ
No existing compressor can reduce hard to compress
datasets by more than a factor of 2.
Objective 1: Reduce hard to compress datasets by
one order of magnitude
Objective 2: Add user-required error controls (error
bound, shape of error distribution, spectral behavior
of error function, etc. etc.)
NCAR
atmosphere
simulation
output
(1.5 TB)
WRF
hurricane
simulation
output
Advanced
Photon Source
mouse brain
data
What we need to
compress
(bit map of 128
floating point
numbers):
Random noise
Franck Cappello
40
Lossy compression: Atmospheric simulation
Franck Cappello
Latest SZ
41
Characterizing compression error
0.0001
0.001
0.01
0.1
1 1/N 6/N 11/N 16/N 21/N 26/N 31/N 36/N 41/N 46/N
Amplitude
Frequency 0
2e-07
4e-07
6e-07
8e-07
1e-06
1.2e-06
1.4e-06
1.6e-06
1.8e-06
2e-06
0 20 40 60 80 100
120
140
160
180
200MaximumCompressionError
Variables
SZ(max error)
SZ(avg error)
ZFP(max error)
ZFP(avg error)
Error distribution
Spectral behavior Laplacian (derivatives)
Autocorrelation of errors
Respect of error bounds Error propagation
Franck Cappello
42
Z-checker: Analysis of data reduction error
• Community tool to enable comprehensive assessment of lossy data reduction error:
• Collection of data quality criteria from applications
• Community repository for datasets, reduction quality requirements, compression
performance
• Modular design enables contributed analysis modules (C and R) and format
readers (ADIOS, HDF5, etc.)
• Off-line/on-line parallel statistical, spectral, point-wise distortion analysis with static
& dynamic visualization
Franck Cappello, Julie Bessac, Sheng Di
43
Z-Checker computations
• Normalized root mean squared error
• Peak signal to noise ratio
• Distribution of error
• Pearson correlation between raw and reduced datasets
• Power spectrum distortion
• Auto-correlation of compression error
• Maximum error
• Point-wise error bound (relative or absolute)
• Preservation of derivatives
• Structural similarity (SSIM) index
44
Science-driven optimizations
• Information-theoretically derived methods like SZ,
Isabella, ZFP make for good generic capabilities
• If scientists can provide additional details on how to
determine features of interest, we can use those to
drive further optimizations. E.g., if they can select:
• Regions of high gradient
• Regions near turbulent flow
• Particles with velocities > two standard deviations
• How can scientists help define features?
45
Multilevel compression techniques
A hierarchical reduction scheme produces
multiple levels of partial decompression of the
data so that users can work with reduced
representations that require minimal storage
whilst achieving user-specified tolerance
Compression vs. user-specified toleranceResults for turbulence dataset: extremely large,
inherently non-smooth, resistant to compression Mark Ainsworth
46
Manifold learning for change detection and
adaptive sampling
Low dimensional manifold projection
of different state of MD trajectories
• A single molecular dynamics
trajectory can generate 32 PB
• Use online data analysis to detect
relevant or significant events
• Project MD trajectories to manifold
space (dimensionality reduction) across
time into two dimensional space
• Change detection on manifold space is
more robust than original full coordinate
space as it removes local vibrational
noise
• Apply adaptive sampling strategy based
on accumulated changes of trajectories
Shinjae Yoo
47
Critical points extracted
with topology analysis
Tracking blobs in XGC fusion simulations
Blobs, regions of high turbulence that can
damage the Tokamak, can run along the
edge wall down toward the diverter and
damage it. Blob extraction and tracking
enables the exploration and analysis of
high-energy blobs across timesteps.
• Access data with ADIOS I/O in high
performance
• Precondition the input data with robust PCA
• Detect blobs as local extrema with topology
analysis
• Track blobs over time with combinatorial
feature flow field method
Extracting, tracking, and visualizing blobs in large 5D gyrokinetic Tokamak simulations
Hanqi Guo, Tom Peterka
Tracking graph that visualizes the dynamics of blobs
(birth, merge, split, and death) over time
Data preconditioning
with robust PCA
48
Reduction for visualization
“an extreme scale simulation … calculates
temperature and density over 1000 of time
steps. For both variables, a scientist would like
to visualize 10 isosurface values and X, Y, and Z
cut planes for 10 locations in each dimension.
One hundred different camera positions are
also selected, in a hemisphere above the
dataset pointing towards the data set. We will
run the in situ image acquisition for every time
step. These parameters will produce: 2
variables x 1000 time steps x (10 isosurface
values + 3 x 10 cut planes) x 100 camera
positions x 3 images (depth, float, lighting)
= 2.4 x 107 images.”
J. Ahrens et al., SC’14
103 time steps x
1015 B state per
time step = 1018 B
2.4 x 107 images x
1MB/image
(megapixel, 4B) =
2.4 x 1012 B
49
Fusion whole device model
XGC GENEInterpolator
100+ PB
PB/day on
Titan today;
10+ PB/day
in the future
10 TB/day on
Titan today;
100+ TB/day
in the future
Analysis
Analysis
Analysis
Read 10-100 PB
per analysis
http://bit.ly/2fcyznK
50
XGC GENEInterpolator
Reduction Reduction
XGC
Viz.
XGC
output
GENE
Viz.
GENE
output
Comparative
Viz.
NVRAM
PFS
TAPE
Fusion whole device model
http://bit.ly/2fcyznK
51
Integrates multiple technologies:
•ADIOS staging (DataSpaces) for coupling
•Sirius (ADIOS + Ceph) for storage
•ZFP, SZ, Dogstar for reduction
•VTK-M services for visualization
•TAU for instrumenting the code
•Cheetah + Savanna to test the different
configurations (same node, different node,
hybrid-combination) to determine where to
place the different services
•Flexpath for staged-write from XGC to storage
•Ceph + ADIOS to manage storage hierarchy
•Swift for workflow automation
XGC GENEInterpolator
Reduction Reduction
XGC
Viz.
XGC
output
GENE
Viz.
GENE
output
TAU TAU
Comparative
Viz.
NVRAM
PFS
TAPE
Performance
Viz.
Cheetah +
Savanna drive
codesign experiments
Fusion whole device model
52
Codesign questions to be addressed
• How can we couple multiple codes? Files, staging on the same
node, different nodes, synchronous, asynchronous?
• How we can test different placement strategies for memory
optimization, performance optimizations?
• What are the best reduction technologies to allow us to capture
all relevant information during a simulation? E.g., Performance
vs. accuracy.
• How can we create visualization services that work on the
different architectures and use the data models in the codes?
• How do we manage data across storage hierarchies?
53
Savannah: Swift workflows coupled with ADIOS
Z-Check
dup
Multi-node workflow components communicate over ADIOS
Application data
Cheetah
Experiment
configuration
and dispatch
User monitoring and
control of multiple
pipeline instances
Co-design data
Store
experiment
metadata
Chimbuko
captures co-design
performance data
Other co-design
output
(e.g., Z-Checker)
CODAR
campaign
definition
Analysis
ADIOS output
Job launch
Science
App
Reduce
Co-design experiment architecture
54
Tasks demands new systems capabilities
Single
program
Multiple
program
Offline
analysis
Online
analysis
A few or many tasks:
• Loosely or tightly coupled
• Hierarchical or not
• Static or dynamic
• Fail-stop or recoverable
• Shared state
• Persistent and transient state
• Scheduled or data driven
Multiple
simulations
+ analyses
Simulation
+ analysis
Multiple
simulations
55
Challenge: Enable isolation, fault
tolerance, and composability for
ensembles of scientific
simulation/analysis pipelines
Defined MPIX_Comm_launch() call to enable
vendors to support dynamic workflow
pipelines, in which parallel applications of
various sizes are coupled in complex ways.
Key use case: ADIOS-based in situ analysis.
Integrated this feature with Swift/T, a scalable,
MPI-based workflow system. Allows ease of
development when coupling existing codes.
Working to have this mode of operation
supported in Cray OS.
Codesign of MPI interfaces in support of HPC workflows
Depiction of workflow of simulation analysis pipelines. Clusters of
boxes are MPI programs passing output data downstream. An
algorithm such as parameter optimization controls progress. Our
launch feature was scaled to 192 nodes with a challenging workload
for performance analysis, and the feature is in use by the CODES
network simulation team for its resilience capabilities.
Dorier, Wozniak, and Ross. Supporting Task-level Fault-tolerance in HPC Workflows by Launching MPI Jobs inside MPI Jobs. WORKS @ SC, 2017.
56Justin Wozniak and Jonathan Ozik
EMEWS: Extreme-scale
Model Exploration
With Swift
Many ways to extend:
- Hyperband
Li et al., arXiv:1603.0656
- Population-based training
Jagerberg, arXiv:1711.09846
EMEWS hyperparameter optimization
57
Co-evolution of HPC applications and systems …
… demand new application, software, and hardware …
… resulting in exciting new computer science challenges
foster@anl.gov
Thanks to US Department of Energy and CODAR team
58
Extra slides

More Related Content

What's hot

High Performance Data Analytics and a Java Grande Run Time
High Performance Data Analytics and a Java Grande Run TimeHigh Performance Data Analytics and a Java Grande Run Time
High Performance Data Analytics and a Java Grande Run TimeGeoffrey Fox
 
Accelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceAccelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceIan Foster
 
Taming Big Data!
Taming Big Data!Taming Big Data!
Taming Big Data!Ian Foster
 
Bionimbus Cambridge Workshop (3-28-11, v7)
Bionimbus Cambridge Workshop (3-28-11, v7)Bionimbus Cambridge Workshop (3-28-11, v7)
Bionimbus Cambridge Workshop (3-28-11, v7)Robert Grossman
 
AI at Scale for Materials and Chemistry
AI at Scale for Materials and ChemistryAI at Scale for Materials and Chemistry
AI at Scale for Materials and ChemistryIan Foster
 
Health & Status Monitoring (2010-v8)
Health & Status Monitoring (2010-v8)Health & Status Monitoring (2010-v8)
Health & Status Monitoring (2010-v8)Robert Grossman
 
Comparing Big Data and Simulation Applications and Implications for Software ...
Comparing Big Data and Simulation Applications and Implications for Software ...Comparing Big Data and Simulation Applications and Implications for Software ...
Comparing Big Data and Simulation Applications and Implications for Software ...Geoffrey Fox
 
An Overview of Bionimbus (March 2010)
An Overview of Bionimbus (March 2010)An Overview of Bionimbus (March 2010)
An Overview of Bionimbus (March 2010)Robert Grossman
 
Open Science Data Cloud - CCA 11
Open Science Data Cloud - CCA 11Open Science Data Cloud - CCA 11
Open Science Data Cloud - CCA 11Robert Grossman
 
Large Scale On-Demand Image Processing For Disaster Relief
Large Scale On-Demand Image Processing For Disaster ReliefLarge Scale On-Demand Image Processing For Disaster Relief
Large Scale On-Demand Image Processing For Disaster ReliefRobert Grossman
 
My Other Computer is a Data Center: The Sector Perspective on Big Data
My Other Computer is a Data Center: The Sector Perspective on Big DataMy Other Computer is a Data Center: The Sector Perspective on Big Data
My Other Computer is a Data Center: The Sector Perspective on Big DataRobert Grossman
 
Open Science Data Cloud (IEEE Cloud 2011)
Open Science Data Cloud (IEEE Cloud 2011)Open Science Data Cloud (IEEE Cloud 2011)
Open Science Data Cloud (IEEE Cloud 2011)Robert Grossman
 
Classification of Big Data Use Cases by different Facets
Classification of Big Data Use Cases by different FacetsClassification of Big Data Use Cases by different Facets
Classification of Big Data Use Cases by different FacetsGeoffrey Fox
 
Coding the Continuum
Coding the ContinuumCoding the Continuum
Coding the ContinuumIan Foster
 
The Discovery Cloud: Accelerating Science via Outsourcing and Automation
The Discovery Cloud: Accelerating Science via Outsourcing and AutomationThe Discovery Cloud: Accelerating Science via Outsourcing and Automation
The Discovery Cloud: Accelerating Science via Outsourcing and AutomationIan Foster
 
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...Robert Grossman
 
Computing Outside The Box June 2009
Computing Outside The Box June 2009Computing Outside The Box June 2009
Computing Outside The Box June 2009Ian Foster
 
What is the "Big Data" version of the Linpack Benchmark? ; What is “Big Data...
What is the "Big Data" version of the Linpack Benchmark?; What is “Big Data...What is the "Big Data" version of the Linpack Benchmark?; What is “Big Data...
What is the "Big Data" version of the Linpack Benchmark? ; What is “Big Data...Geoffrey Fox
 

What's hot (20)

High Performance Data Analytics and a Java Grande Run Time
High Performance Data Analytics and a Java Grande Run TimeHigh Performance Data Analytics and a Java Grande Run Time
High Performance Data Analytics and a Java Grande Run Time
 
Accelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceAccelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy Science
 
Taming Big Data!
Taming Big Data!Taming Big Data!
Taming Big Data!
 
Bionimbus Cambridge Workshop (3-28-11, v7)
Bionimbus Cambridge Workshop (3-28-11, v7)Bionimbus Cambridge Workshop (3-28-11, v7)
Bionimbus Cambridge Workshop (3-28-11, v7)
 
AI at Scale for Materials and Chemistry
AI at Scale for Materials and ChemistryAI at Scale for Materials and Chemistry
AI at Scale for Materials and Chemistry
 
Health & Status Monitoring (2010-v8)
Health & Status Monitoring (2010-v8)Health & Status Monitoring (2010-v8)
Health & Status Monitoring (2010-v8)
 
Comparing Big Data and Simulation Applications and Implications for Software ...
Comparing Big Data and Simulation Applications and Implications for Software ...Comparing Big Data and Simulation Applications and Implications for Software ...
Comparing Big Data and Simulation Applications and Implications for Software ...
 
An Overview of Bionimbus (March 2010)
An Overview of Bionimbus (March 2010)An Overview of Bionimbus (March 2010)
An Overview of Bionimbus (March 2010)
 
Open Science Data Cloud - CCA 11
Open Science Data Cloud - CCA 11Open Science Data Cloud - CCA 11
Open Science Data Cloud - CCA 11
 
Reproducible Research and the Cloud
Reproducible Research and the CloudReproducible Research and the Cloud
Reproducible Research and the Cloud
 
Large Scale On-Demand Image Processing For Disaster Relief
Large Scale On-Demand Image Processing For Disaster ReliefLarge Scale On-Demand Image Processing For Disaster Relief
Large Scale On-Demand Image Processing For Disaster Relief
 
My Other Computer is a Data Center: The Sector Perspective on Big Data
My Other Computer is a Data Center: The Sector Perspective on Big DataMy Other Computer is a Data Center: The Sector Perspective on Big Data
My Other Computer is a Data Center: The Sector Perspective on Big Data
 
Open Science Data Cloud (IEEE Cloud 2011)
Open Science Data Cloud (IEEE Cloud 2011)Open Science Data Cloud (IEEE Cloud 2011)
Open Science Data Cloud (IEEE Cloud 2011)
 
Classification of Big Data Use Cases by different Facets
Classification of Big Data Use Cases by different FacetsClassification of Big Data Use Cases by different Facets
Classification of Big Data Use Cases by different Facets
 
Coding the Continuum
Coding the ContinuumCoding the Continuum
Coding the Continuum
 
The Discovery Cloud: Accelerating Science via Outsourcing and Automation
The Discovery Cloud: Accelerating Science via Outsourcing and AutomationThe Discovery Cloud: Accelerating Science via Outsourcing and Automation
The Discovery Cloud: Accelerating Science via Outsourcing and Automation
 
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...
Lessons Learned from a Year's Worth of Benchmarking Large Data Clouds (Robert...
 
Computing Outside The Box June 2009
Computing Outside The Box June 2009Computing Outside The Box June 2009
Computing Outside The Box June 2009
 
What is the "Big Data" version of the Linpack Benchmark? ; What is “Big Data...
What is the "Big Data" version of the Linpack Benchmark?; What is “Big Data...What is the "Big Data" version of the Linpack Benchmark?; What is “Big Data...
What is the "Big Data" version of the Linpack Benchmark? ; What is “Big Data...
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
 

Similar to Computing Just What You Need: Online Data Analysis and Reduction at Extreme Scales

Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...Ian Foster
 
Big Data Processing Beyond MapReduce by Dr. Flavio Villanustre
Big Data Processing Beyond MapReduce by Dr. Flavio VillanustreBig Data Processing Beyond MapReduce by Dr. Flavio Villanustre
Big Data Processing Beyond MapReduce by Dr. Flavio VillanustreHPCC Systems
 
H2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupH2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupSri Ambati
 
Time to Science/Time to Results: Transforming Research in the Cloud
Time to Science/Time to Results: Transforming Research in the CloudTime to Science/Time to Results: Transforming Research in the Cloud
Time to Science/Time to Results: Transforming Research in the CloudAmazon Web Services
 
Azure Databricks for Data Scientists
Azure Databricks for Data ScientistsAzure Databricks for Data Scientists
Azure Databricks for Data ScientistsRichard Garris
 
Best pratices at BGI for the Challenges in the Era of Big Genomics Data
Best pratices at BGI for the Challenges in the Era of Big Genomics DataBest pratices at BGI for the Challenges in the Era of Big Genomics Data
Best pratices at BGI for the Challenges in the Era of Big Genomics DataXing Xu
 
Rpi talk foster september 2011
Rpi talk foster september 2011Rpi talk foster september 2011
Rpi talk foster september 2011Ian Foster
 
Introduction to Machine Learning with H2O and Python
Introduction to Machine Learning with H2O and PythonIntroduction to Machine Learning with H2O and Python
Introduction to Machine Learning with H2O and PythonJo-fai Chow
 
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...inside-BigData.com
 
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...QuantUniversity
 
WWV2015: Jibes Paul van der Hulst big data
WWV2015: Jibes Paul van der Hulst big dataWWV2015: Jibes Paul van der Hulst big data
WWV2015: Jibes Paul van der Hulst big datawebwinkelvakdag
 
GlobusWorld 2019 Opening Keynote
GlobusWorld 2019 Opening KeynoteGlobusWorld 2019 Opening Keynote
GlobusWorld 2019 Opening KeynoteGlobus
 
Linking Scientific Instruments and Computation
Linking Scientific Instruments and ComputationLinking Scientific Instruments and Computation
Linking Scientific Instruments and ComputationIan Foster
 
Data council sf amundsen presentation
Data council sf    amundsen presentationData council sf    amundsen presentation
Data council sf amundsen presentationTao Feng
 
So Long Computer Overlords
So Long Computer OverlordsSo Long Computer Overlords
So Long Computer OverlordsIan Foster
 
Get Started with Data Science by Analyzing Traffic Data from California Highways
Get Started with Data Science by Analyzing Traffic Data from California HighwaysGet Started with Data Science by Analyzing Traffic Data from California Highways
Get Started with Data Science by Analyzing Traffic Data from California HighwaysAerospike, Inc.
 
eScience: A Transformed Scientific Method
eScience: A Transformed Scientific MethodeScience: A Transformed Scientific Method
eScience: A Transformed Scientific MethodDuncan Hull
 
Big Data Lessons from the Cloud
Big Data Lessons from the CloudBig Data Lessons from the Cloud
Big Data Lessons from the CloudMapR Technologies
 

Similar to Computing Just What You Need: Online Data Analysis and Reduction at Extreme Scales (20)

Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...Computing Just What You Need: Online Data Analysis and Reduction  at Extreme ...
Computing Just What You Need: Online Data Analysis and Reduction at Extreme ...
 
Stream Processing
Stream Processing Stream Processing
Stream Processing
 
Big Data Processing Beyond MapReduce by Dr. Flavio Villanustre
Big Data Processing Beyond MapReduce by Dr. Flavio VillanustreBig Data Processing Beyond MapReduce by Dr. Flavio Villanustre
Big Data Processing Beyond MapReduce by Dr. Flavio Villanustre
 
H2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupH2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User Group
 
Shikha fdp 62_14july2017
Shikha fdp 62_14july2017Shikha fdp 62_14july2017
Shikha fdp 62_14july2017
 
Time to Science/Time to Results: Transforming Research in the Cloud
Time to Science/Time to Results: Transforming Research in the CloudTime to Science/Time to Results: Transforming Research in the Cloud
Time to Science/Time to Results: Transforming Research in the Cloud
 
Azure Databricks for Data Scientists
Azure Databricks for Data ScientistsAzure Databricks for Data Scientists
Azure Databricks for Data Scientists
 
Best pratices at BGI for the Challenges in the Era of Big Genomics Data
Best pratices at BGI for the Challenges in the Era of Big Genomics DataBest pratices at BGI for the Challenges in the Era of Big Genomics Data
Best pratices at BGI for the Challenges in the Era of Big Genomics Data
 
Rpi talk foster september 2011
Rpi talk foster september 2011Rpi talk foster september 2011
Rpi talk foster september 2011
 
Introduction to Machine Learning with H2O and Python
Introduction to Machine Learning with H2O and PythonIntroduction to Machine Learning with H2O and Python
Introduction to Machine Learning with H2O and Python
 
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...
The Transformation of HPC: Simulation and Cognitive Methods in the Era of Big...
 
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...
QuTrack: Model Life Cycle Management for AI and ML models using a Blockchain ...
 
WWV2015: Jibes Paul van der Hulst big data
WWV2015: Jibes Paul van der Hulst big dataWWV2015: Jibes Paul van der Hulst big data
WWV2015: Jibes Paul van der Hulst big data
 
GlobusWorld 2019 Opening Keynote
GlobusWorld 2019 Opening KeynoteGlobusWorld 2019 Opening Keynote
GlobusWorld 2019 Opening Keynote
 
Linking Scientific Instruments and Computation
Linking Scientific Instruments and ComputationLinking Scientific Instruments and Computation
Linking Scientific Instruments and Computation
 
Data council sf amundsen presentation
Data council sf    amundsen presentationData council sf    amundsen presentation
Data council sf amundsen presentation
 
So Long Computer Overlords
So Long Computer OverlordsSo Long Computer Overlords
So Long Computer Overlords
 
Get Started with Data Science by Analyzing Traffic Data from California Highways
Get Started with Data Science by Analyzing Traffic Data from California HighwaysGet Started with Data Science by Analyzing Traffic Data from California Highways
Get Started with Data Science by Analyzing Traffic Data from California Highways
 
eScience: A Transformed Scientific Method
eScience: A Transformed Scientific MethodeScience: A Transformed Scientific Method
eScience: A Transformed Scientific Method
 
Big Data Lessons from the Cloud
Big Data Lessons from the CloudBig Data Lessons from the Cloud
Big Data Lessons from the Cloud
 

More from Ian Foster

Global Services for Global Science March 2023.pptx
Global Services for Global Science March 2023.pptxGlobal Services for Global Science March 2023.pptx
Global Services for Global Science March 2023.pptxIan Foster
 
The Earth System Grid Federation: Origins, Current State, Evolution
The Earth System Grid Federation: Origins, Current State, EvolutionThe Earth System Grid Federation: Origins, Current State, Evolution
The Earth System Grid Federation: Origins, Current State, EvolutionIan Foster
 
Better Information Faster: Programming the Continuum
Better Information Faster: Programming the ContinuumBetter Information Faster: Programming the Continuum
Better Information Faster: Programming the ContinuumIan Foster
 
ESnet6 and Smart Instruments
ESnet6 and Smart InstrumentsESnet6 and Smart Instruments
ESnet6 and Smart InstrumentsIan Foster
 
A Global Research Data Platform: How Globus Services Enable Scientific Discovery
A Global Research Data Platform: How Globus Services Enable Scientific DiscoveryA Global Research Data Platform: How Globus Services Enable Scientific Discovery
A Global Research Data Platform: How Globus Services Enable Scientific DiscoveryIan Foster
 
Foster CRA March 2022.pptx
Foster CRA March 2022.pptxFoster CRA March 2022.pptx
Foster CRA March 2022.pptxIan Foster
 
Big Data, Big Computing, AI, and Environmental Science
Big Data, Big Computing, AI, and Environmental ScienceBig Data, Big Computing, AI, and Environmental Science
Big Data, Big Computing, AI, and Environmental ScienceIan Foster
 
Data Tribology: Overcoming Data Friction with Cloud Automation
Data Tribology: Overcoming Data Friction with Cloud AutomationData Tribology: Overcoming Data Friction with Cloud Automation
Data Tribology: Overcoming Data Friction with Cloud AutomationIan Foster
 
Research Automation for Data-Driven Discovery
Research Automation for Data-Driven DiscoveryResearch Automation for Data-Driven Discovery
Research Automation for Data-Driven DiscoveryIan Foster
 
Scaling collaborative data science with Globus and Jupyter
Scaling collaborative data science with Globus and JupyterScaling collaborative data science with Globus and Jupyter
Scaling collaborative data science with Globus and JupyterIan Foster
 
Team Argon Summary
Team Argon SummaryTeam Argon Summary
Team Argon SummaryIan Foster
 
Thoughts on interoperability
Thoughts on interoperabilityThoughts on interoperability
Thoughts on interoperabilityIan Foster
 
Going Smart and Deep on Materials at ALCF
Going Smart and Deep on Materials at ALCFGoing Smart and Deep on Materials at ALCF
Going Smart and Deep on Materials at ALCFIan Foster
 
Software Infrastructure for a National Research Platform
Software Infrastructure for a National Research PlatformSoftware Infrastructure for a National Research Platform
Software Infrastructure for a National Research PlatformIan Foster
 
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...Ian Foster
 
Globus Auth: A Research Identity and Access Management Platform
Globus Auth: A Research Identity and Access Management PlatformGlobus Auth: A Research Identity and Access Management Platform
Globus Auth: A Research Identity and Access Management PlatformIan Foster
 
Streamlined data sharing and analysis to accelerate cancer research
Streamlined data sharing and analysis to accelerate cancer researchStreamlined data sharing and analysis to accelerate cancer research
Streamlined data sharing and analysis to accelerate cancer researchIan Foster
 
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...Ian Foster
 
building global software/earthcube->sciencecloud
building global software/earthcube->sciencecloudbuilding global software/earthcube->sciencecloud
building global software/earthcube->sciencecloudIan Foster
 

More from Ian Foster (19)

Global Services for Global Science March 2023.pptx
Global Services for Global Science March 2023.pptxGlobal Services for Global Science March 2023.pptx
Global Services for Global Science March 2023.pptx
 
The Earth System Grid Federation: Origins, Current State, Evolution
The Earth System Grid Federation: Origins, Current State, EvolutionThe Earth System Grid Federation: Origins, Current State, Evolution
The Earth System Grid Federation: Origins, Current State, Evolution
 
Better Information Faster: Programming the Continuum
Better Information Faster: Programming the ContinuumBetter Information Faster: Programming the Continuum
Better Information Faster: Programming the Continuum
 
ESnet6 and Smart Instruments
ESnet6 and Smart InstrumentsESnet6 and Smart Instruments
ESnet6 and Smart Instruments
 
A Global Research Data Platform: How Globus Services Enable Scientific Discovery
A Global Research Data Platform: How Globus Services Enable Scientific DiscoveryA Global Research Data Platform: How Globus Services Enable Scientific Discovery
A Global Research Data Platform: How Globus Services Enable Scientific Discovery
 
Foster CRA March 2022.pptx
Foster CRA March 2022.pptxFoster CRA March 2022.pptx
Foster CRA March 2022.pptx
 
Big Data, Big Computing, AI, and Environmental Science
Big Data, Big Computing, AI, and Environmental ScienceBig Data, Big Computing, AI, and Environmental Science
Big Data, Big Computing, AI, and Environmental Science
 
Data Tribology: Overcoming Data Friction with Cloud Automation
Data Tribology: Overcoming Data Friction with Cloud AutomationData Tribology: Overcoming Data Friction with Cloud Automation
Data Tribology: Overcoming Data Friction with Cloud Automation
 
Research Automation for Data-Driven Discovery
Research Automation for Data-Driven DiscoveryResearch Automation for Data-Driven Discovery
Research Automation for Data-Driven Discovery
 
Scaling collaborative data science with Globus and Jupyter
Scaling collaborative data science with Globus and JupyterScaling collaborative data science with Globus and Jupyter
Scaling collaborative data science with Globus and Jupyter
 
Team Argon Summary
Team Argon SummaryTeam Argon Summary
Team Argon Summary
 
Thoughts on interoperability
Thoughts on interoperabilityThoughts on interoperability
Thoughts on interoperability
 
Going Smart and Deep on Materials at ALCF
Going Smart and Deep on Materials at ALCFGoing Smart and Deep on Materials at ALCF
Going Smart and Deep on Materials at ALCF
 
Software Infrastructure for a National Research Platform
Software Infrastructure for a National Research PlatformSoftware Infrastructure for a National Research Platform
Software Infrastructure for a National Research Platform
 
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...
Accelerating the Experimental Feedback Loop: Data Streams and the Advanced Ph...
 
Globus Auth: A Research Identity and Access Management Platform
Globus Auth: A Research Identity and Access Management PlatformGlobus Auth: A Research Identity and Access Management Platform
Globus Auth: A Research Identity and Access Management Platform
 
Streamlined data sharing and analysis to accelerate cancer research
Streamlined data sharing and analysis to accelerate cancer researchStreamlined data sharing and analysis to accelerate cancer research
Streamlined data sharing and analysis to accelerate cancer research
 
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...
Science Services and Science Platforms: Using the Cloud to Accelerate and Dem...
 
building global software/earthcube->sciencecloud
building global software/earthcube->sciencecloudbuilding global software/earthcube->sciencecloud
building global software/earthcube->sciencecloud
 

Recently uploaded

Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodJuan lago vázquez
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...DianaGray10
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...Martijn de Jong
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingEdi Saputra
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobeapidays
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MIND CTI
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 

Recently uploaded (20)

Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 

Computing Just What You Need: Online Data Analysis and Reduction at Extreme Scales

  • 1. 1 Computing Just What You Need: Online Data Analysis and Reduction at Extreme Scales Ian Foster Argonne National Lab & University of Chicago December 21, 2017 HiPC, Jaipur, India https://www.researchgate.net/publication/317703782 foster@anl.gov
  • 2. 2 Earth to be paradise; distance to lose enchantment “If, as it is said to be not unlikely in the near future, the principle of sight is applied to the telephone as well as that of sound, earth will be in truth a paradise, and distance will lose its enchantment by being abolished altogether.” — Arthur Mee, 1898
  • 3. 3 Cooperating at a distance 1994 1999, 2003 2017
  • 4. 4 Automating the research data lifecycle. Reducing barriers to cooperation at a distance globus.org
  • 5. 5 Automating research data lifecycle 5 major services 13 national labs use Globus 340PB transferred 10,000 active endpoints 50 Bn files processed 75,000 registered users 99.5% uptime 65+ institutional subscribers 1 PB largest single transfer to date 3 months longest continuously managed transfer 300+ federated campus identities 12,000 active users/year
  • 6. 6 Transferring 1PB in a day Argonne → NCSA • Cosmology simulation on Mira @ Argonne produces 1 PB in 24 hours • Data streamed to Blue Waters for analytics • Application reveals feasibility of real-time streaming at scale Without checksums With checksums
  • 7. 7 Time to discovery Simula on me Ultra scale Where does HPC fit in the research lifecycle?
  • 8. 8 The challenges of managing data and computation at the 1018 scale US Department of Energy
  • 9. 9 US Exascale Computing Program: Using codesign and integration to achieve capable exascale US Department of Energy
  • 10. 10 A. C. Bauer et al., EuroVis 2016 Computation 125 PB/s Node memory 4.5 PB/s Interconnect (largest cross-sectional b/w) 24 TB/s Storage 1.4 TB/s Interconnect 24 TB/s Node memory 4.5 PB/s Changing storage geography a major challenge for exascale
  • 11. 11 Disk didn’t use to be the problem ~1980-2000 Patterson, CACM, 2004
  • 12. 12 Disks are getting larger, but not faster ~1980-2000 Patterson, CACM, 2004 https://www.backblaze.com/blog/hard-drive-cost-per-gigabyte/
  • 14. 14
  • 15. 15 Exascale climate goal: Ensembles of 1km models at 15 simulated years/24 hours Full state once per model day  260 TB every 16 seconds  1.4 EB/day
  • 16. 16 Time to discovery Simula on me Ultra scale
  • 17. 17 Time to discovery Simula on me Ultra scale Data Space tools: Popula on, naviga on, manipula on, dissemina on Leadership class facility Smaller systems Leadership class facility Smaller systems Ultra scale
  • 18. 18 Time to discovery Simula on me Ultra scale Data Space tools: Popula on, naviga on, manipula on, dissemina on Leadership class facility Smaller systems Leadership class facility Smaller systems Ultra scale
  • 19. 19 Time to discovery Simula on me Ultra scale Data Space tools: Popula on, naviga on, manipula on, dissemina on Leadership class facility Smaller systems Leadership class facility Smaller systems Ultra scale
  • 20. 20 Time to discovery Simula on me Ultra scale Data Space tools: Popula on, naviga on, manipula on, dissemina on Leadership class facility Smaller systems Leadership class facility Smaller systems Ultra scale
  • 21. 21 The need for online data analysis and reduction Traditional approach: Simulate, output, analyze Write simulation output to secondary storage; read back for analysis Decimate in time when simulation output rate exceeds output rate of computer Online: y = F(x) Offline: a = A(y), b= B(y), …
  • 22. 22 The need for online data analysis and reduction Traditional approach: Simulate, output, analyze Write simulation output to secondary storage; read back for analysis Decimate in time when simulation output rate exceeds output rate of computer Online: y = F(x) Offline: a = A(y), b= B(y), … New approach: Online data analysis & reduction Co-optimize simulation, analysis, reduction for performance and information output Substitute CPU cycles for I/O, via online data (de)compression and/or analysis a) Online: a = A(F(x)), b = B(F(x)), … b) Online: r = R(F(x)) Offline: a = A’(r), b = B’(r), or a = A(U(r)), b = B(U(r)) [R = reduce, U = un-reduce]
  • 23. 23 Exascale computing at Argonne by 2021 Precision medicine Data from sensors and scientific instruments Simulation and modeling of materials and physical systems Support for three types of computing: Traditional: HPC simulation and modeling Learning: Machine learning, deep learning, AI Data: Data analytics, data science, big data [Artists impression]
  • 24. 25 Real-time analysis and experimental steering • Current protocols process and validate data only after experiment, which can lead to undetected errors and prevents online steering • Process data streamed from beamline to supercomputer; control feedback loop makes decisions during experiment • Tests in TXM beamline (32- ID@APS) in cement wetting experiment (2 experiments, each with 8 hours of data acquisition time) Sustained # Projections/seconds CircularBufferSize Reconstruction Frequency Image Quality w.r.t. Streamed Projections SimilarityScore # Streamed Projections Reconstructed Image Sequence Tekin Bicer et al., eScience 2017
  • 25. 26 Deep learning for precision medicine https://de.mathworks.com/company/newsletters/articles/cancer-diagnostics-with-deep-learning-and-photonic-time-stretch.html
  • 27. Using learning to optimize simulation studies
  • 28. Simulation data Learning methods New capabilities New simulations Using learning to optimize simulation studies Logan Ward and Ben Blaiszik
  • 29. 30 Synopsis: Applications are changing Single program Multiple program Offline analysis Online analysis A few or many tasks: • Loosely or tightly coupled • Hierarchical or not • Static or dynamic • Fail-stop or recoverable • Shared state • Persistent and transient state • Scheduled or data driven Multiple simulations + analyses Simulation + analysis Multiple simulations
  • 30. 31 Many interesting codesign problems Big simulation Machine learning Deep learning Streaming Online analysis Online reduction Heterogeneity Prog. models - Many task - Streaming Libraries - Analysis, reduction - Communications System software - Fault tolerance - Resource mgmt Complex nodes - Many core - Accelerators - Heterogeneous NVRAM Networks - Internal - External Node configuration Internal networks External networks Memory hierarchy Storage systems Heterogeneity Operating policies
  • 31. 32 Reduction comes with challenges • Handling high entropy • Performance – no benefit otherwise • Not only error in variable: Ε ≡ 𝑓 − 𝑓 • Must also consider impact on derived quantities: Ε ≡ (𝑔𝑙 𝑡 (𝑓 𝑥, 𝑡 ) − 𝑔𝑙 𝑡 ( 𝑓𝑙 𝑡 ( 𝑥, 𝑡 ) S. Klasky
  • 32. 33 Key research challenge: How to manage the impact of errors on derived quantities? Where did it go??? S. Klasky Reduction comes with challenges
  • 33. 34 CODAR: Codesign center for Online Data Analysis and Reduction • Infrastructure development and deployment • Enable rapid composition of application and “data services” (data reduction methods, data analysis methods, etc.) • Support CODAR-developed and other data services • Method development: new reduction & analysis routines • Motif-specific: e.g., finite difference mesh vs. particles vs. finite elements • Application-specific: e.g., reduced physics to understand deltas • Application engagement • Understand data analysis and reduction requirements • Integrate, deploy, evaluate impact https://codarcode.github.io codar-info@cels.anl.gov
  • 34. 35 Cross-cutting research questions What are the best data analysis and reduction algorithms for different application classes, in terms of speed, accuracy, and resource needs? How can we implement those algorithms to achieve scalability and performance portability? What are tradeoffs in analysis accuracy, resource needs, and overall application performance between using various data reduction methods online prior to offline data reconstruction and analysis vs. performing more online data analysis? How do tradeoffs vary with hardware & software choices? How do we effectively orchestrate online data analysis and reduction to reduce associated overheads? How can hardware and software help with orchestration?
  • 35. 36 Prototypical data analysis and reduction pipeline CODAR runtime Reduced output and reconstruction info I/O system CODAR data API Running simulation Multivariate statistics Feature analysis Outlier detection Application-aware Transforms Encodings Error calculation Refinement hints CODARdataAPI Offlinedataanalysis Simulation knowledge: application, models, numerics, performance optimization, … CODAR data analysis CODAR data reduction CODAR data monitoring
  • 36. 37 Overarching data reduction challenges • Understanding the science requires massive data reduction • How do we reduce • The time spent in reducing the data to knowledge? • The amount of data moved on the HPC platform? • The amount of data read from the storage system? • The amount of data stored in memory, on storage system, moved over WAN? • Without removing the knowledge. • Requires deep dives into application post processing routines and simulations • Goal is to create both (a) co-design infrastructure and (b) reduction and analysis routines • General: e.g., reduce Nbytes to Mbytes, N<<M • Motif-specific: e.g., finite difference mesh vs. particles vs. finite elements • Application-specific: e.g. reduced physics allows us to understand deltas
  • 37. 38 HPC floating point compression • Current interest is with lossy algorithms, some use preprocessing • Lossless may achieve up to ~3x reduction • ISABELA • SZ • ZFP • Linear auditing • SVD • Adaptive gradient methods Compress each variable separately: Several variables simultaneously: • PCA • Tensor decomposition • …
  • 38. 39 Lossy compression with SZ No existing compressor can reduce hard to compress datasets by more than a factor of 2. Objective 1: Reduce hard to compress datasets by one order of magnitude Objective 2: Add user-required error controls (error bound, shape of error distribution, spectral behavior of error function, etc. etc.) NCAR atmosphere simulation output (1.5 TB) WRF hurricane simulation output Advanced Photon Source mouse brain data What we need to compress (bit map of 128 floating point numbers): Random noise Franck Cappello
  • 39. 40 Lossy compression: Atmospheric simulation Franck Cappello Latest SZ
  • 40. 41 Characterizing compression error 0.0001 0.001 0.01 0.1 1 1/N 6/N 11/N 16/N 21/N 26/N 31/N 36/N 41/N 46/N Amplitude Frequency 0 2e-07 4e-07 6e-07 8e-07 1e-06 1.2e-06 1.4e-06 1.6e-06 1.8e-06 2e-06 0 20 40 60 80 100 120 140 160 180 200MaximumCompressionError Variables SZ(max error) SZ(avg error) ZFP(max error) ZFP(avg error) Error distribution Spectral behavior Laplacian (derivatives) Autocorrelation of errors Respect of error bounds Error propagation Franck Cappello
  • 41. 42 Z-checker: Analysis of data reduction error • Community tool to enable comprehensive assessment of lossy data reduction error: • Collection of data quality criteria from applications • Community repository for datasets, reduction quality requirements, compression performance • Modular design enables contributed analysis modules (C and R) and format readers (ADIOS, HDF5, etc.) • Off-line/on-line parallel statistical, spectral, point-wise distortion analysis with static & dynamic visualization Franck Cappello, Julie Bessac, Sheng Di
  • 42. 43 Z-Checker computations • Normalized root mean squared error • Peak signal to noise ratio • Distribution of error • Pearson correlation between raw and reduced datasets • Power spectrum distortion • Auto-correlation of compression error • Maximum error • Point-wise error bound (relative or absolute) • Preservation of derivatives • Structural similarity (SSIM) index
  • 43. 44 Science-driven optimizations • Information-theoretically derived methods like SZ, Isabella, ZFP make for good generic capabilities • If scientists can provide additional details on how to determine features of interest, we can use those to drive further optimizations. E.g., if they can select: • Regions of high gradient • Regions near turbulent flow • Particles with velocities > two standard deviations • How can scientists help define features?
  • 44. 45 Multilevel compression techniques A hierarchical reduction scheme produces multiple levels of partial decompression of the data so that users can work with reduced representations that require minimal storage whilst achieving user-specified tolerance Compression vs. user-specified toleranceResults for turbulence dataset: extremely large, inherently non-smooth, resistant to compression Mark Ainsworth
  • 45. 46 Manifold learning for change detection and adaptive sampling Low dimensional manifold projection of different state of MD trajectories • A single molecular dynamics trajectory can generate 32 PB • Use online data analysis to detect relevant or significant events • Project MD trajectories to manifold space (dimensionality reduction) across time into two dimensional space • Change detection on manifold space is more robust than original full coordinate space as it removes local vibrational noise • Apply adaptive sampling strategy based on accumulated changes of trajectories Shinjae Yoo
  • 46. 47 Critical points extracted with topology analysis Tracking blobs in XGC fusion simulations Blobs, regions of high turbulence that can damage the Tokamak, can run along the edge wall down toward the diverter and damage it. Blob extraction and tracking enables the exploration and analysis of high-energy blobs across timesteps. • Access data with ADIOS I/O in high performance • Precondition the input data with robust PCA • Detect blobs as local extrema with topology analysis • Track blobs over time with combinatorial feature flow field method Extracting, tracking, and visualizing blobs in large 5D gyrokinetic Tokamak simulations Hanqi Guo, Tom Peterka Tracking graph that visualizes the dynamics of blobs (birth, merge, split, and death) over time Data preconditioning with robust PCA
  • 47. 48 Reduction for visualization “an extreme scale simulation … calculates temperature and density over 1000 of time steps. For both variables, a scientist would like to visualize 10 isosurface values and X, Y, and Z cut planes for 10 locations in each dimension. One hundred different camera positions are also selected, in a hemisphere above the dataset pointing towards the data set. We will run the in situ image acquisition for every time step. These parameters will produce: 2 variables x 1000 time steps x (10 isosurface values + 3 x 10 cut planes) x 100 camera positions x 3 images (depth, float, lighting) = 2.4 x 107 images.” J. Ahrens et al., SC’14 103 time steps x 1015 B state per time step = 1018 B 2.4 x 107 images x 1MB/image (megapixel, 4B) = 2.4 x 1012 B
  • 48. 49 Fusion whole device model XGC GENEInterpolator 100+ PB PB/day on Titan today; 10+ PB/day in the future 10 TB/day on Titan today; 100+ TB/day in the future Analysis Analysis Analysis Read 10-100 PB per analysis http://bit.ly/2fcyznK
  • 50. 51 Integrates multiple technologies: •ADIOS staging (DataSpaces) for coupling •Sirius (ADIOS + Ceph) for storage •ZFP, SZ, Dogstar for reduction •VTK-M services for visualization •TAU for instrumenting the code •Cheetah + Savanna to test the different configurations (same node, different node, hybrid-combination) to determine where to place the different services •Flexpath for staged-write from XGC to storage •Ceph + ADIOS to manage storage hierarchy •Swift for workflow automation XGC GENEInterpolator Reduction Reduction XGC Viz. XGC output GENE Viz. GENE output TAU TAU Comparative Viz. NVRAM PFS TAPE Performance Viz. Cheetah + Savanna drive codesign experiments Fusion whole device model
  • 51. 52 Codesign questions to be addressed • How can we couple multiple codes? Files, staging on the same node, different nodes, synchronous, asynchronous? • How we can test different placement strategies for memory optimization, performance optimizations? • What are the best reduction technologies to allow us to capture all relevant information during a simulation? E.g., Performance vs. accuracy. • How can we create visualization services that work on the different architectures and use the data models in the codes? • How do we manage data across storage hierarchies?
  • 52. 53 Savannah: Swift workflows coupled with ADIOS Z-Check dup Multi-node workflow components communicate over ADIOS Application data Cheetah Experiment configuration and dispatch User monitoring and control of multiple pipeline instances Co-design data Store experiment metadata Chimbuko captures co-design performance data Other co-design output (e.g., Z-Checker) CODAR campaign definition Analysis ADIOS output Job launch Science App Reduce Co-design experiment architecture
  • 53. 54 Tasks demands new systems capabilities Single program Multiple program Offline analysis Online analysis A few or many tasks: • Loosely or tightly coupled • Hierarchical or not • Static or dynamic • Fail-stop or recoverable • Shared state • Persistent and transient state • Scheduled or data driven Multiple simulations + analyses Simulation + analysis Multiple simulations
  • 54. 55 Challenge: Enable isolation, fault tolerance, and composability for ensembles of scientific simulation/analysis pipelines Defined MPIX_Comm_launch() call to enable vendors to support dynamic workflow pipelines, in which parallel applications of various sizes are coupled in complex ways. Key use case: ADIOS-based in situ analysis. Integrated this feature with Swift/T, a scalable, MPI-based workflow system. Allows ease of development when coupling existing codes. Working to have this mode of operation supported in Cray OS. Codesign of MPI interfaces in support of HPC workflows Depiction of workflow of simulation analysis pipelines. Clusters of boxes are MPI programs passing output data downstream. An algorithm such as parameter optimization controls progress. Our launch feature was scaled to 192 nodes with a challenging workload for performance analysis, and the feature is in use by the CODES network simulation team for its resilience capabilities. Dorier, Wozniak, and Ross. Supporting Task-level Fault-tolerance in HPC Workflows by Launching MPI Jobs inside MPI Jobs. WORKS @ SC, 2017.
  • 55. 56Justin Wozniak and Jonathan Ozik EMEWS: Extreme-scale Model Exploration With Swift Many ways to extend: - Hyperband Li et al., arXiv:1603.0656 - Population-based training Jagerberg, arXiv:1711.09846 EMEWS hyperparameter optimization
  • 56. 57 Co-evolution of HPC applications and systems … … demand new application, software, and hardware … … resulting in exciting new computer science challenges foster@anl.gov Thanks to US Department of Energy and CODAR team