SlideShare a Scribd company logo
1 of 32
Download to read offline
This	
  work	
  was	
  performed	
  under	
  the	
  auspices	
  of	
  the	
  U.S.	
  
Department	
  of	
  Energy	
  by	
  Lawrence	
  Livermore	
  National	
  
Laboratory	
  under	
  Contract	
  DE-­‐AC52-­‐07NA27344	
  and	
  
supported	
  by	
  Office	
  of	
  Science	
  Advanced	
  Scientific	
  
Computing	
  Research	
  Program.	
  
ICME	
  2015	
  
Cheyenne	
  Mountain	
  Resort	
  
Colorado	
  Springs,	
  Colorado	
  
3	
  June,	
  	
  02015	
  
James	
  Belak	
  (belak@llnl.gov)	
  
www.exmatex.org	
  
	
  
*	
  co-­‐PI:	
  Tim	
  Germann	
  (tcg@lanl.gov)	
  
LLNL-­‐PRES-­‐673227	
  
'There	
  is	
  not	
  pure	
  science	
  and	
  
applied	
  science	
  but	
  only	
  
science	
  and	
  the	
  applications	
  of	
  
science’	
  
	
  	
  	
  	
  	
  	
  	
  	
  	
  	
  	
  	
  	
  	
  -­‐	
  Louis	
  Pasteur	
  	
  
‘People	
  who	
  are	
  serious	
  about	
  
software	
  should	
  make	
  their	
  
own	
  hardware.’	
  
	
  -­‐	
  Alan	
  Kay	
  (Xerox	
  PARC)	
  
Initiate	
  
Use-­‐inspired	
  research	
  and	
  development	
  
involves	
  a	
  dynamic	
  interaction	
  between	
  both	
  
considerations	
  of	
  use	
  and	
  fundamental	
  
understanding.	
  	
  	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  It	
  begs	
  the	
  question:	
  what	
  are	
  the	
  computations	
  that	
  enable	
  ICME:	
  
§  Databases	
  (integrated,	
  quality,	
  provenance,…)	
  
§  Knowledge-­‐driven	
  organization	
  of	
  data	
  and	
  databases,	
  e.g.	
  enable	
  
communication	
  between	
  materials	
  engineer	
  and	
  process	
  engineer,	
  do	
  
what	
  google	
  does	
  to	
  organize	
  the	
  web	
  
§  Expert	
  Systems:	
  Automated	
  Materials	
  Selection	
  from	
  Database	
  
§  Inverse	
  problem	
  analysis	
  (e.g.	
  Bayesian,	
  material	
  from	
  property),	
  plus	
  
other	
  techniques	
  from	
  Data	
  Analytics	
  
§  Direct	
  Numerical	
  Simulation:	
  identify	
  the	
  controlling	
  phenomena	
  and	
  do	
  
the	
  relevant	
  simulation,	
  meso-­‐scale,	
  molecular	
  dynamics,	
  DFT,	
  …	
  	
  
§  Better	
  physics-­‐based	
  continuum	
  models	
  (these	
  use	
  to	
  be	
  called	
  
phenomenological,	
  but	
  nobody	
  understands	
  that	
  any	
  more)	
  
2	
  
Integrated	
  Engineering:	
  Integrating	
  “Computational	
  Materials”	
  
into	
  Engineering	
  or	
  Integrating	
  “Computations”	
  into	
  Materials	
  
Engineering.	
  Integrated	
  Engineering	
  is	
  more	
  than	
  ICME!	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   3	
  
Component	
  
Design	
  
(Integrated	
  Code)	
  
Component	
  
Test	
  
Material	
  
Design	
  
(Models)	
  
Characterization,	
  	
  
Material	
  Test,	
  
Virtual	
  Test	
  
Process	
  
Design	
  
(Models)	
  
Material	
  
Specification	
  
Fabrication	
  
Process	
  
Property	
  
Specification	
  
Process	
  
Specification	
  
Process	
  
Qualification	
  
Property	
  
Qualification	
  
Material	
  
Qualification	
  
Design	
  Specification	
  
Component	
  
Qualification	
  
Component	
  
Fabrication	
  
Objective:	
  Agile	
  Design	
  and	
  Qualification	
  with	
  Minimal	
  Component-­‐scale	
  Test	
  
S&E	
  Use-­‐Case:	
  Workflow	
  for	
  Materials	
  Development	
  
http://extremescaleresearch.labworks.org/events/workshop-­‐future-­‐scientific-­‐
workflows#block-­‐system-­‐main	
  
Advanced	
  
Manufacturing	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  Our	
  goal	
  is	
  to	
  establish	
  the	
  interrelationship	
  
between	
  hardware,	
  middleware	
  (software	
  
stack),	
  programming	
  models,	
  and	
  algorithms	
  
required	
  to	
  enable	
  a	
  productive	
  exascale	
  
environment	
  for	
  multiphysics	
  simulations	
  of	
  
materials	
  in	
  extreme	
  mechanical	
  and	
  
radiation	
  environments.	
  
§  We	
  will	
  exploit,	
  rather	
  than	
  avoid,	
  the	
  greatly	
  
increased	
  levels	
  of	
  concurrency,	
  
heterogeneity,	
  and	
  flop/byte	
  ratios	
  on	
  the	
  
upcoming	
  exascale	
  platforms.	
  	
  
•  Our	
  vision	
  is	
  an	
  uncertainty	
  quantification	
  (UQ)-­‐driven	
  adaptive	
  physics	
  
refinement	
  in	
  which	
  meso-­‐	
  and	
  macro-­‐scale	
  materials	
  simulations	
  spawn	
  micro-­‐
scale	
  simulations	
  as	
  needed.	
  
-  This	
  task-­‐based	
  approach	
  leverages	
  the	
  extensive	
  concurrency	
  and	
  
heterogeneity	
  expected	
  at	
  exascale	
  while	
  enabling	
  fault	
  tolerance	
  within	
  
applications.	
  	
  
-  The	
  programming	
  models	
  and	
  approaches	
  developed	
  to	
  achieve	
  this	
  will	
  
be	
  broadly	
  applicable	
  to	
  a	
  variety	
  of	
  multiscale,	
  multiphysics	
  applications,	
  
including	
  astrophysics,	
  climate	
  and	
  weather	
  prediction,	
  structural	
  
engineering,	
  plasma	
  physics,	
  and	
  radiation	
  hydrodynamics.	
  
APP	
  
SW	
  HW	
  
Func+onal	
  
Exascale	
  
Simula+on	
  
Environment	
  
Domain	
  Workload	
  
4	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   5	
  
Commonly asked questions in exascale: What is the
problem?!
•  Power, energy, and heat dissipation are the central issues"
•  Imagine a computer with billions and billions of cell phone processors
(14MW) or millions and millions of throughput optimized cores,
GPGPUs (20MW)"
•  How do you program it to work on one science problem?"
•  The architecture will be heterogeneous and hierarchical, with very
high flop/byte ratios."
•  Single program multiple data bulk synchronous parallelism will no
longer be viable."
•  Data Movement will be expensive and computation will be cheap"
•  Need to present the physics so the computation occurs where the
data is!"
•  Traditional global checkpoint/restart will be impractical: need local /
micro checkpoint (flash memory?)"
•  Simulation codes will need to become fault tolerant and resilient!
•  Recover from soft and hard errors, and anticipating faults"
•  Ability to drop or replace nodes and keep on running"
•  The curse of silent errors"
3D Stacked
Memory
(Low Capacity, High Bandwidth)
Fat
Core
Fat
Core
Thin Cores / Accelerators
DRAMNVRAM
(High Capacity,
Low Bandwidth)
Coherence DomainCore
Integrated NIC
for Off-Chip
Communication
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Preparation:	
  
Science	
  and	
  Mission	
  
Stakeholder	
  Buy-­‐in	
  
Assemble	
  Team	
  
Implementation	
  
Plan	
  
Development	
  Plan	
  
Cycle	
  Artifacts:	
  
	
  R&D	
  Backlog	
  
	
  Algorithm	
  and	
  
	
  Model	
  
Implementation	
  
	
  Proxy	
  Applications	
  
	
  Architecture	
  
Evaluation	
  
Co-­‐Design	
  
Agile	
  
Development	
  
Cycle	
  Incorporated	
  
Design	
  
Elements	
  
Algorithm	
  
Development	
  
Trade-­‐off	
  
Analysis	
  
Impact	
  
Feedback	
  
Code	
  
Design	
  
Exascale	
  Community:	
  
Release	
  Artifacts:	
  
HW	
  Requirements	
  
SW	
  Constraints	
  
Proxy	
  Applications	
  
Documentation	
  
Software	
  Development:	
  
ASCR	
  X-­‐stack,	
  ASC	
  CSSE	
  
Data/Analysis	
  
Hardware	
  
Development:	
  Vendors,	
  
Fastforward,	
  ASCR	
  
Advanced	
  Architecture	
  
Code	
  
Implementation	
  
Release	
  to	
  
Exascale	
  
Community	
  
Release	
  n	
  
Domain	
  
Science:	
  
Domain	
  Workload	
  
Physical	
  Models	
  
Algorithms	
  
Simulations	
  
Team	
  Roles:	
  
Cycle	
  Master:	
  Co-­‐design	
  PI	
  
	
  Project	
  Team:	
  Labs,	
  Univ’s	
  
	
  Stakeholders:	
  ASCR,	
  ASC,	
  
Vendors	
  
	
  Customers:	
  Scientists,	
  HW+SW	
  
Developers	
  
6	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
System
Software
Proxy
Apps
Application
Co-Design
Hardware
Co-Design
Computer
Science
Co-Design
Vendor
Analysis
Sim Exp
Proto HW
Prog Models
HW Simulator
Tools
Open
Analysis
Models
Simulators
Emulators
HW
Design
Stack
Analysis
Prog models
Tools
Compilers
Runtime
OS, I/O, ...HW Constraints
Domain/Alg
Analysis
SW Solutions
System Design!
Application Design!
Workflow of Co-design between
Application Co-design Centers,
hardware vendors, and the
broader research community"
X-Stack!
Exascale OS/R!
Fast Forward!
Design Forward!
Application Co-
design Centers!
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
DOE	
  Office	
  of	
  Science	
  ASCR	
  Co-­‐design	
  Projects:	
  
Center	
  for	
  Exascale	
  Simulation	
  of	
  Advanced	
  Reactors	
  (CESAR)	
  
http://cesar.mcs.anl.gov	
  
Director:	
  Andrew	
  Siegel	
  (ANL/Uchicago),	
  Deputy	
  Director:	
  Paul	
  Fischer	
  (ANL)	
  
Center	
  for	
  Exascale	
  Simulation	
  of	
  Combustion	
  in	
  Turbulance	
  (ExaCT)	
  
http://exactcodesign.org	
  
Director:	
  Jacqueline	
  Chen	
  (SNL),	
  Deputy	
  Director:	
  John	
  Bell	
  (LBNL)	
  
Exascale	
  Co-­‐Design	
  Center	
  for	
  Materials	
  in	
  Extreme	
  Environments	
  (ExMatEX)	
  
http://exmatex.lanl.gov	
  
Director:	
  Tim	
  Germann	
  (LANL),	
  Deputy	
  Director:	
  Jim	
  Belak	
  (LLNL)	
  
	
  	
  
DOE	
  NNSA	
  ASC	
  Co-­‐design	
  Project:	
  	
  
National	
  Security	
  Applications	
  Co-­‐Design	
  Project	
  (NSApp	
  CDP)	
  
Contacts:	
  Sriram	
  Swaminarayan(LANL),	
  Rob	
  Neely(LLNL),	
  Hoekstra/Heroux(SNL-­‐A)	
  
http://proxyapps.lanl.gov	
  
8	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  Materials	
  Science	
  applications	
  present	
  
their	
  requirements	
  through	
  proxies	
  
•  Real	
  applications	
  have	
  multiple	
  
proxies	
  
•  Proxies	
  have	
  docs,	
  specs,	
  and	
  a	
  
reference	
  implementation	
  
§  Ecosystem	
  members	
  evaluate	
  proxies	
  
and	
  respond	
  with	
  capabilities	
  
•  This	
  informs	
  trade-­‐off	
  analysis	
  
§  Co-­‐Design	
  is	
  more	
  than	
  just	
  
applications	
  and	
  hardware	
  
architecture	
  
•  There	
  is	
  a	
  whole	
  ecosystem	
  to	
  
influence	
  
9	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Code:	
  Qbox/
LATTE	
  
	
  
Motif:	
  Particles	
  
and	
  
wavefunctions,	
  
plane	
  wave	
  DFT,	
  
ScaLAPACK,	
  
BLACS,	
  and	
  
custom	
  parallel	
  
3D	
  FFTs	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  CUBLAS/CUDA	
  
Code:SPaSM/
ddcMD/CoMD	
  
	
  
Motif:	
  Particles,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  
dynamic	
  load	
  
balancing,	
  parity	
  
error	
  recovery,	
  
and	
  in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  SEAKMC	
  
	
  
	
  
Motif:	
  Particles	
  
and	
  defects,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  and	
  
in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  
MPI	
  +	
  Threads	
  
Code:	
  AMPE/GL	
  
	
  
	
  
Motif:	
  Regular	
  and	
  
adaptive	
  grids,	
  
implicit	
  time	
  
integration,	
  real-­‐
space	
  and	
  spectral	
  
methods,	
  complex	
  
order	
  parameter	
  
	
  
Prog.	
  Model:	
  MPI	
  
Code:	
  ParaDiS	
  
	
  
	
  
Motif:	
  
“segments”	
  
Regular	
  mesh,	
  
implicit	
  time	
  
integration,	
  fast	
  
multipole	
  
method	
  
	
  
Prog.	
  Model:	
  
MPI	
  
Code:	
  VP-­‐FFT	
  
	
  
	
  
Motif:	
  Regular	
  
grids,	
  tensor	
  
arithmetic,	
  
meshless	
  image	
  
processing,	
  
implicit	
  time	
  
integration,	
  3D	
  
FFTs.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  ALE3D/
LULESH	
  
	
  
Motif:	
  Regular	
  
and	
  irregular	
  
grids,	
  explicit	
  and	
  
implicit	
  time	
  
integration.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
	
  
Ab-­‐initio	
   Atoms	
   Long-­‐time	
   Microstructure	
   Dislocation	
   Crystal	
   Continuum	
  
Inter-­‐atomic	
  
forces,	
  EOS,	
  
excited	
  states	
  
Defects	
  and	
  
interfaces,	
  
nucleation	
  
Defects	
  and	
  
defect	
  
structures	
  
Meso-­‐scale	
  multi-­‐
phase,	
  multi-­‐grain	
  
evolution	
  
Meso-­‐scale	
  
strength	
  
Meso-­‐scale	
  
material	
  
response	
  
Macro-­‐scale	
  
material	
  
response	
  
10	
  
Jim	
  Belak,	
  ExMatEx	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
~	
  20	
  atoms	
  in	
  each	
  box	
  
⇒  each	
  atom	
  interacts	
  with	
  540	
  other	
  atoms	
  
⇒  However,	
  only	
  ~70	
  atoms	
  lie	
  within	
  cutoff	
  
⇒  Lots	
  of	
  wasted	
  work	
  
⇒  We	
  need	
  a	
  means	
  of	
  rejecting	
  atoms	
  efficiently	
  
even	
  within	
  this	
  reduced	
  set	
  	
  
Halo	
  Region	
  
11	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   12	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  Sam	
  Reeve	
  (Purdue)	
  LAMMPS,	
  CoMD,	
  leanMD	
  
§  Riyaz	
  Haque	
  (UCLA),	
  CNC	
  (Habenero-­‐C)	
  
§  Luc	
  Jualmes	
  (Barcelona)	
  OmpSs,	
  Chapel	
  
§  Sameer	
  Abu	
  Asal	
  (LSU)	
  C++11	
  futures,	
  HPX	
  
§  Aaron	
  Landmehr	
  (Delaware)	
  OCR	
  
§  Sanian	
  Gaffer	
  (NM-­‐State)	
  UPC++	
  
§  Gheorghe-­‐Teodor	
  Bercea	
  (Imperial)	
  PyOP2	
  DSL	
  
§  Zach	
  Rubinstein	
  (Chicago),	
  Troels	
  Henriksen	
  
(Copenhagen)	
  Embedded	
  DSL	
  
Hybrid)Decomposi>on
7/23/14' Charm++':'grainsize'
Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>
" )We)hav
objects)to
o )Each)
assigne
o )Numb
(3D):))
o )14XNu
)
Worker	
  
Work	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   14	
  
Hybrid)Decomposi>on)
7/23/14' Charm++':'grainsize'
Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>al)Deco
" )We)have)man
objects)to)load)
o )Each)diamon
assigned)to)an
o )Number)of)d
(3D):))
o )14XNumber)o
)
MD	
  “MPI”	
  
Worker	
  Decomposition	
  
Charm++	
  
Work	
  Decomposition	
  
c.f.	
  http://charm.cs.uiuc.edu/	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Hybrid)Decomposi>on)
7/23/14' Charm++':'grainsize' 14'
Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>al)Decomp.)
" )We)have)many)
objects)to)load)balance:)
o )Each)diamond)can)be)
assigned)to)any)proc.)
o )Number)of)diamonds)
(3D):))
o )14XNumber)of)Cells)
)
Di	
  
Dj	
  
Dk	
  
I
n
i
t
i
a
t
e	
  
Di	
  
Dj	
  
Dk	
  
Propagate	
  
Atoms	
  
Ti	
  
Tj	
  
Tk	
  
Compute	
  
Forces	
  
Tij	
  
Tjk	
  
Di	
  
Dj	
  
Dk	
  
Di	
  
Dj	
  
Dk	
  
Migrate	
  
Atoms	
  
Ti	
  
Loop	
  until	
  done	
  
Ti	
  
Tj	
  
Tk	
  
Tj	
  
Tk	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Cray	
  T3D,	
  Clock	
  Speed:	
  0.15GHz,	
  Date:	
  1994	
  
Microprocessor	
  Peak	
  Teraflops:	
  0.02,	
  Memory:	
  0.01TB	
  
Processors:	
  128.	
  Power	
  Consumption:	
  41.40	
  kW	
  
	
  
SPMD,	
  MPI,	
  
Geometric	
  domain	
  
decomposition	
  
Diffusion	
  of	
  a	
  
small	
  drug	
  
molecule	
  
through	
  a	
  bio-­‐
membrane	
  
Three	
  Essential	
  Elements:	
  
•  Problem	
  Specification	
  
•  Target	
  Architecture	
  
•  Programming	
  Model	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  20	
  Years	
  Ago	
  
•  Hardware:	
  Linux	
  Beowolf	
  Cluster	
  
•  Software	
  Programming	
  Model:	
  MPI/SPMD	
  
•  Application	
  Code:	
  LAMMPS	
  
§  5+	
  Years	
  in	
  the	
  Future	
  
•  Hardware:	
  Exascale	
  (Petascale	
  in	
  a	
  Rack)	
  
•  Software	
  Programming	
  Model:	
  MPI+X,	
  Q?	
  
Task-­‐based	
  with	
  “Control”	
  of	
  Data	
  Locality	
  
(User	
  and	
  Runtime)	
  
•  Application	
  Codes???	
  
	
  
17	
  
Worker	
  
Hybrid)Decomposi>on
7/23/14' Charm++':'grainsize'
Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>
" )We)hav
objects)to
o )Each)
assigne
o )Numb
(3D):))
o )14XNu
)
Work	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Code:	
  Qbox/
LATTE	
  
	
  
Motif:	
  Particles	
  
and	
  
wavefunctions,	
  
plane	
  wave	
  DFT,	
  
ScaLAPACK,	
  
BLACS,	
  and	
  
custom	
  parallel	
  
3D	
  FFTs	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  CUBLAS/CUDA	
  
Code:SPaSM/
ddcMD/CoMD	
  
	
  
Motif:	
  Particles,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  
dynamic	
  load	
  
balancing,	
  parity	
  
error	
  recovery,	
  
and	
  in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  SEAKMC	
  
	
  
	
  
Motif:	
  Particles	
  
and	
  defects,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  and	
  
in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  
MPI	
  +	
  Threads	
  
Code:	
  AMPE/GL	
  
	
  
	
  
Motif:	
  Regular	
  and	
  
adaptive	
  grids,	
  
implicit	
  time	
  
integration,	
  real-­‐
space	
  and	
  spectral	
  
methods,	
  complex	
  
order	
  parameter	
  
	
  
Prog.	
  Model:	
  MPI	
  
Code:	
  ParaDiS	
  
	
  
	
  
Motif:	
  
“segments”	
  
Regular	
  mesh,	
  
implicit	
  time	
  
integration,	
  fast	
  
multipole	
  
method	
  
	
  
Prog.	
  Model:	
  
MPI	
  
Code:	
  VP-­‐FFT	
  
	
  
	
  
Motif:	
  Regular	
  
grids,	
  tensor	
  
arithmetic,	
  
meshless	
  image	
  
processing,	
  
implicit	
  time	
  
integration,	
  3D	
  
FFTs.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  ALE3D/
LULESH	
  
	
  
Motif:	
  Regular	
  
and	
  irregular	
  
grids,	
  explicit	
  and	
  
implicit	
  time	
  
integration.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
	
  
Ab-­‐initio	
   Atoms	
   Long-­‐time	
   Microstructure	
   Dislocation	
   Crystal	
   Continuum	
  
Inter-­‐atomic	
  
forces,	
  EOS,	
  
excited	
  states	
  
Defects	
  and	
  
interfaces,	
  
nucleation	
  
Defects	
  and	
  
defect	
  
structures	
  
Meso-­‐scale	
  multi-­‐
phase,	
  multi-­‐grain	
  
evolution	
  
Meso-­‐scale	
  
strength	
  
Meso-­‐scale	
  
material	
  
response	
  
Macro-­‐scale	
  
material	
  
response	
  
18	
  
Jim	
  Belak,	
  ExMatEx	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Simulations suggest novel in situ x-ray scattering
experiments using emerging sources such as LCLS!
19	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
100 nm	

Atoms	

Phase	

time!
Coarse
Graining	
  
20	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
•  Each color represents a
different value of the
phase field (solid
orientation)!
•  Free energy describes
how colors interact and
evolve!
•  Accuracy depends on
fidelity of physics in the
equations!
F(P,T) = dx∫ ∇
!
φ
2
+ f (
!
φ, P,T)+…}{
∂
!
φ
∂t
= −Γ
δF
δ
!
φ
+ noise
Evolution Equations!
Thermodynamic representation of
phase (or “color”) everywhere!
!
φ
!
φ(r,t) = pink
!
φ(r,t) = green
!
φ(r,t) = solid
!
φ(r,t) = liquid
21	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
F =
εϕ
2
2
∇ϕ
2
+ f (ϕ,c,T)+ HT[1− p(ϕ)] (∇qi )2
i
∑
$
%
&
'
(
)
1/2*
+
,
,
-
.
/
/
∫ d3
r
∂qi
∂t
= −Mq
δF
δqi
+ζi = Mq ∇⋅ D
∇qi
∇qi
%
&
''
(
)
**− 2λqi
+
,
-
-
.
/
0
0
+ζi
Pusztai et al., have proposed a 3D quaternion-based phase-field model !
• Represents crystal orientation with quaternion order parameter"
• Quaternions are widely used to analyze crystallography of polycrystal interfaces"
• Quaternion algebra is fast, efficient, avoids singularities, …"
Where qi is the quaternion order parameter, Mq is the associated mobility and ζ is the
fluctuation in q."
Free
Energy	

Evolution	

Refs: T. Pusztai, G. Bortel, and L. Granasy, “Phase field theory of polycrystalline solidification in three dimensions,”
Europhys. Lett, 71 (2005) 131-137; Dorr, M.R., Fattebert, J.-L., Wickett, M.E., Belak, J.F., and Turchi, P.E.A., “A Numerical
Algorithm for the Solution of a Phase-field Model of Polycrystalline Materials,” J. Comp. Phys. Vol. 229, 626 (2010). "
We have implemented the Pusztai model in our 3D AMR code !
• Enhance energy functional to represent energetics of grain boundaries"
• Crystal symmetry aware quaternion mathematics"
• Extend energy functional to include elasticity and alloy concentration"
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   23	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Growth of large
grains
Blue: MD
nucleation
Red: Phase-field
evolution
MD	
  nucleated	
  microstructure	
  onto	
  the	
  micro-­‐second	
  
hydro	
  time-­‐scale	
  with	
  the	
  crystallographic	
  quaternion	
  
model	
  
While significant grain coarsening has occurred on the
microsecond scale, the microstructure is far from log-normal
Phase Order Parameter! Quaternion Order Parameter!
24	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
Code:	
  Qbox/
LATTE	
  
	
  
Motif:	
  Particles	
  
and	
  
wavefunctions,	
  
plane	
  wave	
  DFT,	
  
ScaLAPACK,	
  
BLACS,	
  and	
  
custom	
  parallel	
  
3D	
  FFTs	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  CUBLAS/CUDA	
  
Code:SPaSM/
ddcMD/CoMD	
  
	
  
Motif:	
  Particles,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  
dynamic	
  load	
  
balancing,	
  parity	
  
error	
  recovery,	
  
and	
  in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  SEAKMC	
  
	
  
	
  
Motif:	
  Particles	
  
and	
  defects,	
  
explicit	
  time	
  
integration,	
  
neighbor	
  and	
  
linked	
  lists,	
  and	
  
in	
  situ	
  
visualization	
  
	
  
Prog.	
  Model:	
  
MPI	
  +	
  Threads	
  
Code:	
  AMPE/GL	
  
	
  
	
  
Motif:	
  Regular	
  and	
  
adaptive	
  grids,	
  
implicit	
  time	
  
integration,	
  real-­‐
space	
  and	
  spectral	
  
methods,	
  complex	
  
order	
  parameter	
  
	
  
Prog.	
  Model:	
  MPI	
  
Code:	
  ParaDiS	
  
	
  
	
  
Motif:	
  
“segments”	
  
Regular	
  mesh,	
  
implicit	
  time	
  
integration,	
  fast	
  
multipole	
  
method	
  
	
  
Prog.	
  Model:	
  
MPI	
  
Code:	
  VP-­‐FFT	
  
	
  
	
  
Motif:	
  Regular	
  
grids,	
  tensor	
  
arithmetic,	
  
meshless	
  image	
  
processing,	
  
implicit	
  time	
  
integration,	
  3D	
  
FFTs.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
Code:	
  ALE3D/
LULESH	
  
	
  
Motif:	
  Regular	
  
and	
  irregular	
  
grids,	
  explicit	
  and	
  
implicit	
  time	
  
integration.	
  
	
  
Prog.	
  Model:	
  MPI	
  
+	
  Threads	
  
	
  
Ab-­‐initio	
   Atoms	
   Long-­‐time	
   Microstructure	
   Dislocation	
   Crystal	
   Continuum	
  
Inter-­‐atomic	
  
forces,	
  EOS,	
  
excited	
  states	
  
Defects	
  and	
  
interfaces,	
  
nucleation	
  
Defects	
  and	
  
defect	
  
structures	
  
Meso-­‐scale	
  multi-­‐
phase,	
  multi-­‐grain	
  
evolution	
  
Meso-­‐scale	
  
strength	
  
Meso-­‐scale	
  
material	
  
response	
  
Macro-­‐scale	
  
material	
  
response	
  
25	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  To	
  achieve	
  this,	
  we	
  are	
  
developing	
  a	
  UQ-­‐driven	
  
adaptive	
  physics	
  refinement	
  
approach.	
  
§  Coarse-­‐scale	
  simulations	
  
dynamically	
  spawn	
  tightly	
  
coupled	
  and	
  self-­‐consistent	
  
fine-­‐scale	
  simulations	
  as	
  
needed.	
  
§  This	
  task-­‐based	
  approach	
  
naturally	
  maps	
  to	
  exascale	
  
heterogeneity,	
  concurrency,	
  
and	
  resiliency	
  issues.	
  
	
  
•  Task-­‐based	
  embedded	
  Scale-­‐Bridging	
  escapes	
  the	
  traditional	
  synchronous	
  
SPMD	
  paradigm	
  and	
  exploits	
  the	
  heterogeneity	
  expected	
  in	
  exascale	
  hardware.	
  	
  
	
  
26	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   27	
  
FSMs	
  
•  Brute	
  force	
  multi-­‐scale	
  coupling:	
  Full	
  fine	
  scale	
  
model	
  (FSM,	
  e.g.	
  a	
  crystal	
  plasticity	
  model)	
  
run	
  for	
  every	
  zone	
  &	
  time	
  step	
  of	
  coarse	
  scale	
  
mode	
  (CSM,	
  e.g.	
  an	
  ALE	
  code)	
  
•  Adaptive	
  Sampling:	
  	
  
–  Save	
  FSM	
  results	
  in	
  database	
  
–  Before	
  running	
  another	
  FSM,	
  check	
  database	
  
for	
  FSM	
  results	
  similar	
  enough	
  to	
  those	
  
needed	
  that	
  interpolation	
  or	
  extrapolation	
  
suffices	
  
–  Only	
  run	
  full	
  FSM	
  when	
  results	
  in	
  database	
  not	
  
close	
  enough	
  
CSM	
  
Ref:	
  Barton	
  et.al,	
  ‘A	
  call	
  to	
  arms	
  for	
  task	
  parallelism	
  in	
  multi-­‐scale	
  materials	
  modeling,’	
  Int.	
  J.	
  Numer.	
  Meth.	
  Engng	
  2011;	
  86:744–764	
  
§  Heterogeneous,	
  hierarchical	
  MPMD	
  algorithms	
  map	
  naturally	
  to	
  anticipated	
  
heterogeneous,	
  hierarchical	
  architectures	
  
§  Escape	
  the	
  traditional	
  bulk	
  synchronous	
  SPMD	
  paradigm,	
  improve	
  scalability	
  and	
  
reduce	
  scheduling	
  
§  Task-­‐based	
  MPMD	
  approach	
  leverages	
  concurrency	
  and	
  heterogeneity	
  at	
  exascale	
  
while	
  enabling	
  novel	
  data	
  models,	
  power	
  management,	
  and	
  fault	
  tolerance	
  
strategies	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
AS	
  Database	
  
Regions	
  over	
  
which	
  models	
  
may	
  extrapolate	
  
Past	
  fine-­‐scale	
  evaulation	
  
results;	
  approximation	
  models	
  
Queried	
  points	
  
Queried	
  point	
  close	
  
enough	
  for	
  
approximation	
  
Fine-­‐scale	
  evaluation	
  
at	
  this	
  query	
  
Input	
  Space	
  
§  Coarse	
  scale	
  model	
  
queries	
  database	
  for	
  fine-­‐
scale	
  material	
  response	
  
§  If	
  possible,	
  approximate	
  
response	
  from	
  past	
  
evaluations	
  
§  Otherwise	
  perform	
  fine	
  
scale	
  evaluation	
  
§  Fine-­‐scale	
  	
  
evaluations	
  grow	
  	
  
database	
  
28	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
On-demand fine
scale models"
CSM"
DB$"
Adaptive"
Sampler"
FSM"
Subdomain 1"
Subdomain 2"
FSM" FSM"
Node 1!
DB$"
Adaptive"
Sampler"
Subdomain N-1"
Subdomain N"
Node N/2!
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
DB"
On-demand fine
scale models"
CSM"
DB$"
Eventually
consistent
distributed
database"
Adaptive"
Sampler"
FSM"
Subdomain 1"
Subdomain 2"
FSM" FSM"
Node 1!
DB"
DB"
DB$"
Adaptive"
Sampler"
Subdomain N-1"
Subdomain N"
Node N/2!
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
  
§  Our	
  goal	
  is	
  to	
  establish	
  the	
  
interrelationship	
  between	
  hardware,	
  
middleware	
  (software	
  stack),	
  
programming	
  models,	
  and	
  algorithms	
  
required	
  to	
  enable	
  a	
  productive	
  exascale	
  
environment	
  for	
  multiphysics	
  simulations	
  
of	
  materials	
  in	
  extreme	
  mechanical	
  and	
  
radiation	
  environments.	
  
§  We	
  will	
  exploit,	
  rather	
  than	
  avoid,	
  the	
  
greatly	
  increased	
  levels	
  of	
  concurrency,	
  
heterogeneity,	
  and	
  flop/byte	
  ratios	
  on	
  the	
  
upcoming	
  exascale	
  platforms.	
  	
  
•  Our	
  vision	
  is	
  an	
  uncertainty	
  quantification	
  (UQ)-­‐driven	
  adaptive	
  physics	
  
refinement	
  in	
  which	
  meso-­‐	
  and	
  macro-­‐scale	
  materials	
  simulations	
  spawn	
  
micro-­‐scale	
  simulations	
  as	
  needed.	
  
-  This	
  task-­‐based	
  approach	
  leverages	
  the	
  extensive	
  concurrency	
  and	
  
heterogeneity	
  expected	
  at	
  exascale	
  while	
  enabling	
  fault	
  tolerance	
  within	
  
applications.	
  	
  
-  The	
  programming	
  models	
  and	
  approaches	
  developed	
  to	
  achieve	
  this	
  will	
  be	
  
broadly	
  applicable	
  to	
  a	
  variety	
  of	
  multiscale,	
  multiphysics	
  applications,	
  including	
  
astrophysics,	
  climate	
  and	
  weather	
  prediction,	
  structural	
  engineering,	
  plasma	
  
physics,	
  and	
  radiation	
  hydrodynamics.	
  
APP	
  
SW	
  HW	
  
Func+onal	
  
Exascale	
  
Simula+on	
  
Environment	
  
Domain	
  Workload	
  
31	
  
Lawrence Livermore National Laboratory 3	
  	
  June,	
  02015	
   32	
  

More Related Content

What's hot

The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research PlatformLarry Smarr
 
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)Keiichiro Ono
 
Mark_Yashar_Resume_2017
Mark_Yashar_Resume_2017Mark_Yashar_Resume_2017
Mark_Yashar_Resume_2017Mark Yashar
 
Ingredients for Semantic Sensor Networks
Ingredients for Semantic Sensor NetworksIngredients for Semantic Sensor Networks
Ingredients for Semantic Sensor NetworksOscar Corcho
 
Creating a Big Data Machine Learning Platform in California
Creating a Big Data Machine Learning Platform in CaliforniaCreating a Big Data Machine Learning Platform in California
Creating a Big Data Machine Learning Platform in CaliforniaLarry Smarr
 
The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research PlatformLarry Smarr
 
Towards Incidental Collaboratories; Research Data Services
Towards Incidental Collaboratories; Research Data ServicesTowards Incidental Collaboratories; Research Data Services
Towards Incidental Collaboratories; Research Data ServicesAnita de Waard
 
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...Splunk
 
Allinea DDT Resolves an Unsolved Mystery at Argonne National Laboratory
Allinea DDT Resolves an Unsolved Mystery at Argonne National LaboratoryAllinea DDT Resolves an Unsolved Mystery at Argonne National Laboratory
Allinea DDT Resolves an Unsolved Mystery at Argonne National LaboratoryAllineaSoftware
 
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...Keiichiro Ono
 
Next Technology Wave
Next Technology WaveNext Technology Wave
Next Technology WaveFalascoj
 
Assessing Galaxy's ability to express scientific workflows in bioinformatics
Assessing Galaxy's ability to express scientific workflows in bioinformaticsAssessing Galaxy's ability to express scientific workflows in bioinformatics
Assessing Galaxy's ability to express scientific workflows in bioinformaticsPeter van Heusden
 
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...A Semantic-Based Approach to Attain Reproducibility of Computational Environm...
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...Idafen Santana Pérez
 
An Integrated West Coast Science DMZ for Data-Intensive Research
An Integrated West Coast Science DMZ for Data-Intensive ResearchAn Integrated West Coast Science DMZ for Data-Intensive Research
An Integrated West Coast Science DMZ for Data-Intensive ResearchLarry Smarr
 
big_data_casestudies_2.ppt
big_data_casestudies_2.pptbig_data_casestudies_2.ppt
big_data_casestudies_2.pptvishal choudhary
 
Scientific Workflow Systems for accessible, reproducible research
Scientific Workflow Systems for accessible, reproducible researchScientific Workflow Systems for accessible, reproducible research
Scientific Workflow Systems for accessible, reproducible researchPeter van Heusden
 
The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research PlatformLarry Smarr
 
Histolab: an Open Source Python Library for Reproducible Digital Pathology
Histolab: an Open Source Python Library for Reproducible Digital PathologyHistolab: an Open Source Python Library for Reproducible Digital Pathology
Histolab: an Open Source Python Library for Reproducible Digital PathologyAlessia Marcolini
 
Analysis of User Submission Behavior on HPC and HTC
Analysis of User Submission Behavior on HPC and HTCAnalysis of User Submission Behavior on HPC and HTC
Analysis of User Submission Behavior on HPC and HTCRafael Ferreira da Silva
 

What's hot (20)

The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research Platform
 
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)
Cytoscape Tutorial Session 2 at UT-KBRIN Bioinformatics Summit 2014 (4/11/2014)
 
Mark_Yashar_Resume_2017
Mark_Yashar_Resume_2017Mark_Yashar_Resume_2017
Mark_Yashar_Resume_2017
 
Ingredients for Semantic Sensor Networks
Ingredients for Semantic Sensor NetworksIngredients for Semantic Sensor Networks
Ingredients for Semantic Sensor Networks
 
Creating a Big Data Machine Learning Platform in California
Creating a Big Data Machine Learning Platform in CaliforniaCreating a Big Data Machine Learning Platform in California
Creating a Big Data Machine Learning Platform in California
 
The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research Platform
 
Towards Incidental Collaboratories; Research Data Services
Towards Incidental Collaboratories; Research Data ServicesTowards Incidental Collaboratories; Research Data Services
Towards Incidental Collaboratories; Research Data Services
 
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...
Virtual Gov Day - IT Operations Breakout - Jennifer Green, R&D Scientist, Los...
 
Allinea DDT Resolves an Unsolved Mystery at Argonne National Laboratory
Allinea DDT Resolves an Unsolved Mystery at Argonne National LaboratoryAllinea DDT Resolves an Unsolved Mystery at Argonne National Laboratory
Allinea DDT Resolves an Unsolved Mystery at Argonne National Laboratory
 
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...
Workshop: Introduction to Cytoscape at UT-KBRIN Bioinformatics Summit 2014 (4...
 
Next Technology Wave
Next Technology WaveNext Technology Wave
Next Technology Wave
 
Assessing Galaxy's ability to express scientific workflows in bioinformatics
Assessing Galaxy's ability to express scientific workflows in bioinformaticsAssessing Galaxy's ability to express scientific workflows in bioinformatics
Assessing Galaxy's ability to express scientific workflows in bioinformatics
 
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...A Semantic-Based Approach to Attain Reproducibility of Computational Environm...
A Semantic-Based Approach to Attain Reproducibility of Computational Environm...
 
An Integrated West Coast Science DMZ for Data-Intensive Research
An Integrated West Coast Science DMZ for Data-Intensive ResearchAn Integrated West Coast Science DMZ for Data-Intensive Research
An Integrated West Coast Science DMZ for Data-Intensive Research
 
big_data_casestudies_2.ppt
big_data_casestudies_2.pptbig_data_casestudies_2.ppt
big_data_casestudies_2.ppt
 
Scientific Workflow Systems for accessible, reproducible research
Scientific Workflow Systems for accessible, reproducible researchScientific Workflow Systems for accessible, reproducible research
Scientific Workflow Systems for accessible, reproducible research
 
The Pacific Research Platform
The Pacific Research PlatformThe Pacific Research Platform
The Pacific Research Platform
 
Collins seattle-2014-final
Collins seattle-2014-finalCollins seattle-2014-final
Collins seattle-2014-final
 
Histolab: an Open Source Python Library for Reproducible Digital Pathology
Histolab: an Open Source Python Library for Reproducible Digital PathologyHistolab: an Open Source Python Library for Reproducible Digital Pathology
Histolab: an Open Source Python Library for Reproducible Digital Pathology
 
Analysis of User Submission Behavior on HPC and HTC
Analysis of User Submission Behavior on HPC and HTCAnalysis of User Submission Behavior on HPC and HTC
Analysis of User Submission Behavior on HPC and HTC
 

Similar to Belak_ICME_June02015

Exascale Computing Project - Driving a HUGE Change in a Changing World
Exascale Computing Project - Driving a HUGE Change in a Changing WorldExascale Computing Project - Driving a HUGE Change in a Changing World
Exascale Computing Project - Driving a HUGE Change in a Changing Worldinside-BigData.com
 
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy Sciences
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy SciencesDiscovery Engines for Big Data: Accelerating Discovery in Basic Energy Sciences
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy SciencesIan Foster
 
Scientific
Scientific Scientific
Scientific marpierc
 
Discovering new functional materials for clean energy and beyond using high-t...
Discovering new functional materials for clean energy and beyond using high-t...Discovering new functional materials for clean energy and beyond using high-t...
Discovering new functional materials for clean energy and beyond using high-t...Anubhav Jain
 
The U.S. Exascale Computing Project: Status and Plans
The U.S. Exascale Computing Project: Status and PlansThe U.S. Exascale Computing Project: Status and Plans
The U.S. Exascale Computing Project: Status and Plansinside-BigData.com
 
Learning Systems for Science
Learning Systems for ScienceLearning Systems for Science
Learning Systems for ScienceIan Foster
 
Accelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceAccelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceIan Foster
 
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022Monitoring Exascale Supercomputers With Tim Osborne | Current 2022
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022HostedbyConfluent
 
2014 11-13-sbsm032-reproducible research
2014 11-13-sbsm032-reproducible research2014 11-13-sbsm032-reproducible research
2014 11-13-sbsm032-reproducible researchYannick Wurm
 
Software Mining and Software Datasets
Software Mining and Software DatasetsSoftware Mining and Software Datasets
Software Mining and Software DatasetsTao Xie
 
Stories About Spark, HPC and Barcelona by Jordi Torres
Stories About Spark, HPC and Barcelona by Jordi TorresStories About Spark, HPC and Barcelona by Jordi Torres
Stories About Spark, HPC and Barcelona by Jordi TorresSpark Summit
 
Acceleration of XML Parsing through Prefetching
Acceleration of XML  Parsing through PrefetchingAcceleration of XML  Parsing through Prefetching
Acceleration of XML Parsing through PrefetchingRohit Deshpande
 
09 The Extreme-scale Scientific Software Stack for Collaborative Open Source
09 The Extreme-scale Scientific Software Stack for Collaborative Open Source09 The Extreme-scale Scientific Software Stack for Collaborative Open Source
09 The Extreme-scale Scientific Software Stack for Collaborative Open SourceRCCSRENKEI
 
HPC Cluster Computing from 64 to 156,000 Cores 
HPC Cluster Computing from 64 to 156,000 Cores HPC Cluster Computing from 64 to 156,000 Cores 
HPC Cluster Computing from 64 to 156,000 Cores inside-BigData.com
 
Evaluating Machine Learning Algorithms for Materials Science using the Matben...
Evaluating Machine Learning Algorithms for Materials Science using the Matben...Evaluating Machine Learning Algorithms for Materials Science using the Matben...
Evaluating Machine Learning Algorithms for Materials Science using the Matben...Anubhav Jain
 
05 Preparing for Extreme Geterogeneity in HPC
05 Preparing for Extreme Geterogeneity in HPC05 Preparing for Extreme Geterogeneity in HPC
05 Preparing for Extreme Geterogeneity in HPCRCCSRENKEI
 
How HPC and large-scale data analytics are transforming experimental science
How HPC and large-scale data analytics are transforming experimental scienceHow HPC and large-scale data analytics are transforming experimental science
How HPC and large-scale data analytics are transforming experimental scienceinside-BigData.com
 
Towards an Infrastructure for Enabling Systematic Development and Research of...
Towards an Infrastructure for Enabling Systematic Development and Research of...Towards an Infrastructure for Enabling Systematic Development and Research of...
Towards an Infrastructure for Enabling Systematic Development and Research of...Rafael Ferreira da Silva
 

Similar to Belak_ICME_June02015 (20)

AI for Science
AI for ScienceAI for Science
AI for Science
 
Exascale Computing Project - Driving a HUGE Change in a Changing World
Exascale Computing Project - Driving a HUGE Change in a Changing WorldExascale Computing Project - Driving a HUGE Change in a Changing World
Exascale Computing Project - Driving a HUGE Change in a Changing World
 
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy Sciences
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy SciencesDiscovery Engines for Big Data: Accelerating Discovery in Basic Energy Sciences
Discovery Engines for Big Data: Accelerating Discovery in Basic Energy Sciences
 
Scientific
Scientific Scientific
Scientific
 
Discovering new functional materials for clean energy and beyond using high-t...
Discovering new functional materials for clean energy and beyond using high-t...Discovering new functional materials for clean energy and beyond using high-t...
Discovering new functional materials for clean energy and beyond using high-t...
 
The U.S. Exascale Computing Project: Status and Plans
The U.S. Exascale Computing Project: Status and PlansThe U.S. Exascale Computing Project: Status and Plans
The U.S. Exascale Computing Project: Status and Plans
 
Learning Systems for Science
Learning Systems for ScienceLearning Systems for Science
Learning Systems for Science
 
Accelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy ScienceAccelerating Data-driven Discovery in Energy Science
Accelerating Data-driven Discovery in Energy Science
 
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022Monitoring Exascale Supercomputers With Tim Osborne | Current 2022
Monitoring Exascale Supercomputers With Tim Osborne | Current 2022
 
Future of hpc
Future of hpcFuture of hpc
Future of hpc
 
2014 11-13-sbsm032-reproducible research
2014 11-13-sbsm032-reproducible research2014 11-13-sbsm032-reproducible research
2014 11-13-sbsm032-reproducible research
 
Software Mining and Software Datasets
Software Mining and Software DatasetsSoftware Mining and Software Datasets
Software Mining and Software Datasets
 
Stories About Spark, HPC and Barcelona by Jordi Torres
Stories About Spark, HPC and Barcelona by Jordi TorresStories About Spark, HPC and Barcelona by Jordi Torres
Stories About Spark, HPC and Barcelona by Jordi Torres
 
Acceleration of XML Parsing through Prefetching
Acceleration of XML  Parsing through PrefetchingAcceleration of XML  Parsing through Prefetching
Acceleration of XML Parsing through Prefetching
 
09 The Extreme-scale Scientific Software Stack for Collaborative Open Source
09 The Extreme-scale Scientific Software Stack for Collaborative Open Source09 The Extreme-scale Scientific Software Stack for Collaborative Open Source
09 The Extreme-scale Scientific Software Stack for Collaborative Open Source
 
HPC Cluster Computing from 64 to 156,000 Cores 
HPC Cluster Computing from 64 to 156,000 Cores HPC Cluster Computing from 64 to 156,000 Cores 
HPC Cluster Computing from 64 to 156,000 Cores 
 
Evaluating Machine Learning Algorithms for Materials Science using the Matben...
Evaluating Machine Learning Algorithms for Materials Science using the Matben...Evaluating Machine Learning Algorithms for Materials Science using the Matben...
Evaluating Machine Learning Algorithms for Materials Science using the Matben...
 
05 Preparing for Extreme Geterogeneity in HPC
05 Preparing for Extreme Geterogeneity in HPC05 Preparing for Extreme Geterogeneity in HPC
05 Preparing for Extreme Geterogeneity in HPC
 
How HPC and large-scale data analytics are transforming experimental science
How HPC and large-scale data analytics are transforming experimental scienceHow HPC and large-scale data analytics are transforming experimental science
How HPC and large-scale data analytics are transforming experimental science
 
Towards an Infrastructure for Enabling Systematic Development and Research of...
Towards an Infrastructure for Enabling Systematic Development and Research of...Towards an Infrastructure for Enabling Systematic Development and Research of...
Towards an Infrastructure for Enabling Systematic Development and Research of...
 

Belak_ICME_June02015

  • 1. This  work  was  performed  under  the  auspices  of  the  U.S.   Department  of  Energy  by  Lawrence  Livermore  National   Laboratory  under  Contract  DE-­‐AC52-­‐07NA27344  and   supported  by  Office  of  Science  Advanced  Scientific   Computing  Research  Program.   ICME  2015   Cheyenne  Mountain  Resort   Colorado  Springs,  Colorado   3  June,    02015   James  Belak  (belak@llnl.gov)   www.exmatex.org     *  co-­‐PI:  Tim  Germann  (tcg@lanl.gov)   LLNL-­‐PRES-­‐673227   'There  is  not  pure  science  and   applied  science  but  only   science  and  the  applications  of   science’                              -­‐  Louis  Pasteur     ‘People  who  are  serious  about   software  should  make  their   own  hardware.’    -­‐  Alan  Kay  (Xerox  PARC)   Initiate   Use-­‐inspired  research  and  development   involves  a  dynamic  interaction  between  both   considerations  of  use  and  fundamental   understanding.      
  • 2. Lawrence Livermore National Laboratory 3    June,  02015   §  It  begs  the  question:  what  are  the  computations  that  enable  ICME:   §  Databases  (integrated,  quality,  provenance,…)   §  Knowledge-­‐driven  organization  of  data  and  databases,  e.g.  enable   communication  between  materials  engineer  and  process  engineer,  do   what  google  does  to  organize  the  web   §  Expert  Systems:  Automated  Materials  Selection  from  Database   §  Inverse  problem  analysis  (e.g.  Bayesian,  material  from  property),  plus   other  techniques  from  Data  Analytics   §  Direct  Numerical  Simulation:  identify  the  controlling  phenomena  and  do   the  relevant  simulation,  meso-­‐scale,  molecular  dynamics,  DFT,  …     §  Better  physics-­‐based  continuum  models  (these  use  to  be  called   phenomenological,  but  nobody  understands  that  any  more)   2   Integrated  Engineering:  Integrating  “Computational  Materials”   into  Engineering  or  Integrating  “Computations”  into  Materials   Engineering.  Integrated  Engineering  is  more  than  ICME!  
  • 3. Lawrence Livermore National Laboratory 3    June,  02015   3   Component   Design   (Integrated  Code)   Component   Test   Material   Design   (Models)   Characterization,     Material  Test,   Virtual  Test   Process   Design   (Models)   Material   Specification   Fabrication   Process   Property   Specification   Process   Specification   Process   Qualification   Property   Qualification   Material   Qualification   Design  Specification   Component   Qualification   Component   Fabrication   Objective:  Agile  Design  and  Qualification  with  Minimal  Component-­‐scale  Test   S&E  Use-­‐Case:  Workflow  for  Materials  Development   http://extremescaleresearch.labworks.org/events/workshop-­‐future-­‐scientific-­‐ workflows#block-­‐system-­‐main   Advanced   Manufacturing  
  • 4. Lawrence Livermore National Laboratory 3    June,  02015   §  Our  goal  is  to  establish  the  interrelationship   between  hardware,  middleware  (software   stack),  programming  models,  and  algorithms   required  to  enable  a  productive  exascale   environment  for  multiphysics  simulations  of   materials  in  extreme  mechanical  and   radiation  environments.   §  We  will  exploit,  rather  than  avoid,  the  greatly   increased  levels  of  concurrency,   heterogeneity,  and  flop/byte  ratios  on  the   upcoming  exascale  platforms.     •  Our  vision  is  an  uncertainty  quantification  (UQ)-­‐driven  adaptive  physics   refinement  in  which  meso-­‐  and  macro-­‐scale  materials  simulations  spawn  micro-­‐ scale  simulations  as  needed.   -  This  task-­‐based  approach  leverages  the  extensive  concurrency  and   heterogeneity  expected  at  exascale  while  enabling  fault  tolerance  within   applications.     -  The  programming  models  and  approaches  developed  to  achieve  this  will   be  broadly  applicable  to  a  variety  of  multiscale,  multiphysics  applications,   including  astrophysics,  climate  and  weather  prediction,  structural   engineering,  plasma  physics,  and  radiation  hydrodynamics.   APP   SW  HW   Func+onal   Exascale   Simula+on   Environment   Domain  Workload   4  
  • 5. Lawrence Livermore National Laboratory 3    June,  02015   5   Commonly asked questions in exascale: What is the problem?! •  Power, energy, and heat dissipation are the central issues" •  Imagine a computer with billions and billions of cell phone processors (14MW) or millions and millions of throughput optimized cores, GPGPUs (20MW)" •  How do you program it to work on one science problem?" •  The architecture will be heterogeneous and hierarchical, with very high flop/byte ratios." •  Single program multiple data bulk synchronous parallelism will no longer be viable." •  Data Movement will be expensive and computation will be cheap" •  Need to present the physics so the computation occurs where the data is!" •  Traditional global checkpoint/restart will be impractical: need local / micro checkpoint (flash memory?)" •  Simulation codes will need to become fault tolerant and resilient! •  Recover from soft and hard errors, and anticipating faults" •  Ability to drop or replace nodes and keep on running" •  The curse of silent errors" 3D Stacked Memory (Low Capacity, High Bandwidth) Fat Core Fat Core Thin Cores / Accelerators DRAMNVRAM (High Capacity, Low Bandwidth) Coherence DomainCore Integrated NIC for Off-Chip Communication
  • 6. Lawrence Livermore National Laboratory 3    June,  02015   Preparation:   Science  and  Mission   Stakeholder  Buy-­‐in   Assemble  Team   Implementation   Plan   Development  Plan   Cycle  Artifacts:    R&D  Backlog    Algorithm  and    Model   Implementation    Proxy  Applications    Architecture   Evaluation   Co-­‐Design   Agile   Development   Cycle  Incorporated   Design   Elements   Algorithm   Development   Trade-­‐off   Analysis   Impact   Feedback   Code   Design   Exascale  Community:   Release  Artifacts:   HW  Requirements   SW  Constraints   Proxy  Applications   Documentation   Software  Development:   ASCR  X-­‐stack,  ASC  CSSE   Data/Analysis   Hardware   Development:  Vendors,   Fastforward,  ASCR   Advanced  Architecture   Code   Implementation   Release  to   Exascale   Community   Release  n   Domain   Science:   Domain  Workload   Physical  Models   Algorithms   Simulations   Team  Roles:   Cycle  Master:  Co-­‐design  PI    Project  Team:  Labs,  Univ’s    Stakeholders:  ASCR,  ASC,   Vendors    Customers:  Scientists,  HW+SW   Developers   6  
  • 7. Lawrence Livermore National Laboratory 3    June,  02015   System Software Proxy Apps Application Co-Design Hardware Co-Design Computer Science Co-Design Vendor Analysis Sim Exp Proto HW Prog Models HW Simulator Tools Open Analysis Models Simulators Emulators HW Design Stack Analysis Prog models Tools Compilers Runtime OS, I/O, ...HW Constraints Domain/Alg Analysis SW Solutions System Design! Application Design! Workflow of Co-design between Application Co-design Centers, hardware vendors, and the broader research community" X-Stack! Exascale OS/R! Fast Forward! Design Forward! Application Co- design Centers!
  • 8. Lawrence Livermore National Laboratory 3    June,  02015   DOE  Office  of  Science  ASCR  Co-­‐design  Projects:   Center  for  Exascale  Simulation  of  Advanced  Reactors  (CESAR)   http://cesar.mcs.anl.gov   Director:  Andrew  Siegel  (ANL/Uchicago),  Deputy  Director:  Paul  Fischer  (ANL)   Center  for  Exascale  Simulation  of  Combustion  in  Turbulance  (ExaCT)   http://exactcodesign.org   Director:  Jacqueline  Chen  (SNL),  Deputy  Director:  John  Bell  (LBNL)   Exascale  Co-­‐Design  Center  for  Materials  in  Extreme  Environments  (ExMatEX)   http://exmatex.lanl.gov   Director:  Tim  Germann  (LANL),  Deputy  Director:  Jim  Belak  (LLNL)       DOE  NNSA  ASC  Co-­‐design  Project:     National  Security  Applications  Co-­‐Design  Project  (NSApp  CDP)   Contacts:  Sriram  Swaminarayan(LANL),  Rob  Neely(LLNL),  Hoekstra/Heroux(SNL-­‐A)   http://proxyapps.lanl.gov   8  
  • 9. Lawrence Livermore National Laboratory 3    June,  02015   §  Materials  Science  applications  present   their  requirements  through  proxies   •  Real  applications  have  multiple   proxies   •  Proxies  have  docs,  specs,  and  a   reference  implementation   §  Ecosystem  members  evaluate  proxies   and  respond  with  capabilities   •  This  informs  trade-­‐off  analysis   §  Co-­‐Design  is  more  than  just   applications  and  hardware   architecture   •  There  is  a  whole  ecosystem  to   influence   9  
  • 10. Lawrence Livermore National Laboratory 3    June,  02015   Code:  Qbox/ LATTE     Motif:  Particles   and   wavefunctions,   plane  wave  DFT,   ScaLAPACK,   BLACS,  and   custom  parallel   3D  FFTs     Prog.  Model:  MPI   +  CUBLAS/CUDA   Code:SPaSM/ ddcMD/CoMD     Motif:  Particles,   explicit  time   integration,   neighbor  and   linked  lists,   dynamic  load   balancing,  parity   error  recovery,   and  in  situ   visualization     Prog.  Model:  MPI   +  Threads   Code:  SEAKMC       Motif:  Particles   and  defects,   explicit  time   integration,   neighbor  and   linked  lists,  and   in  situ   visualization     Prog.  Model:   MPI  +  Threads   Code:  AMPE/GL       Motif:  Regular  and   adaptive  grids,   implicit  time   integration,  real-­‐ space  and  spectral   methods,  complex   order  parameter     Prog.  Model:  MPI   Code:  ParaDiS       Motif:   “segments”   Regular  mesh,   implicit  time   integration,  fast   multipole   method     Prog.  Model:   MPI   Code:  VP-­‐FFT       Motif:  Regular   grids,  tensor   arithmetic,   meshless  image   processing,   implicit  time   integration,  3D   FFTs.     Prog.  Model:  MPI   +  Threads   Code:  ALE3D/ LULESH     Motif:  Regular   and  irregular   grids,  explicit  and   implicit  time   integration.     Prog.  Model:  MPI   +  Threads     Ab-­‐initio   Atoms   Long-­‐time   Microstructure   Dislocation   Crystal   Continuum   Inter-­‐atomic   forces,  EOS,   excited  states   Defects  and   interfaces,   nucleation   Defects  and   defect   structures   Meso-­‐scale  multi-­‐ phase,  multi-­‐grain   evolution   Meso-­‐scale   strength   Meso-­‐scale   material   response   Macro-­‐scale   material   response   10   Jim  Belak,  ExMatEx  
  • 11. Lawrence Livermore National Laboratory 3    June,  02015   ~  20  atoms  in  each  box   ⇒  each  atom  interacts  with  540  other  atoms   ⇒  However,  only  ~70  atoms  lie  within  cutoff   ⇒  Lots  of  wasted  work   ⇒  We  need  a  means  of  rejecting  atoms  efficiently   even  within  this  reduced  set     Halo  Region   11  
  • 12. Lawrence Livermore National Laboratory 3    June,  02015   12  
  • 13. Lawrence Livermore National Laboratory 3    June,  02015   §  Sam  Reeve  (Purdue)  LAMMPS,  CoMD,  leanMD   §  Riyaz  Haque  (UCLA),  CNC  (Habenero-­‐C)   §  Luc  Jualmes  (Barcelona)  OmpSs,  Chapel   §  Sameer  Abu  Asal  (LSU)  C++11  futures,  HPX   §  Aaron  Landmehr  (Delaware)  OCR   §  Sanian  Gaffer  (NM-­‐State)  UPC++   §  Gheorghe-­‐Teodor  Bercea  (Imperial)  PyOP2  DSL   §  Zach  Rubinstein  (Chicago),  Troels  Henriksen   (Copenhagen)  Embedded  DSL   Hybrid)Decomposi>on 7/23/14' Charm++':'grainsize' Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa> " )We)hav objects)to o )Each) assigne o )Numb (3D):)) o )14XNu ) Worker   Work  
  • 14. Lawrence Livermore National Laboratory 3    June,  02015   14   Hybrid)Decomposi>on) 7/23/14' Charm++':'grainsize' Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>al)Deco " )We)have)man objects)to)load) o )Each)diamon assigned)to)an o )Number)of)d (3D):)) o )14XNumber)o ) MD  “MPI”   Worker  Decomposition   Charm++   Work  Decomposition   c.f.  http://charm.cs.uiuc.edu/  
  • 15. Lawrence Livermore National Laboratory 3    June,  02015   Hybrid)Decomposi>on) 7/23/14' Charm++':'grainsize' 14' Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa>al)Decomp.) " )We)have)many) objects)to)load)balance:) o )Each)diamond)can)be) assigned)to)any)proc.) o )Number)of)diamonds) (3D):)) o )14XNumber)of)Cells) ) Di   Dj   Dk   I n i t i a t e   Di   Dj   Dk   Propagate   Atoms   Ti   Tj   Tk   Compute   Forces   Tij   Tjk   Di   Dj   Dk   Di   Dj   Dk   Migrate   Atoms   Ti   Loop  until  done   Ti   Tj   Tk   Tj   Tk  
  • 16. Lawrence Livermore National Laboratory 3    June,  02015   Cray  T3D,  Clock  Speed:  0.15GHz,  Date:  1994   Microprocessor  Peak  Teraflops:  0.02,  Memory:  0.01TB   Processors:  128.  Power  Consumption:  41.40  kW     SPMD,  MPI,   Geometric  domain   decomposition   Diffusion  of  a   small  drug   molecule   through  a  bio-­‐ membrane   Three  Essential  Elements:   •  Problem  Specification   •  Target  Architecture   •  Programming  Model  
  • 17. Lawrence Livermore National Laboratory 3    June,  02015   §  20  Years  Ago   •  Hardware:  Linux  Beowolf  Cluster   •  Software  Programming  Model:  MPI/SPMD   •  Application  Code:  LAMMPS   §  5+  Years  in  the  Future   •  Hardware:  Exascale  (Petascale  in  a  Rack)   •  Software  Programming  Model:  MPI+X,  Q?   Task-­‐based  with  “Control”  of  Data  Locality   (User  and  Runtime)   •  Application  Codes???     17   Worker   Hybrid)Decomposi>on 7/23/14' Charm++':'grainsize' Object)Based)Paralleliza>on)for)MD:)Force)Decomp.)+)Spa> " )We)hav objects)to o )Each) assigne o )Numb (3D):)) o )14XNu ) Work  
  • 18. Lawrence Livermore National Laboratory 3    June,  02015   Code:  Qbox/ LATTE     Motif:  Particles   and   wavefunctions,   plane  wave  DFT,   ScaLAPACK,   BLACS,  and   custom  parallel   3D  FFTs     Prog.  Model:  MPI   +  CUBLAS/CUDA   Code:SPaSM/ ddcMD/CoMD     Motif:  Particles,   explicit  time   integration,   neighbor  and   linked  lists,   dynamic  load   balancing,  parity   error  recovery,   and  in  situ   visualization     Prog.  Model:  MPI   +  Threads   Code:  SEAKMC       Motif:  Particles   and  defects,   explicit  time   integration,   neighbor  and   linked  lists,  and   in  situ   visualization     Prog.  Model:   MPI  +  Threads   Code:  AMPE/GL       Motif:  Regular  and   adaptive  grids,   implicit  time   integration,  real-­‐ space  and  spectral   methods,  complex   order  parameter     Prog.  Model:  MPI   Code:  ParaDiS       Motif:   “segments”   Regular  mesh,   implicit  time   integration,  fast   multipole   method     Prog.  Model:   MPI   Code:  VP-­‐FFT       Motif:  Regular   grids,  tensor   arithmetic,   meshless  image   processing,   implicit  time   integration,  3D   FFTs.     Prog.  Model:  MPI   +  Threads   Code:  ALE3D/ LULESH     Motif:  Regular   and  irregular   grids,  explicit  and   implicit  time   integration.     Prog.  Model:  MPI   +  Threads     Ab-­‐initio   Atoms   Long-­‐time   Microstructure   Dislocation   Crystal   Continuum   Inter-­‐atomic   forces,  EOS,   excited  states   Defects  and   interfaces,   nucleation   Defects  and   defect   structures   Meso-­‐scale  multi-­‐ phase,  multi-­‐grain   evolution   Meso-­‐scale   strength   Meso-­‐scale   material   response   Macro-­‐scale   material   response   18   Jim  Belak,  ExMatEx  
  • 19. Lawrence Livermore National Laboratory 3    June,  02015   Simulations suggest novel in situ x-ray scattering experiments using emerging sources such as LCLS! 19  
  • 20. Lawrence Livermore National Laboratory 3    June,  02015   100 nm Atoms Phase time! Coarse Graining   20  
  • 21. Lawrence Livermore National Laboratory 3    June,  02015   •  Each color represents a different value of the phase field (solid orientation)! •  Free energy describes how colors interact and evolve! •  Accuracy depends on fidelity of physics in the equations! F(P,T) = dx∫ ∇ ! φ 2 + f ( ! φ, P,T)+…}{ ∂ ! φ ∂t = −Γ δF δ ! φ + noise Evolution Equations! Thermodynamic representation of phase (or “color”) everywhere! ! φ ! φ(r,t) = pink ! φ(r,t) = green ! φ(r,t) = solid ! φ(r,t) = liquid 21  
  • 22. Lawrence Livermore National Laboratory 3    June,  02015   F = εϕ 2 2 ∇ϕ 2 + f (ϕ,c,T)+ HT[1− p(ϕ)] (∇qi )2 i ∑ $ % & ' ( ) 1/2* + , , - . / / ∫ d3 r ∂qi ∂t = −Mq δF δqi +ζi = Mq ∇⋅ D ∇qi ∇qi % & '' ( ) **− 2λqi + , - - . / 0 0 +ζi Pusztai et al., have proposed a 3D quaternion-based phase-field model ! • Represents crystal orientation with quaternion order parameter" • Quaternions are widely used to analyze crystallography of polycrystal interfaces" • Quaternion algebra is fast, efficient, avoids singularities, …" Where qi is the quaternion order parameter, Mq is the associated mobility and ζ is the fluctuation in q." Free Energy Evolution Refs: T. Pusztai, G. Bortel, and L. Granasy, “Phase field theory of polycrystalline solidification in three dimensions,” Europhys. Lett, 71 (2005) 131-137; Dorr, M.R., Fattebert, J.-L., Wickett, M.E., Belak, J.F., and Turchi, P.E.A., “A Numerical Algorithm for the Solution of a Phase-field Model of Polycrystalline Materials,” J. Comp. Phys. Vol. 229, 626 (2010). " We have implemented the Pusztai model in our 3D AMR code ! • Enhance energy functional to represent energetics of grain boundaries" • Crystal symmetry aware quaternion mathematics" • Extend energy functional to include elasticity and alloy concentration"
  • 23. Lawrence Livermore National Laboratory 3    June,  02015   23  
  • 24. Lawrence Livermore National Laboratory 3    June,  02015   Growth of large grains Blue: MD nucleation Red: Phase-field evolution MD  nucleated  microstructure  onto  the  micro-­‐second   hydro  time-­‐scale  with  the  crystallographic  quaternion   model   While significant grain coarsening has occurred on the microsecond scale, the microstructure is far from log-normal Phase Order Parameter! Quaternion Order Parameter! 24  
  • 25. Lawrence Livermore National Laboratory 3    June,  02015   Code:  Qbox/ LATTE     Motif:  Particles   and   wavefunctions,   plane  wave  DFT,   ScaLAPACK,   BLACS,  and   custom  parallel   3D  FFTs     Prog.  Model:  MPI   +  CUBLAS/CUDA   Code:SPaSM/ ddcMD/CoMD     Motif:  Particles,   explicit  time   integration,   neighbor  and   linked  lists,   dynamic  load   balancing,  parity   error  recovery,   and  in  situ   visualization     Prog.  Model:  MPI   +  Threads   Code:  SEAKMC       Motif:  Particles   and  defects,   explicit  time   integration,   neighbor  and   linked  lists,  and   in  situ   visualization     Prog.  Model:   MPI  +  Threads   Code:  AMPE/GL       Motif:  Regular  and   adaptive  grids,   implicit  time   integration,  real-­‐ space  and  spectral   methods,  complex   order  parameter     Prog.  Model:  MPI   Code:  ParaDiS       Motif:   “segments”   Regular  mesh,   implicit  time   integration,  fast   multipole   method     Prog.  Model:   MPI   Code:  VP-­‐FFT       Motif:  Regular   grids,  tensor   arithmetic,   meshless  image   processing,   implicit  time   integration,  3D   FFTs.     Prog.  Model:  MPI   +  Threads   Code:  ALE3D/ LULESH     Motif:  Regular   and  irregular   grids,  explicit  and   implicit  time   integration.     Prog.  Model:  MPI   +  Threads     Ab-­‐initio   Atoms   Long-­‐time   Microstructure   Dislocation   Crystal   Continuum   Inter-­‐atomic   forces,  EOS,   excited  states   Defects  and   interfaces,   nucleation   Defects  and   defect   structures   Meso-­‐scale  multi-­‐ phase,  multi-­‐grain   evolution   Meso-­‐scale   strength   Meso-­‐scale   material   response   Macro-­‐scale   material   response   25  
  • 26. Lawrence Livermore National Laboratory 3    June,  02015   §  To  achieve  this,  we  are   developing  a  UQ-­‐driven   adaptive  physics  refinement   approach.   §  Coarse-­‐scale  simulations   dynamically  spawn  tightly   coupled  and  self-­‐consistent   fine-­‐scale  simulations  as   needed.   §  This  task-­‐based  approach   naturally  maps  to  exascale   heterogeneity,  concurrency,   and  resiliency  issues.     •  Task-­‐based  embedded  Scale-­‐Bridging  escapes  the  traditional  synchronous   SPMD  paradigm  and  exploits  the  heterogeneity  expected  in  exascale  hardware.       26  
  • 27. Lawrence Livermore National Laboratory 3    June,  02015   27   FSMs   •  Brute  force  multi-­‐scale  coupling:  Full  fine  scale   model  (FSM,  e.g.  a  crystal  plasticity  model)   run  for  every  zone  &  time  step  of  coarse  scale   mode  (CSM,  e.g.  an  ALE  code)   •  Adaptive  Sampling:     –  Save  FSM  results  in  database   –  Before  running  another  FSM,  check  database   for  FSM  results  similar  enough  to  those   needed  that  interpolation  or  extrapolation   suffices   –  Only  run  full  FSM  when  results  in  database  not   close  enough   CSM   Ref:  Barton  et.al,  ‘A  call  to  arms  for  task  parallelism  in  multi-­‐scale  materials  modeling,’  Int.  J.  Numer.  Meth.  Engng  2011;  86:744–764   §  Heterogeneous,  hierarchical  MPMD  algorithms  map  naturally  to  anticipated   heterogeneous,  hierarchical  architectures   §  Escape  the  traditional  bulk  synchronous  SPMD  paradigm,  improve  scalability  and   reduce  scheduling   §  Task-­‐based  MPMD  approach  leverages  concurrency  and  heterogeneity  at  exascale   while  enabling  novel  data  models,  power  management,  and  fault  tolerance   strategies  
  • 28. Lawrence Livermore National Laboratory 3    June,  02015   AS  Database   Regions  over   which  models   may  extrapolate   Past  fine-­‐scale  evaulation   results;  approximation  models   Queried  points   Queried  point  close   enough  for   approximation   Fine-­‐scale  evaluation   at  this  query   Input  Space   §  Coarse  scale  model   queries  database  for  fine-­‐ scale  material  response   §  If  possible,  approximate   response  from  past   evaluations   §  Otherwise  perform  fine   scale  evaluation   §  Fine-­‐scale     evaluations  grow     database   28  
  • 29. Lawrence Livermore National Laboratory 3    June,  02015   On-demand fine scale models" CSM" DB$" Adaptive" Sampler" FSM" Subdomain 1" Subdomain 2" FSM" FSM" Node 1! DB$" Adaptive" Sampler" Subdomain N-1" Subdomain N" Node N/2!
  • 30. Lawrence Livermore National Laboratory 3    June,  02015   DB" On-demand fine scale models" CSM" DB$" Eventually consistent distributed database" Adaptive" Sampler" FSM" Subdomain 1" Subdomain 2" FSM" FSM" Node 1! DB" DB" DB$" Adaptive" Sampler" Subdomain N-1" Subdomain N" Node N/2!
  • 31. Lawrence Livermore National Laboratory 3    June,  02015   §  Our  goal  is  to  establish  the   interrelationship  between  hardware,   middleware  (software  stack),   programming  models,  and  algorithms   required  to  enable  a  productive  exascale   environment  for  multiphysics  simulations   of  materials  in  extreme  mechanical  and   radiation  environments.   §  We  will  exploit,  rather  than  avoid,  the   greatly  increased  levels  of  concurrency,   heterogeneity,  and  flop/byte  ratios  on  the   upcoming  exascale  platforms.     •  Our  vision  is  an  uncertainty  quantification  (UQ)-­‐driven  adaptive  physics   refinement  in  which  meso-­‐  and  macro-­‐scale  materials  simulations  spawn   micro-­‐scale  simulations  as  needed.   -  This  task-­‐based  approach  leverages  the  extensive  concurrency  and   heterogeneity  expected  at  exascale  while  enabling  fault  tolerance  within   applications.     -  The  programming  models  and  approaches  developed  to  achieve  this  will  be   broadly  applicable  to  a  variety  of  multiscale,  multiphysics  applications,  including   astrophysics,  climate  and  weather  prediction,  structural  engineering,  plasma   physics,  and  radiation  hydrodynamics.   APP   SW  HW   Func+onal   Exascale   Simula+on   Environment   Domain  Workload   31  
  • 32. Lawrence Livermore National Laboratory 3    June,  02015   32