SlideShare a Scribd company logo
1 of 45
―Set My Data Free: High-Performance CI
              for Data-Intensive Research‖

                             KeynoteSpeaker
                          Cyberinfrastructure Days
                           University of Michigan
                               Ann Arbor, MI
                             November 3, 2010

                                    Dr. Larry Smarr
Director, California Institute for Telecommunications and Information Technology
     Harry E. Gruber Professor, Dept. of Computer Science and Engineering
                        Jacobs School of Engineering, UCSD
                             Follow me on Twitter: lsmarr
Abstract


As the need for large datasets and high-volume transfer grows,
the shared Internet is becoming a bottleneck for cutting-edge
research in universities. What are needed instead are large-
bandwidth "data freeways." In this talk, I will describe some of the
state-of-the-art uses of high-performance CI and how universities
can evolve to support free movement of large datasets.
The Data-Intensive Discovery Era Requires
          High Performance Cyberinfrastructure
• Growth of Digital Data is Exponential
  – ―Data Tsunami‖
• Driven by Advances in Digital Detectors, Computing,
  Networking, & Storage Technologies
• Shared Internet Optimized for Megabyte-Size Objects
• Need Dedicated Photonic Cyberinfrastructure for
  Gigabyte/Terabyte Data Objects
• Finding Patterns in the Data is the New Imperative
  –   Data-Driven Applications
  –   Data Mining
  –   Visual Analytics
  –   Data Analysis Workflows

                           Source: SDSC
Large Data Challenge: Average Throughput to End User
          on Shared Internet is 10-100 Mbps
                                                                        Tested
                                                                     October 2010




                                                           Transferring 1 TB:
                                                           --10 Mbps = 10 Days
                                                           --10 Gbps = 15 Minutes




           http://ensight.eos.nasa.gov/Missions/icesat/index.shtml
The Large Hadron Collider
Uses a Global Fiber Infrastructure To Connect Its Users




• The grid relies on optical fiber networks to distribute data from
  CERN to 11 major computer centers in Europe, North America,
  and Asia
• The grid is capable of routinely processing 250,000 jobs a day
• The data flow will be ~6 Gigabits/sec or 15 million gigabytes a
  year for 10 to 15 years
Next Great Planetary Instrument:
The Square Kilometer Array Requires Dedicated Fiber

                                www.skatelescope.org




          Transfers Of
        1 TByte Images
          World-wide
        Will Be Needed
         Every Minute!   Currently Competing Between
                         Australia and S. Africa
GRAND CHALLENGES IN
              DATA-INTENSIVE SCIENCES
                    OCTOBER 26-28, 2010
        SAN DIEGO SUPERCOMPUTER CENTER , UC SAN DIEGO

Confirmed conference topics and speakers :
  Needs and Opportunities in Observational Astronomy - Alex Szalay, JHU
  Transient Sky Surveys – Peter Nugent, LBNL
  Large Data-Intensive Graph Problems – John Gilbert, UCSB
  Algorithms for Massive Data Sets – Michael Mahoney, Stanford U.
  Needs and Opportunities in Seismic Modeling and Earthquake Preparedness -
  Tom Jordan, USC
  Needs and Opportunities in Fluid Dynamics Modeling and Flow Field Data
  Analysis – Parviz Moin, Stanford U.
  Needs and Emerging Opportunities in Neuroscience – Mark Ellisman, UCSD

  Data-Driven Science in the Globally Networked World – Larry Smarr, UCSD


              Petascale High Performance Computing
                 Generates TB Datasets to Analyze
Growth of Turbulence Data Over Three Decades
(Assuming Double Precision and Collocated Points)
 Year       Authors                Simulation               Points       Size
 1972 Orszag & Patterson       Isotropic Turbulence          323         1 MB
 1987   Kim, Moin & Moser      Plane Channel Flow        192x160x128    120 MB
 1988        Spalart        Turbulent Boundary Layer      432x80x320    340 MB
 1994      Le & Moin          Backward-Facing Step        768x64x192    288 MB
 2000    Freund, Lele &      Compressible Turbulent      640x270x128    845 MB
             Moin                    Jet
 2003    Earth Simulator       Isotropic Turbulence         40963       0.8 TB*
 2006   Hoyas & Jiménez        Plane Channel Flow        6144x633x460   550 GB
                                                              8
 2008      Wu & Moin           Turbulent Pipe Flow         256x5122     2.1 GB
 2009    Larsson & Lele     Isotropic Shock-Turbulence    1080x3842     6.1 GB
 2010      Wu & Moin        Turbulent Boundary Layer     8192x500x256   40 GB

  Turbulent Boundary Layer:
  One-Periodic Direction
  100x Larger Data Sets in 20 Years
             Source: Parviz Moin, Stanford
CyberShake 1.0 Hazard Model
     Need to Analyze Terabytes of Computed Data
• CyberShake 1.0 Computation         Source: Thomas H. Jordan, USC,
                                     Director, Southern California Earthquake Center

 - 440,000 Simulations per Site
 - 5.5 Million CPU hrs (50-Day Run
   on Ranger Using 4,400 cores)
                                                            CyberShake seismogram
 - 189 Million Jobs
 - 165 TB of Total Output Data
                                         CyberShake Hazard Map
 - 10.6 TB of Stored Data                     PoE = 2% in 50 yrs

 - 2.1 TB of Archived Data




                                 LA region
Large-Scale PetaApps Climate Change Run
      Generates Terabyte Per Day of Computed Data
• 155 Year Control Run                                            100x
  –   0.1° Ocean model [ 3600 x 2400 x 42 ]                     Current
  –   0.1° Sea-ice model [3600 x 2400 x 20 ]                   Production
  –   0.5° Atmosphere [576 x 384 x 26 ]
  –   0.5° Land [576 x 384]
• Statistics
  –   ~18M CPU Hours               4x current production
  –   5844 Cores for 4-5 Months
  –   ~100 TB of Data Generated
  –   0.5 to 1 TB per Wall Clock Day Generated




                                                                    10
             Source: John M. Dennis, Matthew Woitaszek, UCAR
The Required Components of
    High Performance Cyberinfrastructure




•   High Performance Optical Networks
•   Scalable Visualization and Analysis
•   Multi-Site Collaborative Systems
•   End-to-End Wide Area CI
•   Data-Intensive Campus Research CI
Australia—The Broadband Nation:
   Universal Coverage with Fiber, Wireless, Satellite
• Connect 93% of All Australian Premises with Fiber
  – 100 Mbps to Start, Upgrading to Gigabit
• 7% with Next Gen Wireless and Satellite
  – 12 Mbps to Start
• Provide Equal Wholesale Access to Retailers
  – Providing Advanced Digital Services to the Nation
  – Driven by Consumer Internet, Telephone, Video
     – ―Triple Play‖, eHealth, eCommerce…
   ―NBN is Australia’s largest nation building project
                    in our history.‖
              - Minister Stephen Conroy



                               www.nbnco.com.au
Globally Fiber to the Premise is Growing Rapidly,
                    Mostly in Asia
                                                            If Couch Potatoes
                                                                  Deserve
                                                              a Gigabit Fiber,
   FTTP                                                           Why Not
Connections                                                      University
 Growing at                                                    Data-Intensive
 ~30%/year                                                     Researchers?




130 Million
Households
 with FTTH
  in 2013

         Source: Heavy Reading (www.heavyreading.com), the market
          research division of Light Reading (www.lightreading.com).
The Global Lambda Integrated Facility--
Creating a Planetary-Scale High Bandwidth Collaboratory

   Research Innovation Labs Linked by 10G GLIF




                                www.glif.is
                                Created in Reykjavik,
                                Iceland 2003

                     Visualization courtesy of
                     Bob Patterson, NCSA.
The OptIPuter Project: Creating High Resolution Portals
Over Dedicated Optical Channels to Global Science Data
                                                                             Scalable
                                                                             Adaptive
                                                                             Graphics
                                                                             Environment
                                                                             (SAGE)




                                                                               Picture
                                                                               Source:
                                                                               Mark
                                                                               Ellisman,
                                                                               David Lee,
                                                                               Jason Leigh
      Calit2 (UCSD, UCI), SDSC, and UIC Leads—Larry Smarr PI
      Univ. Partners: NCSA, USC, SDSU, NW, TA&M, UvA, SARA, KISTI, AIST
      Industry: IBM, Sun, Telcordia, Chiaro, Calient, Glimmerglass, Lucent
Nearly Seamless AESOP OptIPortal

46‖ NEC Ultra-Narrow Bezel 720p LCD Monitors




         Source: Tom DeFanti, Calit2@UCSD;
3D Stereo Head Tracked OptIPortal:
            NexCAVE




  Array of JVC HDTV 3D LCD Screens
    KAUST NexCAVE = 22.5MPixels
 www.calit2.net/newsroom/article.php?id=1584
       Source: Tom DeFanti, Calit2@UCSD
High Definition Video Connected OptIPortals:
Virtual Working Spaces for Data Intensive Research



                                                              NASA Supports
                                                                Two Virtual
                                                                 Institutes




                                                               LifeSize HD

          Calit2@UCSD 10Gbps Link to
NASA Ames Lunar Science Institute, Mountain View, CA

           Source: Falko Kuester, Kai Doerr Calit2;
           Michael Sims, Larry Edwards, Estelle Dodson NASA
U Michigan Virtual Space Interaction Testbed (VISIT)
 Instrumenting OptIPortals for Social Science Research
• Using Cameras Embedded in
  the Seams of Tiled Displays
  and Computer Vision
  Techniques, we can
  Understand how People
  Interact with OptIPortals
   – Classify Attention, Expression,
     Gaze
   – Initial Implementation Based on
     Attention Interaction Design
     Toolkit (J. Lee, MIT)
• Close to Producing Usable
  Eye/Nose Tracking Data using
  OpenCV
                      Leading U.S.
                   Researchers on the
                    Social Aspects of
                     Collaboration


                 Source: Erik Hofer, UMich, School of Information
EVL’s SAGE OptIPortal VisualCasting
              Multi-Site OptIPuter Collaboratory
                                   CENIC CalREN-XD Workshop Sept. 15, 2008
           Total Aggregate VisualCasting Bandwidth for Nov. 18, 2008
EVL-UI Chicago
           Sustained 10,000-20,000 Mbps!
             At Supercomputing 2008 Austin, Texas
             November, 2008                                        Streaming 4k

             SC08 Bandwidth Challenge Entry

                                       Remote:
 On site:
                                       U of Michigan
 SARA (Amsterdam)                      UIC/EVL
U Michigan
 GIST / KISTI (Korea)                  U of Queensland
 Osaka Univ. (Japan)                   Russian Academy of Science
                                       Masaryk Univ. (CZ)
              Requires 10 Gbps Lightpath to Each Site



              Source: Jason Leigh, Luc Renambot, EVL, UI Chicago
Exploring Cosmology With Supercomputers,
       Supernetworks, and Supervisualization
                       Source: Mike Norman, SDSC

Intergalactic Medium on 2 GLyr Scale
                                              • 40963 Particle/Cell
                                                Hydrodynamic
                                                Cosmology
                                                Simulation
                                              • NICS Kraken (XT5)
                                                  – 16,384 cores
                                              • Output
                                                  – 148 TB Movie Output
                                                    (0.25 TB/file)
                                                  – 80 TB Diagnostic
Science: Norman, Harkness,Paschos SDSC
Visualization: Insley, ANL; Wagner SDSC             Dumps (8 TB/file)


               •   ANL * Calit2 * LBNL * NICS * ORNL * SDSC
Project StarGate Goals:
    Combining Supercomputers and Supernetworks
• Create an ―End-to-End‖
  10Gbps Workflow
• Explore Use of OptIPortals as
                                                  OptIPortal@SDSC
  Petascale Supercomputer
  ―Scalable Workstations‖
• Exploit Dynamic 10Gbps
  Circuits on ESnet
• Connect Hardware Resources
  at ORNL, ANL, SDSC
• Show that Data Need Not be
  Trapped by the Network                      Rick Wagner     Mike Norman

  ―Event Horizon‖
                   Source: Michael Norman, SDSC, UCSD
               •   ANL * Calit2 * LBNL * NICS * ORNL * SDSC
Using Supernetworks to Couple End User’s OptIPortal
 to Remote Supercomputers and Visualization Servers
Source: Mike Norman,
Rick Wagner, SDSC                                Argonne NL
                                                               DOE Eureka
                                          100 Dual Quad Core Xeon Servers
                                          200 NVIDIA Quadro FX GPUs in 50
                                              Quadro Plex S4 1U enclosures
                                                               3.2 TB RAM        rendering


                                                   ESnet
 SDSC                                              10 Gb/s fiber optic network
                                                                                     NICS
               visualization
                                                                                     ORNL
 Calit2/SDSC OptIPortal1
 20 30‖ (2560 x 1600 pixel) LCD panels
                                           NSF TeraGrid Kraken   simulation
                                                      Cray XT5
 10 NVIDIA Quadro FX 4600 graphics
                                          8,256 Compute Nodes
 cards > 80 megapixels
                                         99,072 Compute Cores
 10 Gb/s network throughout
                                                   129 TB RAM




                       *ANL * Calit2 * LBNL * NICS * ORNL * SDSC
National-Scale Interactive Remote Rendering
     of Large Datasets Over 10Gbps Fiber Network
        SDSC                ESnet            ALCF
                                 Science Data Network (SDN)
                                 > 10 Gb/s Fiber Optic Network
                                 Dynamic VLANs Configured
                                 Using OSCARS
                                                                 Rendering
     Visualization                                                         Eureka
OptIPortal (40M pixels LCDs)                     100 Dual Quad Core Xeon Servers
10 NVIDIA FX 4600 Cards                                      200 NVIDIA FX GPUs
10 Gb/s Network Throughout                                            3.2 TB RAM
                               Interactive Remote Rendering
  Real-Time Volume Rendering Streamed from ANL to SDSC
        Last Year                                Last Week
High-Resolution (4K+, 15+ FPS)—But:      Now Driven by a Simple Web GUI
• Command-Line Driven                    •Rotate, Pan, Zoom
• Fixed Color Maps, Transfer Functions   •GUI Works from Most Browsers
• Slow Exploration of Data               • Manipulate Colors and Opacity
                                         • Fast Renderer Response Time

                          Source: Rick Wagner, SDSC
NSF’s Ocean Observatory Initiative
            Has the Largest Funded NSF CI Grant




      OOI CI Grant:
30-40 Software Engineers
Housed at Calit2@UCSD




            Source: Matthew Arrott, Calit2 Program Manager for OOI CI
OOI CI is Built on Dedicated
                                 OOI CI
                  Optical Infrastructure Using Clouds
                   Physical Network Implementation
  Source: John Orcutt,
Matthew Arrott, SIO/Calit2
California and Washington Universities Are Testing
     a 10Gbps Connected Commercial Data Cloud
• Amazon Experiment for Big Data
  – Only Available Through CENIC & Pacific NW
    GigaPOP
    – Private 10Gbps Peering Paths
  – Includes Amazon EC2 Computing
    & S3 Storage Services
• Early Experiments Underway
  – Robert Grossman, Open Cloud Consortium
  – Phil Papadopoulos, Calit2/SDSC Rocks
Open Cloud OptIPuter Testbed--Manage and Compute
            Large Datasets Over 10Gbps Lambdas




              CENIC         NLR C-Wave           Dragon


                                                 Open Source SW
                                                  Hadoop
•   9 Racks                        MREN           Sector/Sphere
•   500 Nodes                                     Nebula
•   1000+ Cores                                   Thrift, GPB
•   10+ Gb/s Now                                  Eucalyptus
•   Upgrading Portions to                         Benchmarks
    100 Gb/s in 2010/2011

                                                            28
                  Source: Robert Grossman, UChicago
Terasort on Open Cloud Testbed
Sustains >5 Gbps--Only 5% Distance Penalty!

    Sorting 10 Billion Records (1.2 TB)
           at 4 Sites (120 Nodes)




        Source: Robert Grossman, UChicago
Hybrid Cloud Computing
                   with modENCODE Data
• Computations in Bionimbus Can Span the Community Cloud
  & the Amazon Public Cloud to Form a Hybrid Cloud
• Sector was used to Support the Data Transfer between
  Two Virtual Machines
    – One VM was at UIC and One VM was an Amazon EC2 Instance
• Graph Illustrates How the Throughput between Two Virtual
  Machines in a Wide Area Cloud Depends upon the File Size




Biological data
 (Bionimbus)      Source: Robert Grossman, UChicago
Ocean Modeling HPC In the Cloud:
           Tropical Pacific SST (2 Month Ave 2002)
   MIT GCM 1/3 Degree Horizontal Resolution, 51 Levels, Forced by NCEP2.
     Grid is 564x168x51, Model State is T,S,U,V,W and Sea Surface Height




        Run on EC2 HPC Instance. In Collaboration with OOI CI/Calit2

Source: B. Cornuelle, N. Martinez, C.Papadopoulos COMPAS, SIO
Using Condor and Amazon EC2 on
     Adaptive Poisson-Boltzmann Solver (APBS)
• APBS Rocks Roll (NBCR) + EC2 Roll
  + Condor Roll = Amazon VM
• Cluster extension into Amazon using Condor

                         Local
 Running in Amazon Cloud
                         Cluster       EC2 Cloud
                                       NBCR          NBCR
                                        VM            VM

                                              NBCR
                                               VM




  APBS + EC2 + Condor


          Source: Phil Papadopoulos,
                 SDSC/Calit2
―Blueprint for the Digital University‖--Report of the
   UCSD Research Cyberinfrastructure Design Team
• Focus on Data-Intensive Cyberinfrastructure
                                                                    April 2009

No Data
Bottlenecks
--Design for
Gigabit/s
Data Flows




      http://research.ucsd.edu/documents/rcidt/RCIDTReportFinal2009.pdf
What do Campuses Need to Build to Utilize
                CENIC’s Three Layer Network?




 ~ $14M
Invested
    in
Upgrade




             Now
           Campuses
            Need to
           Upgrade!
                      Source: Jim Dolgonas, CENIC
Current UCSD Optical Core:
           Bridging End-Users to CENIC L1, L2, L3 Services
                                      Quartzite Communications
  To 10GigE cluster
   node interfaces
                                             Core Year 3
                                   Enpoints:
                                            Quartzite    Wavelength
                                   >= 60 endpoints at 10 GigE
                                              Core
                                                          Selective
       .....


                                                           Switch

                                   >= 32 Packet switched Lucent                         To 10GigE cluster
                                                                                       node interfaces and
                                   >= 32 Switched wavelengths                               other switches


To cluster nodes
                   .....
                                   >= 300 Connected endpoints
                                                                 Glimmerglass
                                                                                                  To cluster nodes
                                                                                          .....
                                                                  Production
            GigE Switch with
                                                                    OOO
           Dual 10GigE Upliks
                                                                    Switch
To cluster nodes
                                    Approximately 0.5 TBit/s
                                               32 10GigE

                   .....
                                    Arrive at the ―Optical‖                        GigE Switch with
                                                                     Force10      Dual 10GigE Upliks

                                    Center of Campus.
                                                ...

            GigE Switch with
                                    Switching is a Hybrid of:
                                       To              Packet Switch            CalREN-HPR
                                                                                 Research
           Dual 10GigE Upliks
                                    Packet, Lambda, Circuit --
                                       other
                                       nodes
                                                                                   Cloud
    GigE
                                    OOO and Packet Switches
 10GigE
                                                                                Campus Research
  4 GigE
  4 pair fiber
                                                                                    Cloud
                                                       Juniper T320

                                Source: Phil Papadopoulos, SDSC/Calit2
                                (Quartzite PI, OptIPuter co-PI)
                                Quartzite Network MRI #CNS-0421555;
                                OptIPuter #ANI-0225642
UCSD Campus Investment in Fiber Enables
Consolidation of Energy Efficient Computing & Storage
                                        WAN 10Gb:
           N x 10Gb                    CENIC, NLR, I2




                       Gordon –
                      HPD System

 Cluster Condo

                                                       DataOasis
                                                    (Central) Storage
                  Triton – Petascale
                    Data Analysis
    Scientific
  Instruments


                       Digital Data       Campus Lab            OptIPortal
                       Collections          Cluster         Tile Display Wall


                 Source: Philip Papadopoulos, SDSC/Calit2
The GreenLight Project:
Instrumenting the Energy Cost of Computational Science
• Focus on 5 Communities with At-Scale Computing Needs:
   –   Metagenomics
   –   Ocean Observing
   –   Microscopy
   –   Bioinformatics
   –   Digital Media
• Measure, Monitor, & Web Publish
  Real-Time Sensor Outputs
   – Via Service-oriented Architectures
   – Allow Researchers Anywhere To Study Computing Energy Cost
   – Enable Scientists To Explore Tactics For Maximizing Work/Watt
• Develop Middleware that Automates Optimal Choice
  of Compute/RAM Power Strategies for Desired Greenness
• Partnering With Minority-Serving Institutions
  Cyberinfrastructure Empowerment Coalition

                    Source: Tom DeFanti, Calit2; GreenLight PI
UCSD Biomed Centers
                 Drive High Performance CI




                    National Resource for Network Biology
iDASH: Integrating Data for Analysis, Anonymization, and Sharing
Calit2 Microbial Metagenomics Cluster-
 Next Generation Optically Linked Science Data Server
                               Source: Phil Papadopoulos, SDSC, Calit2




       512 Processors
                                                 ~200TB
         ~5 Teraflops                              Sun
                                  1GbE            X4500
   ~ 200 Terabytes Storage         and           Storage
                                  10GbE
                                Switched         10GbE
                                / Routed
       Several Large Users         Core
        at Univ. Michigan


   4000 Users
From 90 Countries
Calit2 CAMERA Automatic Overflows
           into SDSC Triton



                                    @ SDSC

                                     Triton Resource

@ CALIT2
                    Transparently            CAMERA -
                    Sends Jobs to             Managed
                    Submit Portal            Job Submit
                      on Triton              Portal (VM)
                                       10Gbps
                                                     Direct
                                                     Mount
                                    CAMERA             ==
                                     DATA           No Data
                                                    Staging
Rapid Evolution of 10GbE Port Prices
   Makes Campus-Scale 10Gbps CI Affordable
    • Port Pricing is Falling
    • Density is Rising – Dramatically
    • Cost of 10GbE Approaching Cluster HPC Interconnects
$80K/port
Chiaro
(60 Max)



                 $ 5K
                 Force 10
                 (40 max)                                     ~$1000
                                                              (300+ Max)

                                      $ 500
                                      Arista                  $ 400
                                      48 ports                Arista
                                                              48 ports
2005               2007                2009            2010




            Source: Philip Papadopoulos, SDSC/Calit2
10G Switched Data Analysis Resource:
                    SDSC’s Data Oasis

                OptIPuter                    RCN
                               Colo

                                             CalRe
                   32                        n
 Triton
                              20
                                                24
           32
                                         2
Trestles            12                                           Existing
                                               40
                                                                 Storage
                                                Oasis Procurement (RFP)

 Dash
                   8               • Phase0: > 8GB/s sustained, today             1500 –
                                   • RFP for Phase1: > 40 GB/sec for Lustre      2000 TB
                                                                                   > 40
                                   • Nodes must be able to function as Lustre      GB/s
                                   OSS (Linux) or NFS (Solaris)
                        100        • Connectivity to Network is 2 x 10GbE/Node
Gordon                             • Likely Reserve dollars for inexpensive
                                   replica servers
                         Source: Philip Papadopoulos, SDSC/Calit2
NSF Funds a Data-Intensive Track 2 Supercomputer:
        SDSC’s Gordon-Coming Summer 2011
• Data-Intensive Supercomputer Based on
  SSD Flash Memory and Virtual Shared Memory SW
  – Emphasizes MEM and IOPS over FLOPS
  – Supernode has Virtual Shared Memory:
     – 2 TB RAM Aggregate
     – 8 TB SSD Aggregate
     – Total Machine = 32 Supernodes
     – 4 PB Disk Parallel File System >100 GB/s I/O
• System Designed to Accelerate Access to Massive
  Data Bases being Generated in all Fields of
  Science, Engineering, Medicine, and Social
  Science

            Source: Mike Norman, Allan Snavely SDSC
Academic Research ―OptIPlatform‖ Cyberinfrastructure:
      A 10Gbps ―End-to-End‖ Lightpath Cloud



                                                                              HD/4k Video Cams
                                               HD/4k Telepresence
                                Instruments
           End User                                                                   HPC
           OptIPortal

                   10G
                   Lightpaths




                                  National LambdaRail



Campus
Optical Switch



                   Data Repositories & Clusters          HD/4k Video Images
You Can Download This Presentation
        at lsmarr.calit2.net

More Related Content

What's hot

How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...
How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...
How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...Larry Smarr
 
Supercomputer End Users: the OptIPuter Killer Application
Supercomputer End Users: the OptIPuter Killer ApplicationSupercomputer End Users: the OptIPuter Killer Application
Supercomputer End Users: the OptIPuter Killer ApplicationLarry Smarr
 
Blowing up the Box--the Emergence of the Planetary Computer
Blowing up the Box--the Emergence of the Planetary ComputerBlowing up the Box--the Emergence of the Planetary Computer
Blowing up the Box--the Emergence of the Planetary ComputerLarry Smarr
 
The Pacific Research Platform:a Science-Driven Big-Data Freeway System
The Pacific Research Platform:a Science-Driven Big-Data Freeway SystemThe Pacific Research Platform:a Science-Driven Big-Data Freeway System
The Pacific Research Platform:a Science-Driven Big-Data Freeway SystemLarry Smarr
 
Report to the NAC
Report to the NACReport to the NAC
Report to the NACLarry Smarr
 
Toward a Global Interactive Earth Observing Cyberinfrastructure
Toward a Global Interactive Earth Observing CyberinfrastructureToward a Global Interactive Earth Observing Cyberinfrastructure
Toward a Global Interactive Earth Observing CyberinfrastructureLarry Smarr
 
Toward A National Big Data Superhighway
Toward A National Big Data SuperhighwayToward A National Big Data Superhighway
Toward A National Big Data SuperhighwayLarry Smarr
 
Why Researchers are Using Advanced Networks
Why Researchers are Using Advanced NetworksWhy Researchers are Using Advanced Networks
Why Researchers are Using Advanced NetworksLarry Smarr
 
Looking Back, Looking Forward NSF CI Funding 1985-2025
Looking Back, Looking Forward NSF CI Funding 1985-2025Looking Back, Looking Forward NSF CI Funding 1985-2025
Looking Back, Looking Forward NSF CI Funding 1985-2025Larry Smarr
 
The Energy Efficient Cyberinfrastructure in Slowing Climate Change
The Energy Efficient Cyberinfrastructure in Slowing Climate ChangeThe Energy Efficient Cyberinfrastructure in Slowing Climate Change
The Energy Efficient Cyberinfrastructure in Slowing Climate ChangeLarry Smarr
 
The Pacific Research Platform: a Science-Driven Big-Data Freeway System
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemThe Pacific Research Platform: a Science-Driven Big-Data Freeway System
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemLarry Smarr
 
Remote Telepresence for Exploring Virtual Worlds
Remote Telepresence for Exploring Virtual WorldsRemote Telepresence for Exploring Virtual Worlds
Remote Telepresence for Exploring Virtual WorldsLarry Smarr
 
The Strongly Coupled LambdaCloud
The Strongly Coupled LambdaCloudThe Strongly Coupled LambdaCloud
The Strongly Coupled LambdaCloudLarry Smarr
 
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...Larry Smarr
 
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙Tracy Chen
 
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...Larry Smarr
 
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Larry Smarr
 
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific Applications
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific ApplicationsOptIPuter-A High Performance SOA LambdaGrid Enabling Scientific Applications
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific ApplicationsLarry Smarr
 
High Performance Cyberinfrastructure for Data-Intensive Research
High Performance Cyberinfrastructure for Data-Intensive ResearchHigh Performance Cyberinfrastructure for Data-Intensive Research
High Performance Cyberinfrastructure for Data-Intensive ResearchLarry Smarr
 
Information Technology Infrastructure Committee (ITIC): Report to the NAC
Information Technology Infrastructure Committee (ITIC): Report to the NACInformation Technology Infrastructure Committee (ITIC): Report to the NAC
Information Technology Infrastructure Committee (ITIC): Report to the NACLarry Smarr
 

What's hot (20)

How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...
How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...
How Personal Lightwaves Enable Telepresence: Collapsing the Flat World to a “...
 
Supercomputer End Users: the OptIPuter Killer Application
Supercomputer End Users: the OptIPuter Killer ApplicationSupercomputer End Users: the OptIPuter Killer Application
Supercomputer End Users: the OptIPuter Killer Application
 
Blowing up the Box--the Emergence of the Planetary Computer
Blowing up the Box--the Emergence of the Planetary ComputerBlowing up the Box--the Emergence of the Planetary Computer
Blowing up the Box--the Emergence of the Planetary Computer
 
The Pacific Research Platform:a Science-Driven Big-Data Freeway System
The Pacific Research Platform:a Science-Driven Big-Data Freeway SystemThe Pacific Research Platform:a Science-Driven Big-Data Freeway System
The Pacific Research Platform:a Science-Driven Big-Data Freeway System
 
Report to the NAC
Report to the NACReport to the NAC
Report to the NAC
 
Toward a Global Interactive Earth Observing Cyberinfrastructure
Toward a Global Interactive Earth Observing CyberinfrastructureToward a Global Interactive Earth Observing Cyberinfrastructure
Toward a Global Interactive Earth Observing Cyberinfrastructure
 
Toward A National Big Data Superhighway
Toward A National Big Data SuperhighwayToward A National Big Data Superhighway
Toward A National Big Data Superhighway
 
Why Researchers are Using Advanced Networks
Why Researchers are Using Advanced NetworksWhy Researchers are Using Advanced Networks
Why Researchers are Using Advanced Networks
 
Looking Back, Looking Forward NSF CI Funding 1985-2025
Looking Back, Looking Forward NSF CI Funding 1985-2025Looking Back, Looking Forward NSF CI Funding 1985-2025
Looking Back, Looking Forward NSF CI Funding 1985-2025
 
The Energy Efficient Cyberinfrastructure in Slowing Climate Change
The Energy Efficient Cyberinfrastructure in Slowing Climate ChangeThe Energy Efficient Cyberinfrastructure in Slowing Climate Change
The Energy Efficient Cyberinfrastructure in Slowing Climate Change
 
The Pacific Research Platform: a Science-Driven Big-Data Freeway System
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemThe Pacific Research Platform: a Science-Driven Big-Data Freeway System
The Pacific Research Platform: a Science-Driven Big-Data Freeway System
 
Remote Telepresence for Exploring Virtual Worlds
Remote Telepresence for Exploring Virtual WorldsRemote Telepresence for Exploring Virtual Worlds
Remote Telepresence for Exploring Virtual Worlds
 
The Strongly Coupled LambdaCloud
The Strongly Coupled LambdaCloudThe Strongly Coupled LambdaCloud
The Strongly Coupled LambdaCloud
 
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...
The Jump to Light Speed - Data Intensive Earth Sciences are Leading the Way t...
 
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙
Cloud Computing,雲端運算-中研院網格計畫主持人林誠謙
 
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...
Advanced Global-Scale Networking Supporting Data-Intensive Artificial Intelli...
 
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
 
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific Applications
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific ApplicationsOptIPuter-A High Performance SOA LambdaGrid Enabling Scientific Applications
OptIPuter-A High Performance SOA LambdaGrid Enabling Scientific Applications
 
High Performance Cyberinfrastructure for Data-Intensive Research
High Performance Cyberinfrastructure for Data-Intensive ResearchHigh Performance Cyberinfrastructure for Data-Intensive Research
High Performance Cyberinfrastructure for Data-Intensive Research
 
Information Technology Infrastructure Committee (ITIC): Report to the NAC
Information Technology Infrastructure Committee (ITIC): Report to the NACInformation Technology Infrastructure Committee (ITIC): Report to the NAC
Information Technology Infrastructure Committee (ITIC): Report to the NAC
 

Viewers also liked

NCSA and Telepresence Collaboration
NCSA and Telepresence CollaborationNCSA and Telepresence Collaboration
NCSA and Telepresence CollaborationLarry Smarr
 
Easy and Effective Remote Collaboration
Easy and Effective Remote CollaborationEasy and Effective Remote Collaboration
Easy and Effective Remote CollaborationWrike
 
Building Cross-Functional Scrum-Teams in a Hardware Project
Building Cross-Functional Scrum-Teams in a Hardware ProjectBuilding Cross-Functional Scrum-Teams in a Hardware Project
Building Cross-Functional Scrum-Teams in a Hardware ProjectStephanie Gasche
 
Arkadin WebEx Demo
Arkadin WebEx DemoArkadin WebEx Demo
Arkadin WebEx Demojfz1967
 
Remote Control: Your Guide to Successful Collaboration
Remote Control: Your Guide to Successful CollaborationRemote Control: Your Guide to Successful Collaboration
Remote Control: Your Guide to Successful CollaborationJason Wishard
 
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...Harvard Web Working Group
 
What Is Collaboration?
What Is Collaboration?What Is Collaboration?
What Is Collaboration?Eugene Kim
 
Ten commandments-for-collaborative team work
Ten commandments-for-collaborative team workTen commandments-for-collaborative team work
Ten commandments-for-collaborative team workShabbar Suterwala
 
Creating Collaborative Cultures
Creating Collaborative CulturesCreating Collaborative Cultures
Creating Collaborative CulturesBernice Moore, PhD
 
Collaboration in workplace
Collaboration in workplaceCollaboration in workplace
Collaboration in workplacebelziebub
 
Collaboration Techniques that really work
Collaboration Techniques that really workCollaboration Techniques that really work
Collaboration Techniques that really workleisa reichelt
 
Collaboration PowerPoint slides
Collaboration PowerPoint slidesCollaboration PowerPoint slides
Collaboration PowerPoint slideseisolomon
 
12 Principles of Collaboration
12 Principles of Collaboration12 Principles of Collaboration
12 Principles of CollaborationJacob Morgan
 
How to Successfully Run a Remote Team
How to Successfully Run a Remote TeamHow to Successfully Run a Remote Team
How to Successfully Run a Remote TeamWeekdone.com
 

Viewers also liked (17)

NCSA and Telepresence Collaboration
NCSA and Telepresence CollaborationNCSA and Telepresence Collaboration
NCSA and Telepresence Collaboration
 
The Workplace of Things
The Workplace of ThingsThe Workplace of Things
The Workplace of Things
 
Ui Testing with Ghost Inspector
Ui Testing with Ghost InspectorUi Testing with Ghost Inspector
Ui Testing with Ghost Inspector
 
Easy and Effective Remote Collaboration
Easy and Effective Remote CollaborationEasy and Effective Remote Collaboration
Easy and Effective Remote Collaboration
 
Building Cross-Functional Scrum-Teams in a Hardware Project
Building Cross-Functional Scrum-Teams in a Hardware ProjectBuilding Cross-Functional Scrum-Teams in a Hardware Project
Building Cross-Functional Scrum-Teams in a Hardware Project
 
Arkadin WebEx Demo
Arkadin WebEx DemoArkadin WebEx Demo
Arkadin WebEx Demo
 
Remote Control: Your Guide to Successful Collaboration
Remote Control: Your Guide to Successful CollaborationRemote Control: Your Guide to Successful Collaboration
Remote Control: Your Guide to Successful Collaboration
 
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...
5 Steps to (Remote) Team Bliss: How to Build Thriving, High-Performing (Remot...
 
What Is Collaboration?
What Is Collaboration?What Is Collaboration?
What Is Collaboration?
 
Ten commandments-for-collaborative team work
Ten commandments-for-collaborative team workTen commandments-for-collaborative team work
Ten commandments-for-collaborative team work
 
Creating Collaborative Cultures
Creating Collaborative CulturesCreating Collaborative Cultures
Creating Collaborative Cultures
 
Collaboration
CollaborationCollaboration
Collaboration
 
Collaboration in workplace
Collaboration in workplaceCollaboration in workplace
Collaboration in workplace
 
Collaboration Techniques that really work
Collaboration Techniques that really workCollaboration Techniques that really work
Collaboration Techniques that really work
 
Collaboration PowerPoint slides
Collaboration PowerPoint slidesCollaboration PowerPoint slides
Collaboration PowerPoint slides
 
12 Principles of Collaboration
12 Principles of Collaboration12 Principles of Collaboration
12 Principles of Collaboration
 
How to Successfully Run a Remote Team
How to Successfully Run a Remote TeamHow to Successfully Run a Remote Team
How to Successfully Run a Remote Team
 

Similar to Set My Data Free: High-Performance CI for Data-Intensive Research

An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...
An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...
An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...Larry Smarr
 
Terabit Applications: What Are They, What is Needed to Enable Them?
Terabit Applications: What Are They, What is Needed to Enable Them?Terabit Applications: What Are They, What is Needed to Enable Them?
Terabit Applications: What Are They, What is Needed to Enable Them?Larry Smarr
 
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...Larry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...Larry Smarr
 
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...Larry Smarr
 
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...Larry Smarr
 
A New Global Research Platform – Dedicated 10Gbps Lightpaths
A New Global Research Platform – Dedicated 10Gbps LightpathsA New Global Research Platform – Dedicated 10Gbps Lightpaths
A New Global Research Platform – Dedicated 10Gbps LightpathsLarry Smarr
 
Building a Regional 100G Collaboration Infrastructure
Building a Regional 100G Collaboration InfrastructureBuilding a Regional 100G Collaboration Infrastructure
Building a Regional 100G Collaboration InfrastructureLarry Smarr
 
How Global-Scale Personal Lightwaves are Transforming Scientific Research
How Global-Scale Personal Lightwaves are Transforming Scientific ResearchHow Global-Scale Personal Lightwaves are Transforming Scientific Research
How Global-Scale Personal Lightwaves are Transforming Scientific ResearchLarry Smarr
 
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...Larry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...Larry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
The Future Applications of Australia’s National Broadband Network
The Future Applications of Australia’s National Broadband NetworkThe Future Applications of Australia’s National Broadband Network
The Future Applications of Australia’s National Broadband NetworkLarry Smarr
 
How Fiber Optics are Transforming our World
How Fiber Optics are Transforming our WorldHow Fiber Optics are Transforming our World
How Fiber Optics are Transforming our WorldLarry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyLarry Smarr
 

Similar to Set My Data Free: High-Performance CI for Data-Intensive Research (20)

An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...
An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...
An End-to-End Campus-Scale High Performance Cyberinfrastructure for Data-Inte...
 
Terabit Applications: What Are They, What is Needed to Enable Them?
Terabit Applications: What Are They, What is Needed to Enable Them?Terabit Applications: What Are They, What is Needed to Enable Them?
Terabit Applications: What Are They, What is Needed to Enable Them?
 
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...
High Performance Cyberinfrastructure Enables Data-Driven Science in the Glob...
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Imp...
 
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
 
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...
Bringing 3D, Ultra-Resolution, and Virtual Reality into the Global LambaGrid ...
 
A New Global Research Platform – Dedicated 10Gbps Lightpaths
A New Global Research Platform – Dedicated 10Gbps LightpathsA New Global Research Platform – Dedicated 10Gbps Lightpaths
A New Global Research Platform – Dedicated 10Gbps Lightpaths
 
Building a Regional 100G Collaboration Infrastructure
Building a Regional 100G Collaboration InfrastructureBuilding a Regional 100G Collaboration Infrastructure
Building a Regional 100G Collaboration Infrastructure
 
How Global-Scale Personal Lightwaves are Transforming Scientific Research
How Global-Scale Personal Lightwaves are Transforming Scientific ResearchHow Global-Scale Personal Lightwaves are Transforming Scientific Research
How Global-Scale Personal Lightwaves are Transforming Scientific Research
 
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...
The Missing Link: Dedicated End-to-End 10Gbps Optical Lightpaths for Clusters...
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...
Shrinking the Planet: A New Global Research Platform –Dedicated 10Gbps Lightp...
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
The Future Applications of Australia’s National Broadband Network
The Future Applications of Australia’s National Broadband NetworkThe Future Applications of Australia’s National Broadband Network
The Future Applications of Australia’s National Broadband Network
 
How Fiber Optics are Transforming our World
How Fiber Optics are Transforming our WorldHow Fiber Optics are Transforming our World
How Fiber Optics are Transforming our World
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 
Coupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation EconomyCoupling Australia’s Researchers to the Global Innovation Economy
Coupling Australia’s Researchers to the Global Innovation Economy
 

More from Larry Smarr

My Remembrances of Mike Norman Over The Last 45 Years
My Remembrances of Mike Norman Over The Last 45 YearsMy Remembrances of Mike Norman Over The Last 45 Years
My Remembrances of Mike Norman Over The Last 45 YearsLarry Smarr
 
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019Larry Smarr
 
Panel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsPanel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsLarry Smarr
 
Global Network Advancement Group - Next Generation Network-Integrated Systems
Global Network Advancement Group - Next Generation Network-Integrated SystemsGlobal Network Advancement Group - Next Generation Network-Integrated Systems
Global Network Advancement Group - Next Generation Network-Integrated SystemsLarry Smarr
 
Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...
 Wireless FasterData and Distributed Open Compute Opportunities and (some) Us... Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...
Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...Larry Smarr
 
Panel Discussion: Engaging underrepresented technologists, researchers, and e...
Panel Discussion: Engaging underrepresented technologists, researchers, and e...Panel Discussion: Engaging underrepresented technologists, researchers, and e...
Panel Discussion: Engaging underrepresented technologists, researchers, and e...Larry Smarr
 
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon Moon
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon MoonThe Asia Pacific and Korea Research Platforms: An Overview Jeonghoon Moon
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon MoonLarry Smarr
 
Panel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsPanel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsLarry Smarr
 
Panel: The Global Research Platform: An Overview
Panel: The Global Research Platform: An OverviewPanel: The Global Research Platform: An Overview
Panel: The Global Research Platform: An OverviewLarry Smarr
 
Panel: Future Wireless Extensions of Regional Optical Networks
Panel: Future Wireless Extensions of Regional Optical NetworksPanel: Future Wireless Extensions of Regional Optical Networks
Panel: Future Wireless Extensions of Regional Optical NetworksLarry Smarr
 
Global Research Platform Workshops - Maxine Brown
Global Research Platform Workshops - Maxine BrownGlobal Research Platform Workshops - Maxine Brown
Global Research Platform Workshops - Maxine BrownLarry Smarr
 
Built around answering questions
Built around answering questionsBuilt around answering questions
Built around answering questionsLarry Smarr
 
Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Larry Smarr
 
Democratizing Science through Cyberinfrastructure - Manish Parashar
Democratizing Science through Cyberinfrastructure - Manish ParasharDemocratizing Science through Cyberinfrastructure - Manish Parashar
Democratizing Science through Cyberinfrastructure - Manish ParasharLarry Smarr
 
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;Larry Smarr
 
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Larry Smarr
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardLarry Smarr
 

More from Larry Smarr (20)

My Remembrances of Mike Norman Over The Last 45 Years
My Remembrances of Mike Norman Over The Last 45 YearsMy Remembrances of Mike Norman Over The Last 45 Years
My Remembrances of Mike Norman Over The Last 45 Years
 
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019
Metagenics How Do I Quantify My Body and Try to Improve its Health? June 18 2019
 
Panel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsPanel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving Institutions
 
Global Network Advancement Group - Next Generation Network-Integrated Systems
Global Network Advancement Group - Next Generation Network-Integrated SystemsGlobal Network Advancement Group - Next Generation Network-Integrated Systems
Global Network Advancement Group - Next Generation Network-Integrated Systems
 
Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...
 Wireless FasterData and Distributed Open Compute Opportunities and (some) Us... Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...
Wireless FasterData and Distributed Open Compute Opportunities and (some) Us...
 
Panel Discussion: Engaging underrepresented technologists, researchers, and e...
Panel Discussion: Engaging underrepresented technologists, researchers, and e...Panel Discussion: Engaging underrepresented technologists, researchers, and e...
Panel Discussion: Engaging underrepresented technologists, researchers, and e...
 
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon Moon
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon MoonThe Asia Pacific and Korea Research Platforms: An Overview Jeonghoon Moon
The Asia Pacific and Korea Research Platforms: An Overview Jeonghoon Moon
 
Panel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving InstitutionsPanel: Reaching More Minority Serving Institutions
Panel: Reaching More Minority Serving Institutions
 
Panel: The Global Research Platform: An Overview
Panel: The Global Research Platform: An OverviewPanel: The Global Research Platform: An Overview
Panel: The Global Research Platform: An Overview
 
Panel: Future Wireless Extensions of Regional Optical Networks
Panel: Future Wireless Extensions of Regional Optical NetworksPanel: Future Wireless Extensions of Regional Optical Networks
Panel: Future Wireless Extensions of Regional Optical Networks
 
Global Research Platform Workshops - Maxine Brown
Global Research Platform Workshops - Maxine BrownGlobal Research Platform Workshops - Maxine Brown
Global Research Platform Workshops - Maxine Brown
 
Built around answering questions
Built around answering questionsBuilt around answering questions
Built around answering questions
 
Panel: NRP Science Impacts​
Panel: NRP Science Impacts​Panel: NRP Science Impacts​
Panel: NRP Science Impacts​
 
Democratizing Science through Cyberinfrastructure - Manish Parashar
Democratizing Science through Cyberinfrastructure - Manish ParasharDemocratizing Science through Cyberinfrastructure - Manish Parashar
Democratizing Science through Cyberinfrastructure - Manish Parashar
 
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;
Panel: Building the NRP Ecosystem with the Regional Networks on their Campuses;
 
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...
Open Force Field: Scavenging pre-emptible CPU hours* in the age of COVID - Je...
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and B...
 
Frank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forwardFrank Würthwein - NRP and the Path forward
Frank Würthwein - NRP and the Path forward
 

Set My Data Free: High-Performance CI for Data-Intensive Research

  • 1. ―Set My Data Free: High-Performance CI for Data-Intensive Research‖ KeynoteSpeaker Cyberinfrastructure Days University of Michigan Ann Arbor, MI November 3, 2010 Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E. Gruber Professor, Dept. of Computer Science and Engineering Jacobs School of Engineering, UCSD Follow me on Twitter: lsmarr
  • 2. Abstract As the need for large datasets and high-volume transfer grows, the shared Internet is becoming a bottleneck for cutting-edge research in universities. What are needed instead are large- bandwidth "data freeways." In this talk, I will describe some of the state-of-the-art uses of high-performance CI and how universities can evolve to support free movement of large datasets.
  • 3. The Data-Intensive Discovery Era Requires High Performance Cyberinfrastructure • Growth of Digital Data is Exponential – ―Data Tsunami‖ • Driven by Advances in Digital Detectors, Computing, Networking, & Storage Technologies • Shared Internet Optimized for Megabyte-Size Objects • Need Dedicated Photonic Cyberinfrastructure for Gigabyte/Terabyte Data Objects • Finding Patterns in the Data is the New Imperative – Data-Driven Applications – Data Mining – Visual Analytics – Data Analysis Workflows Source: SDSC
  • 4. Large Data Challenge: Average Throughput to End User on Shared Internet is 10-100 Mbps Tested October 2010 Transferring 1 TB: --10 Mbps = 10 Days --10 Gbps = 15 Minutes http://ensight.eos.nasa.gov/Missions/icesat/index.shtml
  • 5. The Large Hadron Collider Uses a Global Fiber Infrastructure To Connect Its Users • The grid relies on optical fiber networks to distribute data from CERN to 11 major computer centers in Europe, North America, and Asia • The grid is capable of routinely processing 250,000 jobs a day • The data flow will be ~6 Gigabits/sec or 15 million gigabytes a year for 10 to 15 years
  • 6. Next Great Planetary Instrument: The Square Kilometer Array Requires Dedicated Fiber www.skatelescope.org Transfers Of 1 TByte Images World-wide Will Be Needed Every Minute! Currently Competing Between Australia and S. Africa
  • 7. GRAND CHALLENGES IN DATA-INTENSIVE SCIENCES OCTOBER 26-28, 2010 SAN DIEGO SUPERCOMPUTER CENTER , UC SAN DIEGO Confirmed conference topics and speakers : Needs and Opportunities in Observational Astronomy - Alex Szalay, JHU Transient Sky Surveys – Peter Nugent, LBNL Large Data-Intensive Graph Problems – John Gilbert, UCSB Algorithms for Massive Data Sets – Michael Mahoney, Stanford U. Needs and Opportunities in Seismic Modeling and Earthquake Preparedness - Tom Jordan, USC Needs and Opportunities in Fluid Dynamics Modeling and Flow Field Data Analysis – Parviz Moin, Stanford U. Needs and Emerging Opportunities in Neuroscience – Mark Ellisman, UCSD Data-Driven Science in the Globally Networked World – Larry Smarr, UCSD Petascale High Performance Computing Generates TB Datasets to Analyze
  • 8. Growth of Turbulence Data Over Three Decades (Assuming Double Precision and Collocated Points) Year Authors Simulation Points Size 1972 Orszag & Patterson Isotropic Turbulence 323 1 MB 1987 Kim, Moin & Moser Plane Channel Flow 192x160x128 120 MB 1988 Spalart Turbulent Boundary Layer 432x80x320 340 MB 1994 Le & Moin Backward-Facing Step 768x64x192 288 MB 2000 Freund, Lele & Compressible Turbulent 640x270x128 845 MB Moin Jet 2003 Earth Simulator Isotropic Turbulence 40963 0.8 TB* 2006 Hoyas & Jiménez Plane Channel Flow 6144x633x460 550 GB 8 2008 Wu & Moin Turbulent Pipe Flow 256x5122 2.1 GB 2009 Larsson & Lele Isotropic Shock-Turbulence 1080x3842 6.1 GB 2010 Wu & Moin Turbulent Boundary Layer 8192x500x256 40 GB Turbulent Boundary Layer: One-Periodic Direction 100x Larger Data Sets in 20 Years Source: Parviz Moin, Stanford
  • 9. CyberShake 1.0 Hazard Model Need to Analyze Terabytes of Computed Data • CyberShake 1.0 Computation Source: Thomas H. Jordan, USC, Director, Southern California Earthquake Center - 440,000 Simulations per Site - 5.5 Million CPU hrs (50-Day Run on Ranger Using 4,400 cores) CyberShake seismogram - 189 Million Jobs - 165 TB of Total Output Data CyberShake Hazard Map - 10.6 TB of Stored Data PoE = 2% in 50 yrs - 2.1 TB of Archived Data LA region
  • 10. Large-Scale PetaApps Climate Change Run Generates Terabyte Per Day of Computed Data • 155 Year Control Run 100x – 0.1° Ocean model [ 3600 x 2400 x 42 ] Current – 0.1° Sea-ice model [3600 x 2400 x 20 ] Production – 0.5° Atmosphere [576 x 384 x 26 ] – 0.5° Land [576 x 384] • Statistics – ~18M CPU Hours 4x current production – 5844 Cores for 4-5 Months – ~100 TB of Data Generated – 0.5 to 1 TB per Wall Clock Day Generated 10 Source: John M. Dennis, Matthew Woitaszek, UCAR
  • 11. The Required Components of High Performance Cyberinfrastructure • High Performance Optical Networks • Scalable Visualization and Analysis • Multi-Site Collaborative Systems • End-to-End Wide Area CI • Data-Intensive Campus Research CI
  • 12. Australia—The Broadband Nation: Universal Coverage with Fiber, Wireless, Satellite • Connect 93% of All Australian Premises with Fiber – 100 Mbps to Start, Upgrading to Gigabit • 7% with Next Gen Wireless and Satellite – 12 Mbps to Start • Provide Equal Wholesale Access to Retailers – Providing Advanced Digital Services to the Nation – Driven by Consumer Internet, Telephone, Video – ―Triple Play‖, eHealth, eCommerce… ―NBN is Australia’s largest nation building project in our history.‖ - Minister Stephen Conroy www.nbnco.com.au
  • 13. Globally Fiber to the Premise is Growing Rapidly, Mostly in Asia If Couch Potatoes Deserve a Gigabit Fiber, FTTP Why Not Connections University Growing at Data-Intensive ~30%/year Researchers? 130 Million Households with FTTH in 2013 Source: Heavy Reading (www.heavyreading.com), the market research division of Light Reading (www.lightreading.com).
  • 14. The Global Lambda Integrated Facility-- Creating a Planetary-Scale High Bandwidth Collaboratory Research Innovation Labs Linked by 10G GLIF www.glif.is Created in Reykjavik, Iceland 2003 Visualization courtesy of Bob Patterson, NCSA.
  • 15. The OptIPuter Project: Creating High Resolution Portals Over Dedicated Optical Channels to Global Science Data Scalable Adaptive Graphics Environment (SAGE) Picture Source: Mark Ellisman, David Lee, Jason Leigh Calit2 (UCSD, UCI), SDSC, and UIC Leads—Larry Smarr PI Univ. Partners: NCSA, USC, SDSU, NW, TA&M, UvA, SARA, KISTI, AIST Industry: IBM, Sun, Telcordia, Chiaro, Calient, Glimmerglass, Lucent
  • 16. Nearly Seamless AESOP OptIPortal 46‖ NEC Ultra-Narrow Bezel 720p LCD Monitors Source: Tom DeFanti, Calit2@UCSD;
  • 17. 3D Stereo Head Tracked OptIPortal: NexCAVE Array of JVC HDTV 3D LCD Screens KAUST NexCAVE = 22.5MPixels www.calit2.net/newsroom/article.php?id=1584 Source: Tom DeFanti, Calit2@UCSD
  • 18. High Definition Video Connected OptIPortals: Virtual Working Spaces for Data Intensive Research NASA Supports Two Virtual Institutes LifeSize HD Calit2@UCSD 10Gbps Link to NASA Ames Lunar Science Institute, Mountain View, CA Source: Falko Kuester, Kai Doerr Calit2; Michael Sims, Larry Edwards, Estelle Dodson NASA
  • 19. U Michigan Virtual Space Interaction Testbed (VISIT) Instrumenting OptIPortals for Social Science Research • Using Cameras Embedded in the Seams of Tiled Displays and Computer Vision Techniques, we can Understand how People Interact with OptIPortals – Classify Attention, Expression, Gaze – Initial Implementation Based on Attention Interaction Design Toolkit (J. Lee, MIT) • Close to Producing Usable Eye/Nose Tracking Data using OpenCV Leading U.S. Researchers on the Social Aspects of Collaboration Source: Erik Hofer, UMich, School of Information
  • 20. EVL’s SAGE OptIPortal VisualCasting Multi-Site OptIPuter Collaboratory CENIC CalREN-XD Workshop Sept. 15, 2008 Total Aggregate VisualCasting Bandwidth for Nov. 18, 2008 EVL-UI Chicago Sustained 10,000-20,000 Mbps! At Supercomputing 2008 Austin, Texas November, 2008 Streaming 4k SC08 Bandwidth Challenge Entry Remote: On site: U of Michigan SARA (Amsterdam) UIC/EVL U Michigan GIST / KISTI (Korea) U of Queensland Osaka Univ. (Japan) Russian Academy of Science Masaryk Univ. (CZ) Requires 10 Gbps Lightpath to Each Site Source: Jason Leigh, Luc Renambot, EVL, UI Chicago
  • 21. Exploring Cosmology With Supercomputers, Supernetworks, and Supervisualization Source: Mike Norman, SDSC Intergalactic Medium on 2 GLyr Scale • 40963 Particle/Cell Hydrodynamic Cosmology Simulation • NICS Kraken (XT5) – 16,384 cores • Output – 148 TB Movie Output (0.25 TB/file) – 80 TB Diagnostic Science: Norman, Harkness,Paschos SDSC Visualization: Insley, ANL; Wagner SDSC Dumps (8 TB/file) • ANL * Calit2 * LBNL * NICS * ORNL * SDSC
  • 22. Project StarGate Goals: Combining Supercomputers and Supernetworks • Create an ―End-to-End‖ 10Gbps Workflow • Explore Use of OptIPortals as OptIPortal@SDSC Petascale Supercomputer ―Scalable Workstations‖ • Exploit Dynamic 10Gbps Circuits on ESnet • Connect Hardware Resources at ORNL, ANL, SDSC • Show that Data Need Not be Trapped by the Network Rick Wagner Mike Norman ―Event Horizon‖ Source: Michael Norman, SDSC, UCSD • ANL * Calit2 * LBNL * NICS * ORNL * SDSC
  • 23. Using Supernetworks to Couple End User’s OptIPortal to Remote Supercomputers and Visualization Servers Source: Mike Norman, Rick Wagner, SDSC Argonne NL DOE Eureka 100 Dual Quad Core Xeon Servers 200 NVIDIA Quadro FX GPUs in 50 Quadro Plex S4 1U enclosures 3.2 TB RAM rendering ESnet SDSC 10 Gb/s fiber optic network NICS visualization ORNL Calit2/SDSC OptIPortal1 20 30‖ (2560 x 1600 pixel) LCD panels NSF TeraGrid Kraken simulation Cray XT5 10 NVIDIA Quadro FX 4600 graphics 8,256 Compute Nodes cards > 80 megapixels 99,072 Compute Cores 10 Gb/s network throughout 129 TB RAM *ANL * Calit2 * LBNL * NICS * ORNL * SDSC
  • 24. National-Scale Interactive Remote Rendering of Large Datasets Over 10Gbps Fiber Network SDSC ESnet ALCF Science Data Network (SDN) > 10 Gb/s Fiber Optic Network Dynamic VLANs Configured Using OSCARS Rendering Visualization Eureka OptIPortal (40M pixels LCDs) 100 Dual Quad Core Xeon Servers 10 NVIDIA FX 4600 Cards 200 NVIDIA FX GPUs 10 Gb/s Network Throughout 3.2 TB RAM Interactive Remote Rendering Real-Time Volume Rendering Streamed from ANL to SDSC Last Year Last Week High-Resolution (4K+, 15+ FPS)—But: Now Driven by a Simple Web GUI • Command-Line Driven •Rotate, Pan, Zoom • Fixed Color Maps, Transfer Functions •GUI Works from Most Browsers • Slow Exploration of Data • Manipulate Colors and Opacity • Fast Renderer Response Time Source: Rick Wagner, SDSC
  • 25. NSF’s Ocean Observatory Initiative Has the Largest Funded NSF CI Grant OOI CI Grant: 30-40 Software Engineers Housed at Calit2@UCSD Source: Matthew Arrott, Calit2 Program Manager for OOI CI
  • 26. OOI CI is Built on Dedicated OOI CI Optical Infrastructure Using Clouds Physical Network Implementation Source: John Orcutt, Matthew Arrott, SIO/Calit2
  • 27. California and Washington Universities Are Testing a 10Gbps Connected Commercial Data Cloud • Amazon Experiment for Big Data – Only Available Through CENIC & Pacific NW GigaPOP – Private 10Gbps Peering Paths – Includes Amazon EC2 Computing & S3 Storage Services • Early Experiments Underway – Robert Grossman, Open Cloud Consortium – Phil Papadopoulos, Calit2/SDSC Rocks
  • 28. Open Cloud OptIPuter Testbed--Manage and Compute Large Datasets Over 10Gbps Lambdas CENIC NLR C-Wave Dragon Open Source SW  Hadoop • 9 Racks MREN  Sector/Sphere • 500 Nodes  Nebula • 1000+ Cores  Thrift, GPB • 10+ Gb/s Now  Eucalyptus • Upgrading Portions to  Benchmarks 100 Gb/s in 2010/2011 28 Source: Robert Grossman, UChicago
  • 29. Terasort on Open Cloud Testbed Sustains >5 Gbps--Only 5% Distance Penalty! Sorting 10 Billion Records (1.2 TB) at 4 Sites (120 Nodes) Source: Robert Grossman, UChicago
  • 30. Hybrid Cloud Computing with modENCODE Data • Computations in Bionimbus Can Span the Community Cloud & the Amazon Public Cloud to Form a Hybrid Cloud • Sector was used to Support the Data Transfer between Two Virtual Machines – One VM was at UIC and One VM was an Amazon EC2 Instance • Graph Illustrates How the Throughput between Two Virtual Machines in a Wide Area Cloud Depends upon the File Size Biological data (Bionimbus) Source: Robert Grossman, UChicago
  • 31. Ocean Modeling HPC In the Cloud: Tropical Pacific SST (2 Month Ave 2002) MIT GCM 1/3 Degree Horizontal Resolution, 51 Levels, Forced by NCEP2. Grid is 564x168x51, Model State is T,S,U,V,W and Sea Surface Height Run on EC2 HPC Instance. In Collaboration with OOI CI/Calit2 Source: B. Cornuelle, N. Martinez, C.Papadopoulos COMPAS, SIO
  • 32. Using Condor and Amazon EC2 on Adaptive Poisson-Boltzmann Solver (APBS) • APBS Rocks Roll (NBCR) + EC2 Roll + Condor Roll = Amazon VM • Cluster extension into Amazon using Condor Local Running in Amazon Cloud Cluster EC2 Cloud NBCR NBCR VM VM NBCR VM APBS + EC2 + Condor Source: Phil Papadopoulos, SDSC/Calit2
  • 33. ―Blueprint for the Digital University‖--Report of the UCSD Research Cyberinfrastructure Design Team • Focus on Data-Intensive Cyberinfrastructure April 2009 No Data Bottlenecks --Design for Gigabit/s Data Flows http://research.ucsd.edu/documents/rcidt/RCIDTReportFinal2009.pdf
  • 34. What do Campuses Need to Build to Utilize CENIC’s Three Layer Network? ~ $14M Invested in Upgrade Now Campuses Need to Upgrade! Source: Jim Dolgonas, CENIC
  • 35. Current UCSD Optical Core: Bridging End-Users to CENIC L1, L2, L3 Services Quartzite Communications To 10GigE cluster node interfaces Core Year 3 Enpoints: Quartzite Wavelength >= 60 endpoints at 10 GigE Core Selective ..... Switch >= 32 Packet switched Lucent To 10GigE cluster node interfaces and >= 32 Switched wavelengths other switches To cluster nodes ..... >= 300 Connected endpoints Glimmerglass To cluster nodes ..... Production GigE Switch with OOO Dual 10GigE Upliks Switch To cluster nodes Approximately 0.5 TBit/s 32 10GigE ..... Arrive at the ―Optical‖ GigE Switch with Force10 Dual 10GigE Upliks Center of Campus. ... GigE Switch with Switching is a Hybrid of: To Packet Switch CalREN-HPR Research Dual 10GigE Upliks Packet, Lambda, Circuit -- other nodes Cloud GigE OOO and Packet Switches 10GigE Campus Research 4 GigE 4 pair fiber Cloud Juniper T320 Source: Phil Papadopoulos, SDSC/Calit2 (Quartzite PI, OptIPuter co-PI) Quartzite Network MRI #CNS-0421555; OptIPuter #ANI-0225642
  • 36. UCSD Campus Investment in Fiber Enables Consolidation of Energy Efficient Computing & Storage WAN 10Gb: N x 10Gb CENIC, NLR, I2 Gordon – HPD System Cluster Condo DataOasis (Central) Storage Triton – Petascale Data Analysis Scientific Instruments Digital Data Campus Lab OptIPortal Collections Cluster Tile Display Wall Source: Philip Papadopoulos, SDSC/Calit2
  • 37. The GreenLight Project: Instrumenting the Energy Cost of Computational Science • Focus on 5 Communities with At-Scale Computing Needs: – Metagenomics – Ocean Observing – Microscopy – Bioinformatics – Digital Media • Measure, Monitor, & Web Publish Real-Time Sensor Outputs – Via Service-oriented Architectures – Allow Researchers Anywhere To Study Computing Energy Cost – Enable Scientists To Explore Tactics For Maximizing Work/Watt • Develop Middleware that Automates Optimal Choice of Compute/RAM Power Strategies for Desired Greenness • Partnering With Minority-Serving Institutions Cyberinfrastructure Empowerment Coalition Source: Tom DeFanti, Calit2; GreenLight PI
  • 38. UCSD Biomed Centers Drive High Performance CI National Resource for Network Biology iDASH: Integrating Data for Analysis, Anonymization, and Sharing
  • 39. Calit2 Microbial Metagenomics Cluster- Next Generation Optically Linked Science Data Server Source: Phil Papadopoulos, SDSC, Calit2 512 Processors ~200TB ~5 Teraflops Sun 1GbE X4500 ~ 200 Terabytes Storage and Storage 10GbE Switched 10GbE / Routed Several Large Users Core at Univ. Michigan 4000 Users From 90 Countries
  • 40. Calit2 CAMERA Automatic Overflows into SDSC Triton @ SDSC Triton Resource @ CALIT2 Transparently CAMERA - Sends Jobs to Managed Submit Portal Job Submit on Triton Portal (VM) 10Gbps Direct Mount CAMERA == DATA No Data Staging
  • 41. Rapid Evolution of 10GbE Port Prices Makes Campus-Scale 10Gbps CI Affordable • Port Pricing is Falling • Density is Rising – Dramatically • Cost of 10GbE Approaching Cluster HPC Interconnects $80K/port Chiaro (60 Max) $ 5K Force 10 (40 max) ~$1000 (300+ Max) $ 500 Arista $ 400 48 ports Arista 48 ports 2005 2007 2009 2010 Source: Philip Papadopoulos, SDSC/Calit2
  • 42. 10G Switched Data Analysis Resource: SDSC’s Data Oasis OptIPuter RCN Colo CalRe 32 n Triton 20 24 32 2 Trestles 12 Existing 40 Storage Oasis Procurement (RFP) Dash 8 • Phase0: > 8GB/s sustained, today 1500 – • RFP for Phase1: > 40 GB/sec for Lustre 2000 TB > 40 • Nodes must be able to function as Lustre GB/s OSS (Linux) or NFS (Solaris) 100 • Connectivity to Network is 2 x 10GbE/Node Gordon • Likely Reserve dollars for inexpensive replica servers Source: Philip Papadopoulos, SDSC/Calit2
  • 43. NSF Funds a Data-Intensive Track 2 Supercomputer: SDSC’s Gordon-Coming Summer 2011 • Data-Intensive Supercomputer Based on SSD Flash Memory and Virtual Shared Memory SW – Emphasizes MEM and IOPS over FLOPS – Supernode has Virtual Shared Memory: – 2 TB RAM Aggregate – 8 TB SSD Aggregate – Total Machine = 32 Supernodes – 4 PB Disk Parallel File System >100 GB/s I/O • System Designed to Accelerate Access to Massive Data Bases being Generated in all Fields of Science, Engineering, Medicine, and Social Science Source: Mike Norman, Allan Snavely SDSC
  • 44. Academic Research ―OptIPlatform‖ Cyberinfrastructure: A 10Gbps ―End-to-End‖ Lightpath Cloud HD/4k Video Cams HD/4k Telepresence Instruments End User HPC OptIPortal 10G Lightpaths National LambdaRail Campus Optical Switch Data Repositories & Clusters HD/4k Video Images
  • 45. You Can Download This Presentation at lsmarr.calit2.net