For the current overview of Open Cirrus
Upcoming SlideShare
Loading in...5
×

Like this? Share it with your network

Share
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
    Be the first to like this
No Downloads

Views

Total Views
658
On Slideshare
658
From Embeds
0
Number of Embeds
0

Actions

Shares
Downloads
10
Comments
0
Likes
0

Embeds 0

No embeds

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
    No notes for slide

Transcript

  • 1. Open Cirrus™ Cloud Computing Testbed A joint initiative sponsored by HP, Intel, and Yahoo! http://opencirrus.org 6/22/2009 For more info, contact Dave O’Hallaron, [email_address] Intel Open Cirrus team: David O’Hallaron, Michael Kozuch, Michael Ryan, Richard Gass, James Gurganus, Milan Milenkovic, Eric Gayles, Virginia Meade
  • 2. Open Cirrus™ Cloud Computing Testbed
    • Shared: research , applications , infrastructure (11K cores), data sets
    • Global services : sign on, monitoring, store. Open src stack (prs, tashi, hadoop)
    • Sponsored by HP, Intel, and Yahoo! (with additional support from NSF )
      • 9 sites currently, target of around 20 in the next two years.
  • 3. Open Cirrus Goals
    • Goals
        • Foster new systems and services research around cloud computing
        • Catalyze open-source stack and APIs for the cloud
    • How are we unique?
        • Support for systems research and applications research
        • Federation of heterogeneous datacenters
  • 4. Process
    • Central Management Office, oversees Open Cirrus
        • Currently owned by HP
    • Governance model
        • Research team
        • Technical team
        • New site additions
        • Support (legal (export, privacy), IT, etc.)
    • Each site
        • Runs its own research and technical teams
        • Contributes individual technologies
        • Operates some of the global services
    • E.g.
        • HP site supports portal and PRS
        • Intel site developing and supporting Tashi
        • Yahoo! contributes to Hadoop
  • 5. Intel BigData Open Cirrus site http://opencirrus.intel-research.net 45 Mb/s T3 to Internet Rack of 5 3u storage nodes Node: 12TB disk (12x1TB) Rack of 40 blade compute/ storage nodes Node: 8 Core2 cores (4x2), 8GB RAM, 0.3TB disk (2x150GB) 1 Gb/s (x4 p2p) Switch 48 Gb/s 1 Gb/s (x5 p2p) Rack of 15 2u compute/ storage nodes Node: 8 Core2 cores (2x4), 8GB RAM, 6TB disk (6x1TB) x3 Switch 48 Gb/s Rack of 40 blade compute/ storage nodes 1 Gb/s (x4 p2p) Switch 48 Gb/s 10 nodes : 8 Core2 cores (4x2), 8GB RAM, 0.3TB disk (2x150GB) 20 nodes: 1 Xeon core, 6GB RAM, 366GB disk (36+300GB) 10 nodes : 4 Xeon cores (2x2), 4GB RAM, 150 GB disk (2x75GB) 1 Gb/s (x15 p2p) Rack of 15 1u compute/ storage nodes Node : 8 Core2 cores (4x2), 8GB RAM, 2TB disk (2x1TB) x2 Switch 48 Gb/s 1 Gb/s (x4) 1 Gb/s (x4) 1 Gb/s (x4) 1 Gb/s (x4) Nodes/cores: 40/140 40/320 30/240 45/360 [155/1060] RAM (GB): 240 320 240 360 [1160] Storage (TB) 11 12 60 270 60 [413] Spindles: 80 80 60 270 60 [550] 1 Gb/s (x15 p2p) Switch 48 Gb/s [totals]
  • 6. Open Cirrus Sites 11,780 1,703 1,029 25 TB 2.6 PB Total Storage, Tashi 1 Gb/s -- -- 2TB 64 128 1,024 CMU Site Characteristics #Cores #Srvrs Public Memory Storage Spindles Network Focus HP 1,024 256 178 3.3TB 632TB 1152 10G internal 1Gb/s x-rack Hadoop, Cells, PRS, scheduling IDA 2,400 300 100 4.8TB 43TB+ 16TB SAN 600 1Gb/s Apps based on Hadoop, Pig Intel 1,060 155 145 1.16TB 353TB local 60TB attach 550 1Gb/s Tashi, PRS, MPI, Hadoop KIT 2,048 256 128 10TB 1PB 192 1Gb/s Apps with high throughput UIUC 1,024 128 64 2TB ~500TB 288 1Gb/s Datasets, cloud infrastructure Yahoo (M45) 3,200 480 400 2.4TB 1.2PB 1600 1Gb/s Hadoop on demand
  • 7. Testbed Comparison Testbeds Open Cirrus IBM/Google TeraGrid PlanetLab EmuLab Open Cloud Consortium Amazon EC2 LANL/NSF cluster Type of research Systems & services Data-intensive applications research Scientific applications Systems and services Systems Interoperab. across clouds using open APIs Commer. use Systems Approach Federation of hetero- geneous data centers A cluster supported by Google and IBM Multi-site hetero clusters super comp A few 100 nodes hosted by research instit. A single-site cluster with flexible control Multi-site heteros clusters, focus on network Raw access to virtual machines Re-use of LANL’s retiring clusters Participants HP, Intel, IDA, KIT, UIUC, Yahoo! CMU IBM, Google, Stanford, U.Wash, MIT Many schools and orgs Many schools and orgs University of Utah 4 centers Amazon CMU, LANL, NSF Distribution 6 sites 1,703 nodes 11,780 cores 1 site 11 partners in US > 700 nodes world-wide >300 nodes univ@Utah 480 cores, distributed in four locations 1 site 1 site 1000s of older, still useful nodes
  • 8. Open Cirrus Stack Compute + network + storage resources Power + cooling Management and control subsystem Physical Resource set (PRS) service Credit: John Wilkes (HP)
  • 9. Open Cirrus Stack PRS service PRS clients, each with their own “physical data center” Research Tashi NFS storage service HDFS storage service
  • 10. Open Cirrus Stack PRS service Virtual clusters (e.g., Tashi) Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 11. Open Cirrus Stack PRS service BigData App Hadoop
    • Application running
    • On Hadoop
    • On Tashi virtual cluster
    • On a PRS
    • On real hardware
    Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 12. Open Cirrus Stack PRS service BigData app Hadoop Experiment/ save/restore Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 13. Open Cirrus Stack PRS service BigData App Hadoop Experiment/ save/restore Platform services Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 14. Open Cirrus Stack PRS service BigData App Hadoop Experiment/ save/restore Platform services User services Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 15. Open Cirrus Stack PRS BigData App Hadoop Experiment/ save/restore Platform services User services Research Tashi NFS storage service HDFS storage service Virtual cluster Virtual cluster
  • 16. Open Cirrus stack - PRS
    • PRS service goals
        • Provide mini-datacenters to researchers
        • Isolate experiments from each other
        • Stable base for other research
    • PRS service approach
        • Allocate sets of physical co-located nodes, isolated inside VLANs.
    • PRS code from HP being merged into Tashi Apache project.
        • Running on HP site
        • Being ported to Intel site
        • Will eventually run on all sites
  • 17. Open Cirrus Stack - Tashi
    • An open source Apache Software Foundation project sponsored by Intel (with CMU, Yahoo, HP)
        • Infrastructure for cloud computing on Big Data
        • http://incubator.apache.org/projects/tashi
    • Research focus:
        • Location-aware co-scheduling of VMs, storage, and power.
        • Seamless physical/virtual migration.
    • Joint with Greg Ganger (CMU), Mor Harchol-Balter (CMU), Milan Milenkovic (CTG)
  • 18. Tashi High-Level Design Cluster Manager Storage Service Virtualization Service Scheduler Cluster nodes are assumed to be commodity machines Services are instantiated through virtual machines Data location and power information is exposed to scheduler and services CM maintains databases and routes messages; decision logic is limited Most decisions happen in the scheduler; manages compute/storage/power in concert The storage service aggregates the capacity of the commodity nodes to house Big Data repositories. Node Node Node Node Node Node
  • 19. Open Cirrus Stack - Hadoop
    • An open-source Apache Software Foundation project sponsored by Yahoo!
        • http://wiki.apache.org/hadoop/ProjectDescription
    • Provides a parallel programming model (MapReduce), a distributed file system, and a parallel database (HDFS)
  • 20. How do users get access to Open Cirrus sites?
      • Project PIs apply to each site separately.
      • Contact names, email addresses, and web links for applications to each site will be available on the Open Cirrus Web site (which goes live Q209)
        • http://opencirrus.org
      • Each Open Cirrus site decides which users and projects get access to its site.
      • Developing a global sign on for all sites (Q2 09)
        • Users will be able to login to each Open Cirrus site for which they are authorized using the same login and password.
  • 21. What kinds of research projects are Open Cirrus sites looking for?
    • Open Cirrus is seeking research in the following areas (different centers will weight these differently):
        • Datacenter federation
        • Datacenter management
        • Web services
        • Data-intensive applications and systems
    • The following kinds of projects are generally not of interest:
        • Traditional HPC application development.
        • Production applications that just need lots of cycles.
        • Closed source system development.
  • 22. Open Cirrus Systems Research at Intel
    • Tashi
        • Open source software infrastructure for cloud computing on big data (IRP, with Apache Software Foundation, CMU, Yahoo, HP)
          • http://incubator.apache.org/projects/tashi
        • Research focus:
            • Location-aware co-scheduling of VMs, storage, and power
            • Seamless physical/virtual migration
        • Joint with Greg Ganger (CMU), Mor Harchol-Balter (CMU), Milan Milenkovic (CTG)
    • Sprout
        • Software infrastructure for parallel video stream processing (IRP, IRS, ESP project)
        • Central parallelization infrastructure for ESP and SLIPStream projects
  • 23. Open Cirrus Application Research at Intel
    • ESP (Everyday Sensing and Perception)
        • Detection of everyday activities from video
    • SLIPStream (with CMU)
        • Parallel event and object detection in massive videos
        • Robot sensing and perception
        • Gesture based game and computer interfaces
        • Food Recognizing Interactive Electronic Nutrition Display (FRIEND)
    • NeuroSys - Parallel brain activity analysis
        • Interactive functional MRI (with CMU)
    • Parallel machine learning (with CMU)
        • Parallel belief propagation on massive graphical models
        • Automatically converting movies from 2D to 3D
    • Stem cell tracking (with UPitt/CMU)
    • Parallel dynamic physical rendering (with CMU)
    • Log-based architecture (with CMU)
    • Autolab autograding handin service for the world (with CMU)
  • 24. Summary
    • Intel is collaborating with HP and Yahoo! to provide a cloud computing testbed for the research community
    • Primary goals are to
        • Foster new systems research around cloud computing
        • Catalyze open-source stack and APIs for the cloud
    • Opportunities for Intel and Intel customers