SlideShare a Scribd company logo
1 of 42
Attie Juyn & Wilhelm van Belkum igh H erformance P omputing C & Computing GRID
Agenda ,[object Object],[object Object],[object Object]
Background ,[object Object],[object Object],[object Object],[object Object]
To establish an Institutional HPC Level 1 : (Entry Level) Personal workstation Level 2 : Departmental  Compute Cluster Level 3 : Institutional HPC Level 4  Nat./Int. HPC
University Strategy ,[object Object],[object Object],[object Object]
The Challenge: to innovate ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
HPC (IT) success criteria ,[object Object],[object Object],[object Object],[object Object],[object Object],& Performance NWU HPC  management strategy NWU HPC  design
Enabling factors ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Project milestones ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
Management principles ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
 
The New World Order Source 2006 UC Regents  Mainframe Mini Computer PC Cluster & Grids Vector Supercomputer
Technical goals Build a Institutional  H igh  P erformance  C omputing facility,  based on Beowulf cluster principals,  coexisting and linking  existing departmental cluster,  the National and International computational Grids
Beowulf cluster ,[object Object],[object Object]
History of Clusters -  The first Beowulf ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],Source 2006 UC Regents
The Evolved Cluster Compute Nodes Admin User Job Queue Source Cluster Resources, Inc. Resource Manager Scheduler License Manager Myrinet Identity Manager Allocation Manager Resource Manager Scheduler Departmental Cluster
Cluster and Grid software landscape
Grid/Cluster Stack or Framework EGEE Chinese USA EU MPI PVM LAM MPICH Parallel Serial Application Resource Manager Rocks Oscar MPI PVM LAM MPICH Parallel Serial Application Resource Manager Oscar Torque Rocks Hardware (Cluster or SMP) CentOS Solaris RedHat UNICOS AIX Scientific Linux Windows Mac OS X HP UX Other Operating System Security GLOBUS CROWNGrid gLite UNICORE Grid Workload Manager:  Scheduler, Policy Manager, Integration Platform Load Leveler PBSpro PBS SGE Condor(G) LSF SLURM Cluster Workload Manager:  Scheduler, Policy Manager, Integration Platform Nimrod MOAB MAUI Portal CLI GUI Application Users Admin
Departmental Computer Cluster
CHPC (May 2007) iQudu” (isiXhosa name for Kudu  “ Tshepe” (Sesotho name for ‘Springbok’) and Impala  ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
The #1 and #13 in world (2007) BlueGene/L - eServer Blue Gene Solution (IBM/212992 Power cores) DOE/NNSA/LLNL - USA  MareNostrum - BladeCenter JS21 Cluster, PPC 970, 2.3 GHz, Myrinet (IBM 10240 Power cores) Barcelona Supercomputer Centre – Spain (63.83 teraFLOP) 478.2 trillion floating operations per second (teraFLOPS) on LINPACK  The #4 and #40 in world (2008)
As of November 2008 #1 : Roadrunner Roadrunner - BladeCenter QS22/LS21 Cluster,  12,240 x PowerXCell 8i 3.2 Ghz 6,562 Dual-Core Opteron 1.8 GHz DOE/NNSA/LANL - United States 1.105   PetaFlop
Reliability & Availability of HPC
HPC (IT) success criteria ,[object Object],[object Object],[object Object],[object Object],[object Object],& Performance NWU HPC  management strategy NWU HPC  design
Introducing  - Utility Computing Swap & migrating of Hardware (First Phase) Dynamic load shifting on RM level (Second Phase) Grid Workload Manager Condor, MOAB Utility Computing Data Center RM HPC RM
Grid/Cluster Stack or Framework EGEE Chinese USA EU Hardware (Cluster or SMP) MPI PVM LAM MPICH Parallel Serial Application Resource Manager Rocks Oscar MPI PVM LAM MPICH Parallel Serial Application Resource Manager Oscar Torque Rocks CentOS Solaris RedHat UNICOS AIX Scientific Linux Windows Mac OS X HP UX Other Operating System GLOBUS CROWNGrid gLite UNICORE Grid Workload Manager:  Scheduler, Policy Manager, Integration Platform Load Leveler PBSpro PBS SGE Condor(G) LSF SLURM Cluster Workload Manager:  Scheduler, Policy Manager, Integration Platform Nimrod MOAB MAUI Security Portal CLI GUI Application Users Admin
HP BL460c 8*3GHz Xeon  12G L2, 1333Mhz FSB 10G memory (96GFlop) HP Modular Cooling System G2 Up to 4 HP C7000, 512 CPU cores 5.12 TFlop HP Blc Virtual Connect Ethernet D-Link X-stack DSN3200 10.5TB RAID5, 80 000 I/O per second HP C7000 Up to 16 HP2x220c (3.072TFlop) 1024 CPU cores HP2x220c (12.288TFlop) BL2x220c 16*3GHz Xeon 192GFlop HP C7000  Up to 16 HP460c (1.536TFlop)
HP ProLiant BL460c Integrated Lights Out 2  Standard Blade Edition Management 2 integrated Multifunction Gigabit NICs Networking 2 mezzanine expansion slots Mezzanine Slots BL460c  ,[object Object],[object Object],[object Object],Memory ,[object Object],[object Object],Internal Storage Up to two Dual & Quad-Core Intel Xeon processors Processor
BL460c Internal View Embedded Smart Array Controller integrated on drive backplane 8 Fully Buffered  DIMM Slots DDR II 667Mhz ,[object Object],[object Object],[object Object],Two hot-plug SAS/SATA drive bays QLogic QMH2462 2-pt 4Gb FC HBA NC512m  2-pt 10GbE-KX4Netxen 4x DDR InfiniBand 2-pt 4X DDR (20Gb) Mellanox
HP ProLiant BL2x220c G5 32 server blades in 10U enclosure 16 server blades in 6U enclosure *2 blades per HH enclosure bay Density Integrated Lights Out 2  Standard Blade Edition Management 2 integrated Gigabit NICs per board Networking 1 PCIe mezzanine expansion slot (x8, Type I) per board Mezzanine Slots BL2x220c G5 ,[object Object],[object Object],[object Object],Memory 1 Non Hot-Plug SFF SATA HDD per board Internal Storage Up to two Dual or Quad-Core  Intel® Xeon® processors per board Processor
HP ProLiant BL2x220c G5 Internal View Two Mezzanine Slots Two x8   (both reside on bottom board) 2 x Optional SATA HDDs Top and bottom PCA, side by side 2 x 2 CPUs 2 x 4 DIMM Slots DDR2 533/667MHz 2 x Embedded 1Gb Ethernet Dual-Port NICs Server Board Connectors
10U ,[object Object],[object Object],[object Object],[object Object],[object Object],Servers and other racked equipment ,[object Object],[object Object]
NWU HPC Hardware Spec. ,[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object],[object Object]
NWU HPC/Grid Campus GRID
University Wide Area Network/Internet Total of 45Mbps  34.2Mbps International
SANREN SANREN Vision and the Players InfraCo SEACOM
SA-Grid CHPC NWU C4 UOVS SA-Grid
SEACOM TE-North is a new cable currently being laid across the Mediterranean Sea Cable Laying to start  Oct. 08 Final splicing April 09 Service launch June 09
International Grid
High Performance Computing @ NWU 12/15/2008
igh H erformance P omputing C & GRID Computing orth U U est U U niversity Sustainable Efficient  Reliable  High Availability & Performance @ >3TFlop Scientific Linux
 

More Related Content

What's hot

Maxwell siuc hpc_description_tutorial
Maxwell siuc hpc_description_tutorialMaxwell siuc hpc_description_tutorial
Maxwell siuc hpc_description_tutorial
madhuinturi
 
Session 8,9 PCI Express
Session 8,9 PCI ExpressSession 8,9 PCI Express
Session 8,9 PCI Express
Subhash Iyer
 

What's hot (20)

OpenPOWER Acceleration of HPCC Systems
OpenPOWER Acceleration of HPCC SystemsOpenPOWER Acceleration of HPCC Systems
OpenPOWER Acceleration of HPCC Systems
 
BUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoCBUD17 Socionext SC2A11 ARM Server SoC
BUD17 Socionext SC2A11 ARM Server SoC
 
ODSA Use Case - SmartNIC
ODSA Use Case - SmartNICODSA Use Case - SmartNIC
ODSA Use Case - SmartNIC
 
OpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software StackOpenHPC: A Comprehensive System Software Stack
OpenHPC: A Comprehensive System Software Stack
 
@IBM Power roadmap 8
@IBM Power roadmap 8 @IBM Power roadmap 8
@IBM Power roadmap 8
 
A Dataflow Processing Chip for Training Deep Neural Networks
A Dataflow Processing Chip for Training Deep Neural NetworksA Dataflow Processing Chip for Training Deep Neural Networks
A Dataflow Processing Chip for Training Deep Neural Networks
 
POWER10 innovations for HPC
POWER10 innovations for HPCPOWER10 innovations for HPC
POWER10 innovations for HPC
 
Maxwell siuc hpc_description_tutorial
Maxwell siuc hpc_description_tutorialMaxwell siuc hpc_description_tutorial
Maxwell siuc hpc_description_tutorial
 
Session 8,9 PCI Express
Session 8,9 PCI ExpressSession 8,9 PCI Express
Session 8,9 PCI Express
 
AMD EPYC 7002 World Records
AMD EPYC 7002 World RecordsAMD EPYC 7002 World Records
AMD EPYC 7002 World Records
 
Sparc t4 systems customer presentation
Sparc t4 systems customer presentationSparc t4 systems customer presentation
Sparc t4 systems customer presentation
 
ODSA Workshop: Development Effort Summary
ODSA Workshop: Development Effort SummaryODSA Workshop: Development Effort Summary
ODSA Workshop: Development Effort Summary
 
AMD and the new “Zen” High Performance x86 Core at Hot Chips 28
AMD and the new “Zen” High Performance x86 Core at Hot Chips 28AMD and the new “Zen” High Performance x86 Core at Hot Chips 28
AMD and the new “Zen” High Performance x86 Core at Hot Chips 28
 
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
OpenCAPI-based Image Analysis Pipeline for 18 GB/s kilohertz-framerate X-ray ...
 
AMD Hot Chips Bulldozer & Bobcat Presentation
AMD Hot Chips Bulldozer & Bobcat PresentationAMD Hot Chips Bulldozer & Bobcat Presentation
AMD Hot Chips Bulldozer & Bobcat Presentation
 
POWER9 for AI & HPC
POWER9 for AI & HPCPOWER9 for AI & HPC
POWER9 for AI & HPC
 
PLNOG 13: Maciej Grabowski: HP Moonshot
PLNOG 13: Maciej Grabowski: HP MoonshotPLNOG 13: Maciej Grabowski: HP Moonshot
PLNOG 13: Maciej Grabowski: HP Moonshot
 
PCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA CampPCIe Gen 3.0 Presentation @ 4th FPGA Camp
PCIe Gen 3.0 Presentation @ 4th FPGA Camp
 
Revisit DCA, PCIe TPH and DDIO
Revisit DCA, PCIe TPH and DDIORevisit DCA, PCIe TPH and DDIO
Revisit DCA, PCIe TPH and DDIO
 
DOME 64-bit μDataCenter
DOME 64-bit μDataCenterDOME 64-bit μDataCenter
DOME 64-bit μDataCenter
 

Similar to NWU and HPC

Kindratenko hpc day 2011 Kiev
Kindratenko hpc day 2011 KievKindratenko hpc day 2011 Kiev
Kindratenko hpc day 2011 Kiev
Volodymyr Saviak
 

Similar to NWU and HPC (20)

Seminar Accelerating Business Using Microservices Architecture in Digital Age...
Seminar Accelerating Business Using Microservices Architecture in Digital Age...Seminar Accelerating Business Using Microservices Architecture in Digital Age...
Seminar Accelerating Business Using Microservices Architecture in Digital Age...
 
Exascale Capabl
Exascale CapablExascale Capabl
Exascale Capabl
 
Design installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttuDesign installation-commissioning-red raider-cluster-ttu
Design installation-commissioning-red raider-cluster-ttu
 
Kindratenko hpc day 2011 Kiev
Kindratenko hpc day 2011 KievKindratenko hpc day 2011 Kiev
Kindratenko hpc day 2011 Kiev
 
GIST AI-X Computing Cluster
GIST AI-X Computing ClusterGIST AI-X Computing Cluster
GIST AI-X Computing Cluster
 
Systems Support for Many Task Computing
Systems Support for Many Task ComputingSystems Support for Many Task Computing
Systems Support for Many Task Computing
 
Japan's post K Computer
Japan's post K ComputerJapan's post K Computer
Japan's post K Computer
 
11540800.ppt
11540800.ppt11540800.ppt
11540800.ppt
 
Summit 16: Deploying Virtualized Mobile Infrastructures on Openstack
Summit 16: Deploying Virtualized Mobile Infrastructures on OpenstackSummit 16: Deploying Virtualized Mobile Infrastructures on Openstack
Summit 16: Deploying Virtualized Mobile Infrastructures on Openstack
 
Heterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of SystemsHeterogeneous Computing : The Future of Systems
Heterogeneous Computing : The Future of Systems
 
The Cell Processor
The Cell ProcessorThe Cell Processor
The Cell Processor
 
No[1][1]
No[1][1]No[1][1]
No[1][1]
 
Stream Processing
Stream ProcessingStream Processing
Stream Processing
 
DATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe ConferenceDATE 2020: Design, Automation and Test in Europe Conference
DATE 2020: Design, Automation and Test in Europe Conference
 
Parallel_and_Cluster_Computing.ppt
Parallel_and_Cluster_Computing.pptParallel_and_Cluster_Computing.ppt
Parallel_and_Cluster_Computing.ppt
 
AI Accelerators for Cloud Datacenters
AI Accelerators for Cloud DatacentersAI Accelerators for Cloud Datacenters
AI Accelerators for Cloud Datacenters
 
組み込みから HPC まで ARM コアで実現するエコシステム
組み込みから HPC まで ARM コアで実現するエコシステム組み込みから HPC まで ARM コアで実現するエコシステム
組み込みから HPC まで ARM コアで実現するエコシステム
 
LEGaTO Heterogeneous Hardware
LEGaTO Heterogeneous HardwareLEGaTO Heterogeneous Hardware
LEGaTO Heterogeneous Hardware
 
Programmable Exascale Supercomputer
Programmable Exascale SupercomputerProgrammable Exascale Supercomputer
Programmable Exascale Supercomputer
 
HPC Infrastructure To Solve The CFD Grand Challenge
HPC Infrastructure To Solve The CFD Grand ChallengeHPC Infrastructure To Solve The CFD Grand Challenge
HPC Infrastructure To Solve The CFD Grand Challenge
 

Recently uploaded

Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 

Recently uploaded (20)

Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - English
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptx
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structure
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
Fostering Friendships - Enhancing Social Bonds in the Classroom
Fostering Friendships - Enhancing Social Bonds  in the ClassroomFostering Friendships - Enhancing Social Bonds  in the Classroom
Fostering Friendships - Enhancing Social Bonds in the Classroom
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 

NWU and HPC

  • 1. Attie Juyn & Wilhelm van Belkum igh H erformance P omputing C & Computing GRID
  • 2.
  • 3.
  • 4. To establish an Institutional HPC Level 1 : (Entry Level) Personal workstation Level 2 : Departmental Compute Cluster Level 3 : Institutional HPC Level 4 Nat./Int. HPC
  • 5.
  • 6.
  • 7.
  • 8.
  • 9.
  • 10.
  • 11.  
  • 12. The New World Order Source 2006 UC Regents Mainframe Mini Computer PC Cluster & Grids Vector Supercomputer
  • 13. Technical goals Build a Institutional H igh P erformance C omputing facility, based on Beowulf cluster principals, coexisting and linking existing departmental cluster, the National and International computational Grids
  • 14.
  • 15.
  • 16. The Evolved Cluster Compute Nodes Admin User Job Queue Source Cluster Resources, Inc. Resource Manager Scheduler License Manager Myrinet Identity Manager Allocation Manager Resource Manager Scheduler Departmental Cluster
  • 17. Cluster and Grid software landscape
  • 18. Grid/Cluster Stack or Framework EGEE Chinese USA EU MPI PVM LAM MPICH Parallel Serial Application Resource Manager Rocks Oscar MPI PVM LAM MPICH Parallel Serial Application Resource Manager Oscar Torque Rocks Hardware (Cluster or SMP) CentOS Solaris RedHat UNICOS AIX Scientific Linux Windows Mac OS X HP UX Other Operating System Security GLOBUS CROWNGrid gLite UNICORE Grid Workload Manager: Scheduler, Policy Manager, Integration Platform Load Leveler PBSpro PBS SGE Condor(G) LSF SLURM Cluster Workload Manager: Scheduler, Policy Manager, Integration Platform Nimrod MOAB MAUI Portal CLI GUI Application Users Admin
  • 20.
  • 21. The #1 and #13 in world (2007) BlueGene/L - eServer Blue Gene Solution (IBM/212992 Power cores) DOE/NNSA/LLNL - USA MareNostrum - BladeCenter JS21 Cluster, PPC 970, 2.3 GHz, Myrinet (IBM 10240 Power cores) Barcelona Supercomputer Centre – Spain (63.83 teraFLOP) 478.2 trillion floating operations per second (teraFLOPS) on LINPACK The #4 and #40 in world (2008)
  • 22. As of November 2008 #1 : Roadrunner Roadrunner - BladeCenter QS22/LS21 Cluster, 12,240 x PowerXCell 8i 3.2 Ghz 6,562 Dual-Core Opteron 1.8 GHz DOE/NNSA/LANL - United States 1.105 PetaFlop
  • 24.
  • 25. Introducing - Utility Computing Swap & migrating of Hardware (First Phase) Dynamic load shifting on RM level (Second Phase) Grid Workload Manager Condor, MOAB Utility Computing Data Center RM HPC RM
  • 26. Grid/Cluster Stack or Framework EGEE Chinese USA EU Hardware (Cluster or SMP) MPI PVM LAM MPICH Parallel Serial Application Resource Manager Rocks Oscar MPI PVM LAM MPICH Parallel Serial Application Resource Manager Oscar Torque Rocks CentOS Solaris RedHat UNICOS AIX Scientific Linux Windows Mac OS X HP UX Other Operating System GLOBUS CROWNGrid gLite UNICORE Grid Workload Manager: Scheduler, Policy Manager, Integration Platform Load Leveler PBSpro PBS SGE Condor(G) LSF SLURM Cluster Workload Manager: Scheduler, Policy Manager, Integration Platform Nimrod MOAB MAUI Security Portal CLI GUI Application Users Admin
  • 27. HP BL460c 8*3GHz Xeon 12G L2, 1333Mhz FSB 10G memory (96GFlop) HP Modular Cooling System G2 Up to 4 HP C7000, 512 CPU cores 5.12 TFlop HP Blc Virtual Connect Ethernet D-Link X-stack DSN3200 10.5TB RAID5, 80 000 I/O per second HP C7000 Up to 16 HP2x220c (3.072TFlop) 1024 CPU cores HP2x220c (12.288TFlop) BL2x220c 16*3GHz Xeon 192GFlop HP C7000 Up to 16 HP460c (1.536TFlop)
  • 28.
  • 29.
  • 30.
  • 31. HP ProLiant BL2x220c G5 Internal View Two Mezzanine Slots Two x8 (both reside on bottom board) 2 x Optional SATA HDDs Top and bottom PCA, side by side 2 x 2 CPUs 2 x 4 DIMM Slots DDR2 533/667MHz 2 x Embedded 1Gb Ethernet Dual-Port NICs Server Board Connectors
  • 32.
  • 33.
  • 35. University Wide Area Network/Internet Total of 45Mbps 34.2Mbps International
  • 36. SANREN SANREN Vision and the Players InfraCo SEACOM
  • 37. SA-Grid CHPC NWU C4 UOVS SA-Grid
  • 38. SEACOM TE-North is a new cable currently being laid across the Mediterranean Sea Cable Laying to start Oct. 08 Final splicing April 09 Service launch June 09
  • 40. High Performance Computing @ NWU 12/15/2008
  • 41. igh H erformance P omputing C & GRID Computing orth U U est U U niversity Sustainable Efficient Reliable High Availability & Performance @ >3TFlop Scientific Linux
  • 42.  

Editor's Notes

  1. In summary we determined that the following would need to be address in any HPC to be successful.
  2. In the beginning their where only one big shark (The Mainframe) The next era of supper computing came with the introduction vector Supercomputer the likes of Cray etc.. The next step was compacting into Mini Computer Al the previous approaches where based on SMP closely coupled in one box And then came the modes Personal Computer. Not very strong on it’s own, but connecting a lot of them together make one Big fish So we suited-up got our best fishing rods and decides to go fishing... For one of these new Big Fish Becoming part of the New World Order.
  3. At the previous HPC conference we came, we saw, we determined that, as Institutional IT, the time was right. The University wanted to become a major player in the New World order. This would not be the first try at this, in 1991 we implemented the SP, but the time was not right (Previous part of the presentation) In the mean time the University also ventured into clustering with three Departmental clusters (FSK, Chemistry, BWI) So what do we want to do Technically that would be different. We what to implement the H in HPC > 1 TFlop configuration. Beowulf approach -> open source , commodity of the shelf hardware,
  4. So what is a Beowulf Cluster ?
  5. How did the first Beowulf cluster look like Note the amount of time it took the assemble the cluster 8 months, taking into account Moore’s law this would makeable influenced the effective production life of the cluster.
  6. The light dotted lines show the originator of software. The issue for us is choice of Cluster software as to allow integration into grids The major issues is on scheduler level and making the HPC appear as a CE in the grid.
  7. Concept framework source Cluster Resources, Inc. Show what did we decide, representing the previous slides in a layer approach simmalar to ISO layers We started with hardware. -> ? HW, OS Resource Manager, Cluster schedulers and finally the Grid workload manager.
  8. Based the Barcelona picture we did put I a requisition for a new building to house the new NWU HPC.. But we are still waiting… OK the real reason is. Reason for showing #13. When slides where setup Barcelona was #5 dropped down in #13 in less than 6 months We needs to build have a strategy that is sustainable with fast upgraded path.
  9. We started looking around to determine what is the major issues that HPC have and found that Reliability and Availability is a major factor.
  10. In summary we determined that the following would need to be address in any HPC to be successful.
  11. The first strategy that we will used to extend the capacity and lifecycle of HPC technology will be to: Utilize the characteristics of Data center vs. that of HPC Implement new high performance CPU in HPC and migrate technology to data center As first phase to manual hardware load management through swapping of blades between HPC and data center to match peak demands extend concept to later do this dynamically on Resource Manger level in the long run (also referred to as utility computing) We needed to a strategy to make the HPC cost effective
  12. So looking at the technologies that we were already using in the data center Why start here ? Cost effectiveness, training people on new technology that is only used in HPC would reduce cost effectiveness. Take note Modular fast extension with less work
  13. HP Confidential – may only be shown to customers under NDA and may not be left behind with a customer under any circumstance. [Enter any extra notes here; leave the item ID line at the bottom] Avitage Item ID: {{C97F4853-0C33-430E-AE0B-9F33E6E58879}}
  14. HP Confidential – may only be shown to customers under NDA and may not be left behind with a customer under any circumstance. [Enter any extra notes here; leave the item ID line at the bottom] Avitage Item ID: {{A59D56E3-21C4-498D-B6C4-605439F2D290}}
  15. Show how does the NWU HPC configuration look like.
  16. What is the Spec’s 256 Cores
  17. Addressing the Reliability and Availability
  18. Institutional facility how do we like this. The limitation still is speed. Brining on SANREN.
  19. Monday, 31 March 2008 : The four sites are the main campuses of Wits, UJ, and two of UJ’s satellite campuses, Bunting and Doornfontein says Christiaan Kuun , SANReN Project Manager at the Meraka Institute
  20. How will SANREN be used for the National GRID But what about International Grid. -> SEACOM
  21. SEACOM PROJECT UPDATE - 14 Aug 2008 Construction on-schedule with major ground and sea-based activities proceeding over the next eight months 14 August 2008 – The construction of SEACOM’s 15,000 km fibre optic undersea cable, linking southern and east Africa, Europe and south Asia, is on schedule and set to go live as planned in June 2009 . Some 10,000 km of cable has been manufactured to date at locations in the USA and Japan and Tyco Telecommunications (US) Inc., the project contractors, will begin shipping terrestrial equipment this month with the cable expected to be loaded on the first ship in September 2008. Laying of shore end cables for each landing stations will also proceed from September. This process will comprise the cable portions at shallow depths ranging from 15 to 50m where large vessels are not able to operate. From October 2008, the first of three Reliance Class vessels will start laying the actual cable. The final splicing, which involves connecting all cable sections together, will happen in April 2009, allowing enough time for testing of the system before the commercial launch in June 2009. The final steps of the Environmental Social Impact Assessment (ESIA) process are well advanced and all small archeological, marine and ecological studies, which required scuba diving analysis, have been completed, as well as social consultations with the affected parties. The cable, including repeaters necessary to amplify the signal, will be stored in large tanks onboard the ships. The branching units necessary to divert the cable to the planned landing stations will be connected into the cable path on the ship just prior to deployment into the sea. The cable will then be buried under the ocean bed with the help of a plow along the best possible route demarcated through the marine survey. The connectivity from Egypt to Marseille, France, will be provided through Telecom Egypt’s TE-North fibre pairs that SEACOM has purchased on the system. TE-North is a new cable currently being laid across the Mediterranean Sea. Brian Herlihy, SEACOM President, said: “ We are very happy with the progress made over the past five months. Our manufacturing and deployment schedule is on target and we are confident that we will meet our delivery promises in what is today an incredibly tight market underpinned by sky-rocketing demand for new cables resulting in worldwide delivery delays. “The recently announced executive appointments combined with the project management capabilities already existent within SEACOM position us as a fully fledged telecoms player. We are able to meet the African market’s urgent requirements for cheap and readily available bandwidth within less than a year. ” The cable will go into service long before the 2010 FIFA World Cup kicks-off in South Africa and SEACOM has already been working with key broadcasters to meet their broadband requirements. The team is also trying to expedite the construction in an attempt to assist with the broadcasting requirements of the FIFA Confederations Cup scheduled for June 2009. SEACOM, which is privately funded and over three quarter African owned, will assist communication carriers in south and east Africa through the sale of wholesale international capacity to global networks via India and Europe. The undersea fibre optic cable system will provide African retail carriers with equal and open access to inexpensive bandwidth, removing the international infrastructure bottleneck and supporting east and southern African economic growth. SEACOM will be the first cable to provide broadband to countries in east Africa which, at the moment, rely entirely on expensive satellite connections.
  22. The result of SEACOM and SANREN…
  23. The Timeline vision in terms of production quality National & International GRID
  24. In Summary : NWU‘s HPC will consists of …