Your SlideShare is downloading. ×
0
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Clemson: Solving the HPC Data Deluge
Upcoming SlideShare
Loading in...5
×

Thanks for flagging this SlideShare!

Oops! An error has occurred.

×
Saving this for later? Get the SlideShare app to save on your phone or tablet. Read anywhere, anytime – even offline.
Text the download link to your phone
Standard text messaging rates apply

Clemson: Solving the HPC Data Deluge

360

Published on

In this presentation from the Dell booth at SC13, Boyd Wilson from Clemson describes how Big Data gets handled for HPC at the University. …

In this presentation from the Dell booth at SC13, Boyd Wilson from Clemson describes how Big Data gets handled for HPC at the University.

"As science drives a rapidly growing need for storage, existing environments face increasing pressure to expand capabilities while controlling costs. Many researchers, scientists and engineers find that they are outgrowing their current system, but fear their organizations may be too small to cover the cost and support needed for more storage. Join these experts for a lively discussion on how you can take control and solve the HPC data deluge."

Watch the video presentation: http://insidehpc.com/2013/12/03/panel-discussion-solving-hpc-data-deluge/

Published in: Technology, Education
0 Comments
0 Likes
Statistics
Notes
  • Be the first to comment

  • Be the first to like this

No Downloads
Views
Total Views
360
On Slideshare
0
From Embeds
0
Number of Embeds
0
Actions
Shares
0
Downloads
9
Comments
0
Likes
0
Embeds 0
No embeds

Report content
Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
No notes for slide

Transcript

  • 1. Clemson     HPC  Storage   Dell  Panel  SC13     Boyd  Wilson   So,ware  CTO   Clemson  University      
  • 2. Outline   •  Palme9o  Cluster   •  Wide  Area  Storage  Across  the  Innova@on   PlaAorm   •  Collec@ve  Cluster     (Real-­‐Time  Data  Aggrega@on  and  Analy@cs  Cluster)     •  Performance  Numbers   •  Research  DMZ/Network    
  • 3. Palmetto  Storage   Primary  Research  Cluster  at  Clemson   •  1972  nodes   •  22928  Cores   •  998400  Cuda  Cores   •  396  TF  (only  benchmarked  newest  GPU  nodes)   •  ~120  +  TF  addi@onal  not  benchmarked.   •  Condominium  Model   •  Home  Storage  SAMQFS  backed  by  SL8500  (6PB)   •  Scratch  OrangeFS  
  • 4. Palmetto  Storage   MX  Nodes   1622  Nodes   96  TF   10G  MX   96  IB  Nodes   with     FDR  IB  Nodes   200  Nodes   400  Nvidia  K20     396  TF   FDR  IB   10G  Eth   Scratch   •  32  R510   •  16  R720   •  512TB  OrangeFS   (v2.8.8)   Home/Archive   •  SAMQFS  over  NFS   •  120TB  Disk   •  6PB  Tape   NFS     SAM  QFS  Home   and  Archive  on   SL8500  
  • 5. Palmetto  Scratch     Next  Steps   MX  Nodes   1622  Nodes   96  TF   10G  IPoMX   Mul@ple  10G  Eth   WebDAV     Campus  Data  Access   •  •  •  •  •  32  Dell  R720   520TB  Scratch   OrangeFS   WebDAV  to  OrangeFS   Hadoop  over  OrangeFS  with   MyHadoop   FDR  IPoIB   FDR  IB  Nodes   200  Nodes   400  Nvidia  K20   GPU   396  TF   Mul@ple  10G  Eth  /  100  G   ScienceDMZ     Innova@on  PlaAorm   Data  Access  
  • 6. Clemson  –  USC  100Gb  tests   12  Dell  R720   OrangeFS   Servers   OrangeFS   Clients   •  File  Write  37Gb/s   •  Server  Hw  problems  &  network  packet  loss  during  tests   •  Perfsonar  49Gb/s  ini@al   •  Later  retest  ~70Gb/s  with  tuning   •  Addi@onal  File  tes@ng  planned   (Ini@al  tes@ng  systems  had  to  move  to  produc@on)  
  • 7. SC13  Demo   OrangeFS   Clients   OrangeFS   Clients   SC13  Floor   •  Clemson   •  USC   •  I2   •  Omnibond   16  Dell  R720   OrangeFS   Servers  
  • 8. The  “Collective”  Cluster   Palme9o   •  12  R720   •  170TB   •  D3  based  Vis  Toolkit   called  SocalTap   •  Social  Media   Aggrega@on  Via  GNIP   •  Elas@c  Search   •  Hadoop  MapReduce   •  OrangeFS   •  WebDAV  to  OrangeFS   Mul@ple  10G  Eth     WebDAV   Campus  Data  Access   Social  Data  Input   ScienceDMZ     Innova@on  PlaAorm   Data  Access  
  • 9. OrangeFS  on  Dell  R720s   •  16  Dell  R720  Servers  Connected  with  10Gb/s  Ethernet   •  32  Clients  reached  nearly  12GB/s  read  and  8GB/s  write   #  Write   iozone  -­‐i  0  -­‐c  -­‐e  -­‐w  -­‐r  $RS  -­‐s  4g  -­‐t  $NUM_PROCESSES  -­‐+n  -­‐+m  $CLIENT_LIST   #  Read   iozone  -­‐i  1  -­‐c  -­‐e  -­‐w  -­‐r  $RS  -­‐s  4g  -­‐t  $NUM_PROCESSES  -­‐+n  -­‐+m  $CLIENT_LIST  
  • 10. MapReduce  over  OrangeFS   •  *25%  improvement  with  OrangeFS  running  on  Separate  nodes   from  Map  Reduce     •  8  Dell  R720  Servers  Connected  with  10Gb/s  Ethernet   •  Remote  Case  adds  an  additional  8  Identical  Servers  and   does  all  OrangeFS  work  Remotely  and  only  Local  work  is   done  on  Compute  Node  (Traditional  HPC  Model)  
  • 11. MapReduce  over  OrangeFS   •  16  Dell  R720  Servers  Connected  with  10Gb/s  Ethernet   •  Remote  Clients  are  Dell  R720s  with  single  SAS  disks  for   local  data  (vs.  12  disk  arrays  in  the  previous  test).  
  • 12. Clemson  Research  Network   Internet/I2/NLR PerfSonar PerfSonar Collaborator PerfSonar CLight Science(DMZ( Perimeter&F/W I2&InnovaJon&PlaKorm Internet F/W&(ACL)&and&Route&Filter DMZ Campus gg ed & Peer&Link Clemson 10 0G i g&T a PerfSonar Tr un k PerfSonar Innova@on( PlaAorm Palme>oNet Host&Firewall Brocade(MLx32( Core((Router CC7NIE Fibre(Channel Dell&Z9000 SamQFS Dell&S4810 Top&of&Rack Dell&S4810 Dell&S4810

×