Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

Proactively Managing Your Data Center Infrastructure

1,281 views

Published on

Attached is the presentation from our Proactively Manage Data Center Infrastructure Webinar - to view the webinar with audio, go here:http://blog.eecnet.com/proactive-manage-data-center/

  • Be the first to comment

Proactively Managing Your Data Center Infrastructure

  1. 1. Proactively Managing Your Data Center Infrastructure<br />POWER -- COOLING -- MAINTENANCE<br />
  2. 2. Webinar Agenda<br /><ul><li>Data Center Power
  3. 3. Data Center Cooling
  4. 4. Other Issues Found in Data Centers
  5. 5. Maintaining Data Centers
  6. 6. Summary</li></li></ul><li>Today’s Presenters <br />James Stark, P.E.<br />Design / Build Manager<br />Lealand Chittim<br />Director of Operations<br />
  7. 7. Key Issues<br />Power – <br />More than 50% of power going into a typical data center goes to the power and cooling systems –NOT to the IT loads<br />Every kW saved in a data center saves about $1,000 per year<br /> Cooling – <br />Legacy data centers can be overcooled by up to 2.6 times of the required capacity<br />For each raise in set point – you can save 4% off the cost of running the unit<br />Maintenance <br />Preventive – routine scheduled maintenance<br />Predictive – Infrared scanning and battery trending<br />
  8. 8. Polling Question<br />Which of the three key points just listed affect you most or are of most interested to you?<br />
  9. 9.
  10. 10.
  11. 11. Common Power Problems Found in Legacy Data Centers<br />Insufficient power capacity to support higher load densities – can lead to excessive branch circuits and overloading.<br />Insufficient quantity of breaker positions. Distribution panels & PDU’s were not specified for dual power supply servers and 3-phase power supplies.<br />Unexpected load drop with dual power supply servers. Electrical design has single power path and/or single points of failure.<br />
  12. 12. Common Power Problems Found in Legacy Data Centers<br />Dropped load due to PDU failure or tripped circuit breakers. Mismanagement of load utilization on redundant PDU’s and circuits.<br />UPS dropped load during utility power outage. UPS batteries not monitored or maintained properly. Generator not maintained properly.<br />
  13. 13. Legacy vs. Modern Data Center Power <br />Legacy Data Centers<br />Power design based on total kW<br />N or N+1 UPS redundancy<br />N UPS distribution redundancy<br />Benchmark is 100 watts/sq ft or 3 kW/rack<br />Modern Data Centers<br />Power design based on kW plus distribution<br /><ul><li>2N UPS
  14. 14. 2N UPS distribution
  15. 15. Benchmark is:</li></ul>Med density 5–10 kW/rack<br />High density > 10 kw/rack<br />
  16. 16. Legacy Data Center Power <br />Power design based ontotal kW<br />N or N+1 UPS redundancy<br />N UPS distribution redundancy<br />Benchmark is 100 watts/sq ft or 3 kW/rack<br />
  17. 17. Modern Data Center Power <br />Power design based on kW plus distribution<br /><ul><li>2N UPS
  18. 18. N UPS distribution
  19. 19. Benchmark is:</li></ul>Med density 5–10 kW/rack<br />High density > 10 kW/rack<br />
  20. 20. Polling Question<br />Do you have growth restriction due to the amount of available power?<br />
  21. 21.
  22. 22.
  23. 23. Solving Power Issues<br />Insufficient power capacity<br />As rack power densities increase, UPS branch circuits become overloaded – replace with larger capacity circuits.<br />Imbalanced/overloaded power distribution units (PDU) – circuit tracing, load balancing and additional PDU’s may be required.<br /><ul><li>Insufficient quantity of breaker positions
  24. 24. As racks require more power, replacing single-phase circuits with larger capacity, three-phase circuits can reduce breaker space requirements.
  25. 25. Additional PDU’s and/or panel boards may need to be installed.</li></li></ul><li>Solving Power Issues <br />Loads dropped despite power supply redundancy<br />Trace and properly label circuits to avoid human error.<br />Sites with single power paths may have to be upgraded with redundant power paths.<br />Redundant power paths and equipment can become overloaded if not managed properly<br />Loads should be metered either real-time or at maintenance intervals.<br />Site documentation should be kept up-to-date.<br />Model failure scenarios to predict system reaction to a failure.<br />Proper maintenance, record keeping and trending are key components to avoiding preventable equipment failures<br />
  26. 26. Common Cooling Issues Found in Legacy Data Centers<br />Traditional floor mounted CRAC unit designs are insufficient for loads exceeding 5kW per rack.<br />CRAC units misaligned with hot aisles - leads to hot air mixing, cooling ineffectiveness and inefficiency.<br />
  27. 27. Common Cooling Issues Found in Legacy Data Centers<br />Sub-floor air plenums choked off due to excessive cabling, limiting airflow through perforated floor tiles.<br />Improper location of high density rack - Causes hot spots due to insufficient supply airflow.<br />
  28. 28. Common Cooling Issues Found in Legacy Data Centers<br />Overcooling - excessive number of CRAC units running to cool “the hot spots”.<br />Perforated floor tiles located in hot aisle and near CRAC units. Allows mixing of hot and cold air – reduces cooling system effectiveness and efficiency.<br />Lack of blanking panels in data racks allows hot air to re-circulate to rack fronts and server inlets.<br />Before<br />Installing<br />Blanking <br />panels<br />After<br />Installing<br />Blanking <br />panels<br />
  29. 29. Legacy vs. Modern Cooling Issues<br />Legacy Data Centers<br /><ul><li>Cooling design based on total tonnage plus CRAC unit redundancy
  30. 30. Supply air delivery through raised access floor or ductwork
  31. 31. CRAC units placed within 35’ supply air throw radius </li></ul>Modern Data Centers<br /><ul><li>Cooling design based on close coupled cooling and engineered heat rejection capability
  32. 32. In row or overhead systems
  33. 33. Hot or cold aisle containment </li></li></ul><li>Typical Legacy Data Center Layout<br />Power equipment in Red<br />Cooling equipment in Blue<br />
  34. 34. Legacy vs. Modern Cooling Issues<br />Legacy Data Centers<br /><ul><li>Cooling design based on total tonnage plus CRAC unit redundancy
  35. 35. Supply air delivery through raised access floor or ductwork
  36. 36. CRAC units placed within 35’ supply air throw radius </li></ul>Modern Data Centers<br /><ul><li>Cooling design based on close coupled cooling and engineered heat rejection capability
  37. 37. In row or overhead systems
  38. 38. Hot or cold aisle containment </li></ul>23<br />
  39. 39. Typical ModernData Center Layout<br />Power equipment in Red<br />Cooling equipment in Blue<br />
  40. 40. Polling Question<br />If you have cooling issues, please identify the related cause(s) below?<br />25<br />
  41. 41.
  42. 42.
  43. 43. Solving Cooling Issues<br />High density IT racks (> 5kW):<br />Isolated racks can be treated with specialized cooling solutions (fan-tray tiles).<br />Larger quantities of high density racks can be configuredin “high density cooling zones” and cooled with high density solutions (fan-tray tiles, in-row cooling, etc.)<br /><ul><li>Air mixing:
  44. 44. IT racks should be arranged in Hot Row – Cold Row layouts, with CRAC units positioned to efficiently capture the heated return air.
  45. 45. Perforated tiles should not be located near CRAC units.
  46. 46. Blanking panels should be installed in racks to eliminate hot air re-circulation.</li></li></ul><li>Solving Cooling Issues<br />Sub-floor plenums should be planned to reduce cable congestion. Designated cable pathways and wire management solutions can help to limit congestion. Abandoned cables should be routinely removed from the plenum.<br />Overcooling: The tendency to overcool the data center is common. Targeted cooling solutions for high-density loads and hot spots can save on infrastructure costs and operating costs.<br />
  47. 47. Common Problems Found in Any Data Center<br />Stranded Capacity<br />Mismanagement of power - cooling – space utilization <br />Upgrade of one infrastructure without looking at the big picture<br />Unprotected EPO buttons<br />Accidental actuation will shut the room down!<br />Dry chemical fire extinguishers<br />Dry chemical agent will destroy computer equipment – Clean agents only.<br />30<br />
  48. 48. Common Problems Found in Any Data Center<br />Excessive debris and dust<br />Contaminate computer equipment<br />Cause false fire suppression system alarms<br />Wall / ceiling / floor penetrations<br />Compromise cooling system effectiveness & fire suppression system integrity<br />Generator impairment<br />Not in “auto”<br />Battery failure condition<br />31<br />
  49. 49. Maintaining Data Center Infrastructure<br />Preventive Maintenance<br />Neglecting to maintain equipment or implementing improper maintenance procedures are two of the biggest reasons why mission critical equipment fails. <br />Predictive Maintenance<br />Maintain critical equipment at high performance levels and prevent failures that could cause unplanned downtime with catastrophic results.<br />32<br />
  50. 50. Maintaining Data Center Infrastructure<br />The management of a company’s assets is based on maximizing those assets (ROI) and getting the maximum benefit from assets based on cost <br />Asset management includes Preventive Maintenance<br />In today’s critical environments, IT service continuity is a high priority<br />Downtime is measured<br />Emphasis on planning, preparedness, and adoption of best practice standards<br />
  51. 51. Preventive Maintenance (PM) Program<br /><ul><li>Key to any successful asset management program
  52. 52. Allows equipment to run more efficiently and can minimize energy consumption up to 5% less
  53. 53. Corrective repairs typically cost 2 to 4 times more than preventative maintenance due to their inherent inefficiencies</li></li></ul><li>Preventive Maintenance (PM) Program<br /><ul><li>Most Original Equipment Manufacturers (OEM’s) require a PM program to continue warranty of the product.
  54. 54. Can help identify problems before they become headlines.</li></ul>Pictured here is a damaged UPS inverter assembly, the damage happened when one of the aging capacitors blew; this is the type of damage that can occur if your UPS isn’t properly maintained.<br />
  55. 55. Polling Question<br />With Maintenance being key to preventing down time, which type of service contract do you prefer?<br />
  56. 56.
  57. 57.
  58. 58. Preventive vs. Reactive cost<br /><ul><li>A ratio of 85% planned to 15% unplanned work/cost as a target value is considered to be a good maintenance practice.
  59. 59. Current EEC customer with $954k annualized PM cost.</li></li></ul><li>Predictive Maintenance<br />With no room for downtime, data centers need a strategy to ensure that supporting assets are reliable. <br />Infrared Thermography <br />An effective means of identifying potential problems before an incident or failure occurs.<br />Monitoring<br />Battery Trending - the battery is the most vulnerable part of any UPS, regardless of capacity, topology or brand. <br />
  60. 60. Hourly Impact / System Outages<br />A Meta Group study identified the hourly impact on different organizations to system outages. <br />Source Meta Practice<br />
  61. 61. Maintenance Summary<br /><ul><li>Preventive maintenance is conducted to keep equipment working.
  62. 62. Preventive maintenance activities include partial or complete overhauls at specified periods and recording trends.
  63. 63. The ideal preventive/predictive maintenance program would prevent equipment failure beforeit occurs.</li></li></ul><li>Summary<br />Proactive Management of your IT Environment <br />Pay attention to potential problem areas before the problems exist<br />Diagnose potential problems through maintenance & systems management<br />Electronic Environments – Data Center Experts<br />43<br />
  64. 64. Thank-You<br />Recorded version will be available on www.eecnet.com<br />Questions? Contact<br />Jim Stark, Design/Build Manager508-229-1414 / jstark@eecnet.com<br />Lea Chittim, Operations Manager508-229-1425 / lchittim@eecnet.com<br />44<br />

×