Peter M. Curtis

Maintaining Mission Critical Systems in a 24/7 Environment


Скачать книгу

provides best practices for sustainable design. BREEAM has a set of standard categories in which a building can be assessed under called schemes. The array of schemes they provide covers a wide range of facilities, including prisons, schools, offices, and industrial factories. There is incredible flexibility with this system because if there is a building that does not fall under standard categories, BREEAM will develop a customized criterion to evaluate the building. When trying to attain BREEAM certification, the ratings that can be attained include Pass, Good, Very Good, and Excellent.

      To avoid downtime, facilities managers also must understand the trends that affect the utilization of data centers, given the rapid evolution of technology and its requirements on power distribution systems. The degree of power reliability in a data center will impact the design of the facility infrastructure, the technology plant, system architecture, and end‐user connectivity.

      Today, data centers are pushed to the limit. Servers are crammed into racks, and their high‐performance processors all add up to outrageous power consumption. In the early 2000s, data center power consumption increased by about 25% each year, according to Hewlett Packard. The load in DC’s has leveled off in my mind. They were designed for 200 or 300W/sq. ft. at one time, and they never came close to that rating. 100W to 200W/sq. ft. is more common today and offers sufficient power and cooling for these loads. At the same time, processor performance has gone up 500%, and as equipment footprints shrink, the free floor area is populated with more hardware. However, because the smaller equipment is still rejecting the same amount of heat, cooling densities are growing dramatically and are rapidly consuming more floor space. Traditional design using watts per square foot has continued to grow and can also be calculated as transactions per watt. All this increased processing power generates heat, but if the data center gets too hot, all applications grind to a halt.

      Many data center designers (and their clients) would like to build for a 20‐year life cycle, yet the reality is that most cannot realistically look beyond 2 to 5 years. As companies push to wring more data‐crunching ability from the same real estate, the lynchpin technology of future data centers will not necessarily involve greater processing power or more servers, but improved heat dissipation and better airflow management.

      To combat high temperatures and maintain the current trend toward more powerful processors, engineers are reintroducing old technology: liquid cooling, which was used to cool mainframe computers decades ago. To successfully reintroduce liquid into computer rooms, standards will need to be developed, another arena where standardization can promote reliable solutions that mitigate risk for the industry.

      The large footprint now required for reliable power without planned downtime also affects the planning and maintenance of data center facilities. Over the past two decades, the cost of the facility relative to the computer hardware it houses has not grown proportionately. Budget priorities that favor computer hardware over facilities improvement can lead to insufficient performance. The best way to ensure a balanced allocation of capital is to prepare a business analysis that shows the costs associated with the risk of downtime.

      There is no such thing as plug and play when critical infrastructure is deployed, or existing systems are overhauled to support a company's changing business mission. Reliability is not guaranteed simply by installing new equipment or even building an entirely new data center. An aggressive and rigorous design, failure mode analysis, testing/commissioning process, and operations plan proportional to the facility's criticality level are a necessity and not an option.

      Commissioning is a systematic process of ensuring, through documented verification, that all building systems perform according to the design intent and to the future owner's operational needs. The goal is to provide the owner with a safe and reliable installation. A commissioning agent who serves as the owner's representative usually manages the commissioning process. The commissioning agent's role is to facilitate a highly interactive process of verifying that the project is installed correctly and operating as designed. This is achieved through coordination with the owner, design team, construction team, equipment vendors, and third‐party commissioning provider during the various phases of the project. ASHRAE’s Commissioning Guideline 0‐2500 is a recognized model and a good resource that explains this process in detail and can be applied to critical systems.

      Prior to installation at the site, all major equipment should undergo factory acceptance testing that is witnessed by an independent test engineer familiar with the equipment and the testing procedures. However, relying on the factory acceptance test is not sufficient. Once the equipment is delivered, set in place, wired, and functional testing completed, integrated system testing begins. The integrated system test verifies and certifies that all components work together as a fully integrated system. This is the time to resolve all potential equipment problems. There is no “one size fits all” formula.

      Before a new data center or renovation within an existing building goes on‐line, it is crucial to ensure that the systems are burned‐in and failure scenarios are tested no matter the schedule, milestones, and pressures. You won't have a chance to do this phase over, so get it right the first time. A tremendous amount of coordination is required to fine‐tune and calibrate each component. For example, critical circuit breakers must be tested and calibrated prior to exposing them to any critical electrical load. After all tests are complete, results must be compiled for all equipment and the certified test reports prepared, establishing a benchmark for all future testing.

      Scheduling time to educate staff during systems integration testing is not considered part of the