Guide To Data Center Automation
Guide To Data Center Automation
Guide To Data Center Automation
Data Center
Automation
sponsored by
Introduction to Realtimepublishers
by Don Jones, Series Editor
For several years, now, Realtime has produced dozens and dozens of high-quality books that just
happen to be delivered in electronic format—at no cost to you, the reader. We’ve made this
unique publishing model work through the generous support and cooperation of our sponsors,
who agree to bear each book’s production expenses for the benefit of our readers.
Although we’ve always offered our publications to you for free, don’t think for a moment that
quality is anything less than our top priority. My job is to make sure that our books are as good
as—and in most cases better than—any printed book that would cost you $40 or more. Our
electronic publishing model offers several advantages over printed books: You receive chapters
literally as fast as our authors produce them (hence the “realtime” aspect of our model), and we
can update chapters to reflect the latest changes in technology.
I want to point out that our books are by no means paid advertisements or white papers. We’re an
independent publishing company, and an important aspect of my job is to make sure that our
authors are free to voice their expertise and opinions without reservation or restriction. We
maintain complete editorial control of our publications, and I’m proud that we’ve produced so
many quality books over the past years.
I want to extend an invitation to visit us at http://nexus.realtimepublishers.com, especially if
you’ve received this publication from a friend or colleague. We have a wide variety of additional
books on a range of topics, and you’re sure to find something that’s of interest to you—and it
won’t cost you a thing. We hope you’ll continue to come to Realtime for your educational needs
far into the future.
Until then, enjoy.
Don Jones
i
Table of Contents
Introduction to Realtimepublishers.................................................................................................. i
An Introduction to Data Center Automation ...................................................................................1
Information Technology Infrastructure Library...............................................................................2
Benefits of ITIL ...................................................................................................................2
Improving Levels of Service....................................................................................3
Reducing IT Costs....................................................................................................3
Enforcing Well-Defined Processes ..........................................................................3
ITIL Framework Content Organization ...............................................................................3
ITIL Compliance..................................................................................................................6
ITIL Content and Resources ................................................................................................6
The Business Value of Data Center Automation.............................................................................7
Basic Benefits of IT .............................................................................................................7
Calculating the Value of IT..................................................................................................8
Identifying Costs......................................................................................................8
Discovering Business Benefits ................................................................................8
Communicating Strategic Business Value...............................................................9
Improving the Business Value of IT....................................................................................9
The Value of Data Center Automation ....................................................................9
Implementing Charge-Backs .................................................................................10
Enabling Better Decisions......................................................................................10
Service Provider.............................................................................................................................11
Benefits of Operating IT as a Service Provider .................................................................11
Implement the Service Provider Model .............................................................................11
Identifying Customers’ Needs ...............................................................................12
Determining “Product Pricing”..............................................................................12
Identifying Service Delivery Details .................................................................................12
Measuring Service Levels......................................................................................13
Prioritizing Projects ...............................................................................................13
Network Configuration Management ............................................................................................13
NCM Tasks ........................................................................................................................14
Configuration Management Challenges ............................................................................14
NCM Solutions ..................................................................................................................15
Benefits of Automating NCM............................................................................................15
ii
Table of Contents
iii
Table of Contents
iv
Table of Contents
v
Table of Contents
vi
Table of Contents
vii
Table of Contents
viii
Table of Contents
ix
Table of Contents
x
Table of Contents
xi
Copyright Statement
Copyright Statement
© 2006 Realtimepublishers.com, Inc. All rights reserved. This site contains materials that
have been created, developed, or commissioned by, and published with the permission
of, Realtimepublishers.com, Inc. (the “Materials”) and this site and any such Materials are
protected by international copyright and trademark laws.
THE MATERIALS ARE PROVIDED “AS IS” WITHOUT WARRANTY OF ANY KIND,
EITHER EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE,
TITLE AND NON-INFRINGEMENT. The Materials are subject to change without notice
and do not represent a commitment on the part of Realtimepublishers.com, Inc or its web
site sponsors. In no event shall Realtimepublishers.com, Inc. or its web site sponsors be
held liable for technical or editorial errors or omissions contained in the Materials,
including without limitation, for any direct, indirect, incidental, special, exemplary or
consequential damages whatsoever resulting from the use of any information contained
in the Materials.
The Materials (including but not limited to the text, images, audio, and/or video) may not
be copied, reproduced, republished, uploaded, posted, transmitted, or distributed in any
way, in whole or in part, except that one copy may be downloaded for your personal, non-
commercial use on a single computer. In connection with such use, you may not modify
or obscure any copyright or other proprietary notice.
The Materials may contain trademarks, services marks and logos that are the property of
third parties. You are not permitted to use these trademarks, services marks or logos
without prior written consent of such third parties.
Realtimepublishers.com and the Realtimepublishers logo are registered in the US Patent
& Trademark Office. All other product or service names are the property of their
respective owners.
If you have any questions about these terms, or if you would like information about
licensing materials from Realtimepublishers.com, please contact us via e-mail at
[email protected].
xii
The Reference Guide to Data Center Automation
[Editor's Note: This eBook was downloaded from Realtime Nexus—The Digital Library. All
leading technology guides from Realtimepublishers can be found at
http://nexus.realtimepublishers.com.]
1
The Reference Guide to Data Center Automation
Through each of the topics in this guide, we’ll cover important terms and concepts that will
enable IT departments to perform more tasks with fewer resources. The importance and value of
automating standard IT operations can be significant in data centers of any size. The goal is to
significantly lower IT operational expenses while at the same time improving the end-user
experience. Whether you’re a CIO or IT manager looking for ways to improve efficiency or a
member of the down-in-the-trenches IT staff, you’ll find valuables concepts, methods, and
techniques for better managing your computing infrastructure.
Benefits of ITIL
Many IT organizations tend to operate in an ad-hoc and reactive fashion. They often respond to
issues after they occur, leading to problems such as downtime and lower quality of service
(QoS). In many cases, this scenario is understandable as IT organizations are often faced with
providing increased levels of service with insufficient staff and minimal budgets. Many
organizations either cannot afford to provide additional resources to IT departments or cannot
justify the reasons to increase investments.
On the surface, this problem might seem very difficult to solve. However, one approach—
increasing overall efficiency—can improve IT service delivery without requiring significant
expenditures. It is in this arena where the implementation of IT management best practices
comes in.
The recommendations included in ITIL were developed based on studies of methods used by
successful IT organizations worldwide. These approaches to solving common IT problems have
been compiled and organized into a set of recommendations. Although implementing ITIL
practices can take time and effort, most organizations will find that the potential benefits clearly
justify the cost. The following sections look at some of the potential ways in which
implementing ITIL practices can benefit IT operations.
2
The Reference Guide to Data Center Automation
Reducing IT Costs
Many IT departments suffer from inefficiencies that lead to increased costs. Problems caused by
lack of communication, poor issue tracking, and ad-hoc changes can add up quickly. Often, IT
managers are unaware of the true costs of purchasing capital assets, configuring and deploying
new equipment, and maintaining this equipment. ITIL best practices include methods for
calculating true costs and for translating this information into business-related terms. This
information can then be used to make a strong case for investments in automation and other
labor-saving technologies.
3
The Reference Guide to Data Center Automation
4
The Reference Guide to Data Center Automation
5
The Reference Guide to Data Center Automation
ITIL Compliance
In some cases, organizations might find that they’re already following at least some of the ITIL
practices (regardless of whether they have consciously done so). Using ITIL’s methodology and
recommendations can give structure to these efforts. In other cases, IT departments may be able
to benefit greatly from implementing the features of the framework.
Unlike some other business-related standards, there is no official certification or testing process
that can “approve” an organization’s use of ITIL. It is up to organizations to determine the best
methods and approaches for implementing these practices in their environments. There are,
however, voluntary compliance certificates. These are known as the Foundation Certificate, the
Practitioner’s Certificate, and the Manager’s Certificate (see Table 1 for more information
about the certifications).
6
The Reference Guide to Data Center Automation
7
The Reference Guide to Data Center Automation
Identifying Costs
An important consideration for IT management is to be able to calculate and clearly
communicate the real costs and benefits of the services that they provide. This identification
usually starts with determining the Total Cost of Ownership (TCO) of a specific portion of the
infrastructure. Often, when business leaders think of the costs related to increasing capacity, they
think only of capital expenditures (such as the purchase price of a workstation or a server). In
most environments, however, this cost represents only a very small portion of the total cost. IT
departments must add in network-related costs, labor costs (for installation, configuration, and
management), software licensing costs, and depreciation.
Often, just the act of collecting this information can provide visibility into an IT department’s
purpose and structure. It can also be very useful for identifying areas of improvement. Most
importantly, however, when true costs are communicated, other areas of the business can begin
to understand how their operations affect the overall finances of the company.
8
The Reference Guide to Data Center Automation
9
The Reference Guide to Data Center Automation
For example, a common challenge for most IT environments is that of keeping systems up to
date. Managing security patches and other software changes can easily use up large amounts of
time. Furthermore, the process tends to be error-prone: It’s easy for systems administrators to
accidentally overlook one or a few systems. Through the use of data center automation, the same
tasks can be performed in much less time with far less involvement from IT staff. This provides
numerous benefits, including freeing systems administrators to work on other tasks. Often,
automation increases the server-to-administrator ratio and reduces the amount of time required to
perform operations. Other benefits include improved consistency, the enforcement of policies
and processes, and improved security. Additionally, by implementing best practices (such as
those provided with the ITIL), efficiency and operational reliability can improve.
The bottom line is that data center automation can significantly improve the business value of IT.
By reducing costs and removing data center-related bottlenecks, data center automation enables
IT and business leaders to focus on more important tasks. The entire organization will be able to
react more quickly and surely to changes, providing both strategic and tactical advantages to the
entire enterprise.
Implementing Charge-Backs
A major problem for some IT organizations is that various departments often compete for
infrastructure resources such as new servers or workstations. IT managers are often in the
difficult position of deciding which projects are approved based on their limited resources and
budgets. This can lead to an adversarial relationship and to some less-than-ideal decisions.
One potential solution is to implement a system of charge-backs. In this system, the IT
department would pass costs for various projects back to the departments that request them. The
charges would affect these departments’ bottom lines. The idea is that business leaders will be
much more judicious in their decisions when they directly experience the costs to the business.
Although implementing and managing charge-backs can increase administration overhead, the
overall cost savings can justify it. Of course, in order for this system to be successful,
cooperation from the entire organization must be obtained.
10
The Reference Guide to Data Center Automation
Service Provider
Modern organizations often rely upon many vendors and outside resources to meet business
objectives. For example, a marketing group might recruit outside talent to develop a Web site or
to work on creative aspects of a new campaign. Alternatively, engineering groups might rely on
outsourcing to contractors or consultants to build a portion of a product. IT departments,
however, are often seen as cost centers that provide only basic infrastructure services. By treating
IT departments as service providers, however, a strategic relationship can be established, and IT
can be seen as a business partner.
11
The Reference Guide to Data Center Automation
12
The Reference Guide to Data Center Automation
Prioritizing Projects
All businesses are constrained with limits on their amount of production, and IT departments are
no exception. Based on labor capacity and technical constraints, only some of the proposed
projects might prove to be feasible. In the traditional IT arrangement, departments often have to
compete for infrastructure resources. Often IT departments are faced with the difficult situation
of deciding which projects should continue and which simply cannot be taken on.
However, when IT works as a service provider, the vendor and customer can work together to
determine what is best for the business overall. If a particular implementation is extremely
costly, both can decide to hold off until more resources become available. However, if multiple
projects are similar and efficiency can be gained by combining them, the business will see an
overall benefit.
13
The Reference Guide to Data Center Automation
NCM Tasks
The act of managing the components of a network can place a significant burden on IT staff. The
process starts with the deployment of new routers, switches, firewalls, and other devices. New
hardware has to be purchased and configured before it’s brought online. The deployment must be
tested, and network administrators must verify that it is working according to the network
guidelines. And that’s just the beginning.
Maintenance operations include regularly updating to the latest available security patches. Other
routine maintenance functions involve changing passwords and updating configurations.
Business-related changes can often require significant upgrades or modifications to the network
infrastructure and adding capacity is a regular task in growing organizations. The goal of
configuration is to rapidly respond to change requests that range from opening a single firewall
port to redesigning entire subnets—without introducing new problems to the environment.
14
The Reference Guide to Data Center Automation
NCM Solutions
Automated NCM solutions can help address many of the challenges related to maintaining a
network infrastructure. The key feature of an automated NCM solution is that all modifications
are made through the system. Ideally, network administrators do not have direct access to the
actual device configurations themselves. All modifications must occur based on a specific
workflow and changes are tracked for later review (see Figure 2).
15
The Reference Guide to Data Center Automation
16
The Reference Guide to Data Center Automation
17
The Reference Guide to Data Center Automation
Server Provisioning
Most IT users recognize that one of the most important—and visible—functions of their IT
departments is setting up new computers. Server provisioning is the process of readying a server
for production use. It generally involves numerous tasks, beginning with the purchase of server
hardware and the physical racking of the equipment. Next is the important (and tedious) task of
installing and configuring the operating system (OS). This step is followed by applying security
patches and OS updates, installing any required applications, and performing security
configuration.
When done manually, the entire process can be time consuming and error prone. For example, if
a single update is overlooked, the server may be vulnerable to security exploits. Furthermore,
even in the smallest IT environments, the task of server provisioning is never really “done”—
changes in business and technical requirements often force administrators to repurpose servers
with new configuration settings and roles.
18
The Reference Guide to Data Center Automation
• Technology refreshes—Even the fastest and most modern servers will begin to show their
age in a matter of just a few years. Organizations often have standards for technology
refreshes that require them to replace a certain portion of the server pool on a scheduled
basis. Migrating the old configuration to new hardware can be difficult and time
consuming when done manually.
• Support for remote sites—It’s often necessary to support remote branch offices and other
sites that might require new servers. Sometimes, the servers can be installed and
configured by the corporate IT department and then be physically shipped. In other cases,
IT staff might have to physically travel between sites. The costs and inefficiencies of this
process can add up quickly.
• Business-related costs—As users and business units await new server deployments, there
are often hidden costs associated with decreases in productivity, lost sales opportunities,
and associated business inefficiencies. These factors underscore the importance of quick
and efficient server provisioning.
Clearly, there is room for improvement in the manual server-provisioning process.
Server-Provisioning Methods
Many OS vendors are aware of the pain associated with deploying new servers. They have
included numerous tools and technologies that can make the process easier and smoother, but
these solutions also have their limitations. To address the challenges of server provisioning, there
are two main approaches that are typically used.
Scripting
The first is scripting. This method involves creating a set of “answer files” or scripts that are
used to provide configuration details to the OS installation process. Ideally, the entire process
will be automated—that is, no manual intervention is required. However, there are some
drawbacks to this approach. First, the process of installing an OS can take many hours because
all the hardware has to be detected and configured, drivers must be loaded, hard disks must be
formatted, and so on. The second problem is that the scripts must be maintained over time, and
they tend to be “fragile.” When hardware and software vendors make even small specification
changes, new drivers or versions might be required.
Imaging
The other method of automating server provisioning is known as imaging. As its name suggests,
this approach involves performing a base installation of an OS (including all updates and
configuration), then simply making identical copies of the hard disks. The disk duplication may
be performed through dedicated hardware devices or through software. The major problems with
this approach include the creation and maintenance of images. As the hardware detection portion
of OS installation is bypassed, the images must be created for each hardware platform on which
the OS will be deployed. Hardware configuration changes often require the creation of new
images. Another problem is in managing settings that must be unique, including OS security
identifiers (SIDs), network addresses, computer names, and other details. Both approaches
involve some important tradeoffs and neither is an ideal solution for IT departments.
19
The Reference Guide to Data Center Automation
20
The Reference Guide to Data Center Automation
Return on Investment
IT departments are often challenged to do more with less. They’re posed with the difficult
situation of having to increase service levels with limited budgets. This reality makes the task of
determining which investments to make far more important. The right decisions can dramatically
decrease costs and improve service; the worst decisions might actually increase overall costs. In
many ways, IT managers just know the benefits of particular technologies or implementations.
We can easily see how automation can reduce the time and effort required to perform certain
tasks. But the real challenge is related to how this information can be communicated to others
within the organization.
The basic idea is that one must make an investment in order to gain a favorable return. And most
investments involve at least some risk. Generally, there will be a significant time between when
you choose to make an investment, and when you see the benefits of that venture. In the best
case, you’ll realize the benefits quickly and there will be a clear advantage. In the worst case, the
investment may never pay off. The following sections explore how Return on Investment (ROI)
can be calculated and how it can be used to make better IT decisions.
Calculating ROI
Although there are many ways in which ROI can be determined, the basic concepts remain the
same: The main idea is to compare the anticipated benefit of an investment with its expected
cost. Terms such as “benefit” and “cost” can be ambiguous, but this section will show the
various types of information you’ll need in order to calculate those numbers.
21
The Reference Guide to Data Center Automation
Calculating Costs
IT-related costs can come from many areas. The first, and perhaps easiest to calculate, is related
to capital equipment purchases. This area includes the “hard costs” spent on workstations,
servers, network devices, and infrastructure equipment. The actual amounts spent can be divided
into meaningful values through metrics such as “average IT equipment cost per user.” In addition
to hardware, software might be required. Based on the licensing terms with the vendor, costs
may be one-time, periodic, or usage-based.
For most environments, a large portion of IT spending is related to labor—the effort necessary to
keep an environment running efficiently and in accordance with business requirements. These
costs might be measured in terms of hours spent on specific tasks. For example, managing
security updates might require, on average, 10 hours per server per year. Well-managed IT
organizations can often take advantage of tracking tools and management reports to determine
these costs. In some cases, independent analysis can help.
When considering an investment in an IT project, both capital and labor costs must be taken into
account. IT managers should determine how much time and effort will be required to make the
change, and what equipment will be required to support it. In addition, costs related to down time
or any related business disruptions must be factored in. This might include, for example, a
temporary loss of productivity while a new accounting application is implemented. There will
likely be some “opportunity costs” related to the change: Time spent on this proposed project
might take attention away from other projects. All these numbers combined can help to identify
the total cost of a proposal.
Calculating Benefits
So far, we’ve looked at the downside—the fact that there are costs related to making changes.
Now, let’s look at factors to take into account when determining potential benefits. An easy
place to start is by examining cost reductions related to hardware and software. Perhaps a new
implementation can reduce the number of required servers, or it can help make more efficient use
of network bandwidth. These benefits can be easy to enumerate and total because most IT
organizations already have a good idea of what they are. It can sometimes be difficult for IT
managers to spot areas for improvement in their own organizations. A third party can often shed
some light on the real costs and identify areas in which the IT teams stand to benefit most.
Other benefits are more difficult to quantify. Time savings and increases in productivity are
important factors that can determine the value of a project. In some cases, metrics (such as sales
projections or engineering quality reports) are readily available. If it is expected that the project
will yield improvements in these areas, the financial benefits can be determined. Along with
these “soft” benefits are aspects related to reduced downtime, reduced deployment times, and
increased responsiveness from the IT department.
22
The Reference Guide to Data Center Automation
Measuring Risk
Investment-related risks are just part of the game—there is rarely a “sure thing” when it comes to
making major changes. Common risks are related to labor and equipment cost overruns. Perhaps
designers and project managers underestimated the amount of effort it would require to
implement a new system. Or capacity estimates for new hardware were too optimistic. These
factors can dramatically reduce the potential benefit of an investment.
Although it is not possible to identify everything that could possibly go wrong, it’s important to
take into account the likelihood of cost overruns and the impacts of changing business
requirements. Some of these factors might be outside the control of the project itself, but they
can have an impact on the overall decision.
23
The Reference Guide to Data Center Automation
ROI numbers can also be very helpful for communicating IT decisions throughout an
organization. When non-technical management can see the benefits of changes such as
implementing automated processes and tools, this insight can generate buy-in and support for IT
initiatives. For example, setting up new network services might seem disruptive at first, but if
business leaders understand the cost savings, they will be much more likely to support the effort
24
The Reference Guide to Data Center Automation
Calculating ROI for some IT initiatives can be difficult. For example, security is one area in
which costs are difficult to determine. Although it would be useful if the IT industry had
actuarial statistics (similar to those used in, for example, the insurance industry), such data can
be difficult to come by. In these situations, IT managers should consider using known numbers,
such as the costs of downtime and damages caused by data loss, to help make their ROI-related
case. And it’s important to keep in mind that in most ROI calculations, subjectivity is
inevitable—you can’t always predict the future with total accuracy, and sometimes you must just
take your best guess.
ROI Analysis
Based on the numbers predicted, the implementation of automation tools seems to be a good
investment. The return is a substantial cost savings, and the results will be realized in a brief
period of time. There is an additional benefit to making improvements in automation—time that
IT staff spends on various routine operations can be better spent on other tasks that make more
efficient use of their time and skills. For example, time that is freed by automating security patch
deployment can often increase resources for testing patches. That might result in patches being
deployed more quickly, and fewer problems with the patch deployment process. The end result is
a better experience for the entire organization. In short, data center automation provides an
excellent potential ROI, and is likely to be a good investment for the organization as a whole.
25
The Reference Guide to Data Center Automation
Benefits of a CAB
The main benefits of creating a CAB are related to managing a major source of potential IT
problems—changes to the existing environment. IT changes can often affect the entire
organization, so the purpose of the CAB is to determine which changes should occur and to
specify how and when they should be performed. The CAB can define a repeatable process that
ensures that requests have gone through an organized process and ad-hoc modifications are not
allowed. Through the CAB review process, some types of problems such as “collisions” caused
by multiple related changes being made by different people can be reduced.
26
The Reference Guide to Data Center Automation
The CAB roles that are most important from a process standpoint are the members who perform
the review of the change request. In simple cases, there may only be a single approver. But, for
larger changes, it’s important to have input from both the technical and business sides of the
organization. The specific individuals might be business unit managers, IT managers, or people
who have specific expertise in the type of change being requested.
The next set of roles involves those who actually plan for, test, and implement the change. These
individuals may or may not be a portion of the CAB. In either case, however, it is the
responsibility of those who perform the changes to communication with CAB members to
coordinate changes with all the people that are involved.
As with many other organizational groups, it’s acceptable for one person to fill multiple roles.
However, as changes get more complex and have greater effects throughout the organization, it
is important for IT groups to work with the business units they support.
Ideally, the CAB will have established a uniform process for requesting changes. The request
should include details related to why the change is being requested, who will be affected by the
change, anticipated benefits, possible risks, and details related to what changes should occur.
Changes should be categorized based on various criteria, such as the urgency of the change
request. Organizations that must deal with large numbers of changes can also benefit from
automated systems that create and store requests in a central database.
27
The Reference Guide to Data Center Automation
When the CAB receives a new request, it can start the review process. It’s a good practice for the
CAB members to meet regularly to review new requests and discuss the status of those that are
in progress. During the review process, the CAB determines which requests should be
investigated further.
Implementing Changes
If the potential benefits are difficult to overlook, and the risk is acceptable, the next step is to
implement the changes. An organization should follow a standardized change process, and the
CAB should be responsible for ensuring that the processes are followed. Often, at least the
service desk should be aware of what changes are occurring and any potential impacts. This will
allow them to respond to calls more efficiently and will help identify which issues are related to
the change.
During the implementation portion of the process, good communication can help make for a
smoother ride. For quick and easy changes, all that might be required is an email reminder of the
change and its intended affects. For larger changes, regular status updates might be better. As
with the rest of the process, it’s very important that technical staff work with the affected
business units in a coordinated way.
28
The Reference Guide to Data Center Automation
Reviewing Changes
Although it might be tempting to “close out” a request as soon as a change is made, the
responsibilities of the CAB should include reviewing changes after they’re complete. The goal is
not only to determine whether the proper process was followed but also to look for areas of
improvement within the procedures. The documentation generated by this review (even if it’s
only a brief comment) can be helpful for future reference.
29
The Reference Guide to Data Center Automation
30
The Reference Guide to Data Center Automation
31
The Reference Guide to Data Center Automation
Information to Track
The IT industry includes dozens of standards related to hardware, software, and network
configuration. A CMDB solution may provide support for many kinds of data, with the goal of
being able to track the interaction between the devices in the environment. That raises the
question of what information should be tracked.
Server Configuration
Server configurations can be complex and can vary significantly based on the specific OS
platform and version. The CMDB should be able to track the hardware configuration of server
computers, including such details as BIOS revisions, hard disk configurations, and any health-
related monitoring features that might be available. In addition, the CMDB should contain details
about the OS and which applications are installed on the computer. Finally, important
information such as the network configuration of the server should be recorded.
Desktop Configuration
One of the most critical portions of an IT infrastructure generally exists outside the data center.
End-user workstations, notebook computers, and portable devices all must be managed.
Information about the network configuration, hardware platform, and applications can be stored
within the CMDB. These details can be very useful for performing routine tasks, such as security
updates, and for ensuring that the computers adhere to the corporate computing policies.
Network Configuration
From a network standpoint, routers, switches, firewalls, and other devices should be documented
with the CMDB. Ideally, all important details from within the router configuration files will be
included in the data. As network devices often have to interact, network topology details
(including routing methods and inter-dependencies) should also be documented. Wherever
possible, network administrators should note the purpose of various device configurations within
the CMDB.
32
The Reference Guide to Data Center Automation
Software Configuration
Managing software can be a time-consuming and error-prone process in many environments.
Fortunately, the use of a CMDB can help. By keeping track of which software is installed on
which machines, and how many copies of the software are in use concurrently, systems
administrators and support staff can easily report on important details such as OS versions,
license counts, and security configurations. Often, organizations will find that they have
purchased too many licenses or that many users are relying on outdated versions of software.
33
The Reference Guide to Data Center Automation
• Security features—The CMDB will contain numerous details related to the design and
implementation of the network environment. In the wrong hands, this information can be
a security liability. To help protect sensitive data, the CMDB solution should provide a
method for implementing role-based security access. This setup will allow administrators
to control who has access to which information.
• Flexibility and extensibility—In an ideal world, you would set up your entire IT
environment at once and never have to change it. In reality, IT organizations frequently
need to adapt to changing business and technical requirements. New technologies, such
as blade servers and virtual machines, can place new requirements on tracking solutions.
A CMDB solution should be flexible enough to allow for documenting many different
types of devices and should support expandability for new technologies and device types.
The solution may even allow developers to create definitions of their own devices.
• Generation of reports—The main purpose of the CMDB is to provide information to IT
staff, so the solution should have a strong and flexible reporting engine. Features to look
for include the ability to create and save custom report definitions, and the ability to
automatically publish and distribute reports via email or an intranet site.
• Customizability/Application Programming Interface (API)—Although the pre-built
reports and functionality included with a CMDB tool can meet many of users’
requirements, at some point, it might become necessary to create custom applications that
leverage the data stored in the CMDB. That is where a well-document and supported API
can be valuable. Developers should be able to use the API to programmatically return and
modify data. One potential application of this might be to integrate the CMDB with
organizations’ other IT systems.
Overall, through the use of a CMDB, IT organizations can better track, manage, and report on all
the important components of the IT infrastructure.
34
The Reference Guide to Data Center Automation
Auditing
The process of auditing involves systematic checks and examinations to ensure that a specific
aspect of a business is functioning as expected. In the financial world, auditing requires a review
of accounting records, and verification of the information that is recorded. The purpose is to
ensure that the details are consistent and that rules are being followed. From an IT standpoint,
auditing should be an important aspect of operations.
35
The Reference Guide to Data Center Automation
36
The Reference Guide to Data Center Automation
37
The Reference Guide to Data Center Automation
Performing Audits
The process of performing an audit involves comparing the actual configuration of devices and
settings against their expected settings. For an IT department, a typical example might be a
security audit. The expected values will include details related to server patch levels, firewall
rules, and network configuration settings. The employees that actually perform the audit can
include members of the internal staff, including systems and network administrators and IT
management. The goal for internal staff should be to remain completely objective, wherever
possible. Alternatively, organizations can choose to employ outside professionals and consultants
to provide the audit. This method often leads to better accuracy, especially if the consultants
specialize in IT auditing.
Auditing can be performed manually by inspecting individual devices and settings, but there are
several potential problems with this method. First and foremost, the process can be tedious and
time consuming, even in small IT environments. Second, the process leaves much room for error,
as it’s easy to overlook a device or setting. Finally, performing routine audits can be difficult,
especially in large environments in which changes are frequent and thousands of devices must be
examined. Figure 6 shows an example of a manually generated auditing report. Although this
report is far from ideal, it does show the types of information that should be evaluated.
38
The Reference Guide to Data Center Automation
Automating Auditing
When performed manually, the processes related to designing, preparing for, and performing
auditing functions can add a significant burden to IT staff. IT staff must be sure to define
relevant auditing criteria, and they must work diligently to ensure that process and configuration
requirements are always being met. Additionally, the process of performing audits can be
extremely time consuming and therefore are generally performed only when absolutely required.
Fortunately, there are several ways in which data center automation tools and technologies can
help automate the auditing process. One of the most important is having a Configuration
Management Database (CMDB). A CMDB can centrally store all the details related to the
hardware, software, and network devices in an IT environment, so it serves as a ready source
against which expected settings can be compared. Asset tracking functionality provides IT
managers with the power of knowing where all of their hardware and software investments are
(or should be).
Change and configuration management tools can also help by allowing IT staff to quickly and
automatically make changes even in large environments. Whenever a change is made, it can be
recorded to the audit log. Furthermore, by restricting who can make changes and organizing the
change process, data center automation tools can greatly alleviate the burden of performing
auditing manually.
Although auditing can take time and effort to implement, the investment can quickly and easily
pay off. And, through the use of data center automation tools, the entire process can be managed
without additional burden to IT staff.
39
The Reference Guide to Data Center Automation
Customers
One of the many critical success factors for service-related organizations is customer service.
Businesses often go to great lengths to ensure that they understand their customers’ needs and
invest considerable time and effort in researching how to better serve them. The customer
experience can be a “make-or-break” factor in the overall success of the business.
Although the term “customer” is usually used to refer to individuals that work outside of an
organization, IT departments can gain insight into the users and business processes they support
by viewing them as customers. This shift in service delivery perspective can help improve
overall performance of IT departments and operations for an organization as a whole.
Identifying Customers
An important aspect of service delivery is to define who customers are. In the business world,
marketing organizations often spend considerable time, effort, and money in order to make this
determination. They understand the importance of defining their target markets. IT departments
can take a similar approach. Many IT departments tend to be reactive in that they respond to
requests as they come in. These requests may range from individual user needs (such as
password reset requests) to deployments of new enterprise applications (such as the deployment
of a new CRM application).
The first step in identifying customers is to attempt to group them together. End users might
form one group and represent typical desktop and workstation users from any department.
Another group might be mid-level management, who tend to frequently request new computer
installations or changes to existing ones. Finally, upper-level management often focuses on
strategic initiatives, many of which will require support from the IT department. Figure 7
provides an example of some of these groups.
40
The Reference Guide to Data Center Automation
41
The Reference Guide to Data Center Automation
42
The Reference Guide to Data Center Automation
Overall, there are numerous benefits that stand to be gained by having IT departments treat users
and other business units as customers. By identifying groups of users, determining their needs,
and developing products and services, IT organizations can take advantage of the many best
practices utilized by successful companies. Doing so will translate into a better alignment
between IT departments and other areas of the organization and can help to reduce costs.
43
The Reference Guide to Data Center Automation
Measuring Costs
Costs related to the management of IT hardware, software, and network devices can come from
many areas. Figure 9 illustrates the types of costs that might be associated with a typical IT
purchase.
The numbers are hypothetical approximations and that they will vary significantly based on the size
and amount of automation in various data center environments.
Misc. / Other
Initial Capital
Costs
Costs
6%
22%
Labor Costs
47% Infrastructure
Costs
25%
44
The Reference Guide to Data Center Automation
45
The Reference Guide to Data Center Automation
Measuring TCO
There are many challenges that IT organizations will face when trying to calculate TCO for the
devices they support. The main problem is in determining cost-related numbers. Some of this
information can come from reports by IT staff, but that data is often incomplete. Asset
management tools can greatly help keep track of “hard costs,” especially those related to new
purchases. These tools generally allow factoring in finance costs, operating costs, and
depreciation—all of which can be important for determining TCO.
A good source for labor-related costs can be an automated Help desk solution and change and
configuration management tools. IT staff can easily report on the amount of time they’ve spent
on specific issues by using these tools.
46
The Reference Guide to Data Center Automation
Reporting Requirements
An old management saying states that “If you can’t measure it, you can’t manage it.” The idea is
that, without knowing what is occurring within the business, managers will be unable to make
educated decisions. This idea clearly applies to IT environments, where major changes happen
frequently and often at a pace that is much faster than that of other areas of the business. That is
where reporting comes in—the goal is for IT management to be able to gain the insight they need
to make better decisions. It is useful to know what types of reports can be useful, and how these
reports can be generated.
Configuration Reports
Configuration reports show IT managers the current status of the hardware, software, and
network environments that they support. Details might include the configuration of specific
network devices such as routers or firewalls, or the status of particular servers. Basic
configuration information can be obtained manually through the use of tools such as the
Windows System Information Application (as shown in Figure 10).
47
The Reference Guide to Data Center Automation
Figure 10: Viewing configuration details using the Windows System Information tool.
These reports can be very helpful by allowing IT managers to identify underutilized resources,
and for spotting any potential capacity or performance problems. They are also instrumental in
ensuring that all systems are kept up to date with security and application patches. Reporting
solutions should be able to track assets that are located in multiple sites (including those that are
hidden away in closets at small branch offices) to ensure that nothing is overlooked.
48
The Reference Guide to Data Center Automation
Generating Reports
Once an organization has determined the requirements for its reports, it can start looking at how
the reports can be generated. There are many ways in which report creation and generation can
be simplified.
49
The Reference Guide to Data Center Automation
50
The Reference Guide to Data Center Automation
Convergence Examples
In typical IT environments, there are many examples of devices that blur the line between
network and server operations. Dedicated network appliances—such as network-attached storage
(NAS) devices, firewalls, proxy servers, caching devices, and embedded Web servers—all rely
on an underlying OS. For example, although some NAS devices are based on a proprietary
network operating system (OS), many devices include optimized versions of Windows (such as
the Windows Storage Server) or Linux platforms. Figure 12 shows an example of this
configuration.
In many of these systems, there are clear advantages to this type of configuration. For example,
several major firewall solutions run on either Windows or Linux platforms. The benefit is that
systems administrators can gain the usability features of the underlying OS while retaining the
desired functionality. From a management standpoint, however, this configuration might require
a change to the standard paradigm—to ensure that the device is performing optimally, network
and systems administrators must share these responsibilities.
51
The Reference Guide to Data Center Automation
By highlighting these requirements, IT staff can better visualize all the network and server
infrastructure components that are required to support a specific application.
52
The Reference Guide to Data Center Automation
53
The Reference Guide to Data Center Automation
54
The Reference Guide to Data Center Automation
Developing SLAs
There are several important points to include in a complete SLA. Of course, it begins with a
description of what level of service will be provided. At this point, the more detailed the
information, the better it will be for both sides. Details should include processes that will be used
to manage and maintain SLAs. For example, if a certain level is not being met, points of contact
should be established on the IT and business sides.
In many cases, IT departments might find that many different service level requirements overlap.
For example, several departments might require high availability of Virtual Private Network
(VPN) services in order to support traveling users and remote branch offices. This can help IT
managers prioritize initiatives to best meet their overall goals. In this example, by adding better
monitoring and redundancy features into the VPN, all areas of the organization can benefit.
55
The Reference Guide to Data Center Automation
Enforcing SLAs
When dealing with outside parties, an agreement is often only as strong as the terms of any
guarantee or related penalties. Because most IT departments tend to be located in-house, it’s
generally not appropriate to add financial penalties. Thus, the enforceability of SLAs will be up
to the professionalism of the management team. When goals are not being met, reasons should
be sought out and the team should work together to find a solution. SLAs should be seen as
flexible definitions, and business leaders should expect to adjust them regularly. As with other
performance metrics, organizations might choose to attach salary and performance bonuses
based on SLAs.
Perhaps the biggest challenge is that of prioritization. Given a lack of labor resources, what is
more important: uptime for the CRM application or the deployment of new Engineering servers?
To help in these areas, IT managers might want to schedule regular meetings, both inside and
outside of the IT department, to be sure that everyone in the organization understands the
challenges.
56
The Reference Guide to Data Center Automation
Examples of SLAs
The actual details of SLAs for organizations will differ based on specific business needs.
However, there are some general categories that should be considered. One category is that of
application, hardware, and service uptime. Based on the importance of particular portions of the
IT infrastructure, availability and uptime goals can be developed. Other types of SLAs might
focus on deployment times or issue resolution times.
Table 5 provides some high-level examples of the types of SLAs that might be developed by
an organization. The examples focus on numerical metrics, but it’s also important to keep in
mind that “soft metrics” (such as overall satisfaction with the Service Desk) might also be
included.
SLA Area Metrics Goal Notes/Terms
CRM Application Percent availability 99.9% Excludes planned downtime for
Uptime availability maintenance operations and
downtime due to unrelated
network issues; major application
updates might require additional
planned downtime
Service Desk: Level 1 Issue Resolution Time 4 business Include definition of “Level 1
Issue Resolution hours Issues”
Service Desk: Level 2 Issue Resolution Time 8 business Time is measured from original
Issue Resolution hours submission of issue to the
Service Desk; include definition
of “Level 2 Issues”
Engineering: New Time to deployment 3 days Time is measured from when
Server Deployments formal change request has been
(Physical machine) approved; SLA applies only to
servers that will be hosted within
the data center
Engineering: New Time to deployment 2 hours Virtual machines must use one
Server Deployments of the three standard
(Virtual machine) configuration profiles; time is
measured from when formal
change request has been
approved.
Now that we’ve looked at some examples, let’s see how IT organizations can keep track of
SLAs.
57
The Reference Guide to Data Center Automation
58
The Reference Guide to Data Center Automation
59
The Reference Guide to Data Center Automation
Configuration Management
Keeping track of network configuration files is an important first step to enabling the failover
process. In the event of a failover, restoring this information will help bring a network back to a
usable state. Whenever configuration changes are made, network administrators must be sure
that the change is recorded and replicated to any backup or standby devices.
60
The Reference Guide to Data Center Automation
Figure 15: Maintaining a failover configuration using data center automation tools.
Developing a network business continuity plan is no small task for most IT departments.
Through the use of data center automation solutions, however, this critical task can be made
much more manageable.
61
The Reference Guide to Data Center Automation
Remote Administration
In modern IT environments, systems and network administrators are often tasked with managing
increasing numbers of devices without additional time and resources. In addition, the systems
might be spread out over numerous sites. Centralized management can help meet these needs by
increasing overall efficiency. IT staff should be able to manage devices that are located across
the world just as easily as they can manage the computing devices on their desks. Remote
administration can be used to improve systems and network administration in an IT environment.
62
The Reference Guide to Data Center Automation
Figure 16: Configuring video settings in the Windows XP Remote Desktop client.
63
The Reference Guide to Data Center Automation
• File transfers—In addition to controlling remote computers, Help desk staff and systems
administrations might need a quick and easy way to transfer files. In some cases, transfers
can be handled outside of the remote administration solution by using standard network
file transfer methods. In other cases, such as when a connection is made to a remote
office or across multiple firewalls, a built-in solution that uses the same protocol and
connection as the remote connection can be helpful.
• Shadowing support—For training and troubleshooting purposes, the ability to “shadow” a
connection can be helpful. In this method, the remote user might have view-only
privileges on the remote device. Or, a trainer might be able to demonstrate an operation
on a remote computer without worrying about interruptions from a user.
In addition to these basic features, let’s look closer at details related to security.
64
The Reference Guide to Data Center Automation
65
The Reference Guide to Data Center Automation
Technical Challenges
Regardless of the operating system (OS) platform or the applications that are supported, all
servers must be kept up to date by systems administrators. Common tasks that must be
performed include installing security patches, managing changes to system and network
configurations, and taking an inventory of installed services. These operations are fairly simple
to perform on one or a few servers, but in most data center environments, IT staff members must
manage dozens or hundreds of machines.
Technical challenges include the actual deployment of updates and configuration changes.
Performing this task manually is time-consuming and tedious, even when using remote
administration features. Also, it’s far too easy for systems administrators to accidentally overlook
one or a few machines. In the case of implementing security patches, the result could be serious
security vulnerabilities.
Other challenges are related to actually performing configuration changes. IT departments should
ensure that changes are made consistently, that they adhere to best practices, and that any
modifications are tracked and documented. It’s also important to ensure that only authorized
administrators are making changes and to track who made modifications. Although most systems
administrators would agree to this process, in the real world, it can be difficult to spend the time
and attention required to follow these steps every time.
66
The Reference Guide to Data Center Automation
Process-Related Challenges
It’s important for IT departments to implement and enforce processes related to change and
configuration management. The goal is to ensure that all changes are valid and authorized and to
avoid problems that might appear due to inappropriate modifications to server configurations.
Unfortunately, ensuring communications between IT staff, management, and the users they
support can be difficult. The result is that some changes can cause unexpected problems due to a
lack of coordination.
IT management should also consider “quality assurance” processes and auditing of server
configurations. Ideally, management would be able to quickly and easily view up-to-date details
related to the configuration of all servers in the environment, regardless of location. This can
help identify machines whose configurations are outdated or not in compliance with IT policies.
67
The Reference Guide to Data Center Automation
Figure 17: Making configuration changes using data center automation tools.
68
The Reference Guide to Data Center Automation
Reporting
One of the most visible benefits of automating the server configuration management process is
the ability to generate on-demand reports. The information provided can range from software
installation details to security configurations to server uptime and availability reports. All
configuration and change data is stored in a central CMDB, so systems administrators and IT
managers can quickly obtain the information they need to make better decisions.
Reporting might also be required in order to demonstrate compliance with various regulatory
requirements. A process that was formerly time-consuming and inaccurate can be reduced to a
few simple steps. Better yet, individuals from areas outside of the IT department can view details
that are relevant to performing their jobs.
IT Processes
Processes define a consistent set of steps that should be followed in order to complete a
particular task. From an IT standpoint, processes can range from details about Service Desk
escalations to communicating with end users. The goal of IT processes is to improve overall
operations within the IT department and the organization as a whole.
It’s often a fact that the implementation of processes requires additional effort and may add steps
to some jobs. The steps can be time-consuming and may result in resistance or non-compliance.
That raises the challenge: Processes must be worth far more than the “trouble” they cause in
order to be considered worthwhile. This section will look at details related to what makes a good
process, how you can enforce processes, and the benefits of automating process management.
69
The Reference Guide to Data Center Automation
70
The Reference Guide to Data Center Automation
The reasoning behind processes should be well-understood. IT staff will be much more likely to
adhere to processes that they understand and agree with. Managers should avoid implementing
unnecessarily rigid rules: Processes should not attempt to describe every possible action an
employee must take. Instead, implementers should be given some leeway in determining the best
method by which to complete smaller portions of the tasks. Presenting processes as flexible and
evolving guidelines can go a long way toward ensuring compliance.
Overall, the key goals are that those who follow processes should clearly understand the benefits.
Without buy-in, the process will be seen as a chore that is forced by management.
71
The Reference Guide to Data Center Automation
Managing Exceptions
An unfortunate fact related to working in the real world is that most rules will have at least
occasional exceptions. For example, in an emergency downtime situation, you might not have
enough time to walk through all the steps in a change and configuration management process.
Clearly, in that case, resolving the problem as quickly as possible is the most important factor.
However, the goal should be for exceptions to be relatively rare. If exceptions do occur
frequently, it’s probably worth considering adding them to the current process or developing a
new process.
Examples of IT Processes
By now, it’s likely that you’re either considering updating existing procedures or putting new
processes in place. That raises the question of which operations can benefit most from well-
defined processes. In general, it’s best to focus on tasks that require multiple steps and multiple
operations in order to be completed. The tasks should happen frequently enough so that the
process will be used regularly. Other characteristics include business goals that are often not met
due to miscommunications or inconsistent ways of handling the tasks that are involved.
Some specific examples of IT processes that organizations might either have in place or might be
considering are shown in Table 6.
Business Process Possible Steps Notes
Change and Formal documentation of change Standard forms for
Configuration requests and approval by a CAB communicating changes can be
Management helpful
IT Purchasing Requests for multiple quotes (if Different processes or approval
possible), cost justification, ROI/TCO levels might apply based on the
analysis, and approvals from senior cost and business area related to
management the purchase
Server Server configuration review, security The server should be based on
Deployments configuration checklist, and one of the predefined supported
management acceptance of new configurations
configuration
Service Desk Documentation of new requests, At any given point in time, the
prioritization based on relevant issue must be “owned” by a
Service Level Agreements (SLAs), specific individual
and escalation of process details
72
The Reference Guide to Data Center Automation
73
The Reference Guide to Data Center Automation
This relatively simple situation highlights the importance of understanding the entire
infrastructure that is required to manage an application. For IT departments that support multiple
sites throughout the world and dozens of different line-of-business applications, the overall
problems can be far more complex.
74
The Reference Guide to Data Center Automation
Identifying Interdependencies
Infrastructure components that are shared by multiple applications can be identified after taking
an inventory of the application requirements. Often, the results can help provide greater insight
into operations. Figure 20 provides an example of a shared component that might be used by
multiple applications.
75
The Reference Guide to Data Center Automation
Figure 20: Shared application infrastructure requirements for a modern, distributed application.
76
The Reference Guide to Data Center Automation
77
The Reference Guide to Data Center Automation
78
The Reference Guide to Data Center Automation
79
The Reference Guide to Data Center Automation
80
The Reference Guide to Data Center Automation
Figure 22: Automating configuration management using data center automation tools.
Overall, the process of developing and implementing a business continuity plan for servers will
be a major undertaking for IT staff and management. However, through the use of data center
automation tools, the process can be significantly simplified, and administration overhead can be
minimized. The end result is increased protection of critical data and services at a reasonable
overall cost.
81
The Reference Guide to Data Center Automation
Configuration Management
Over time, servers and network equipment will likely need to be updated to meet changing
business needs. For example, when a router is reconfigured, network address information may
need to change on multiple servers. Alternatively, the organization might implement stricter
security policies that must then be applied to hundreds of devices. In very small and simple
network environments, it might be reasonable to perform these changes manually. In most IT
environments, the process of manually making changes is one that is tedious and leaves a lot of
room for error.
Data center automation solutions can ease the process of making changes on even hundreds of
devices. The process generally involves a member of the IT staff specifying the change that
should be made. Assuming that the staffer has the appropriate permissions, the actual
modifications can be scheduled or applied immediately. Often, the task can be completed in a
matter of minutes, and all that is left for the administrator to do is verify the change log.
82
The Reference Guide to Data Center Automation
Monitoring Performance
All modern OSs require some standard maintenance operations in order to perform at their best.
Actions such as deleting unnecessary files and performing defragmentation can help keep
systems performing optimally. For certain types of applications, such as databases, other tasks
such as index defragmentation or consistency checks might be required. By implementing
automated monitoring solutions, administrators can often be alerted to potential problems before
users experience them. And, many types of problems can be fixed automatically, requiring no
manual intervention at all.
83
The Reference Guide to Data Center Automation
Delegating Responsibility
An important best practice to keep in mind is that of delegating responsibility. Without
coordination between members of the IT team, statements like, “I thought you were going to take
care of that last week,” can lead to significant problems. Data center automation solutions can
allow IT managers to create and configure schedules for their staff members, and can assign
specific duties. This can make it much easier to handle vacation schedules and to ensure that no
area of the environment is left completely uncovered at any time.
84
The Reference Guide to Data Center Automation
Asset Management
The goal of asset management is to track the fixed assets that an organization owns and controls.
From a general standpoint, asset management can include everything ranging from racks to
servers and from buildings to storage devices. IT departments are often tasked with managing
budgets and keeping track of inventory, even in highly distributed environments. Without proper
technology and processes in place, it can quickly become difficult to find and manage all of these
assets. The following sections focus on what to track and how to develop a process that will
allow for easily maintaining information about hardware, software, and other important aspects
of a distributed IT environment.
85
The Reference Guide to Data Center Automation
86
The Reference Guide to Data Center Automation
In addition to basic information, IT departments should consider capturing details related to the
initial and ongoing costs for the device, details about its purpose, and any configuration
information that can help in troubleshooting and management.
87
The Reference Guide to Data Center Automation
It’s likely that IT departments will need to take into account other types of devices, as well. For
example, if a business uses specialized hardware in a testing lab, that hardware should be
included. Additionally, IT departments should take into account assets that are committed to
remote sites.
88
The Reference Guide to Data Center Automation
For organizations that have implemented the IT Infrastructure Library (ITIL) best practices, the
Software Asset Management topic can be particularly useful. For more information, see the ITIL Web
site at http://www.itil.co.uk/.
89
The Reference Guide to Data Center Automation
Automated Discovery
One of the largest barriers related to implementing asset management is the difficulty associated
with collecting data about all the devices that must be supported in an IT environment. In some
cases, this task might be done manually by physically or remotely connecting to each device and
recording details. Of course, apart from the tedium of the process, it’s easy for certain devices to
be overlooked altogether.
Many asset management solutions can leverage an automated discovery feature to
programmatically scan the network and find devices and nodes that are currently active. The
process can often be performed very quickly and can include details about devices located
throughout a distributed environment. Furthermore, routine audits can be performed to ensure
that devices are still available and to track any changes that might have occurred.
Reporting
The key goal of asset management is to facilitate reporting. IT managers should be able to
generate on-demand information about hardware, software, and network devices, as needed.
Many asset management solutions will provide the ability to create real-time reports. Products
often allow for Web-based report design and customization. By making asset-related information
available to managers throughout the organization, IT departments can better ensure that they are
meeting overall business needs. Overall, by developing an asset management approach and
selecting an appropriate data center automation tool, IT organizations can realize the benefits of
tracking the devices they support with minimal time and effort.
90
The Reference Guide to Data Center Automation
Flexible/Agile Management
In just about any IT environment, changes are frequent and inevitable. Successful businesses
must often make significant modifications to business and technical processes to keep pace with
customer demands and increasing competition. In business and IT terms, agility refers to the
ability to quickly and efficiently adapt to changes. The faster an IT organization can react to
changes, the better aligned it will be with business units—and that will translate to overall
success for the entire enterprise.
91
The Reference Guide to Data Center Automation
92
The Reference Guide to Data Center Automation
Automating IT Management
Obviously, all these requirements related to automating IT management can necessitate a
significant amount of expertise, time, and effort. As with many other areas of improving IT
efficiency, data center automation tools can significantly help IT departments increase their
flexibility. Especially when budgets and personnel resources are limited, investments in
automation can decrease the overhead related to changes.
Specific areas from which organizations can benefit include change and configuration
management, server and network provisioning deployment, automatic updates, asset
management, and reporting. For example, there are significant benefits to storing all IT-related
information in a centralized Configuration Management Database (CMDB). The combined data
can help IT and business leaders quickly identify which systems might need to be updated to
accommodate business changes.
Overall, the process of making an IT department more flexible and agile can provide tremendous
advantages throughout an entire organization. By quickly adapting to changing needs, the role of
IT can transform from a rate-of-change limitation to a strategic advantage. And, through the use
of data center automation technology and best practices, IT organizations can quickly work
towards the features that can help make them agile.
93
The Reference Guide to Data Center Automation
Policy Enforcement
Well-managed IT departments are characterized by having defined, repeatable processes that are
communicated throughout the organization. However, sometimes that alone isn’t enough—it’s
important for IT managers and systems administrators to be able to verify that their standards are
being followed throughout the organization.
Types of Policies
Policies can take many forms. For example, one common policy is related to password strength
and complexity. These requirements usually apply to all users within the organization and are
often enforced using technical features in operating systems (OSs) and directory services
solutions. Other types of policies might define response times for certain types of issues or
specify requirements such as approvals before important changes are made. Some policies are
mandated by organizations outside of the enterprise’s direct control. The Health Insurance
Portability and Accountability Act (HIPAA), the Sarbanes-Oxley Act, and related governmental
regulations fall into this category.
Defining Policies
Simply defined, policies specify how areas within an organization are expected to perform their
responsibilities. For an IT department, there are many ways in which policies can be used. On
the technical side, IT staff might create a procedure for performing system updates. The
procedure should include details of how downtime will be scheduled and any related technical
procedures that should be followed. For example, the policy might require systems
administrators to verify system backups before performing major or risky changes.
On the business and operations side, the system update policy should include details about who
should be notified of changes, steps in the approvals process, and the roles of various members
of the team, such as the service desk and other stakeholders.
94
The Reference Guide to Data Center Automation
Figure 27: The typical areas of an organization that should be involved in creating policies.
IT departments should ensure that policies such as those that apply to passwords, email usage,
Internet usage, and other systems and services are congruent with the needs of the entire
organization. In some cases, what works best for IT just doesn’t fit with the organization’s
business model, so compromise is necessary. The greater the “buy-in” for a policy initiative, the
more likely it is to be followed.
95
The Reference Guide to Data Center Automation
Communicating Policies
Policies are most effective when all members of the organization understand them. In many
cases, the most effective way to communicate a policy is to post it on an intranet or other shared
information site. Doing so will allow all staff to view the same documentation, and it will help
encourage updates when changes are needed.
Policy Scope
Another consideration related to defining policies is determining how detailed and specific
policies should be. In many cases, if policies are too detailed, they may defeat their purpose—
either IT staffers will ignore them or will feel stifled by overly rigid requirements. In those cases,
productivity will suffer. Put another way, policy for the sake of policy is generally a bad idea.
When writing policies, major steps and interactions should be documented. For example, if a
policy requires a set of approvals to be obtained, details about who must approve the action
should be spelled out. Additional information such as contact details might also be provided.
Ultimately, however, it will be up to the personnel involved to ensure that everything is working
according to the process.
96
The Reference Guide to Data Center Automation
97
The Reference Guide to Data Center Automation
Server Monitoring
In many IT departments, the process of performing monitoring is done on an ad-hoc basis. Often,
it’s only after numerous users complain about slow response times or throughput when accessing
a system that IT staff gets involved. The troubleshooting process generally requires multiple
steps. Even in the best case, however, the situation is highly reactive—users have already run
into problems that are affecting their work. Clearly, there is room for improvement in this
process.
Note that the process can be repeated, based on the needs of the environment. The key point is that
solid performance-related information is required in order to support the process.
98
The Reference Guide to Data Center Automation
Monitoring Availability
If asked about the purpose of their IT departments, most managers and end users would specify
that it is the task of the IT department to ensure that systems remain available for use. Ideally, IT
staff would be alerted when a server or application becomes unavailable, and would be able to
quickly take the appropriate actions to resolve the situation.
There are many levels at which availability can be monitored. Figure 30 provides an overview of
these levels. At the most basic level, simple network tests (such as a PING request) can be used
to ensure that a specific server or network device is responding to network requests. Of course,
it’s completely possible that the device is responding, but that it is not functioning as requested.
Therefore, a higher-level test can verify that specific services are running.
Tests can also be used to verify that application infrastructure components are functioning
properly. On the network side, route verifications and communications tests can ensure that the
network is running properly. On the server side, isolated application components can be tested by
using procedures such as test database transactions and HTTP requests to Web applications. The
ultimate (and most relevant) test is to simulate the end-user experience. Although it can
sometimes be challenging to implement, it’s best to simulate actual use cases (such as a user
performing routine tasks in a Web application). These tests will take into account most aspects of
even complex applications and networks and will help ensure that systems remain available for
use.
99
The Reference Guide to Data Center Automation
Monitoring Performance
For most real-world applications, it’s not enough for an application or service to be available.
These components must also respond within a reasonable amount of time in order to be useful.
As with the monitoring of availability, the process of performance monitoring can be carried out
at many levels. The more closely a test mirrors end-user activity, the more relevant will be the
performance information that is returned. For complex applications that involve multiple servers
and network infrastructure components, it’s best to begin with a real-world case load that can be
simulated. For example, in a typical Customer Relationship Management (CRM) application,
developers and systems administrators can work together to identify common operations (such as
creating new accounts, running reports, or updating customers’ contact details). Each set of
actions can be accompanied by expected response times.
All this information can help IT departments proactively respond to issues, ideally before users
are even aware of them. As businesses increasingly rely on their computing resources, this data
can help tremendously.
100
The Reference Guide to Data Center Automation
Although tools such as the Windows System Monitor utility can help monitor one or a few
servers, it quickly becomes difficult to manage monitoring for an entire environment. Therefore,
most systems administrators will use these tools only when they must troubleshoot a problem in
a reactive way. Also, it’s very easy to overlook critical systems when implementing monitoring
throughout a distributed environment. Overall, there are many limitations to the manual
monitoring process. In the real world, this means that most IT departments work in a reactive
way when dealing with their critical information systems.
101
The Reference Guide to Data Center Automation
102
The Reference Guide to Data Center Automation
Overall, through the use of data center automation tools, IT departments can dramatically
improve visibility into their environments. They can quickly and easily access information that
will help them more efficiently troubleshoot problems, and they report on the most critical aspect
of their systems: availability and performance.
Change Tracking
An ancient adage states, “The only constant is change.” This certainly applies well to most
modern IT environments and the businesses they support. Often, as soon as systems are
deployed, it’s time to update them or make modifications to address business needs. And keeping
up with security patches can take significant time and effort. Although the ability to quickly
adapt can increase the agility of organizations as a whole, with change comes the potential for
problems.
103
The Reference Guide to Data Center Automation
Establishing Accountability
It’s no secret that most IT staffers are extremely busy keeping up with their normal tasks.
Therefore, it should not be surprising that network and systems administrators will forget to
update change-tracking information. When performed manually, policy enforcement generally
becomes a task for IT managers. In some cases, frequent reminders and reviews of policies and
processes are the only way to ensure that best practices are being followed.
104
The Reference Guide to Data Center Automation
Table 8 shows an example of a simple, spreadsheet-based audit log. Although this system is
difficult and tedious to administer, it does show the types of information that should be collected.
Unfortunately, it does not facilitate advanced reporting, and it can be difficult to track changes
that affect complex applications that have many dependencies.
105
The Reference Guide to Data Center Automation
By using a Configuration Management Database (CMDB), all change and configuration data can
be stored in a single location. When performing troubleshooting, systems and network
administrators can quickly run reports to help isolate any problems that might have occurred due
to a configuration change. IT managers can also generate enterprise-wide reports to track which
changes have occurred. Overall, automation can help IT departments implement reliable change
tracking while minimizing the amount of overhead incurred.
106
The Reference Guide to Data Center Automation
Unauthorized Changes
In stark contrast to authorized changes that have the best of intentions, network-related changes
might also be committed by unauthorized personnel. In some cases, a junior-level network
administrator might open a port on a firewall at the request of a user without thoroughly
considering the overall ramifications. In worse situations, a malicious attacker from outside the
organization might purposely modify settings to weaken overall security.
107
The Reference Guide to Data Center Automation
Of course, there are obvious drawbacks to this manual process. The main issue is that the
information is only useful when all members of the network administration team place useful
information in the “system.” When data is stored in spreadsheets or other files, it’s also difficult
to ensure that the information is always up to date.
108
The Reference Guide to Data Center Automation
Once a change is ready to be deployed, the network automation utility can take care of
committing the changes automatically. Hundreds of devices can be updated simultaneously or
based on a schedule. Best of all, network administrators need not connect to any of the devices
directly, thereby increasing security.
Notification Management
It’s basic human nature to be curious about how IT systems and applications are performing, but
it can become a mission-critical concern whenever events related to performance or availability
occurs. In those cases, it’s the responsibility of the IT department to ensure that problems are
addressed quickly and that any affected members of the business are notified of the status.
109
The Reference Guide to Data Center Automation
Creating Notifications
In many IT environments, IT departments are notorious for delivering vague, ambiguous, and
overly technical communications. The goal for the content of notifications is to make them
concise and informative in a way that users and non-technical management can understand.
110
The Reference Guide to Data Center Automation
• What to expect—The notification should include details about the current and expected
effects of the problem. In some cases, systems and network administrators might need to
reboot devices or cause additional downtime in unrelated systems. If time windows are
known, it’s a good idea to include those details as well.
• Any required actions—If users are expect to carry out any particular tasks or make
changes to their normal processes, this information should be spelled out in the
notification. If emergency processes are in place, users should be pointed to the
documentation. If not, a point-person (such as a department manager) should be specified
to make the determinations related to what users should do.
• Which users and systems are affected—Some recipients of notifications might be
unaware of the problem altogether. The fact that they’re receiving a notification might
indicate that they should be worried. If it’s likely that some recipients will be able to
safely ignore the message, this should also be stated clearly. The goal is to minimize any
unnecessary disruption to work.
• Reassurance—This might border on the “public relations” side of IT management, but
it’s important for users to believe that their IT departments are doing whatever is possible
to resolve the situation quickly. The notification might include contact information for
reporting further problems, and can refer users to any posted policies or processes that
might be relevant to the downtime.
Although this might seem like a lot of information to include, in many cases, it can be summed
up in just a few sentences. The important point is for the message to be concise and informative.
111
The Reference Guide to Data Center Automation
The messages themselves can use a uniform format based on a predefined template. Fields for
common information such as “Affected Systems,” “Summary,” and “Details” can also be
defined. This can make it much easier for Service desk staff to respond to common queries about
applications. Finally, the system can keep track of who was notified about particular issues, and
when a response was taken. Overall, automated notifications can go a long way toward keeping
IT staff and users informed of both expected and unexpected downtime and related issues. The
end result is a better “customer satisfaction” experience for the entire organization.
112
The Reference Guide to Data Center Automation
Server Virtualization
Virtualization refers to the abstraction between the underlying physical components of an IT
architecture and how it appears to users and other devices. The term virtualization can be applied
to network devices, storage environments, databases, other portions of an IT infrastructure, and
servers. Simply put, server virtualization is the ability to run multiple independent operating
systems (OSs) concurrently on the same hardware.
Understanding Virtualization
The concept of running multiple “virtual machines” on a single computer can be traced back to
the days of mainframes. In that architecture, many individual computing environments or
sessions can be created on a single large computer. Although each session runs in what seems
like an isolated space, the underlying management software and hardware translates users’
requests and commands so that users can access the same physical hardware. The benefits
include scalability (many virtual machines can run simultaneously on the same hardware) and
manageability (most administration is handled centrally and client-side hardware requirements
are minimal).
Virtualization Architecture
For modern computing environments, virtualization solutions can be quickly and easily installed
on standard hardware. Figure 35 shows a generic example of one way in which virtualization can
be implemented.
113
The Reference Guide to Data Center Automation
At the bottom of the figure is the actual physical hardware—the CPU, memory, hard disks,
network adapters, and other components that make up the complete system. Running atop the
hardware is the OS, which includes device drivers that interact with physical system
components. Moving up the stack, within the OS is a virtualization management layer. This layer
allows for the creation of multiple independent virtual machine environments. The virtualization
layer may run as an application or as a service (depending on the product). Finally, at the top of
the “stack” are the virtual machines. It is at this level that multiple OSs can run simultaneously.
The job of the virtualization layer is to translate and coordinate calls from within each virtual
machine to and from the underlying hardware. For example, if the Linux-based OS within a
virtual machine requests access to a file, the virtualization management application translates the
request and redirects it to the actual file that represents a virtual hard drive on the host file
system. Figure 36 shows an example of how a Microsoft Virtual Server 2005-based
virtualization stack might look.
114
The Reference Guide to Data Center Automation
Figure 36: An example of a virtualization configuration using Microsoft Virtual Server 2005 R2.
Virtualization Terminology
Virtualization provides new ways in which to refer to standard computer resources, so it’s
important to keep in mind some basic terminology. The physical computer on which the
virtualization platform is running is known as the host computer and the primary OS is referred
to as the host OS. The OSs that run on top of the virtualization platform are known as guest OSs.
An additional concept to keep in mind is the virtual hard disk. From the perspective of the guest
OS, these files appear to be actual physical hard disks. However, physically, they’re stored as
files within the host OS file system.
Finally, another major advantage of virtual machines is that they can be “rolled back” to a
previous state. This is done by keeping track of all write operations and storing them in a file that
is separate from the primary virtual hard disk.
115
The Reference Guide to Data Center Automation
The following sections focus on the type of virtualization described in Figure 36.
Benefits of Virtualization
The list of benefits related to working with virtual machines is a long one. Let’s take a brief look
at some of the most relevant advantages from the standpoint of data center management:
• Increased hardware utilization—By allowing multiple virtual machines to run
concurrently on a single server, overall resource utilization can be dramatically improved.
This benefit can lead to dramatic cost reductions in data center environments, without
significant costs for upgrading current hardware.
• Hardware independence—One of the major challenges related to managing data center
environments is dealing with heterogeneous hardware configurations. Although it’s easy
to physically relocate an array of hard disks to another machine, chances are good that
OS and device driver differences will prevent it from working smoothly (if at all). On a
given virtualization platform, however, virtual machines will use a standardized virtual
environment that will stay constant regardless of the physical hardware configuration.
• Load-balancing and portability—Guest OSs are designed for compatibility with the
virtualization platform (and not the underlying hardware), so they can easily be moved
between host computers. This process can allow users and systems administrators to
easily make copies of entire virtual machines or to rebalance them based on overall server
load. Figure 37 provides an illustration. This method allows systems administrators to
optimize performance as business and performance needs change over time. In addition,
it’s far easier than manually moving applications or reallocating physical servers.
116
The Reference Guide to Data Center Automation
117
The Reference Guide to Data Center Automation
Virtualization Scenarios
Earlier, we mentioned how virtualization can help data center administrators in the area of server
consolidation. This, however, is only one of the many ways in which this technology can be
used. Others include:
• Agile management—As virtual machines can be created, reconfigured, copied, and
moved far more easily than can physical servers, virtualization technology can help IT
departments remain flexible enough to accommodate rapid changes.
• Support for legacy applications—IT departments are commonly stuck with supporting
older servers because applications require OSs that can’t run on newer hardware. The
result is higher support costs and decreased reliability. By placing these application
within a virtual machine, the application can be moved to newer hardware while still
running on an older OS.
• Software development and testing—Developers and testers often require the ability to
test their software in many configurations. Virtual machines can easily be created for this
purpose. It’s easy to copy virtual machines to make, for example, changes to the service
pack level. Additionally, whenever a test is complete, the virtual machine can be reverted
to its original state to start the process again.
• Training—Dozens of virtual machines can be hosted on just a few physical servers, and
trainers can easily roll back changes before or after classes. Students can access their
virtual machines using low-end client terminals or even over the Internet. Usually, it’s far
easier to maintain a few host servers than it is to maintain dozens of client workstations.
Limitations of Virtualization
Despite the many benefits and applications of virtualization technology, there are scenarios in
which this approach might not be the perfect solution. The first and foremost concern for most
systems administrators is that of performance. All virtualization solutions will include some level
of overhead due to the translation of hardware calls between each virtual machine and physical
hardware device. Furthermore, virtual machines are unaware of each other, so competition for
resources such as CPU, memory, disk, and network devices can become quite high. Overall, for
many types of applications and services, organizations will likely find that the many benefits of
virtualization will outweigh the performance hit. The key point is that IT departments should do
as much performance testing as possible before rolling out virtualized applications.
There are additional considerations to keep in mind. For example, for physical servers that are
currently running at or near capacity, it might make more sense to leave those systems as they
are. The same goes for complex multi-tier applications that may be optimized for a very specific
hardware configuration. Additionally, for applications require custom hardware that is not
supported by the virtualization platform (for example, 3-D video acceleration), running within a
virtual machine will not be an option. Over time, virtualization solutions will include increasing
levels of hardware support, but in the mean time, it’s important to test and verify your
requirements before going live with virtualization.
118
The Reference Guide to Data Center Automation
119
The Reference Guide to Data Center Automation
Technical Issues
In some ways, technology has come to the rescue: network bandwidth is more readily available
(and at a lower cost) than it has been in the past, and establishing physical network connectivity
is usually fairly simple. In other ways, improvements in technology have come with a slew of
new problems. Storage requirements often grow at a pace that far exceeds the capacity of
devices. In addition, almost all employees of modern organizations have grown accustomed to
high-bandwidth, low-latency network connections regardless of their locations. IT departments
must meet these demands while working within budget and resource constraints.
Perhaps one of the most pertinent issues related to remote office management is that of network
bandwidth. Usually the total amount of bandwidth is constrained, and factors such as latency
must be taken into account. This process has often lead to remote office systems being less
frequently updated. Servers sitting in a wiring closet of a branch office are often neglected and
don’t get the attention they deserve. The result is systems that are likely out of compliance with
IT policies and standards.
Personnel Issues
Ideally, organizations would be able to place senior-level systems and network administrators at
each remote office. Unfortunately, cost considerations almost always prohibit this. Therefore,
certain tasks must be performed manually (and often by less-trained individuals). Common tasks
include the installation of security updates or the management of backup media. Dedicated
technical staff is not available, so it’s common for these important operations to be overlooked or
to be performed improperly. Even when using remote management tools, some tasks cannot
easily be accomplished from a remote location.
Business Issues
Functions served by remote offices can be mission critical for many of an organization’s
operations. From a business standpoint, new initiatives and changes in standard operating
procedures must apply through the entire organization. The concept of “out of sight, out of
mind” simply is not acceptable for remote locations. All of the hardware, software, and network
devices that are under IT’s supervision must be maintained to ensure overall reliability and
security.
120
The Reference Guide to Data Center Automation
121
The Reference Guide to Data Center Automation
Patch Management
One of the least glamorous but still important tasks faced by systems and network administrators
is that of keeping their hardware and software up to date. The benefits of applying patches for all
devices within an environment can range from reducing security vulnerabilities to ensuring
reliability and uptime. More importantly, the cost of not diligently testing and applying updates
can be extremely high.
122
The Reference Guide to Data Center Automation
Obtaining Updates
It’s important for IT staff to be aware of new updates and patches as soon as possible after
they’re released. Although many vendors provide newsletters and bulletins related to updates,
most IT environments must continuously monitor many sources for this information. This
requirement makes it very likely that some updates will be overlooked.
Testing Updates
A sad-but-true fact about working in IT is that sometimes the “cure” can be worse than the
disease. Software and hardware vendors are usually under a tremendous amount of pressure to
react to vulnerabilities once they’re discovered, and it’s possible that these updates will introduce
new bugs or may be incompatible with certain system configurations. This reality highlights the
need for testing an update. Developers and systems administrators should establish test
environments that can be used to help ensure that a patch does not have any unintended effects.
123
The Reference Guide to Data Center Automation
Deploying Updates
Assuming that a patch has passed the testing process, it’s time to roll out the update to systems
throughout the environment. Ideally, it will be possible to deploy all the changes simultaneously.
More likely, however, the need for system reboots or downtime will force IT departments to
work within regularly scheduled downtime windows.
Auditing Changes
Once patches have been deployed, it’s important to verify that all systems have been updated.
Due to technical problems or human error, it’s possible that some systems were not correctly
patched. When done manually, this portion of the process often requires the tedious step of
logging into each server or manually running a network scanning tool.
124
The Reference Guide to Data Center Automation
The process begins with the detection of new patches. Ideally, the system will automatically
download the appropriate files. If systems administrators determine that the update is relevant
and that it should be tested, they can instruct the solution to deploy the update to a test set of
servers. They can then perform any required testing. If the update passes the tests, they can
instruct the automated patch management system to update the relevant devices. Patches are then
applied and verified based on the organization’s rules. The entire process is often reduced to a
small fraction of the total time of performing these steps manually.
125
The Reference Guide to Data Center Automation
Network Provisioning
Perhaps the most critical portion of modern IT environments is the underlying network
infrastructure. Almost all applications, workstations, and servers depend on connectivity in order
to get their jobs done. In the “old days” of computing, networks were able to remain largely
static. Although new switches may be added occasionally to support additional devices, the
scope of the changes was limited. In current environments, the need to react to rapidly changing
business and technical needs has made the process of network provisioning increasingly
important.
Common operations include the deployment of new devices and making network-wide changes.
Additional tasks include making sure that devices are configured as expected and that they meet
the organization’s business and technical requirements. Figure 41 provides an overview of the
types of tasks that are required to perform network provisioning. Let’s take a look at some of
these requirements in more detail, and how using an automated network provisioning solution
can help.
126
The Reference Guide to Data Center Automation
127
The Reference Guide to Data Center Automation
128
The Reference Guide to Data Center Automation
All these layers work together to form the links in an organization’s armor. For example, before
an employee or consultant can access a specific database application, the employee will first
have to have access to a physical network port. He or she will then be verified at the network and
server levels, and finally at the application level. The user must meet all these challenges in order
to be able to access the application.
129
The Reference Guide to Data Center Automation
Security Protocols
For managing authentication in a distributed network environment, one of the most common
protocols is Kerberos. This protocol allows computer systems to be able to positively identify a
user in a secure way. It can help avoid security problems such as the interception of security
credentials through the use of encryption. Generally, Kerberos is implemented at the server or
the application level. However, network devices and other components can also take advantage
of it.
There are also several other authentication methods that can be used. Older versions of the
Microsoft Windows platform use the NTLM authentication protocol and method. Although this
method is less secure than Kerberos, NTLM is a widely supported standard that might still be
required to support down-level clients and servers. Also, numerous Lightweight Directory
Access Protocol (LDAP)-compliant solutions can integrate with or replace AD. Remote
Authentication Dial-In User Service (RADIUS), which was originally developed for the purpose
of authenticating remote users, can help centralize security for mobile users and remote
locations.
Authentication Mechanisms
The goal of authentication is to ensure that a specific user is who he or she claims to be. By far,
the most common authentication mechanism is through the use of a login and password
combination. Although this method meets basic security requirements, it has numerous
drawbacks. First, users are forced to memorize these pieces of information, and handling lost
passwords is a tedious and time-consuming process. Additionally, passwords can be shared or
stolen, making it possible that a person is not actually being positively identified. So much is
dependent on having the right credentials that this method leaves much room for improvement.
Newer authentication mechanisms include biometrics and the use of specialized security devices.
Biometric devices are most commonly based on the use of fingerprints or voice identification to
identify individuals. Other methods such as retinal scans are available (though they’re most
commonly seen in spy movies). Security devices such as an encryption card or “fob” can also be
used to verify individuals’ identities, especially for remote access. All of these methods involve a
certain level of management overhead, and IT departments must be able to keep track of security
principals, regardless of the method used.
130
The Reference Guide to Data Center Automation
Authorization
Figuring out how administrators can control access to a system is only part of the security
puzzle. Just as important is defining what exactly these users can do. Restrictions can range from
determining which files and folders can be accessed to limiting the time of day during which a
user can log on. Authorization is the process of granting permission to security principals (such
as users or computers) in order to granularly manage what tasks they can perform.
Overall, by integrating the management of overall security, IT departments and organizations can
be sure that all their systems remain coordinated and that only authorized users can access the
network.
131
The Reference Guide to Data Center Automation
Business Processes
An important characteristic of successful businesses is a strong alignment of the efforts between
multiple areas of the organization. This arrangement rarely occurs by itself—instead, it requires
significant time and effort from organizational leaders. The end result is often the creation of
processes that define how all areas of the enterprise should work together to reach common
goals.
132
The Reference Guide to Data Center Automation
133
The Reference Guide to Data Center Automation
Developing Processes
When it comes to deciding who should be involved in developing processes, the general rule of
thumb is the more, the better. Although it might be tempting for managers to take a top-down
approach to defining processes or for a single business manager to document the details, it’s
much better to solicit the input of all those that are involved. Many operations and tasks have
effects that are felt outside of the immediate realm of a single department. Therefore, it’s
important to ensure coordination with other portions of the business.
Specifically, there are several roles that should be represented during the creation of a process.
Business leaders from all areas of the organization should be welcome. Additionally,
stakeholders whose jobs will be directly affected by the process should drive the process. This
might include employees ranging from hands-on staff members to executive management
(depending on the scope of the process). An organized process for implementing ideas and
reviewing documentation drafts can go a long way toward keeping the development process
humming along. At the risk of sounding like a half-baked management fad, it’s often helpful to
have a process for creating processes.
Specific details include the owner of the document—the individual or group that is responsible
for defining and maintaining the process. Other details include who is affected by the process,
and the roles that might be required. The actual steps of the process can be defined in a variety of
ways. Although text might be useful as a basis, flowcharts and other visual aids can help
illustrate and summarize the main points very effectively.
134
The Reference Guide to Data Center Automation
135
The Reference Guide to Data Center Automation
Most of the information in this section is adaptable to organizations of just about any size.
136
The Reference Guide to Data Center Automation
In addition, this table could include details about any service level agreements (SLAs) that the IT
department has created as well as target issue resolution times. Of course, manual judgment will
always be required on the part of service desk staff. Still, the goal should be to capture and route
important information as accurately as possible.
137
The Reference Guide to Data Center Automation
138
The Reference Guide to Data Center Automation
Notice that in this document, there are many decision points and branching logic that will affect
the path to issue resolution. The major areas of ownership start at the left and begin with the
reporting of an issue (which can be from any area of the organization). The Level-1 staff is
responsible for categorizing the issues and determining the next steps. The issue may be resolved
at this level or it may be moved on to other members of the staff. At all points, the issue is owned
by an individual or a group. In this particular flowchart, it is ultimately the responsibility of the
Level-1 staff to ensure that an issue is closed.
Although this flowchart may not be perfect, it is easy to read and provides a simple overview of
many portions of the process. Most IT organizations will also want to accompany the flowchart
with additional details such as definitions of terms and steps involved in procedures.
For more information about ITIL, see the ITIL Web site at http://www.itil.co.uk.
Numerous third-party products and software solutions are also available. Some products are very
customizable, while others introduce their own suggested workflows, terminology, and best
practices.
When evaluating potential service desk solutions, IT organizations should start by looking at
their overall needs. For example, some solutions might better lend themselves to the support of
customers that are external to an organization (by allowing for fee-based support and related
features); others might be more appropriate for internal IT service desks. In some cases, an
enterprise might decide to build its own service desk solution. Although doing so can lead to a
system that is well-aligned with business goals, the time, cost, and maintenance effort required
might not lead to a strong enough business case for this approach.
Regardless of the approach and the technology selected, the implementation of an organized
service desk process is an excellent example of how IT organizations can benefit from the
implementation of business processes.
139
The Reference Guide to Data Center Automation
140
The Reference Guide to Data Center Automation
An adage related to technical project management specifies that organizations can choose to
define two of the following: scope, timeliness, and quality. For example, if the project deadline is
most important, followed by quality, then it’s quite possible that the scope (the list of included
features and functionality) might need to be reduced (see Figure 47).
141
The Reference Guide to Data Center Automation
Overall, the goal of the Executive Action Committee is to better align IT with the needs of the
organization. By ensuring that input is gained from throughout the organization and by
prioritizing the projects that can provide the most “bang for the buck,” enterprises can be sure to
maximize the value of their IT investments.
142
The Reference Guide to Data Center Automation
Authentication Mechanisms
By far, the most commonly used method of computer-based authentication is through the use of
a login and password combination. Although this method is relatively easy to implement, it
comes with significant burdens. Users are responsible for generating and remembering their own
passwords. They should choose strong passwords, but they’re often required to enter them
multiple times per day.
From an IT standpoint, devices such as routers and security accounts for use by applications and
services also often have passwords. Creating and maintaining these passwords can be a difficult
and time-consuming process. From a security standpoint, it can also be difficult to determine
whether a password has been shared, compromised, or used in an authorized way. All too often,
“secrets” are shared. Considering that organizations often have many thousands of passwords
and accounts, this can be a major security-related liability.
143
The Reference Guide to Data Center Automation
144
The Reference Guide to Data Center Automation
145
The Reference Guide to Data Center Automation
Centralized Security
So far, we’ve looked at several authentication mechanisms (with a focus on password-based
authentication). Let’s explore the process of creating and managing security credentials in a
network environment. We’ll focus on the importance of implementing a centralized user
authentication system, but first let’s look at an alternative (and the many problems it can cause).
The most important aspect of decentralized security is that there are many security databases
within the organization. Each one is independent of the others and contains its own
authentication information. For example, every computer might have a separate account named
“SysAdmin.” Although it’s technically possible to manually synchronize the login information
(that is, to ensure that the same usernames and passwords are used on each machine), the process
is tedious and error-prone. Furthermore, maintaining even a few of these systems can quickly
become difficult and time consuming. The end result is often that security is not maintained:
Simple passwords are used, login information is changed infrequently, and passwords are often
written down or recorded in some other way.
146
The Reference Guide to Data Center Automation
Although simply setting up a decentralized security environment can be painful, the real risks are
in the areas of manageability. For example, what will happen if a password is compromised?
Even if IT staff can scramble to update the passwords on multiple devices, there is still a large
window of vulnerability. The new password also has to be communicated to the users that need
it—an inherently risky proposition. What if one or more devices are overlooked and continue to
run with the exposed authentication information? And this doesn’t even take into account the
effort that might be required to ensure that other computers and services that rely upon the login
are properly updated.
In case all of this isn’t incentive enough to see the drawbacks of decentralized security, let’s look
at one more motivator before moving on: Imagine the difficulty that end users will experience if
they must manually log on to each device or application on the network. The decrease in
productivity and frustration might be tantamount to not having a network at all. By now, it’s
probably obvious that decentralized security is not a very effective approach—even for the
smallest of IT organizations.
It’s easy to see how this method can alleviate much of the pain of maintaining many separate
security databases. IT administrators that are responsible for maintaining security can create
accounts in the security database. And, if a password or other user setting must be changed, it
can be done centrally.
147
The Reference Guide to Data Center Automation
148
The Reference Guide to Data Center Automation
149
The Reference Guide to Data Center Automation
150