Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
0% found this document useful (0 votes)
297 views

Implementation and Maintenance Student Guide

Uploaded by

hher
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
297 views

Implementation and Maintenance Student Guide

Uploaded by

hher
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 157

snapshot

Implementation &
Maintenance
Student Guide
2|

Copyright

Information in this document, including URL and other website references, represents the current view
of CommVault Systems, Inc. as of the date of publication and is subject to change without notice to you.

Descriptions or references to third party products, services or websites are provided only as a convenience
to you and should not be considered an endorsement by CommVault. CommVault makes no
representations or warranties, express or implied, as to any third party products, services or websites.

The names of actual companies and products mentioned herein may be the trademarks of their respective
owners. Unless otherwise noted, the example companies, organizations, products, domain names, e-mail
addresses, logos, people, places, and events depicted herein are fictitious.

Complying with all applicable copyright laws is the responsibility of the user. This document is intended
for distribution to and use only by CommVault customers. Use or distribution of this document by any
other persons is prohibited without the express written permission of CommVault. Without limiting the
rights under copyright, no part of this document may be reproduced, stored in or introduced into a
retrieval system, or transmitted in any form or by any means (electronic, mechanical, photocopying,
recording, or otherwise), or for any purpose, without the express written permission of CommVault
Systems, Inc.

CommVault may have patents, patent applications, trademarks, copyrights, or other intellectual property
rights covering subject matter in this document. Except as expressly provided in any written license
agreement from CommVault, this document does not give you any license to CommVault’s intellectual
property.

COMMVAULT MAKES NO WARRANTIES OF ANY KIND, EXPRESS OR IMPLIED, AS TO THE INFORMATION


CONTAINED IN THIS DOCUMENT.

©1999-2015 CommVault Systems, Inc. All rights reserved

CommVault, CommVault and logo, the “CV” logo, CommVault Systems, Solving Forward, SIM, Singular
Information Management, Simpana, CommVault Galaxy, Unified Data Management, QiNetix, Quick
Recovery, QR, CommNet, GridStor, Vault Tracker, InnerVault, QuickSnap, QSnap, Recovery Director,
CommServe, CommCell, IntelliSnap, ROMS, Simpana OnePass, CommVault Edge and CommValue, are
trademarks or registered trademarks of CommVault Systems, Inc. All other third party brands, products,
service names, trademarks, or registered service marks are the property of and used to identify the
products or services of their respective owners. All specifications are subject to change without notice.

All right, title and intellectual property rights in and to the Manual is owned by CommVault. No rights
are granted to you other than a license to use the Manual for your personal use and information. You
may not make a copy or derivative work of this Manual. You may not sell, resell, sublicense, rent, loan or
lease the Manual to another party, transfer or assign your rights to use the Manual or otherwise exploit
or use the Manual for any purpose other than for your personal use and reference. The Manual is
provided "AS IS" without a warranty of any kind and the information provided herein is subject to
change without notice.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Table of Contents | 3

Table of Contents

Table of Contents .................................................................................................................3


Introduction .............................................................................................................................7
Preliminaries .................................................................................................................................... 8
Education Advantage ....................................................................................................................... 9
CommVault® Advantage ................................................................................................................. 10
Education Services Forums ............................................................................................................. 11
Customer Education Lifecycle ......................................................................................................... 12
CommVault Certification ................................................................................................................ 13
Classroom Resources ...................................................................................................................... 14
CVLab and On Demand Lab Environment ........................................................................................ 15
Educational Resources .................................................................................................................... 16
Course Objective ............................................................................................................................ 17
Course Modules ............................................................................................................................. 18
Common Technology Engine ........................................................................................................... 19
Module 1 – Designing a CommCell® Environment .................................................................... 21
Topics ............................................................................................................................................ 22
Common Technology Engine Architecture ....................................................................................... 23
CommCell® Architecture Overview ........................................................................................................... 24
CommServe® Server .................................................................................................................................. 25
Indexing Structure ..................................................................................................................................... 27
Common Technology Engine Best Practices ............................................................................................. 28
Architecting a Storage Solution ....................................................................................................... 30
Simpana Deduplication................................................................................................................... 33
Understanding Simpana Deduplication .................................................................................................... 34
Deduplication Building Block Guidelines................................................................................................... 37
Deduplication Storage Options ................................................................................................................. 38
Partitioned Deduplication Database ......................................................................................................... 40
Enterprise Building Block Guidelines......................................................................................................... 41

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
4 | Table of Contents

SILO Storage .............................................................................................................................................. 42


Advanced Deduplication Configurations ................................................................................................... 43
Deduplication Best Practices ..................................................................................................................... 45
Designing a Sound Data Protection Strategy.................................................................................... 47
Disaster Recovery Concepts ...................................................................................................................... 48
Business Continuity Concepts ................................................................................................................... 52
Protection Methods .................................................................................................................................. 54
Data Description ........................................................................................................................................ 57
Data Availability......................................................................................................................................... 59
Protected Storage Requirements .............................................................................................................. 61
Designing a Sound Data Protection Strategy.................................................................................... 62
Understanding Client Agents .................................................................................................................... 63
Protecting Virtual Environments ............................................................................................................... 64
The VSA Backup Process............................................................................................................................ 66
Protecting Applications ............................................................................................................................. 68
Snapshot Management ............................................................................................................................. 70
Data Protection Best Practices .................................................................................................................. 73
Module 2 – CommCell® Deployment and Configuration........................................................... 77
Topics ............................................................................................................................................ 78
CommCell Deployment Process ...................................................................................................... 79
New CommCell® Deployment Process ...................................................................................................... 80
Existing CommCell® Upgrade Process ....................................................................................................... 81
CommCell® Disaster Recovery Process ..................................................................................................... 82
Environment Requirements ...................................................................................................................... 85
Installing CommServe Software ...................................................................................................... 86
Installing MediaAgent Software ...................................................................................................... 88
Index Cache Configuration ........................................................................................................................ 90
Library Detection and Configuration ............................................................................................... 92
Client Agent Deployment Methods ................................................................................................. 94
Standard Installation Methods - Interactive ............................................................................................. 95
CommCell® Console Push Install ............................................................................................................... 97

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Table of Contents | 5

Custom Installation Methods .................................................................................................................... 99


Deployment Best Practices ........................................................................................................... 101
Module 3 – Data Security and Network Control .................................................................... 104
Topics .......................................................................................................................................... 105
Storage Policy Design ................................................................................................................... 106
Storage-Based Design Strategy ............................................................................................................... 107
Business Based Design Strategy .............................................................................................................. 108
Deduplication’s Impact on Policy Design ................................................................................................ 109
How Many Storage Policies do I really need? ......................................................................................... 111
Understanding Alternate paths ............................................................................................................... 112
Advanced Storage Policy Features .......................................................................................................... 113
Storage Policy Design Best Practices ....................................................................................................... 116
Advanced Job Control................................................................................................................... 121
Understanding Services ........................................................................................................................... 122
Firewall Configuration ............................................................................................................................. 124
Network Control ...................................................................................................................................... 127
Network Throttling .................................................................................................................................. 130
Configuring Data Encryption ................................................................................................................... 131
Base Folder and Resource Pack Tools ............................................................................................ 133
Module 4 – Performance Tuning........................................................................................... 137
Topics .......................................................................................................................................... 138
Performance ................................................................................................................................ 139
Establishing Benchmarks ......................................................................................................................... 140
Storage Performance .............................................................................................................................. 142
Performance Parameters ........................................................................................................................ 146
Stream Management.................................................................................................................... 148
Data Streams ........................................................................................................................................... 149
Deduplication and Stream Management ................................................................................................ 152
Please Give us your Feedback! ...................................................................................................... 154
Index .................................................................................................................................. 155

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
6 | Table of Contents

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 7

Introduction

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
8 | Introduction

Preliminaries

The value of this course comes from three distinct areas – first, the content of the material which guides
your exploration and understanding of the product. Second, the skill of the instructor to expand on
those areas of interest and to add value from their experience with the product. And lastly, you, the
student whose questions and experiences help not only yourself but others in understanding how
Simpana® software can help you with your data management requirements.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 9

Education Advantage

The CommVault Education Advantage product training portal contains a set of powerful tools to enable
CommVault customers and partners to better educate themselves on the use of the CommVault
software suite. The portal includes:

 Training Self-Assessment Tools


 Curriculum Guidance based on your Role in your CommVault Enterprise
 Management of your CommVault Certifications
 Access to Practice Exams and Certification Preparation Tools
 And more!

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
10 | Introduction

CommVault® Advantage

CommVault® Advantage is your profile as a CommVault consumer and expert. The CommVault
Advantage system captures your certifications, participation in learning events and courses, your Forum
participation, Support interaction and much more. Through your CommVault interactions your awarded
Profile Points are collected and compared with other CommVault consumers worldwide. These Profile
Points allow our users to thoroughly demonstrate their Simpana® software expertise for personal and
professional growth. Login to CommVault Advantage to check your progress and compare yourself to
the Global CommVault community.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 11

Education Services Forums

The CommVault Education Services Forum is designed to disseminate information throughout our
technical community and enhance a student’s learning experience. If you have any questions regarding
our products, service, or certification program we encourage you to post your question or search for
answers on our forum. The CommVault Education Services Forum will not only contain questions from
our customers, it will also contain a wealth of knowledge from our education specialists!

To access the forums, simply log into (or create) your account on the CommVault Education Advantage
website and then click on Education Services Forums on the right-hand navigation pane.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
12 | Introduction

Customer Education Lifecycle

Before customers install CommVault® Simpana® software, they should have a basic understanding of the
product. This learning timeline illustrates the role of product education over the early years of owning
CommVault Simpana software. A lifecycle ranging from the pre-installation review of the "Introduction
to Simpana Software" eLearning module, to the pursuit of Master Certification.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 13

CommVault Certification

CommVault's Certification Program validates expertise and advanced knowledge in topics, including:
CommVault Core Fundamentals, Implementation and Maintenance, Preparing for Disaster Recovery and
more advanced Specialist and Master technologies. Certification is a valuable investment for both a
company and the IT professional. Certified personnel can increase a company's productivity, reduce
operating costs, and increase potential for personal career advancement.

CommVault's Certification Program offers Professional-level, Specialist-level and Master-level


certifications. This Program provides certification based on a career path, and enables advancement
through the program based on an individual’s previous experience and desired area of focus. It also
distinguishes higher-level certification from lower-level certification as a verified proof of expertise.

Key Points
 Certification is integrated with and managed through CommVault's online registration in the
Education Advantage Customer Portal.
 Cost of certification registration is included in the associated training course.
 Practice assessments are given in class at the end of each module.
 Students may take the online certification exam(s) any time after completing the course.
 Previous training course students (as validated by the registrar) can also request an opportunity
to take the online assessment exam at no charge.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
14 | Introduction

Classroom Resources

Course manuals and activity guides will be available for download for Instructor Led Training (ILT) and
Virtual Instructor Led Training (vILT) courses. It is recommended to download these documents the day
prior to attending class to ensure the latest document versions are being used.

Self-paced eLearning courses can be launched directly from the EA page. If an eLearning course is part of
an ILT or vILT course, it is a required prerequisite and should be viewed prior to attending class.

If an ILT or vILT class will be using the CommVault® virtual Lab environment, this button will be used to
launch the lab on the first day of class.

CommVault® certification exams can be launched directly from the EA page. If you are automatically
registered for an exam as part of an ILT or vILT course, will be available on final day of class. There is no
time limit on when the exams need to be taken but it is recommended to take them as soon as you feel
you are ready.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 15

CVLab and On Demand Lab Environment

The CommVault Virtual Lab (CVLab environment) is now available to our global customers. The CVLab
allows you access to a vital learning tool that provides a flexible method for gaining hands-on experience
with the Simpana® software platform. You will have anywhere/anytime access to a powerful lab
environment to practice installations, test configurations, review current version capabilities or review
any lab exercises. The CVLab shares a common console with our Education Advantage (EA) portal and is
accessible 24-hours a day up to the amount of connect time purchased.

The CVLab time can be purchased as standalone on-demand CVLab time, or to extend lab time for
training courses attended. Extending CVLab time must be purchased within 48-hours after class end
time in order to maintain your lab progress from the training course. Whether purchasing on-demand
or extending; CVLab connect time may be purchased in four hour blocks in any quantity. Access will be
available for 90 days from point of purchase and is priced at just one Training Unit per four hour block.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
16 | Introduction

Educational Resources

Ask the Educator Webinar Series


Review our on-demand Ask the Educator series for concept overviews and introductions to Education
Services. These sessions are part of our extensive program of courseware and information assets on the
Education Advantage portal and focuses on CommVault-centric valuable technology topics. Each session
is presented by highly experienced CommVault Educators as a practical technical summary of the
featured Simpana® software functionality.

Learning Bytes
Learning Bytes are small snippets of information on various features and functions of Simpana®
Software. These small bytes of information will help users understand many of the daily tasks of the
CommVault® Administrator. Each Learning Byte will offer information and in many cases, a
demonstration of a function within a CommCell® environment. They are small and easy to understand,
making it easy to grasp the many features of Simpana Software.

Training Units Calculator


CommVault® Systems Training Units enable you to stay current with the Simpana® software suite
through training programs that build your technical expertise. Training Units are an easy purchase
option that enable a single transaction, ensure access to courseware, and learning adapted to the needs
of your business for twelve-months from the purchase date of the training unit.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 17

Course Objective

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
18 | Introduction

Course Modules

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Introduction | 19

Common Technology Engine

The CommCell® environment is the logical management boundary for all components that protect,
move, store and manage the movement of data and information. All activity within the CommCell
environment is centrally managed through the CommServe® server. Users log on to the CommCell®
Console Graphical User Interface (GUI) which is used to manage and monitor the environment. Agents
are deployed to clients to protect production data by communicating with the file system or application
requiring protection. The data is processed by the agents and protected through MediaAgents to disk,
tape or cloud storage. Clients, MediaAgents and libraries can be in local or remote locations. All local
and remote resources can be centrally configured and managed through the CommCell console. This
allows for centralized and decentralized organizations to manage all data movement activities through a
single interface. All production data protected by agents, all MediaAgents and all libraries that are
controlled by a CommServe server is referred to as the CommCell environment.

Physical Architecture
A physical CommCell® environment is made up of one CommServe® server, one or more MediaAgents
and one or more Clients. The CommServe server is the central component of a CommCell environment.
It hosts the CommServe database which contains all metadata for the CommCell environment. All
operations are executed and managed through the CommServe server. MediaAgents are the
workhorses which move data from source to destination. Sources can be production data or protected
data and destinations can be disk, cloud or removable media libraries.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
20 | Introduction

Clients are production systems requiring protection and will have one or more Agents installed directly
on them or on a proxy server to protect the production data.

Logical Architecture
CommVault’s logical architecture is defined in two main areas. The first area depicts the logical
management of production data which is designed in a hierarchal tree structure. Production data is
managed using Agents. These agents interface natively with the file system or application and can be
configured based on specific functionality of data being protected. Data within these agents are grouped
into a data set (backup set, replication set, or archive set). These data sets represent all data the Agent is
designed to protect. Within the data set, one or more subclients can be used to map to specific data.
The flexibility of subclients is that data can be grouped into logical containers which can then be
managed independently in the CommVault protected environment.

The second area depicts managing data in CommVault protected storage. This is facilitated through the
use of storage policies. Storage policies are policy containers which contain one or more rule sets for
managing one or more copies of protected data. The first rule set is the primary copy. This copy
manages data being protected from the production environment. Additional secondary copies can be
created with their own rule sets. These rule sets will manage additional copies of data which will be
generated from existing copies within the CommVault protected environment. The rule sets define what
data will be protected (subclients), where it will reside (data path), how long it will be kept for
(retention), encryption options, and media management options.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 21

Module 1 – Designing a CommCell® Environment

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
22 | Designing a Com mCell ® Environment

Topics

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 23

Common Technology Engine Architecture

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
24 | Designing a Com mCell ® Environment

CommCell® Architecture Overview

The heart of any Simpana deployment is the CommServe® server. All activity is managed from this
central point and all backup and restore activity must be initiated from the CommServe server. A
Microsoft SQL database is used to store all CommServe configuration and job history data.

Data movement is conducted from source to destination using MediaAgents. One or more MediaAgents
can be used to move data providing greater flexibility and scalability.

Production data is managed by installing agents on physical hosts, virtual hosts or on proxy hosts. The
agent communicates with the file system or application being protected and uses native APIs and / or
scripting to conduct data protection operations. Physical and virtual hosts with agents installed are
referred to as clients.

Libraries are used to store protected data. CommVault software supports a wide range of library
configurations.

The CommServe server, MediaAgents, libraries and clients that communicate with one another make up
the CommCell® environment.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 25

CommServe® Server

Within a CommCell® environment there can only be one active CommServe server. For high availability
and failover there are several methods that can be implemented. The following information explains
each of these methods.

Hot / Cold Standby


A hot or cold standby CommServe server consists of a physical or virtual machine with the CommServe
software pre-installed. The DR backup Export process directs metadata exports to the standby
CommServe server. In the event that the production CommServe server is not available the standby
CommServe server can quickly be brought online.

Virtualization
Some customers with virtual environments are choosing to virtualize the production CommServe server.
A virtualized CommServe server has an advantage of using the hypervisor’s high availability functionality
(when multiple hypervisors are configured in a cluster) and reduces costs since separate CommServe
hardware is not required. Although this method could be beneficial, it should be properly planned and
implemented. If the virtual environment is not properly scaled the CommServe server could become a
bottleneck when conducting data protection jobs. In larger environments where jobs run throughout
the business day, CommServe server activity could have a negative performance impact on production
servers.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
26 | Designing a Com mCell ® Environment

When virtualizing the CommServe server it is still critical to run the CommServe DR backup. In the event
of a disaster the CommServe server may still have to be reconstructed on a physical server. Do not rely
on the availability of a virtual environment in the case of a disaster. Follow normal CommVault best
practices in protecting the CommServe metadata.

Clustering
The CommServe® server can be deployed in a clustered configuration. This will provide high availability
for environments where CommCell® operations run 24/7. A clustered CommServe server is not a DR
solution and a standby CommServe server must be planned for at a DR site. Clustering the CommServe
server is a good solution in large environments where performance and availability are critical.

Another benefit for using a clustered CommServe server is when using Simpana data archiving. Archiving
operations can be configured to create stub files which allow end users to initiate recall operations. For
the end user recall to complete successfully the CommServe server must be available.

CommServe DR Backup Process


By default every day at 10:00 AM the CommServe DR backup process is executed. This process will first
dump the CommServe SQL database to the <install path>\CommVault\Simpana\CommServeDR folder.
An Export process will then copy the folder contents to a user defined drive letter or UNC path. A
Backup phase will then back up the DR Metadata, registry hive and user defined log files to a location
based on the storage policy associated with the backup phase of the DR process. All processes,
schedules and export/backup location are customizable in the DR Backup Settings applet in Control
Panel.

CommServe DR IP Address
A CommCell® license is bound to the IP address of the CommServe server. In situations where a standby
CommServe server with a different IP address is going to be used, it must be included in the CommCell
license information.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 27

Indexing Structure

Simpana software uses a distributed indexing structure that provides for enterprise level scalability and
automated index management. This works by using the CommServe database to only retain job based
metadata which will keep the database relatively small. Job and detailed index information will be kept
on the MediaAgent protecting the job in the index cache location and automatically copied to media
containing the job.

Job summary data maintained in the CommServe database will keep track of all data chunks being
written to media. As each chunk completes it is logged in the CommServe database. This information
will also maintain media identities where the job was written to which can be used when recalling off-
site media back for restores. This data will be held in the database for as long as the job exists. This
means even if the data has exceeded defined retention rules, the summary information will still remain
in the database until the job has been overwritten or physically destroyed.

The detailed index information for jobs is maintained in the MediaAgent’s index cache. This information
will contain each object protected, what chunk the data is in, and the chunk offset defining the exact
location of the data within the chunk. The index files are stored in the index cache and after the data is
protected to media, an archive index operation is conducted to write the index to the media. This
method automatically protects the index information, eliminating the need to perform separate index
backup operations.

The archived index can also be used if the index cache is not available, when restoring the data at
alternate locations, or if the indexes have been pruned from the index cache location.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
28 | Designing a Com mCell ® Environment

Common Technology Engine Best Practices

CommServe® Server
The CommServe metadata database is the most critical component within the CommCell®
infrastructure. If the data becomes corrupt, the CommServe server disk crashes or you are faced with a
full site disaster situation, having the metadata backup readily accessible is critical.

Consider the following key points for proper metadata protection:


 Wherever you send protected data a copy of the DR Metadata should be included.
 If you have a standby CommServe server or a dedicated DR site with network accessibility set
the Export phase of the metadata backup to be written to that location.
 Make sure you properly secure the metadata backup since all configuration, security, licensing
and encryption key information is kept in the database.
 If data protection copies to tape are typically performed during the day and finish after 10 AM
consider setting a second schedule of the DR backup to ensure the most up to date metadata is
sent off-site with the data tapes.
 If tapes are sent off-site prior to 10 AM consider changing the default DR backup schedule to
ensure the most up to date metadata is sent off-site with backup tapes.
 If your environment is using the Erase Data feature make sure the metadata backup is going to a
dedicated DR storage policy or a storage policy with the Erase Data option deselected.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 29

Index Cache Settings


All object level data protection jobs will use indexes for all operations. These indexes are maintained in
the index cache. Improper configuration of the index cache can result in job failures and long delays in
browse and recovery operations.

Consider the following when designing and configuring the index cache:
 Index cache should be located on dedicated high speed disks, preferably solid state disks.
 Do NOT put the index cache on the system drive. Use a dedicated drive (recommended) or a
dedicated partition (for smaller environments). During MediaAgent installation the default path
for the index cache is the system drive. The location of the cache can be changed by selecting:
right-clicking the MediaAgent and selecting the properties Catalog tab.
 Size the index cache appropriately based on the size of your environment and the estimated
number of objects that will be protected. It is much better to overestimate than underestimate
index cache size. Sizing guidelines are available in the CommVault Online Documentation.
 The default retention time for the index cache is 15 days. If you will be frequently browsing for
data older than 15 days increase this setting and allocate enough disk space for the index cache.
 Index files are automatically backed up to media after each data protection job so there is no
need to perform backups of the index cache location.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
30 | Designing a Com mCell ® Environment

Architecting a Storage Solution

All data storage devices associated/configured with the Simpana® software are referred to as
Libraries. All data destined to and from a library must pass through a MediaAgent (or a NAS Filer). Data
written by the Simpana software is OS independent.

The most common supported library types are listed below. For a list of specific vendor devices consult
the Hardware Compatibility List on CommVault's Maintenance Advantage website. For a list of all
supported library types consult the CommVault Online Documentation.

Disk library - A disk library uses disk media configured for read/write access as one or more mount
paths. The disk library is a logical entity and does not represent a specific hardware entity. The storage
capacity of a disk library is determined by the total storage space in its mount paths.

Tape library- Tape libraries are made up of one or more tape devices with a library controller and
internal media storage. A Tape library can have mixed media and shared access with one or more
MediaAgents (on NAS Filers) in the same CommCell® group.

Blind Library - A blind library is a tape library without a barcode reader, and is the opposite of a sighted
library which has a barcode reader. A blind library must have all its drives (and media) of the same type.
Once configured, a blind library cannot be configured as a sighted library.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 31

IP Library - An IP Library provides LAN-based Media Management for multiple applications. Media
inventory, pools, and device loading/unloading is all managed by the library software. IP Libraries are
the only libraries that can be shared between CommCell® entities. Example: STK ACSLS or ADICS SDLC
libraries.

Stand-alone Tape Library - A single tape device with no library controller or internal storage that is
accessible from a MediaAgent. Stand-alone Tape drives can be pooled together for a multi-stream job or
single stream failover configuration.

NAS NDMP Library - A tape library attached to a NAS Filer for NDMP data storage. The library control
and drives in a NAS NDMP library can be dynamically shared between multiple devices (NAS file servers
and MediaAgents) if these devices are connected to the library in a SAN environment. The device
initially having library control (media changer) would be the first configured device.

Virtual Tape Library - A software representation of a tape library using disk storage. Virtual tape libraries
are supported, but not recommended because a normal disk library provides many more features and
capabilities.

Plug & Play Library - Plug and Play (PnP) storage devices (e.g., FireWire, USB, SATA storage devices, etc.)
can be used for storage instead of tapes. Once configured, PnP disks are treated like tapes in a Stand-
Alone drive. PnP libraries are useful in locations where it is hard to configure and manage tapes due to
operational issues. Only one PnP library can be configured per MediaAgent. Although multiple drives can
be configured, only single-streamed jobs are supported. (Multiple drives provide the ability to span
across multiple media for a single-streamed job.)

Cloud Library - A Cloud library uses online storage devices — cloud storage devices — as storage targets.
Cloud libraries provide a pay-as-you-go capability for network storage. Data is transferred through
secured channels using HTTPS protocol.

Removable Disk Drives - Removable Disk Drives can be configured as stand-alone drives. All operations
supported by stand-alone drives are supported by such devices. Removable disks differ from PnP disks in
that they are drive enclosure devices that retain a persistent drive letter in the Operating System,
regardless of whether or not a disk media is loaded into the enclosure.

Storage Connections
Direct Attached Storage (DAS)
Direct Attached Storage (DAS) means the production storage location is directly attached (not SAN) to
the production server. In situations where many production servers use DAS, there is no single point of
failure. The primary disadvantages are higher administrative overhead and depending on budget
limitations, lower quality storage being used instead of high quality enterprise class disks (typically
found in SAN/NAS storage).

For some applications such as Exchange 2010 using DAG (Database Availability Groups), Direct Attached
Storage may be a valid solution. The main point is that although the storage trend over the past several

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
32 | Designing a Com mCell ® Environment

years has been to storage consolidation, DAS storage should still be considered for certain production
applications.

One key disadvantage regarding DAS protection is that backup operations will likely require data to be
moved over a network. This problem can be reduced by using dedicated backup networks. Another
disadvantage is that DAS is not as efficient as SAN or NAS when moving large amounts of data.

Network Attached Storage (NAS)


Network Attached Storage (NAS) has made a strong comeback over the past few years by taking
advantage of its versatility. Where NAS was once only used as file stores they are now considered good
options for databases and virtual machines. NAS versatility includes the ability to attach Fibre or iSCSI
connections along with traditional NAS NFS/CIFS shares and has a primary advantage of device
intelligence using specifically designed operating systems to control and manage disks and disk access.
From a high availability and disaster recovery aspect, disk cloning or mirroring and replication provide
sound solutions. Simpana's IntelliSnap® integration with supported hardware provides simple yet
powerful snapshot management capabilities.

One key disadvantage of NAS is that it typically requires network protocols when performing data
protection operations. This disadvantage can be greatly reduced through the use of snapshots and proxy
based backup operations.

Storage Area Network (SAN)


Storage Area Networks (SAN) are very commonly implemented for the most mission critical systems
within an environment. The ability to consolidate storage using efficient data movement protocols, Fibre
channel and iSCSI provide flexibility and performance.

One key disadvantage of SAN is the complexity of configuring and managing SAN networks. Typically,
specialized training is required and all hardware must be fully compatible for proper operation. Since
SAN storage lacks the operating system that NAS storage has, it relies on a host system for data
movement. Depending on the configuration, the load of data movement can be offloaded to a proxy
and by adding Host Bus Adapters (HBA) connected to a dedicated backup SAN data can be more
efficiently backed up.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 33

Simpana Deduplication

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
34 | Designing a Com mCell ® Environment

Understanding Simpana Deduplication

The deduplication process contains the following key components:


 Storage policy
 Deduplication blocks
 Signature hash
 MediaAgent
 Deduplication Database (DDB)
 Disk storage
 Optional client side signature cache

Storage Policy
All deduplication activity is centrally managed through a storage policy. Configuration settings are
defined in the policy, the location of the deduplication database is set through the policy, and the disk
library which will be used is also defined in the policy.

Deduplication blocks and Signature Generation


When data protection jobs are executed, the data is sent to the Simpana® agent from the file system or
application the agent is responsible for protecting. Even though the data may be files or application
data, we will process the data as deduplication blocks. The deduplication process starts by performing a
calculation to generate a Signature Hash. This is a 512 bit value that uniquely represents the data within
the block. This hash will then be used to determine if the block already exists in storage.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 35

The block size that will be used is determined in the Storage Policy Properties in the Advanced tab.
CommVault® recommends using the default value of 128 KB but the value ranges from 32 KB to 512 KB.
Higher block sizes for large databases is recommended.

Signature Hash Comparison


The block signature hash is used to determine if the block exists in storage by comparing the hash
against other hashes in the deduplication database. By default, signature hashes are generated on the
client. This is preferred since the processing of block signatures can be distributed to many different
systems. This is required when using Simpana client side deduplication. For underpowered clients that
will not be using client side deduplication, a subclient can be optionally configured to generate
signatures on the MediaAgent.

Deduplication can be configured for storage side deduplication or client (source) side deduplication.
Depending on how deduplication is configured, the process will work as follows:

Storage Side Deduplication. Once the signature hash is generated on the block, the block and the hash
are both sent to the MediaAgent. The MediaAgent with a local or remotely hosted deduplication
database will compare the hash within the database. If the hash does not exist that means the block is
unique. The block will be written to disk storage and the hash will be logged in the database. If the hash
already exists in the database that means the block already exists on disk. The block and hash will be
discarded but the metadata of the data being protected will be written to the disk library.

Client Side Deduplication Once the signature is generated on the block, only the hash will be sent to the
MediaAgent. The MediaAgent with a local or remotely hosted deduplication database will compare the
hash within the database. If the hash does not exist that means the block is unique. The MediaAgent will
request the block to be sent from the client to the MediaAgent which will then write the data to disk. If
the hash already exists in the database that means the block already exists on disk. The MediaAgent will
inform the client to discard the block and only metadata will be written to the disk library.

Client Side Disk Cache An optional configuration for low bandwidth environments is the client side disk
cache. This will maintain a local cache for deduplicated data. Each subclient will maintain its own cache.
The signature is first compared in the local cache. If the hash exists the block is discarded. If the hash
does not exist in the local cache, it is sent to the MediaAgent. If the hash does not exist in the
deduplication database, the MediaAgent will request the block to be sent to the MediaAgent. Both the
local cache and the deduplication database will be updated with the new hash. If the block does exist
the MediaAgent will request the block to be discarded.

Deduplication Database
The deduplication database is the primary component of Simpana’s deduplication process. It maintains
all signature hash records for a deduplicated storage policy. Each storage policy will have its own
deduplication database. Optionally, a global deduplication storage policy can be used to link multiple
storage policies to a single deduplication database by associating storage policy copies to a global
deduplication storage policy.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
36 | Designing a Com mCell ® Environment

Deduplication Store
Each storage policy copy configured with a deduplication database will have its own deduplication store.
Quite simply a deduplication store is a group of folders used to write deduplicated data to disk. Each
store will be completely self-contained. Data blocks from one store cannot be written to another store
and data blocks in one store cannot be referenced from a different deduplication database for another
store. This means that the more independent deduplication storage policies you have, the more
duplicate data will exist in disk storage.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 37

Deduplication Building Block Guidelines

CommVault recommends using building block guidelines for scalability in large environments. There are
two layers to a building block, the physical layer and the logical layer.

For the physical layer, each building block will consist of one or more MediaAgents, one disk library and
one deduplication database.

For the logical layer, each building block will contain one or more storage policies. If multiple storage
policies are going to be used within a single building block, a general recommendation is to associate the
storage policies with a single global deduplication policy.

Best practices for the deduplication database:


 Put the DDB on the same server as the MediaAgent.
 The DDB volume needs to be on dedicated high performance disks.
 DDB disk volume performance must meet IOPs required for qualifying as a disk for DDB use.
 Dedicated storage adaptors - make sure MA sees the drives.
 Do not use virus scan on DDB volume.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
38 | Designing a Com mCell ® Environment

Deduplication Storage Options

There are three methods that disk library data paths can be configured when using deduplication: Direct
Attached Storage or DAS, Storage Area Network or SAN and Network Attached Storage or NAS.

Direct attached storage is when the disk library is physically attached to the MediaAgent. In this case
each building block will be completely self-contained. This provides for high performance but limits
resiliency. If the MediaAgent controlling the building block fails, data stored in the disk library cannot be
recovered until the MediaAgent is repaired or replaced.

Keep in mind that, in this case, all the data in the disk library is still completely indexed and recoverable,
even if the index cache is lost. Once the MediaAgent is reconstructed, data from the disk library can be
restored.

Storage Area Networks or SANs are very common in many data centers. SAN storage can be zoned and
presented to MediaAgents using either Fibre Chanel or iSCSI. In this case the zoned storage is presented
directly to the MediaAgent providing Read / Write access to the disks.

When using SAN storage, each building block should use a dedicated MediaAgent, deduplication
database and disk library. Although the backend disk storage in the SAN can reside on the same disk
array, logically in the Simpana software it should be configured as two separate libraries.

This provides for fast and protocol efficient movement of data but, as in the case of Direct Attached
Storage, if the building block MediaAgent fails, data cannot be restored. When using SAN storage either

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 39

the MediaAgent can be rebuilt or the disk library can be re-zoned to a different MediaAgent. If the disk
library is rezoned, it must be reconfigured in the Simpana software to the MediaAgent that has access to
the LUN.

Network Attached Storage has an advantage in that the path to the storage is directly through the NAS
hardware. This means that by using CIFS or NFS, UNC paths can be configured for a disk library to read
and write directly to storage. When using NAS storage as a disk library, it is still recommended to
configure two separate disk libraries in the Simpana software. In this case the library can be configured
as a shared library, where both MediaAgents can see all storage. Separate building blocks should still be
used for each MediaAgent providing Read / Write access to a disk library but Read Only access can also
be granted to all libraries on the NAS storage. In this case, if a MediaAgent fails, any other MediaAgent
with access to the library can conduct restore operations.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
40 | Designing a Com mCell ® Environment

Partitioned Deduplication Database

Partitioned deduplication is a highly scalable and resilient solution that allows the deduplication
database to be partitioned. It works by dividing signatures between multiple databases to increase the
capacity of a single building block. If two dedupe partitions are used, it effectively doubles the size of the
deduplication store.

In this example, two dedupe partitions have been configured, each on a separate MediaAgent.
Signatures are generated on the Client and depending on the signature generated it will be directed to
one of the two partitions for processing. Although either MediaAgent can process signature lookups, the
data for the client will always use its default MediaAgent path. This allows all unique deduplication
blocks to be protected through a single MediaAgent although duplicate blocks may have been protected
by either of the MediaAgents.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 41

Enterprise Building Block Guidelines

When designing storage policy and building block architecture, another consideration is that certain
data types do not deduplicate well against other data types. A prime example would be file system data
and database data. In this case, different building blocks and storage policies can be configured to
manage different data types. In this example a global deduplication storage policy has been configured
with a block size of 128 KB. Two data management storage policies have been configured, one with a 30
day retention and the other with a 90 day retention. All deduplication blocks from both storage policies
will deduplicate based on the global deduplication policy setting, but will be retained based on the data
management storage policy retention.

A second building block using a dedicated storage policy has been configured for database backups. In
this example a 256 KB block size has been configured and the storage policy has retention of 14 days.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
42 | Designing a Com mCell ® Environment

SILO Storage

Consider all the data that is protected within one fiscal quarter within an organization. Traditionally a
quarter end backup would be preserved for long term retention. Let’s assume that quarter end backup
of all data requires 10 LTO 5 tapes. Unfortunately with this strategy the only data that could be
recovered would be what existed at the time of the quarter end backup. Anything deleted prior to the
backup within the specific quarter would be unrecoverable unless it existed in a prior quarter end
backup. This results in a single point in time that data can be recovered. Now let’s consider those same
10 tapes containing every backup that existed within the entire quarter. Now any point in time within
the entire quarter can be recovered. That is what SILO storage can do.

SILO storage allows deduplicated data to be copied to tape without rehydrating the data. This means the
same deduplication ratio that is achieved on disk can also be achieved to tape. As data on disk storage
gets older the data can be pruned to make space available for new data. This allows disk retention to be
extended out for very long periods of time by moving older data to tape.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 43

Advanced Deduplication Configurations

Compression
It is recommended for most data types to enable compression during the deduplication process.
Compression can be enabled in the storage policy primary copy or in the subclient properties. By default
compression is enabled for a deduplication storage policy. You can turn compression off in the storage
policy copy or you can override the use of compression in the subclient properties.

Some applications such as Oracle and SQL may perform application level compression. Data should not
be compressed twice so the decision to use CommVault compression or application level compression
should be considered before the first backup of data. For databases, the best deduplication ratios are
achieved by backing up full databases redundantly over time. If a short term retention will be used for
the databases, it may be more efficient to use application compression and disable Simpana
deduplication. For medium to longer term retention on databases where three or more full database
backups will be retained, consider disabling application compression, enabled Simpana compression and
using deduplication.

Client Side Disk Cache


Along with configuring Client Side Deduplication, a Client Side Disk Cache can be created. Each subclient
will contain their own disk cache which will hold signatures for data blocks related to the subclient. The
default cache size is 4GB. The Client Side Disk Cache is recommended for slow networks such as WAN
backups. For any networks that are 1Gbps or higher using this option will not improve backup
performance.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
44 | Designing a Com mCell ® Environment

In this example a signature is generated on a deduplication block. The signature is then compared in the
local client disk cache first. If the block does not exist in the disk cache, the signature is then sent to the
MediaAgent and compared in the deduplication database. If the block does not exist, both the client
disk cache and the deduplication database are updated and the block is written to the library.

Variable Content Alignment


Variable Content Alignment can be used in some situations to improve deduplication ratios for large
data files such as database dumps. Enabling this option will read block data and align the blocks to
correspond to prior data blocks that have been deduplicated. By aligning the content prior to
performing the hash process, better deduplication ratios may be attained. This will however require
more processing power on the Client. Since Simpana deduplication is content aware, enabling this
option will not provide better deduplication for average file data. This option is only recommended for
large file system data such as database dumps or PST files with low incremental rates of change.

Fragmentation Considerations
Since CommVault stores data in the disk library in chunks, when blocks are deleted from disk it causes
empty spaces within the chunk. For Windows MediaAgents, the sparse file attribute is used to allow
empty spaces within the chunk to be used to store new blocks. Since Windows uses a write next
mechanism when writing data to disk, the empty spaces will only be allocated to new data when the
disk starts to reach full capacity. If new data is written to the empty spaces, fragmentation could occur.
This could negatively affect performance for auxiliary copy and restore operations. Scheduled
fragmentation analysis operations can be configured for the disk library. This will analyze each mount
path to determine the level of chunk fragmentation that exists. If fragmentation levels are too high,
defragmentation operations can be run by using third party file level defrag tools. When performing
defragmentation operations on a mount path, the mount path should be placed in an offline state.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 45

Deduplication Best Practices

General Guidelines
 Carefully plan your environment before implementing deduplication policies.
 Consider current protection and future growth into your storage policy design. Scale your
deduplication solution accordingly so the deduplication infrastructure can scale with your
environment.
 Not all data should be deduplicated. Consider a non-deduplicated policy for certain data types
such as log files. Note that in Simpana v10, log files will not be deduplicated even if they are
associated with a deduplication enabled storage policy when using Simpana database agents.
 Non-deduplicated data should be stored in a separate disk library. This will ensure accurate
deduplication statistics which can assist in estimating future disk requirements.

Deduplication Database
 Ensure there is adequate disk space for the deduplication database.
 Use dedicated dedupe databases with local disk access on each MediaAgent.
 Ensure the deduplication database is properly protected.

Disk Library Considerations


 It is recommended to use dedicated disk libraries for each MediaAgent.
 If using a shared disk library with multiple MediaAgents use NAS disk storage as opposed to SAN.
 Disk libraries should be divided into 2-8 TB mount paths.
 Use network paths as opposed to drive letters. Drive letters will limit the total number of mount
paths that can be added.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
46 | Designing a Com mCell ® Environment

Deduplication Store
 If DDB performance is poor, consult CommVault support before sealing the store
 When using SILO storage consider sealing stores at specific time intervals e.g. monthly or
quarterly to consolidate the time period to tape media.
 For WAN backups you can seed active stores to reduce data blocks that must be retransmitted
when a store is sealed. Use the option Use Store Priming option with Source-Side Deduplication
to seed new active stores with data blocks from sealed stores.

Block Size & block Processing


 Use the recommended 128 KB block size for all object level and virtual machine data protection
jobs.
 Use compression for object level and virtual machine data protection jobs. A general rule is to
use CommVault compression and deduplication for all protected data and not use native
application level compression for best deduplication ratios.

Performance
 Use DASH Full backup operations for best performance for full data protection operations.
 Use DASH Copy for auxiliary copy jobs for best auxiliary copy performance.
 Ensure the deduplication database is on high speed solid state disk technology.
 Ensure MediaAgents hosting a dedupe database have adequate memory (at least 32GB).

Global Deduplication
 Consider using global dedupe policies as a base for other object level policy copies. This will
provide greater flexibility in defining retention policies when protecting object data.
 Use global deduplication storage policies to consolidate remote office backup data in one
location.
 Use this feature when like data types (File data and or virtual machine data) need to be
managed by different storage policies but in the same disk library.

SILO storage
 SILO storage is for long term data preservation and not short term disaster recovery.
 Recovery time will be longer if data is in tape SILO so for short term fast data recovery use
traditional auxiliary copy operations.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 47

Designing a Sound Data Protection Strategy

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
48 | Designing a Com mCell ® Environment

Disaster Recovery Concepts

Disaster recovery or ‘DR’ is much more than backing up data and sending it off-site. Like other areas of
technology, disaster recovery has been refined to a science encompassing all aspects of data protection,
data preservation and data recovery. This science has been molded to a point where several key
concepts and definitions are commonly used when planning, testing and implementing DR plans. The
following information provides a high level overview of each of these concepts.

Service Level Agreement (SLA)


A Service Level Agreement defines a guaranteed “response time” or “resolution time” for various
incidents that may occur within your enterprise. It is a contract between a business owner and the IT
department. An overall SLA takes into account different objectives like Recovery Time Objective (RTO)
and Recovery Point Objective (RPO). RTO and RPO will be applied on two different levels (business level
and system level) and then integrated and articulated in an organizational DR Plan within an enterprise.
It is important to apply different measures of RTO and RPO on these levels to determine the priority or
sequence of recovery since resources such as power, hardware and bandwidth may be scarce when a
disaster strikes.

Recovery Time Objective (RTO)


A Recovery Time Objective defines the time to recover a business system. Depending on the level of
disaster and defined SLAs the RTO may be based on recovery time from point of disaster or from the
point where the recovery process begins. This will be determined by the level of disaster and should be
quantified by business system owners and other business units. Technologies such as clustering,

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 49

virtualization and disk replication/mirroring are implemented with the intention to reduce and in some
cases eliminate system outages. These systems provide a level of high availability that, when planned
right, can guarantee a high level of up time. However, it is important to properly understand the type of
disasters that may occur and how they might affect RTO.

Recovery Point Objective (RPO)


A Recovery Point Objective defines the frequency of recovery points that will be created and
maintained. Another way of looking at an RPO is that it defines the acceptable amount of data loss that
can occur. If backups are being conducted daily it will result in a 24 hour RPO. If tapes are being sent off-
site weekly, then the RPO can potentially increase to seven days. Mission critical systems such as
databases will typically conduct transaction log backups at short intervals (10-20 minutes) which results
in shorter RPOs. Snapshots, synchronous or asynchronous replication and off-site vaulting replication
technologies are also commonly implemented to shrink RPO times. RPO values are just as important, if
not more than, RTO values. Not meeting an RTO could cost your company money in lost production but
not meeting an RPO could result in data loss that may never be able to be recovered.

Gap Analysis
Gap Analysis is a process in which business units define SLA values for various business systems and then
pass them along to technical teams. The technical teams conduct tests to establish current capabilities
to meet SLAs. Gap analysis is then performed to see if the established SLAs can be met. If not the
technical team must address shortcomings and adapt to better meet the business unit’s requirements.
In some cases procedural adjustments can be made to better meet business’s needs. In other cases
additional investments must be made to meet SLA requirements. If the business unit’s needs cannot be
met or budget limitations prevent gap reduction then the business units must redefine their SLAs to be
more in line with the realistic capabilities of the technical teams.

Another key point regarding gap analysis is that each business unit will always think that their systems
are the most important. Fairly determining system priority and properly defining SLAs is sometimes a
better fit for outside consultants or auditors. If outside consultants are to be used it is important that
they do not represent specific products and technologies as they will sometimes push what they want
and not provide the best solution for your situation. Auditors can be a big benefit as their knowledge of
compliance requirements such as Sarbanes-Oxley can be used to push through technology upgrades and
change legacy processes that impede progress towards providing a sound disaster recovery strategy.

Risk Assessment
Risk Assessment is a companywide coordinated effort to address the likelihood of a disaster, the effect it
may have on business and the cost involved in preparing for it. Risks such as air conditioner leaks, fire,
hacking or sabotage are disaster situations that every company should be prepared to deal with. Major
disasters such as tornado, hurricane, volcanic eruption or terrorist attack are more complicated
disasters that, depending on the nature of a business may or may not be considered in a DR plan. This
may sound contrary to what a DR course should state, but the truth is that location, disaster probability,
nature of the business and data being protected will all factor in to planning a sound DR strategy.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
50 | Designing a Com mCell ® Environment

If you work for a small company on the outskirts of Mt. Rainier, the potential of a volcanic eruption and
the cost in defining short SLAs, which may be defined for an air condition leak, may not be worth the
money and effort when the likelihood of an eruption is very small. In this case the cost associated with
meeting short SLAs for an eruption would be substantially greater than an air condition leak. On the
other hand if you work for a major bank in the same location, short SLAs would most likely be required.
The point here is not that a DR plan should not be put in place, but rather the SLAs for the various levels
of disaster should be realistically weighed on a cost/benefit scale before investing in meeting SLA
requirements. Not all disasters are created equal so risk assessment should be considered at various
disaster levels: business system outage, limited site disaster, site disaster and regional disaster.

Total Cost of Ownership (TCO) & Return on Investment (ROI)


Total Cost of Ownership is a well-known concept in the business world. Regarding DR planning and
implementation TCO must be considered when assessing risk, defining SLAs and conducting gap analysis.
If it is determined that short SLAs are to be defined, the TCO of designing and maintaining a DR plan
including: hot DR site, bandwidth, personnel (including training and turnover), production hardware and
hardware at the DR site must all be factored into the equation. Putting a monetary value on achieving
unrealistic SLAs can often be enough to readdress DR goals.

Where the TCO can usually be quantified with various calculations, Return on Investment (ROI) is not as
easy to quantify. If two months after implementing a DR plan, disaster strikes, the ROI would be
wonderful. If disaster never strikes then ROI may be thought of as being nothing. The truth is that ROI
can be quantified when put into perspective. The piece of mind that a sound DR plan brings to a
company can be factored into the ROI. Many companies who implement sound DR plans may receive a
break on insurance, pass security and DR audits and even have an increase in customer and investor
confidence. These factors should not be taken lightly and depending on the company and services they
provide a sound DR plan can even be used in advertising. In overall planning of a DR strategy TCO and
ROI should be taken into account to properly define SLAs.

Cost Reduction vs. Risk Reduction


Companies are always seeking out ways to reduce costs. In some cases reducing costs results in a
compromise in effective DR planning. In other cases cost reduction can actually benefit DR planning.
Consider the virtualization of data centers as a cost savings measure that actually benefits DR planning.
It would be impractical to request 100 standby servers at a DR site for most companies but to request
four servers to host virtual machines may be in the budget. The choice to terminate a contract with a
dedicated DR facility might at first seem to be a negative towards DR planning but if the company has
another facility a few towns over, it may be a perfect location for a new (and cheaper) DR facility. With
bandwidth becoming considerably cheaper and Simpana features such as deduplication, DASH Full and
DASH Copy, a branch office can be quickly and inexpensively converted into a warm or hot DR site.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 51

In some cases cost reduction can have a negative effect on DR. Consider deduplication, being the big
concept in data protection. When blocks are deduplicated they are only stored once. In this case the
cost reduction in disk storage is countered by an increased risk in a corrupt block affecting the ability to
recover data. This is the concept of cost reduction vs. risk reduction. Saving money in disk storage
results in an increased risk. Another example is implementing archiving solutions where data is moved
to secondary storage to free up space in production. Like deduplication, this results in data being stored
in one location which may increase risk. Using technologies such as Deduplication and archiving can be
methods of reducing risk without increasing cost. When the Simpana software is configured properly
and CommVault best practices are followed, cost and risk reduction can both be achieved.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
52 | Designing a Com mCell ® Environment

Business Continuity Concepts

The concept of Business Continuity (BC) is the holistic approach of defining guidelines and procedures
for the continuation of a business in the face of any disaster situation. In this case disaster may or may
not even involve technical aspects or require DR planning. Business continuity is beyond the scope of an
IT department and beyond the scope of this course, but it is extremely important to consider in regards
to DR planning. A DR strategy may be perfectly planned and executed but without proper BC plans and
procedures the effort of IT may be in vein. The primary point to consider here is that on the technical
end of things you may not have the ability to design a BC strategy but you do have the power to
influence. In some cases influence may include ensuring that DR aspects of a BC plan are properly being
addressed such as facilities, chain of command, communication, and power sources. In other cases
influence might be making upper management aware that they need to create a BC plan as some
companies may have no idea of how important BC planning is.

Consider the following critical BC points and questions as they relate to DR planning:
 Facilities – How secure is the main data center? Is the air conditioner right on top of the data
center? How reliable is the power source? Is there a generator? How often is it tested? How
much fuel does it have?
 Chain of command – Who is in charge when the person in charge is not there? Who’s next on
the list? Who on the management team do you contact if you need to make substantial
emergency purchase? What are ALL methods to contact ALL people in the chain?
 Communication – Who is our cell phone provider and what are their contingency plans in the
event of disaster? Who is responsible for communicating with them? In the case of disaster how
will management communicate with employees on status updates?
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 53

 Contingencies – What happens when DR plans need to be changed? How does the company
deal with extended outages such as utilities where the ability to restore power or
communication is out of the company’s hands.
 Continuation of business – how will employees work if there is no facility to work from? How
will they access resources? How will they communicate?

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
54 | Designing a Com mCell ® Environment

Protection Methods

There are several primary protection methods used in modern data centers. Each of these technologies
have their advantages and disadvantages. It is important to understand that not all technologies are
created equal and a holistic approach should be considered when designing a data protection strategy
to meet SLAs.

Traditional Backups
Traditional backups to disk or tape protect data by backing up each object to protected storage. This is
the tried and true method that has been used for decades so it is the most reliable protection
technology. The main advantages when using traditional backups is that each item protected is a
complete separate copy that is backed up to separate media. When using tape media the backup
becomes portable. Many modern backup solutions incorporate traditional backups to disk storage which
is then replicated to a DR site. CommVault’s deduplication and DASH Copy is an example of using
traditional backups with a scheduled replication (DASH Copy) where only changed blocks are
transmitted to the DR location. Traditional backups and restores are usually slower than some modern
protection technologies which can have a negative effect on SLAs. This performance bottleneck is more
severe when millions of items require protection such as large file repositories. Traditional backups are
still the most common and cost effective data protection technology.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 55

Archiving
Data Archiving is not technically a data protection technology but can be used to improve SLAs.
Archiving removes infrequently accessed data from production disks and moves it to less expensive
secondary storage. The archived data can be recalled by end users or Simpana administrators. By
removing data from the production environment, backup and restores complete faster since less data
needs to be moved improving RTO and RPO.

The Simpana OnePass™ Agent is a comprehensive solution included in Simpana® product suite that
incorporates traditional backup and archiving into a single operation. It enables the movement of data
to a secondary storage location and uses this data to meet both data protection and storage
management archiving business objectives. Secure data recovery is available to both administrators and
end-users via a platform-independent web-based console, file stub recovery and a tightly integrated
Outlook add-in. Policy-driven selective stubbing and deletion from front-end storage provides storage
management archiving without the need to process the data a second time.

Snapshots
Snapshots are logical point in time views of source volumes that can be conducted almost
instantaneously. This allows for shortened RPOs since the snapshots can be conducted more frequently
throughout the day. A snapshot is not truly considered a DR protection strategy since the protected data
is not physically moved to separate media. Advanced snapshot technologies allow for data to be
mirrored or vaulted to separate physical disks which can be located at off-site DR locations. Snapshot
technologies are used to meet strict SLA requirements but are considerably more expensive to
implement requiring dedicated hardware. Simpana’s Continuous Data Replicator (CDR) is a software
based snapshot and replication technology which is a cost effective alternative to hardware snapshots.
For supported hardware and CDR, Intellisnap™ technology can be used to conduct and manage
snapshots.

Edge Data Protection (Desktop Laptop Option)


Edge data protection is a deduplication enhanced desktop / laptop data protection solution for
protecting end user data. Edge works by using Simpana’s client side deduplication solution to backup
block changes using an automatic scheduling mechanism. It allows users connected to specific networks,
locally or remote, to backup data during specific time intervals based on network and system resource
availability.

Replication
Replication technology is used to replicate block or object changes from a source volume to a
destination volume. Replication methods can use synchronous or asynchronous replication to
synchronize source and destination volumes using a one-to-one, one-to-many (fan out), or many-to-one
(fan in) replication strategy. Production data can be replicated providing fast SLAs for high availability.
Backup data or snapshot data can be replicated providing for a more complete DR solution. A
disadvantage of replication is that if corruption occurs at the source it may be replicated to the
destination so replication should be used along with point in time snapshots.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
56 | Designing a Com mCell ® Environment

Image Level Backups


Image level backups are faster than traditional backups because they read all blocks from a production
disk in sequential order. When backing up large file volumes image level backups can have a
performance advantage over traditional backup methods. Once the backup is complete indexes are
generated from the disks metadata to allow for granular restores or full system recovery. For large file
volumes, the time it takes to generate indexes can be slow causing a bottleneck but the data is backed
up from the production system faster than traditional backup methods.

For some file system agents such as Linux, the image level technology is now included within the file
system agent. The option to use block level backup can be enabled in the subclient to conduct an image
level type backup instead of a traditional file read method.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 57

Data Description

Data Description is based on the business data residing in the production environment. This data in
some cases can be an entire server, in other cases a business system may span multiple servers and in
other cases business data requiring different protection may exist on a single server. The key aspect of
describing data should be its business value and not its physical location.

The Client/Host is the system through which the specified data set will be accessed. In the case of
shared or distributed storage there may be more than one client per data set. Identifying the client(s)
marks the first transition point for data movement. Data will be read from primary storage through the
client host onto protected storage. Its path from the client to protected storage will be determined by
the placement of MediaAgents in the final storage design.

The location of the data will help determine whether some protection options (such as Snap or
replication) are possible and it will also determine any possible resource/data path sharing
requirements. Several sets of data located on the same shared storage device but under different client
management can present potential performance problems.

Volume information is, of course, essential to sizing protected storage, but it’s also essential to
determine data movement resource requirements and potentially the need for parallel data movement
to meet operation window requirements.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
58 | Designing a Com mCell ® Environment

The dynamics of the data is its daily change rate and annual growth rate. Both are key data points for
storage and storage policy design. Daily change quantifies both modified and new data as the minimum
data volume that requires protection. This impacts the rate of protected storage growth and resources
required to move the new data into storage. Annual growth rate helps determine future storage
capacity which must be accounted for in any storage design.

In many IT organizations dependency information is captured on documents that are not dynamic and
typically are not updated with any enterprise change. Additionally, dependency details are not readily
available for the people who need them. Questions like "what server is your application on? What
database does it use? What other applications does it depend on – or depend on it?" are being asked by
operation managers, enterprise architects, change managers, service support specialists, and of course –
Protected Storage architects. Dependencies are essential design information, yet many IT organizations
still don't understand how to document, manage, and operate these dependencies.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 59

Data Availability

Data Availability is the speed and ease of access to data. Understanding Data Availability options,
capabilities, and limitations is essential to designing protected storage. Production disk is the primary
data availability media/location. Data at that level is instantaneously and transparently available to both
applications and users. It’s where the data is originally written and read from.

In Hierarchical Storage Management (HSM) availability terms, production disk is considered to be


“Online”. Removable media still contained within an automated library is considered to be “Near line”.
Access may be delayed, but it is still automated. Offline data requires manual intervention in order to
bring it to Near line or Online status. Offline data can be as simple as tapes sitting on a shelf next to the
tape library or tapes stored away in a third party vault which may require much more time to bring
online.

Service Level Agreement


A Service Level Agreement (SLA) specifies the Maximum or Mean Time to Recover (MTTR) for data in
protected storage. As data ages, the MTTR may increase. An example for file level or message recovery
might be:

FS/MB data recovery requests < 14 days old must be recovered within (20) minutes of request.

FS/MB data recovery requests > 14 days old but < 1 year old must be recovered within (24) hours of
request. All versions of the data must be recoverable.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
60 | Designing a Com mCell ® Environment

FS/MB data recovery requests > 1 year old but < 7 years old can be recovered within (24) hours of
request. The data recovered will be the last monthly full iteration of the data.

Business Continuity
Business Continuity (BC) is the immediate availability of data as may be required to minimize the
interruption of day-to-day business. This usually involves loss of a file, folder, disk, or server and is
normally satisfied by restore from on-site backup data on disk storage. BC requirements are usually
specified in media type and length of availability.

Disaster Recovery
Disaster Recovery (DR) provides for protection against loss of both production and on-site backup data
and usually implies loss of a critical business function. DR requirements are usually specified in
frequency and duration of data movement off-site.

Archive
Archive implies long term availability of data that has value to the company. It can also mean
movement of less frequently accessed data to less expensive storage. This data may be historical
records required by legal, industry, or company requirements. The requirement for recall of archived
data may be transparent (on-site disk storage) or limited (vaulted off-site storage) Archive requirements
are usually specified in levels of availability and/or retention.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 61

Protected Storage Requirements

Protected Storage Requirements will determine the following:


 How many copies of data will be required
 Where will the copies be stored
 How long will the data be retained for
 Will data require encryption or content indexing

Using this chart, Storage Policies can be configured in an efficient manner. A chart such as the one above
created in a spreadsheet program can be sorted by fields to determine common requirements such as
storage location and retention. This can simplify the process of creating Storage Policies.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
62 | Designing a Com mCell ® Environment

Designing a Sound Data Protection Strategy

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 63

Understanding Client Agents

The Simpana product suite uses iDataAgents or ‘Agents’ to communicate with file systems and
applications that require protection. Any server with an Agent installed on it is referred to as a Client.
Each Agent contains code that is used to communicate directly with the system requiring protection.
The Agent will communicate using APIs or scripting that is native to the file system or application. For
example: A Windows 2008 file system can use VSS to protect file data so the Windows Agent will have
the option to enable VSS during backup operations.

The Agent will then have a data set defined. The data set is a complete representation of all data the
Agent is responsible to protect. Within the data set, subclients are used to define the actual data
requiring protection. By default, a Default Subclient is used to define ALL data requiring protection
within the backup set.

Additional subclients can be created to define specific content requiring protection. When content is
defined within the user defined subclient, it will automatically be excluded from the default subclient.
An example for a custom subclient could be defining a specific drive containing user data where VSS will
be initiated for the drive during backup jobs to ensure all open files are protected.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
64 | Designing a Com mCell ® Environment

Protecting Virtual Environments

Virtual Server Agent (VSA)


The Simpana Virtual Server Agent can be used to protect virtual machines by communicating with the
hypervisor’s APIs to backup the virtual machine disks. The options available for backup include full,
incremental, differential, and synthetic full. The option to Enable Granular Recovery is set by default to
enable file level recovery from the backup of the virtual machine disks.

For VMware, Simpana supports VMware vStorage API method (VADP) and VMware Consolidated
Backup method (VCB). To support VADP backups, Change Block Tracking (CBT) should be enabled for
best performance. CBT provides for more efficient incremental backup jobs by tracking block changes
within the VMDK files. By default, change block tracking is enabled for all VSA subclients.

Agent based protection


The traditional method for protecting virtual machines is to install a file system or application agent
within the virtual machine itself.

There are advantages and disadvantages to using this method:


 The first advantage is it is simple to deploy. When agents are installed in the virtual machine the
Simpana software will treat the machine as if it is a physical client. All of the functionality for
managing physical clients would be the same for the virtual clients
 The second advantage of installing an agent in a virtual machine, is the application specific
functionality becomes available. For example using MAPI to conduct granular mailbox backups.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 65

Another example would be conducting transaction log backups for a virtualized database
application.

There are also several disadvantages:


 Installing agents within the virtual machine results in all data being granularly backed up and
restored. This can be a slow process if there are many files within the machine.
 The backup and recovery process will also require all data to be moved over the network which
can become a bottleneck.
 A third disadvantage is that all processing during the backup or recovery process will be
conducted on the hypervisor. This could potentially become a bottleneck if too many virtual
machines are being backed up at the same time.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
66 | Designing a Com mCell ® Environment

The VSA Backup Process

This is a conceptual overview of how the VSA backup process works For VMWare:
 The backup is initiated by the Simpana software for each virtual machine that will be backed up.
 The VSA communicates with hypervisor with the list of virtual machines that have been defined
within the subclient contents of the virtual server agent.
 All virtual machines will have their disks quiesced. For windows virtual machines, VSS will be
enabled on all of the disks to provide a consistent point in time backup of each disk.
 Once the disks for the virtual machines are quiesced, the hypervisor conducts a software
snapshot which will be used to back up the VM.
 VSA backs up the virtual machines either through the physical hypervisor or a physical proxy.
With VMware, if the VSA is installed on a physical host it will be used as a proxy to back up the
VMs. If a virtual proxy is being used, the VMs will be backed up through the virtual proxy on the
physical hypervisor. For Hyper-V the VMs will be backed up through the physical hypervisor.
 When the backup process runs, virtual disks are indexed to provide granular recovery of files
and folders within the virtual machine.
 Once the backup is complete, the hypervisor releases the software snapshot. The disk within the
virtual machines are unquiesced and any transactions that were recorded while the disks were
in the quiescent state are replayed.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 67

VSS and its role in VM backups


For Windows virtual machines, Microsoft’s Volume Shadow Copy Service, or VSS, can be used to provide
consistent point in time backups of disk volumes.

VSS is Windows’ built-in infrastructure for application backups. A native Windows service, VSS facilitates
creating a consistent view of application data during the course of a backup. It relies on coordination
between VSS requestors, writers, and providers to quiesce – or “quiet” – a disk volume so that a backup
can be successfully obtained without data corruption.

In order for this to work the VMware Tools VSS component must be enabled. The Virtual Server agent
requests VMware tools to initiate a VSS snapshot in the Guest OS. All registered VSS writers in the
Guest OS get the request and they prepare its application to be backed up committing all transactions.
Once all VSS writers are finished they communicate back to your backup software which then initiates a
VMware snapshot.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
68 | Designing a Com mCell ® Environment

Protecting Applications

Simpana® software supports most major applications through the use of agents installed on the
application servers or on proxy servers with access to data. For unsupported applications, scripts can be
used to properly quiesce application databases and then back them up as file data.

Virtualized Applications
Virtualized applications pose a challenge when it comes to data protection. Issues such as disk I/O
activity, application type and application state at the time of backup can significantly affect the backup
process. There are several methods that can be used to protect virtualized applications.

Simpana application iDataAgents – An iDataAgent installed in the VM will directly communicate with
the application running in the VM. Prior to the snap operation the agent will communicate with the
application to properly quiesce databases. For large databases this is the preferred method for providing
application consistent point in time snap and backup operations. Using application agents in the VM also
provide database and log backup operations and a simplified restore method using the standard browse
and recovery options in the CommCell® GUI.

VSA and VSS aware applications – Some application such as Microsoft SQL and Exchange are VSS aware.
When VSS is initiated on the virtual machine it will attempt to quiesce the VSS aware application to
provide an application consistent snapshot. For smaller application databases with low I/O this process
should be adequate for consistent snapshots. For larger databases with high transaction I/O this method
is not recommended.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 69

Scripting database shutdowns – Using external scripts which can be inserted in the Pre/Post processes
of a subclient, application data can be placed in an offline state to allow for a consistent point-in-time
snap and backup operation. This will require the application to remain in the offline state for the entire
time of the snapshot operation. When the VM is recovered the application will have to be restarted
after the restore operation completes. This method is only recommended when Simpana agents are not
available for the application.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
70 | Designing a Com mCell ® Environment

Snapshot Management

Snapshots are point in time logical views of a volume. The volume block mapping is snapped which
represents a point-in-time view of the block structure when the snap occurred. When existing blocks
need to be overwritten with new blocks the old blocks are preserved. References to these blocks are
recorded to provide a frozen point-in-time snapshot view of the volume. This allows the volume to be
reverted back to any point in which a snapshot was taken. The snapshot can also be mounted off line on
a separate host for mining, testing, backing up or restoring data.

IntelliSnap® Technology provides a single interface to conduct, manage, revert, and backup snapshots.
The following lists the key highlights for the IntelliSnap feature:

Application Awareness – By using Simpana iDataAgents to communicate with hosting applications,


application consistent snapshots can be performed. The application agent will communicate with the
hosting application to quiesce databases prior to the snap occurring. This is a significant benefit when
protecting large databases where traditional backup methods are not adequate to meet protection
windows.

Snapshot backups to reclaim disk cache space – By managing the snapshots, Simpana software can also
be used to backup the snapped data. As older snapshots are backed up to protected storage, the snaps
can be released on the source disk and the space can be freed for new snap operations.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 71

Granular recovery - Snapshots can be mounted for Live Browse and indexed during backup operations
for granular recovery of objects within the snap. Whether using live browse or a restore from a backup,
the method to restore the data is consistent. Using the proper iDataAgent you can browse the snapped
data and select objects for recovery. This process is especially useful when multiple databases or virtual
machines are in the same snap and a full revert cannot be done. In this case just the objects required for
recovery can be selected and restored.

Simplified management – Multiple hardware vendors supported by the IntelliSnap® feature can all be
managed through the Simpana interface. Little additional training is involved since the same subclient
and storage policy strategies used for backing up data are extended when using snapshots. Just a few
additional settings are configured to enable snapshots within the CommCell® environment.

Note: The IntelliSnap® feature is rapidly evolving to incorporate increased capabilities as well as
expanded hardware support. Check with the CommVault Online Documentation for current list of
supported features and supported vendors.

How IntelliSnap® Technology Works


IntelliSnap technology operates by interfacing with hardware APIs to conduct and manage snapshots.
There are several steps required to implement the IntelliSnap feature in a CommCell® environment.

Deploy & Configure Architecture - An IntelliSnap® Architecture consists of the following components:

 Supported hardware array or Simpana Continuous Data Replicator installed on the host server.
 Host server running specific file system and application iDataAgents for the host, VSS provider
(for Windows servers), and a MediaAgent. The host client must also be enabled for IntelliSnap
operations.
 Proxy server to manage and backup snapshots running the same OS as the host server and a
MediaAgent. VMware proxy servers must use a Windows OS and have the Virtual Server Agent
(VSA) installed.

Configure arrays – Array information is set in the Array Management applet in the Control Panel.
Depending on the vendor different information may be required.

Configure storage policies – Storage policies are used to centrally manage snapshots of subclient data
just like backup data. When configuring storage policies, a snapshot copy is added to the policy. For
some vendors, multiple snap copies can be added. NetApp DFM enabled policies currently support
multiple snap mirror/vault copies.

Configure subclients – The IntelliSnap® capability is enabled at the client level in the Advanced tab of
the client properties. Once enabled for the client, subclients will have an IntelliSnap Operations tab that
can be used to enable and configure snapshots for the subclient.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
72 | Designing a Com mCell ® Environment

IntelliSnap® Architecture
IntelliSnap architecture is made up of host servers and proxy servers that work together to provide snap
and optional backup operations.

Host Server Agent Requirements


The host server is the server which has direct access to data on the array. For VMware the host server
will be a proxy server with the Virtual Server Agent (VSA) installed since agents cannot be directly
installed on ESX hosts. Host servers will require the following agents:

 MediaAgent – Provides capabilities to execute array functions and access to snapshots on the
host. It can also be used when backing up snapshots to CommVault protected storage if no
proxy is being used or if the proxy server is unavailable.
 IntelliSnap® – IntelliSnap options are built into iDataAgents and do not require additional
software to be installed on the host. IntelliSnap capabilities are enabled in the Advanced tab of
the client properties. This will add an IntelliSnap Operations tab to subclients to configure snap
operations.
 File System iDataAgent – provides protection for OS data.
 Application iDataAgent – provides application integration to perform application consistent
snapshots for databases.

Proxy Server Agent Requirements


A proxy server can be used to backup snapshots and as an off host proxy to mount snapshots. The use of
a proxy server eliminates load on the host server for snap mining, mounting, and backup operations.
The proxy server requires the following agents:

 OS must be same as host – For a mount or backup operation to be performed the snap must be
mounted on the proxy. In order for the proxy to recognize the file system, the same OS must be
used on the proxy.
 File System iDataAgent – A file system agent is required for backup operations. When a
snapshot is backed up it is treated like a file system backup job.
 MediaAgent – Used for array access, mounting snaps on the proxy and data movement from
array to CommVault protected storage.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 73

Data Protection Best Practices

Meeting Data Protection Windows


 It is critical to meet data protection windows. If windows are not being met then restore
windows may not be met. If data is scheduled to go off-site daily but it takes four days to back
up the data, then the data cannot be sent off-site until the job completes.
 If you are currently meeting protection windows, then there is no need to modify anything.
Improving windows from six to four hours when your window is eight hours just creates more
work and a more complex environment. The following recommendations are intended to
improve performance when protection windows are NOT being met.

Storage policy settings and modification to help meet protection windows:


 Device Streams – Increase device streams to allow for more concurrent jobs streams to write if
adequate resources are available.
 MediaAgent – ensure MediaAgent is properly scaled to accommodate higher stream
concurrency.
 Network – ensure network bandwidth can manage higher traffic.
 Disk Library (Non-Deduplicated) – ensure library can handle higher number of write operations.
Increase the number of mount path writers so the total number of writers across all mount
paths equals the number of device streams.
 Disk Library (Deduplication enabled) – if not using Client Side Deduplication enable it. Each
deduplication database can manage up to 50 concurrent streams. If using Client Side
Deduplication, after the initial full is complete most data processing will be done locally on each

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
74 | Designing a Com mCell ® Environment

Client. This means minimum bandwidth, MediaAgent, and disk resource will be required for data
protection operations.
 Tape Library – If tape write speeds are slow enable multiplexing. Note: enabling multiplexing
can have a positive effect on data protection jobs but may have a negative effect on restore and
auxiliary copy performance.

Data path property settings:


 Increase chunk size to improve performance.
 Increase the block size to improve performance. Note: block size is hardware dependent. Before
changing the block size ensure all NICs, HBAs, switches, routers, MediaAgent OS, and storage
devices at your primary and alternate sites (including DR sites) support the block size setting.

Subclient settings and modifications to help meet protection windows:


 Ensure all data is properly filtered. Use the job history for the client to obtain a list of all objects
being protected. View the failed items log to determine if files are being skipped because they
are open or if they existed at time of scan and not time of backup. This is common with temp
files. Filters should be set to eliminate failed objects as much as possible.
 For file systems and application with granular object access such as Exchange or Domino
consider using data archiving. This will move older and infrequently accessed data to protected
storage which will reduce backup and recovery windows.

File Backup recommendations:


 For backups on Windows operating systems ensure source disks are defragmented.
 Ensure all global and local filters are properly configured.
 If source data is on multiple physical drives increase the number of data readers to multi-stream
protection jobs.
 If source data is on a RAID volume, create subclient(s) for the volume and increase the number
of data readers to improve performance. Enable the Allow Multiple Data Readers within a Drive
or Mount Point option.
 For large volumes containing millions of objects:
o Consider using multiple subclients and stagger scheduling backup operations over a
weekly or even monthly time period.
o For supported hardware consider using the IntelliSnap® feature to snap and backup
volumes using a MediaAgent proxy server.
o Consider using the Simpana Image Level backup agent or enabling block level backup for
supported agents.

Database applications:
 For large databases that are being dumped by application administrators consider using
Simpana database agents to provide multi-streamed backup and restores.
 When using Simpana database agents for instances with multiple databases consider creating
multiple subclients to manage databases.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Designing a CommCell ® Environment | 75

 For large databases consider increasing the number of data streams for backing up database.
Note: For multi-streamed subclient backups of SQL and Sybase databases, the streams cannot
be multiplexed. During auxiliary copy operations to tape if the streams are combined to a tape
they can be pre-staged to a secondary disk target before they can be restored. If the multi-
streamed backup is not first pre-staged to disk, during the restore operation the streams will be
cached to the job results folder on the destination client, so ensure there is adequate space to
cache the job during the restore.
 For MS-SQL databases using file/folder groups, separate subclients can be configured to manage
databases and file/folder groups.

Virtual Machine Protection


General Guidelines:
• Consider using the Simpana Virtual Server Agent (VSA).
• Determine which virtual machines DO NOT require protection and do not back them up.

When using VSA agent to protect VMware environment:


 Ensure enough proxies are being used to handle load.
 Use Simpana Client Side Deduplication and DASH Full backups.
 The data readers setting determines the number of simultaneous snap and backup operations
that will be performed. Increase this number to improve performance. Note: ensure disks where
virtual machines are stored can handle the number of concurrent snapshots or the snapshot
process may fail.

When using file system agents in virtual machines:


Consider having a base vm image that will be used to recreate the virtual machine. Use the
default subclient filters to filter out any volumes and folders that do not require protection.
Note: It is STRONGLY NOT recommended to alter the contents of the default subclient. If you
explicitly map default subclient data the auto-detect feature will be disabled. This means any
new volumes added to the machine will have to be explicitly added to the content of the
subclient.

When protecting applications in VMware environment:


 Use application agents inside the VMs. It is strongly NOT recommended to perform VSA crash
consistent backups of application database data.
 Consider the pros and cons of using Simpana compression and client side deduplication. Using
application level compression may have a better compression ratio but deduplication efficiency
can suffer.

IntelliSnap® Technology and VSA integration:


 Define subclients by data store affinity. When hardware snaps are performed the entire data
store is snapped regardless of whether the VM is being backed up.
 For smaller Exchange or MS-SQL databases (less than 500GB), application consistent snapshots
can be performed using the IntelliSnap feature and VSA.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
76 | Designing a Com mCell ® Environment

 For large databases, install the application agent in the VM and configure the IntelliSnap options
in the subclient. Hardware snapshots will be performed at the database level providing better
scalability and application awareness.

Meeting Restore Requirements


Considerations for tape media:
If streams from different data sets are multiplexed or combined to a tape, only one data set can
be restored at a time. Consider isolating different data set streams to different media using
separate secondary copies for each data set and using the combine to streams option.

Considerations for disk media:


When using Simpana deduplication use the minimum recommended 128 KB block size. Small
block sizes will result in heavier data fragmentation on disk which can reduce restore
performance.

Improving Recovery Time Objectives (RTO):


Filter out data that is not required for data protection operations. The less you backup the less you have
to restore.

 Strongly consider data archiving. It will improve backup and restore performance. Note that
deduplication will improve backups and reduce storage requirements which can actually have a
negative effect on restore performance.
 If a subclient job was multi-streamed you can restore it using multiple streams through the
Restore by Job option.
 Consider assigning different RTOs for different business data. It is not always about restoring
everything. Consider a database server with five databases. Each one can be defined in a
separate subclient. This will allow each database to have a separate RTO so they can be
recovered by priority.

Improving Recovery Point Objectives (RPO):


• Run point in time backups such as incremental or transaction logs more frequently for shorter
RPO.
• Consider prioritizing data for RPO requirements and define the data as a separate subclient and
assign separate schedules. For example a critical database with frequent changes can be
configured in a separate subclient and scheduled to run transaction logs every fifteen minutes.
To provide short off-site RPO windows consider running synchronous copies with the automatic
schedule enabled.
• Consider using hardware snapshots with the IntelliSnap® feature to manage and backup
snapshots.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 77

Module 2 – CommCell® Deployment and Configuration

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
78 | CommCell ® Deploym ent and Conf iguration

Topics

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 79

CommCell Deployment Process

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
80 | CommCell ® Deploym ent and Conf iguration

New CommCell® Deployment Process

The first component to be installed in a new CommCell® environment will be the CommServe®
server. Once it is installed the next step would be to install MediaAgent software and detect and
configure libraries. Policy configuration for storage policies, schedule policies, subclient policies
and global filters should be done prior to installing any client agents. When installing client
agents, options to associate the default subclient for the agent with the policies can be selected
so preconfiguring policies makes the agent deployment process smoother.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 81

Existing CommCell® Upgrade Process

Prior to upgrading a CommCell® environment it is critical to perform a CommServe DR backup. In the


event of problems during the upgrade process, the environment can be rolled back to ensure CommCell
operations can continue. The first component to be upgraded must be the CommServe server. The
upgrade process can be an in-place upgrade or a fresh installation of the CommServe server. It is
recommended that you have the CommServe database inspected by CommVault prior to upgrading.
This can be done by uploading the database dump to cloud.commvault.com. Check the CommVault
Online Documentation for complete instructions for CommServe database inspection.

MediaAgents should be upgraded next and libraries should be tested to ensure everything is functioning
properly. Clients can then be upgraded on an as needed basis. Note that with Simpana® software, client
agents up to two versions back can coexist with a CommServe server and MediaAgents at the latest
version.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
82 | CommCell ® Deploym ent and Conf iguration

CommCell® Disaster Recovery Process

CommServe® Server
The CommServe server must be the first machine recovered before the recovery of any production data
can be accomplished. The speed and method of recovering the CommServe server ultimately depends
on the combination of several factors:

• Which High Availability CommServe server option or Standby CommServe server option was
configured.
• Access to the DR Backup metadata.
• How prepared the production and DR environment is and how practiced and efficient the
Simpana Administrators are at recovering the CommServe server.
• What the effect is for an actual disaster scenario you are confronting (site or regional), or what
practice DR run you are simulating.

CommserveDisasterRecoveryGUI.exe (CSDR tool)


The CommServe Disaster Recovery Tool restores the metadata from the Disaster Recovery Backup file.
The Disaster Recovery Backup file gets created when a Disaster Recovery backup is performed from the
CommCell® Console.

The CommServe Disaster Recovery Tool can be used to rebuild the CommServe server on the same or
different computer, change the name of the CommServe computer, create and maintain a CommServe
server in the hot-site and to update the license.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 83

MediaAgent
If it becomes necessary to build a completely new MediaAgent, as in the case of complete Disaster
Recovery at another location other than the production-site, there are a few things to keep in mind.
Usually, this type of scenario will be using removable media as the primary source of data recovery and
the MediaAgent will be new to the CommCell® environment. This is not a problem for the Simpana®
software. The new MediaAgent will be installed after the CommServe server Disaster Recovery has
taken place and connected to a library where the media has been loaded. Restores can take place from
any library; in fact one of the advanced options of a restore job is to select the desired MediaAgent and
Library.

Some considerations may be licensing issues. With Volume based licensing there is no issue installing an
additional MediaAgent, as with a DR License as well, but if the licensing is per agent it may be necessary
to release the license of an existing MediaAgent to apply a license to a new one.

Rebuilding Disk Libraries


The main reason for changing the hardware of a disk library is to upgrade to better devices. Most disk
libraries will have fault tolerant redundancies built in to avoid major disk failures. Changing the
hardware is a relatively simple process. The original CV_Magnetic folder and all its contents must be
copied to the new location. After moving the physical data, the original mount path can be changed in
the Library and Drive Configuration tool by right-clicking the old mount path and selecting the new
location.

Rebuilding Removable Media Libraries


Replacing a drive
In the event that a drive in a library fails and needs to be replaced, the process is very simple. After the
new drive is installed in the library, the MediaAgent’s operating system must be able to communicate
with the new drive. Within the CommCell® Console, open the library properties page and from the
Attributes tab, choose Enable Auto Drive Replacement when new device is detected during Mount to
enable the option to automatically detect the new drive. A new drive can also be manually detected by
selecting Mark Drive Replaced from the task menu of the replaced drive.

Replacing a Library
There are two reasons to replace a library, the first being an upgrade to newer hardware and the second
being a major hardware malfunction. In either case it is fairly simple to replace the library. Keep in
mind the new library must support the same drive type and must have the same or more drives as the
original. Once the Hardware is connected and can communicate with the MediaAgent’s operating
system you can move the media from the old library to the new one. Then you must use the Library and
Drive Configuration tool in order to configure the new library. For procedures to modify the existing
library to be replaced with the new one, refer to the CommVault Online Documentation.

Changing the MediaAgent hosting a library


There may be instances where it becomes necessary or desired to change the MediaAgent hosting a
library to another MediaAgent. This would involve connecting or presenting the library devices to the

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
84 | CommCell ® Deploym ent and Conf iguration

new MediaAgent first and using the Library and Drive Configuration tool to configure the change of the
MediaAgent host. Choosing both old and new MediaAgents in the Library and Drive Configuration tool
will allow the Administrator to select the option to Change Host when right-clicking on the Library
Controller, and also on the Drive Pool.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 85

Environment Requirements

Minimum system, software, and application requirements are documented in the CommVault Online
Documentation. If you do not see your specific OS or application, contact your CommVault Support
group to check if you can install the software component.

Be sure to read the notes included on the System Requirements page. These notes often contain caveats
or additional information essential to the installation process.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
86 | CommCell ® Deploym ent and Conf iguration

Installing CommServe Software

The CommServe® server is the central management system within a CommCell® environment. All
activity is coordinated and managed by the CommServe server. The CommServe system runs on a
Windows platform and maintains a Microsoft SQL metadata database. This database contains all
environment configuration information, job metadata and security information.

The CommServe server can be installed on a physical host, in a virtual machine or in a clustered
environment. Minimum system requirements, SQL settings, metadata database I/O requirements are
listed in the CommVault Online Documentation and should be checked prior to installation for latest
guidance on deployment. Deployment recommendations are specified for enterprise, data center,
workgroup and express environments.

Before starting the installation, disable any antivirus software running on the server. Ensure the logged
on user performing the installation is a member of the administrators group on the local computer.
Ensure all CommServe software has been downloaded and in the folder location of the install files.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 87

CommServe Installation Steps


The following provides high level steps for installing CommServe® server software:
1. Run the SetupAll executable. Review the CommVault licensing agreement, accept the
agreements and click Next.
2. The installation options provides the ability to create a new CommCell® environment or join an
existing CommCell group. For a new CommCell environment, select Create a New CommCell
group.
3. The installation folder path is set by default to the operating system install drive, Program Files
folder. The default location can be changed to a different location by clicking on the browse
button and selecting the desired location. Visit the CommVault Online Documentation for
destination folder space requirements. These requirements are specific to the operating system
and hardware configuration of the server in which the CommServe server has been installed.
4. After the destination folder is selected, the summary of install options selected will be displayed
for confirmation. Review the install options and click the Install button to proceed with the
installation. The CommServe server installation will now install the Microsoft .NET Framework
4.0 and Microsoft Visual C++ redistributable packages for the selected platforms. A reboot of the
CommServe server may be required after package installation.
5. The installation will set the Microsoft SQL Server administrator password. During the SQL
instance installation the option for database files folder path will be shown.
6. During the installation, a new CommServe database can be created or an existing database can
be used. An existing database in the form of a CommServe database dump or export can be
used when installing a new CommServe server in an existing CommCell environment, if
performing a server upgrade; or when rebuilding a CommServe server in a disaster scenario
where the database dump is available at the DR location. When selecting the Use an Existing
Database option, browse to the location of the database export, select the correct database
.dmp file and click Next.
7. The installation will prompt to create an administrator account. This account will be used to log
into the CommCell Console GUI for configuration and administration tasks. Even if Active
Directory integration will be used to authenticate users, it is strongly recommended to have at
least two administrators with local administrative accounts for DR purposes. The default
administrative account cannot be deleted and a strong password should be used.
8. The software cache is a central repository for Simpana installation packages, service packs and
hotfixes. The cache options can be configured during the CommServe server installation. These
settings can later be changed in the Control Panel.
9. After copying all the installation files as per the selection, an installation completion report will
be displayed indicating the successful completion of the CommServe server.

Service Packs & Patches


It is critical that the production CommServe® server and the target recovery CommServe server be at the
same service pack and patch levels to avoid recovery issues.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
88 | CommCell ® Deploym ent and Conf iguration

Installing MediaAgent Software

A MediaAgent component can be located on the same host as the CommServe component, the same
host as an agent component, or on a separate host by itself. Non-clustered MediaAgent components can
be installed interactively from the installation media or pushed from the CommCell® Console.

Host Name resolution


Proper Host Name resolution, both forward and reverse, are the single most common installation
problems. Verify proper host name resolution of all components before starting an install. IP addresses
can be used in place of a Host Name if DNS is slow or unreliable.

Client(s) access path


If the client data will transit over a LAN, verify the expected access path is available and
addressable. This data path may be different from the control/coordination path used for the
install. Such paths can be set up using Data Interface Pairs.

Data path(s) to Library(ies)


The MediaAgent host must have access to all libraries to which you expect to read/write data from that
MediaAgent. Multiple libraries under a single MediaAgent control are supported.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 89

Index Cache Size and location


During installation you will be asked to select the location for the Index Cache (Default - <software
install path>\IndexCache.) For performance reasons the index cache should be located on a separate
disk of sufficient size. The estimation of index cache size is based on a number of assumptions, which
include the following:

 Average full path name to the indexed object


 Average incremental percentage of the files being backed up
 Job frequency and type
 The retention time for the Index Cache

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
90 | CommCell ® Deploym ent and Conf iguration

Index Cache Configuration

CommVault software uses a two tiered distributed indexing structure providing great resiliency,
availability, and scalability. Job summary data is maintained in the CommServe metadata database and
requires minimal space to retain the data. The job summary information will be maintained as long as
the data is being retained. An index cache maintains detailed indexing information for all objects being
protected. Index data is maintained in the cache based on retention settings of days or disk usage
percentage. Each subclient will have its own index file and new index files are generated during a full
data protection operation. Index files are copied to media automatically at the end of each job.

Sizing the Index Cache


The index cache should be sized based on the need to browse back in time for data to be recovered. The
farther back in time you need to browse, the larger the cache should be. If the index cache is undersized,
index files will be pruned sooner to maintain a default 90% disk capacity. When you attempt to perform
a browse or find operation and the index file is not in the cache it will automatically be restored from
media. If the index file is in disk storage there will be a short delay in recovering the index but if it is on
removable media the time to recover the index can be much longer.

Maintaining the Index Cache


The index cache is self-maintaining based on two configurable parameters: Index Retention Time in
Days and Index Cleanup Percent. Index files will be kept in the cache for a default of 15 days or until the
cache disk reaches 90% disk capacity. A smaller index cache location may result in index files being
pruned before the 15 day time period expires if the cleanup percentage is reached first. Index files will
be pruned from the index based on least recently accessed.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 91

A Minimum Free Space can also be configured to reserve space in the index cache location. The cleanup
percent setting would be based on the allocated space to the index cache. So if you had a 100 GB
partition and wanted to reserve 10 GB of space, the cleanup percent would be based on 90 GB.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
92 | CommCell ® Deploym ent and Conf iguration

Library Detection and Configuration

Using the Library & Drive Configuration Tool


Libraries are either detected (e.g. tape device, library controller) or added (e.g. disk, cloud, IP-based
controller). Essential to both is the ability of the MediaAgent to correctly see/access the device. Prior to
any detection or adding of devices to a MediaAgent, confirm the physical and logical view of the device
from the operating system. If multiple similar devices are involved (e.g. a multi-drive library), all such
devices should be at the same firmware level.

Detection
The system only detects devices for which device drivers are loaded. A detected device may have the
following status:

 Success indicates that the system has all of the information necessary to use the device.
 Partially configured, detect fail - connection error status when the detection fails due to an error
connecting to the MediaAgent
 Partially configured, detect fail - device not found status when the detection fails due to a
missing device

Note: Some devices (e.g., the library associated with a stand-alone drive) have no detection status, since
they are virtual entities and as such have no hardware components that can be detected.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 93

Exhaustive Detection
Modern tape drives have serial numbers which are used by the Simpana software to properly place a
drive physically and logically within a library. Older drives without serial numbers require manual
locating. Exhaustive detection is the process of associating drive numbers to their correct SCSI address.
This is done by mounting a media to each of the drives in the library to obtain the dive’s SCSI address.

Adding
Logical libraries (e.g. Disk, Cloud, PnP) are added by the user allocating assets and/or access to
devices. This usually involves the grouping of devices (mount paths) identified by providing data paths
and user access authority.

A hybrid library requiring both addition and detection would be an IP-based library. The IP address for
the library control is added while the tape devices used by the MediaAgent(s) are detected and logically
associated with the IP-based library.

Configuration
Added or Detected devices can be configured as new libraries or added to existing libraries (e.g. adding
an additional tape drive in an already detected/configured library). Configuration gives the device an
identity within the CommCell® environment and, as appropriate, an association with other devices for
management/control (e.g. tapes drives in an automated library, new mount paths).

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
94 | CommCell ® Deploym ent and Conf iguration

Client Agent Deployment Methods

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 95

Standard Installation Methods - Interactive

Interactive installation can be performed directly from the installation software disc by running the
SetupAll.exe command. Optionally you can copy the installation files to a disk or a network share
accessible to the client and execute the SetupAll.exe command from the client. The user performing the
installation must have administrator privileges on the client to install software.

Any number of components can be selected for installation at the same time. For a new client, the Base
Agent (not visible in the component list) will automatically be selected and the first item installed. The
Base Agent provides files needed for communication with the CommServe® server.

While every effort is made to not require a reboot of the host during or after the installation, the state
of the system at the time of install may require a reboot. If this happens, you will be presented with an
option to not reboot at that time.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
96 | CommCell ® Deploym ent and Conf iguration

Common questions asked during the interactive installation are:


 Installation Path
 Authorized user/password to interact with an application for backup/restore (Application Agent
only)
 Firewall access (if required)
 Computer Group membership
 Default Storage Policy
 Filter Policy
 Include patches/service packs
 Update Schedule

Use a common installation path for all clients if possible, this will help with re-installation and/or a full
system restore should it become necessary.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 97

CommCell® Console Push Install

You can manage the installation of agent and component software packages on client computers or
even on network computers not yet a part of the CommCell® environment, from the CommCell console.

The required software packages can be downloaded or copied to the CommServe Cache Directory and
then pushed to selected computers. Remote Software Cache directories may also be configured and
used to locate installation software closer to their prospective targets or for different access privileges.

Remote caches can be configured for automatic synchronization with the CommServe cache directory.
This entire process is all conveniently managed from the CommCell console.

Remote software cache directories can be created and managed via the Add/remove Software
Configuration applet located in the CommCell console's control panel.

Prior to configuring the installation of software packages to specific computers to build your CommCell
environment, you must copy or download the required software packages to the CommServe cache
directory. The directory is configured to serve as a holding area for software and update packages. To
install from any of the software cache directories the directory must be a shared network directory with
permissions set to write to the directory.

The CommServe cache directory can be populated during install of the CommServe host, by FTP
download, or by using the Copy Software option in the CommCell console.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
98 | CommCell ® Deploym ent and Conf iguration

Client computers that are not in the same domain as the domain in which the CommServe cache is
located must have bidirectional trust in place.

If Authentication for Agent Install is enabled for the CommCell® environment, installation from the
CommCell Console is restricted to only those users belonging to a user group assigned with
Administrative Management capabilities for the CommCell computer or an existing Client computer
within the CommCell environment. However, if it is a new computer, not yet part of the CommCell
group, you must have Administrative Management capabilities for the CommCell group.

During configuration, computers within the domain that are not yet part of the CommCell group can be
selected for installation. Users accessing these computers must have administrative privileges required
for installing software.

Software packages are intelligently pushed to the computers. This means that a windows package
pushed to the domain consisting of both Windows and UNIX computers will only install on the Windows
systems. Additionally, application package software will only install on systems with the prerequisite
software installed.

CopyToCache
For complete information on configuring the cache and using the CopyToCache utility, go to:

http://documentation.commvault.com/commvault/v10/article?p=service_pack/populating_cache.htm

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 99

Custom Installation Methods

Decoupled Install
Decoupled install is performed without involving the CommServe® server until you are ready to add the
Client and/or MediaAgent to the CommCell® environment. Once all necessary physical connections are
established, the computer can be added to the CommCell environment. This feature will be useful when
you want to pre-image computers with the software at a central location and later ship them to the
environment where you plan to use them.

Custom Package Install


Custom Packages enable you to push smaller install packages through a network, which is useful for
reducing WAN/LAN payload while installing remote clients.

Custom packages can be configured from the Simpana Installer's Advanced options. For UNIX systems,
select the cvpkgadd's Advanced options menu choice.

Using Custom packages, you can do the following:


 Select the necessary components to build the custom package. This enables you to build smaller
packages based on the needs of the client computer.
 Create a self-extracting executable file for installing the custom package in the client computer.
This enables you to easily deploy the software with a one-click deployment procedure.
 Save the installation options in the current deployment setup so it can be reused later.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
100 | CommCell ® Deployment and Conf iguration

Installation involves copying the custom package .exe file to the target host and executing to expand the
installation files and start the silent install process. Upon completion, a success message will be
displayed.

For UNIX, copy the folder to the target host and execute the cvpkgadd command.

Restore Only Agent


To install agent software for restore only, select the Restore Only checkbox in Selected Platforms dialog
during the installation. Installing Restore Only agents allows you to install the agent software without
consuming any license. Check the CommVault Online Documentation for step-by-step guides for
installing agents in restore only mode.

Silent Install
A Silent install consists of the following distinct phases:

Recording Mode - In this phase, an install is recorded, saving your install options to an .xml file.

Playback Mode (XML input file) - In this phase, the .xml file is played back by the install program. The
software components are installed as per the recorded options without prompting for any user inputs.
Through this method, the deployment of the software can be automated.

When recording an install, note the following:


 The install program only records your choices, it does not execute the install.
 No license is consumed when recording an install.
 All Agents are selectable when recording an install.
 Review the System Requirements and the Installation procedures for each component you are
installing.
 If you intend to play back this recorded install on multiple computers, choose options that are
applicable to all those computers. For example when specifying an installation folder, be sure it
exists on all target computers.
 Playing back a recorded installation installs the software with the options saved in the .xml file
you created during the record procedures.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 101

Deployment Best Practices

CommServe® server:
 For CommServe server availability consider staging a standby CommServe server.
 In some environments virtualizing the CommServe server can provide high availability.
 In large environments consider clustering the CommServe server for high availability.
 It is ABSOLUTELY CRITICAL that the CommServe database is properly protected. By default every
day at 10 AM a CommServe DR backup job is conducted. This operation can be completely
customized and set to run multiple times a day if required.
 All activity is conducted through the CommServe server, therefore, it is important that
communication between the CommServe server and all CommCell® resources always be
available.
 Do not install the CommServe database on the system drive. Consider a SSD or high
performance 15RPM dedicated disk drive.
 If the CommServe server is configured on a Virtual Machine (VM), then it typically operates at a
range of 60% efficiency as compared to a comparable physical server.
 Follow the CommCell Scalability Guide in the CommVault Online Documentation for the
minimum requirements for a CommServe processor and memory specifications depending on
your scale variables.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
102 | CommCell ® Deployment and Conf iguration

MediaAgents
 If the MediaAgent is configured on a Virtual Machine (VM), then it typically operates at a range
of 60% efficiency as compared to a comparable physical server.
 Follow the CommCell® Scalability Guide in CommVault Online Documentation for the minimum
requirements for a MediaAgent processor and memory specifications depending on your scale
variables.
 There are (3) separate disk drives typically on a MediaAgent that need to be sized and have
sufficient I/O specifications to meet the SLA for data protection and recovery operations as well
retention requirements. They are the disk library, the DDB drive, and the Index Cache drive.
 Format the disk library at minimum 64 KB blocks per sector. The Index Cache and DDB drives
should be left at the default values.

DR MediaAgents
 A DR MediaAgent is installed and preconfigured at a DR location. The most common
implementation of DR MediaAgents is in the use of replica libraries or a secondary disk library
using Simpana deduplication and the DASH Copy feature. By having an active and registered
MediaAgent configured with a library at a DR location RTOs can be more realistically achieved.
Incorporating a DR MediaAgent with a standby CommServe server provides a ‘ready to go’ DR
infrastructure which can expedite recovery procedures in the case of disaster.
 Another use of the DR MediaAgent is the ability to pre-stage recovery operations at a DR
location. This is most commonly implemented in virtual environments. CommVault provides
Online Documentation about the proper implementation on pre-staging the recovery of virtual
machines in DR environments.

Index Cache
The index cache should be sized based on the need to browse back in time for data to be recovered. The
farther back in time you need to browse, the larger the cache should be. If the index cache is undersized,
index files will be pruned sooner to maintain a default 90% disk capacity. When you attempt to perform
a browse or find operation and the index file is not in the cache it will automatically be restored from
media. If the index file is in disk storage there will be a short delay in recovering the index but if it is on
removable media the time to recover the index can be much longer.

To properly size the index cache, consider the following:


 The index file size is based on the number of objects being protected. Estimate 150 bytes per
object. The more objects you are protecting the larger the index files will be.
 Each subclient will contain its own index files within the cache.
 The index cache should be on a dedicated disk or partition with no other data being written to
the disk.
 To reduce the probability of pulling an index file back from media use a large index cache
location.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
CommCell ® Deploym ent and Conf iguration | 103

Maintaining the Index Cache

 The index cache is self-maintaining based on two configurable parameters: Index Retention
Time in Days and Index Cleanup Percent. Index files will be kept in the cache for a default of 15
days or until the cache disk reaches 90% disk capacity. A smaller index cache location may result
in index files being pruned before the 15 day time period expires if the cleanup percentage is
reached first. Index files will be pruned from the index based on least recently accessed.
 A Minimum Free Space can also be configured to reserve space in the index cache location. The
cleanup percent setting would be based on the allocated space to the index cache. So if you had
a 100 GB partition and wanted to reserve 10 GB of space, the cleanup percent would be based
on 90 GB.

Libraries
 For client servers where the source data is in a SAN or DAS environment and target storage can
be made directly accessible to the client, install a MediaAgent on the client server to provide
LAN free backups.
 When backing up to disk storage attached to a network use a dedicated backup network for
library read/write operations. Do not use the same NIC that is receiving data from a client to
write the data to the library.
 If using Fibre Channel SAN storage with an iDataAgent and MediaAgent installed on the same
system use separate HBAs to receive the source data and write the data to storage.
 If using iSCSI ensure the iSCSI initiator and target systems being used are enterprise class.
Consider using a TCP/IP Offload Engine (TOE) NIC card to reduce CPU load on the server. Do not
use the same NIC receiving the data to write the data to storage.
 If considering using a Virtual Tape Library (VTL) carefully weigh the advantages and
disadvantages. Simpana disk features such as deduplication and DASH operations will not work
if disk storage is configured as a VTL.
 If using a shared disk library, where the library will be shared between multiple MediaAgents
and Simpana’s deduplication, it is strongly recommended to use NAS storage instead of SAN
storage.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
104 | Data Securit y and Net work Control

Module 3 – Data Security and Network Control

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 105

Topics

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
106 | Data Securit y and Net work Control

Storage Policy Design

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 107

Storage-Based Design Strategy

This strategy starts with the assumption that protection for the largest data set for a particular data type
is the biggest challenge. For example; if you have hundreds of Oracle databases that drive your business
then their protection should be handled first, even though the databases cross business function lines.
Once you get that storage policy in place, you deal with the next largest data set. This strategy is driven
more by resources than protection requirements.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
108 | Data Securit y and Net work Control

Business Based Design Strategy

As the title says, this strategy approaches your data from the business side. Build the storage policies
you need for your mission critical data/business function first. These storage policies become your core
set. As you review other data sets and business groups look to see which can be incorporated/covered
by existing storage policies and which need new policies. This policy is driven by protection
requirements rather than resources. It often results in the purchase of more storage and data
transmission resources.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 109

Deduplication’s Impact on Policy Design

Block Factor
When using Simpana Deduplication, the dedupe block factor is a primary concern when developing
storage policy strategies. The smaller the block size the more entries are made to the dedupe database.
Currently the database can scale from 500-750 million records. The total volume of data being
protected, which is relatively simple to estimate and the estimated number of unique blocks, which is
certainly not easy to estimate, should be taken into consideration when determining block size. The
following recommendations for block factor settings are based on the following:

128 KB – All object level protection, virtual machines and smaller databases.

128 KB – 512 KB – Current recommendation for database backups depending on size of all database
data managed by the policy. For large databases it is recommended to engage CommVault Professional
Services for proper deployment.

In this case different storage policies should be configured for the different block factors. It is not
recommended to use a single policy for all data when mixed data types are involved since different data
may not deduplicate well in mixed dedupe stores.

Another factor that should be considered is how long the data will be retained. Longer retention will
result in larger databases. Since different data types typically will have different retention settings, it
would require separate storage policies to manage the data so separate dedupe databases will be used.
It is NOT recommended to use global deduplication for long retention or large volume protection.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
110 | Data Securit y and Net work Control

Dissimilar Data Types


Dissimilar data types that require different block settings should have separate policies. In some cases
data types that do use the same block size should still have their own dedicated policies. This is the case
when considering databases since different database applications typically organize and compress data
using specific algorithms. Using the same dedupe policy will not necessarily result in better
deduplication ratios. Also considering databases can contain large amounts of data, using a single
dedupe policy may result in large dedupe databases which can lead to scalability issues. It is
recommended to use different storage policies dedicated to different data types.

Deduplication Database Scaling


A single deduplication database can scale to approximately 750 million records. Based on estimated
dedupe efficiency a single deduplication database can manage between 40 – 60 front end terabytes
(FET), retained for 30 – 90 days. In environments managing large amounts of data, it is recommended to
use multiple deduplication databases hosted on different MediaAgents.

When determining the number of policies that will be needed in large environments, data growth
projections should be considered. Although a single dedupe database may be able to manage all current
data, if the data growth rate is expected to change significantly, you may find yourself scrambling to
redesign your policies at the last minute to accommodate changes in your environment. This will have a
negative effect on deduplication efficiency especially when data is being retained for longer periods of
time.

Global Deduplication on Primary Copy


If different retention settings are required for the primary copy but the disk location and block factor are
the same, a global deduplication policy can be used to achieve a better deduplication ratio. Associating a
global dedupe policy with a primary copy will result in a single dedupe database and dedupe store being
used across multiple copies. Because of this, consider the volume of data that will be protected and
ensure the deduplication database will be able to scale to meet current and future data growth.

The use of global dedupe policies are mainly for consolidating small amounts of data with different
primary retention needs or for consolidating remote location data to a central location. Global
deduplication policies should NOT be used across the board for everything in your datacenter. You can
quickly grow out of the database maximum size which will then require a complete redesign of your
storage policy structure. Realize that policy copies attached to a global dedupe policy cannot be
unattached. New policies will have to be created and the old policies cannot be deleted until ALL data
has aged from the policies.

Global Deduplication for Secondary Copies


The most common implementation method for global deduplication Storage policies is consolidating
remote data to a central disk library. Secondary copies for remote storage policies can be associated
with a global deduplication policy with a data path at the main data center. This will allow multiple
remote locations to be consolidated into a single deduplication store at the main data center. Create
any global deduplication policies prior to creating secondary copies for the remote location.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 111

How Many Storage Policies do I really need?

A general rule of thumb has been – “The more storage policies you have, the more management is
required.” This is not entirely true. Following the rule of thumb you would think the ultimate solution
would be to have just one storage policy. While possible, the problem with this is the potential
complexity of this single storage policy and the efforts needed to handle any additional data/clients.

Storage policies need to reflect your storage organization and business needs. If that means you have 5,
10, or even 100 storage policies then that’s the correct number of storage policies you need.

From the previous design strategies you always start with one storage policy. Within a storage policy
you can add, delete, and modify copies by just moving data around. You can’t move/re-associate
existing data between storage policies

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
112 | Data Securit y and Net work Control

Understanding Alternate paths

Each storage policy copy has a default data path which will be used to perform backup operations. You
can also define alternate data paths in each storage policy copy to ensure the success of backup and
other operations conducted using the storage policy. In addition, alternate data paths provide the
following advantages:

 Automatic switch-over to an alternate data path, when one of the components in the default
data path is not available.
 Utilization of available libraries and drives in the event of failure or non-availability of these
resources.
 Minimizes media utilization by routing backup operations from several subclients to the same
storage policy and hence the same media; instead of creating several storage policies which in
turn utilizes a different media for each subclient.
 Load balancing (round robin) between alternate data paths provides the mechanism to evenly
distribute backup operations between available resources.
 Facility to define a subset of the data paths at the subclient level within the selected storage
policy and its data paths.

Alternate data paths are supported for both the primary and secondary copies associated with storage
policies for all libraries.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 113

Advanced Storage Policy Features

Incremental Storage Policy


An Incremental Storage Policy links two policies together. The main policy will manage all Full backup
jobs. The incremental policy will manage all dependent jobs (incremental, differential or logs). This is
useful when the primary target for full backups needs to be different than dependent jobs. Traditionally
this has been used with database backups where the full backup would go to tape and log backups
would go to disk. When performing log backups multiple times each day, replaying logs from disk during
restore operations is considerably faster than replaying the logs from tape.

Hide Storage Policy


If a storage policy managing protected data is deleted, then all of the data associated with the policy will
be aged and subsequently deleted. If a storage policy is no longer going to be used to protect data, the
option Hide Storage Policy in the General tab of the policy properties can be selected. This will hide the
policy in the storage policy tree and also hide the policy in the subclient drop down box in the Storage
Device tab. In order to hide a storage policy no subclients can be associated with it.

If hidden storage policies need to be visible in the storage policy tree, set the Show hidden storage
policies parameter to 1 in the Service Configuration tab in the Media Management applet.

Copy Precedence
Copy precedence determines the order in which the restore operation will be conducted. By default, the
precedence order specified is based on the order in which the policy copies are created. The default
order can be modified by selecting the copy and moving it down or up. This changes the default order.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
114 | Data Securit y and Net work Control

Precedence can also be specified when performing browse and recovery operations in the Advanced
options of the browse or restore section. When using the browse or restore precedence the selected
copy becomes explicit. This means that if the data is not found in the location the browse or restore
operation will fail.

Erase Data
Erase data is a powerful tool that allows end users or Simpana administrators to granularly mark objects
as unrecoverable within the CommCell® environment. For object level archiving such as files and Email
messages, if an end user deleted a stub, the corresponding object in CommVault protected storage can
be marked as unrecoverable. Administrators can also browse or search for data through the CommCell®
Console and mark the data as unrecoverable.

It is technically not possible to erase specific data from within a job. The way Erase data works is by
logically marking the data unrecoverable. If a browse or find operation is conducted the data will not
appear. In order for this feature to be effective, any media managed by a storage policy with Erase Data
enabled will not be able to be recovered through Media Explorer, Restore by Job, or Catalog.

It is important to note that enabling or disabling this feature cannot be applied retroactively to media
already written. If this option is enabled, then all media managed by the policy cannot be recovered
other than through the CommCell Console. If it is not enabled, then all data managed by the policy can
be recovered through Media Explorer, Restore by Job, or Catalog.

If this feature is going to be used it is recommended to use dedicated storage policies for all data that
may require the Erase Data option to be applied. For data that is known to not require this option,
disable this feature.

Content Indexing
Content indexing allows selected object level data to be indexed for eDiscovery, Records Management,
and compliance purposes. Simpana software allows data to be proactively or retroactively indexed. This
means any jobs being retained in the CommCell® environment can be indexed. Proactive indexing would
relate to recent jobs in storage. Retroactive indexing could be used to index older jobs for investigation
purposes.

Subclients can be defined to protect specific data required for indexing. This allows for several key
advantages when using CommVault content indexing:

 Selected data and users can be defined in specific subclients for investigative purposes.
 Data can be defined in separate subclients for records management policies. This allows data to
be searched based on content and ownership and move relevant information to ERM
(SharePoint), export for 3rd party analysis tools, or moved into separate legal policies for data
preservation.
 The Content Indexing tab allows subclient data to be selected for indexing. This allows for a
policy retaining protected data to selectively index relevant data while adhering to standard
retention policies.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 115

Legal Hold Storage Policy


When using the Simpana Content Indexing and Compliance Search feature, auditors can perform
content searches on end user data. The search results can be incorporated into a legal hold. By
designating a storage policy as a Legal Hold policy, the auditor will have the ability to associate selected
items required for legal hold with designated Legal Hold policies. It is recommended to use dedicated
Legal Hold policies when using this feature.

Legal Hold Storage Policies can also be used with Content Director for records management policies.
This allows content searches to be scheduled and results of the searches can be automatically copied
into a designated Legal Hold Policy.

Subclients Associations
Subclient Properties
In order to protect a subclient, it must be associated with a storage policy. During an iDataAgent install,
a storage policy can be selected for the default subclient. When creating additional subclients you must
select a storage policy. The policy defined to manage the subclient is configured in the Storage Device
tab – Data Storage Policy sub tab. Use the storage policy drop down box to associate the subclient with a
policy.

Storage Policy Level


All subclients for a specific storage policy can be associated with another policy in the Associated
Subclients tab of the Storage Policy Properties. You choose Re-Associate All to change all policies, or you
can use the Shift or Ctrl keys to select specific subclients and choose the Re-Associate button to
associate selected subclients to a new policy.

Policies Level Subclient Association


If subclient associations need to be made for more than one storage policy you can use the Subclient
Associations option by expanding Policies, right-click on Storage Policies and select Subclient
Associations.

The windows will display all subclients for the CommCell® environment. There are several methods that
can be used to associate subclients to storage policies.

Select the subclient and use the drop down box under the storage policy field to select the storage
policy.

You can use the Shift or Ctrl keys to select multiple subclients then use the Change all selected Storage
Policies to drop down box to associate all selected subclients to a specific storage policy.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
116 | Data Securit y and Net work Control

Storage Policy Design Best Practices

Consider these four basic rules for approaching storage policy design:
Rule #1: Keep it Simple
This section will describe several different methods for protecting data. It is designed to provide in-
depth explanations and solutions for the most complex environments. But before overanalyzing and
over-architecting the CommVault environment, use this one simple rule: KEEP IT SIMPLE! If rules 2 – 4
are being satisfied then there is really no reason to change anything. A complex environment leads to
more complex problems.

Rule #2: Meet Protection Requirements


Data protection requirements MUST be met. Though it is true the only reason we protect data is to
recover it, if you are not meeting your windows then you are not protecting data. You cannot recover
something that never finished backing up, so ensure protection windows are being met. In the following
sections methods to improve performance will be discussed. Performance always starts with an
adequately designed physical environment. Before tweaking CommVault software to improve
performance, ensure that Clients, MediaAgents, and networks are scaled appropriately.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 117

Rule #3: Meet Media Management Requirements


In an ideal world data would simply be preserved forever. With the dropping cost of disk storage and
deduplication, most data can be retained longer. As with anything this comes at a price. The best way to
approach media management is to ensure the business end understands your capabilities and
limitations for preserving data. Sometimes a ‘Pie in the Sky’ vision of protecting data can be brought
right down to reality through a little education and a cost association of the business requirements.
Although you understand the capabilities and limitations of your storage, the non-technical people may
not. Provide basic guidance and education so they better understand what you and the Simpana
product suite are capable of doing. You may not have the power to make the final decisions but you do
have the power to influence the decision process.

Rule #4: Meet Recovery Windows


Recovery windows are made up based on Service Level Agreements (SLA).

For data protection and recovery an SLA is made up of three components:

 Protection Windows
 Recovery Time Objectives (RTO)
 Recovery Point Objectives (RPO)

When designing a CommCell® environment focus should always be placed on how data will be
recovered. Does an entire server need to be recovered or only certain critical data on the server require
recovery? What other systems are required for the data to be accessible by users? What is the business
function that the data relies on? What is the associated cost with that system being down for long
periods of time? The following sections will address RTO and RPO and methods for improving recovery
performance.

Requirements for Custom Content Definitions (Subclients)


The Simpana software suite offers powerful features to provide great flexibility in managing data. One of
the most powerful features is the ability to logically address content requiring protection by using
subclients. Subclients allow content to be explicitly defined such as files, folders, mailboxes, document
repositories, or databases. Although most environments only use the Default subclient to protect all
data managed by an agent, custom subclients can provide granular management of data which can be
used to improve performance, make more efficient use of media, or define custom data handling
methods to meet specific protection requirements.

Consider using custom subclients for the following situations:


 When custom retention settings are required for specific data such as a folder, virtual machine
or a database.
 When special storage requirements exist for specific data such as isolating financial data onto
separate media from other data being managed by the agent.
 When special file handling must be performed such as using VSS or Simpana QSnap to protect
open files.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
118 | Data Securit y and Net work Control

 When specific files must be protected and managed independently from other data in the same
location such as PDF and DOC files requiring specific retention or storage requirements.
 When scripts need to be used to place data in a specific state prior to backup such as quiescing a
database before backing it up.

Retention Considerations
Retention Requirements for Contents
Retention requirements should be based on specific contents within a file system or application. All too
often, determining retention requirements is not easy, especially when data owners do not want to
commit to specific numbers.

Considerations for Retention Requirements:


Keep it simple. Unless specific content within an application or file system requires special retention
requirements, don’t over design subclients.

Consider using default retention policies providing several levels of protection. Provide the options to
the data owners and allow them to choose. Also stipulate that if they do not make a choice then a
primary default retention will be used. Also state a deadline in which they must provide their retention
requirements. It is important to note that this is a basic recommendation and you should always follow
policies based on company and compliance guidelines.

Consider defining retention rules for the following:


 Disaster Recovery requirements should be based on the number of Cycles of data that should be
retained. This should also include how many copies (on-site / off-site) for each cycle.
 Data Recovery requirements should be based on how far back in time (days) that data may be
required for recovery.
 Data Preservation/Compliance should be based on the frequency of point-in-time copies
(Monthly, Quarterly, Yearly) and how long the copies should be kept for (Days).

Protection Requirements for specific data types


Managing different data types such as file systems and databases typically require special design
considerations for storage policies. These factors should be considered in the initial design strategy.

Consider special protection requirements for different data types in the following
situations:
 Typically different data types such as databases and files will require different retention settings
which will result in different policies being used to protect the data.
 If the primary storage target is a tape library and multiplexing will be used it is not
recommended to mix database and object level backups to the same media. Using different
storage policies will force different data types to use different media.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 119

Library Considerations
Library and Data Paths
For simplicity of managing a CommCell® environment, different libraries as well as location of the
libraries may require separate storage policies. This will allow for easier policy management, security
configurations, and media management.

Consider the following when determining storage policy strategies for libraries and data
paths:
When using Simpana deduplication, for performance and scalability reasons different policies should be
used for each MediaAgent data path. This will allow the deduplication database to be locally accessible
by each MediaAgent providing better throughput, higher scalability, and more streams to be run
concurrently.

If a shared disk (not using Simpana deduplication) or shared tape library is being used where multiple
Client / MediaAgents have LAN free (Preferred) paths to storage, a single storage policy can be used.
Add each path in the Data Path Properties tab of the Primary Copy. Each Client / MediaAgent will use
the LAN Free path to write to the shared library. This will allow for simplified storage policy
management and the consolidation of data to tape media during auxiliary copy operations.

If a shared disk (not using Simpana deduplication) or tape library is protecting LAN based client data
where multiple MediaAgents can see the library, each data path can be added to the primary copy.
GridStor Round Robin or failover can be implemented to provide data path availability and load
balancing for data protection jobs.

Deduplication Considerations
When using Simpana deduplication careful planning is essential. Whether a policy will use deduplication
or if a policy copy will be associated with a global dedupe policy must be determined during the initial
configuration of the policy copy. Although the block size specified during the policy creation can be
modified, it would result in all signatures becoming unique because of the block size change. If the block
size is going to be modified it is recommended to seal the deduplication store.

Consider the following when using Simpana Deduplication:


 Deduplication gets the best benefit from data that deduplicates well with other data and from
subsequent full backups that protect the same data over time. Not all data will deduplicate
efficiently. For data that will be written to non-deduplicated disk storage, use a separate storage
policy. It is also recommended to use a separate disk library for deduplicated and non-
deduplicated data.
 Use separate storage policies for data types that require different block settings and for data
that does not deduplicate well with other data types. This provides for better performance and
scalability.
 If global deduplication is going to be used for Primary or Secondary copies, create the global
dedupe policy before creating the other policy copies. Associating a policy copy to a global
dedupe policy must be done during the initial creation.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
120 | Data Securit y and Net work Control

 In environments with ‘Big Data’ consider scaling out the environment to accommodate current
data volume and future growth estimates. This may require using dedicated storage policies
which will use independent dedupe databases allowing for higher volumes of data to be
managed. Each database can handle up to 750 million records.
 If certain data will also be placed in SILO storage then dedicated policies should be used to
separate SILO required data from non-SILO required data.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 121

Advanced Job Control

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
122 | Data Securit y and Net work Control

Understanding Services

Several services used by Simpana software are designed to listen for incoming network traffic
on specified network ports; thus the CommServe® server, MediaAgents, and Agents within the
CommCell® environment communicate with each other. Essential CommServe services are
automatically assigned registered static port numbers during installation. MediaAgents, Agents,
and other software components can utilize the same default static port numbers, or any static
port numbers specified during installation.

CommVault® Services are broken down as follows:


 Base Services - common to all hosts
 CommServe® Services - CommServe Only
 MediaAgent Services - MediaAgent Only
 Client/Agent Services - Client/Agent Only

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 123

Common Processes and Associated Log Files

Service / Log Description

CVD (CVD.log Provides the base communication which controls connectivity, firewall
cvfwd.log) access, patch information, Pre/Post process execution and space
checks.
EvMgrC (EvMgrC.log) Used to forward events and conditions from the local machine to the
CommServe® server and is also used to assist in browsing application
data on the local host.
InstallUpdates Used to install updates on the local machine and verify patch information
(UpdateInfo.log) with the local registry.

JobMgr The JobMgr.exe process is responsible for initiating and controlling jobs,
(JobManager.log) and communication with storage resources. It acts as the primary
coordinator for all data movement operations and the JobManager.log is
typically the first log to view when troubleshooting data movement
problems. All starting and stopping of processes during a data
movement operation will be logged in the JobManager.log.

EvMgrS (EvMgrS.log) The EvMgrS is responsible for receiving messages from the EvMgrC
and feeding information to the CommCell® console.

CVMountD (CVMA.log The CVMountD process interacts with hardware storage devices
and attached to the MediaAgent.
SIDBPhysicalDelete.log)

IndexingService The IndexingService process creates a new index or gains access to the
(CreateIndex.log and most recent index. It is also used to prune index files based on the index
UpdateIndex.log) cache retention settings.

ArchiveIndex The ArchiveIndex process is responsible for compacting the index and
(archiveindex.log) writing index to storage. Prior to version 10 it was also responsible for
cleaning up index cache. In version 10 this process is handled by the
IndexingService process.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
124 | Data Securit y and Net work Control

Firewall Configuration

When CommCell® components need to communicate or move data through a firewall, firewall settings
must be configured for each component. This can be done by configuring individual firewall settings for
a specific client or firewall settings can be applied to a client computer group. For example, if a client
needs to communicate with a CommServe® server through a firewall and backup data to a MediaAgent
through a firewall, all three components would require firewall configuration.

There are three primary methods for connecting through a firewall:


 Direct – where the CommCell components communicate directly with each other through a
firewall.
 Through a proxy – where CommCell components use a proxy in a demilitarized Zone or DMZ to
communicate with each other.
 Gateway – where CommCell components communicate through a gateway resource.

Defining Firewall Rules for Client and Client Groups


To configure firewall settings for a client or client group, right-click on the entity in the CommCell®
console, select properties and then click the advanced button. Select the firewall tab and then click
configure firewall settings. Click the advanced radio button to enable full firewall configuration.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 125

There are four configuration tabs available:


 Incoming connections
 Incoming ports
 Outgoing routes
 Options

A fifth tab will show a summary of all options configured for the firewall settings. This summary will be
in the format that will be used to populate the FWConfig.txt file that will be located in the base folder of
all CommCell® components using firewall configurations.

Configuring Incoming Connections


The incoming connections tab is used to determine if other CommCell® components can connect the
client or client group where the firewall settings are being configured. There are three connection
options:

 Open connection – there are no firewall restrictions. In this case, no incoming connections need
to be configured.
 Restricted – there are firewall port restrictions in place and a component on the other side of
the firewall can reach the component that is currently being configured.
 Blocked – there are firewall port restrictions in place and a component on the other side of the
firewall can NOT reach the component that is currently being configured.

Simpana software uses port 8400 as the default communication port for all CommCell traffic. When
firewall settings are enabled for a CommCell component, by default, port 8403 will be used as a listening
port for any inbound connection attempts. Additionally, a dynamic port range can be configured to
provide additional data traffic ports for backup and recovery operations. How these ports will be used is
dependent on a number of factors:

 Communication will be based on the “listen for tunnel connections on port” setting.
 If port 8400 is available on the firewall, once initial communication is made using the listen port,
by default, data transmission will use port 8400 and metadata and communication will use port
8403.

By default, a dynamic port range will not be used for data traffic. This is by design of the network model
Simpana® software uses to transmit data to a MediaAgent. When the MediaAgent setting in the control
tab, “optimize for concurrent LAN backups” is enabled, all data will be tunneled through a single data
port. This means dynamic port ranges are not needed by Simpana software to backup and restore data
through a firewall. In certain situations, performance may be improved by disabling the “enable for
concurrent LAN backup” option and defining a dynamic port range. Keep in mind, that when the LAN
optimization option is disabled, the maximum number of streams a MediaAgent can process will be
limited to 25.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
126 | Data Securit y and Net work Control

Configuring Outgoing Routes


The outgoing routes tab determines how CommCell® components will communicate with each other.

There are three route types:


 Direct
 Via gateway
 Via proxy

For each route type, encryption options can be set by determining the connection protocol that will be
used.

There are three connection protocol options:


 Regular – Session authentication and data will NOT be encrypted.
 Authenticated – Session authentication will be encrypted but data transfer will not be
encrypted.
 Encrypted – Session authentication and data will both be encrypted.

The default option ‘Authenticated’ is the recommended option. If data transfer requires encryption,
consider using client ‘inline’ encryption instead of using the ‘encrypted’ option in the firewall settings.

Configuring Options
When the CommServe® server can reach clients to initiate data protection and recover jobs, it will be
configured as restricted on the clients. If the CommServe server cannot communicate to the client, it will
be configured as blocked and the client will be responsible for establishing connections with the
CommServe server. The keep-alive interval and tunnel Init interval are used to determine how
connections are made and maintained when the CommServe server is blocked from communicating
with clients. For example, a laptop client can initiate a connection with a CommServe server from a
remote location.

Pushing Firewall Settings


Once all firewall settings have been configured, the summary tab will show the firewall output
information which will be pushed to the CommCell® components.

The configuration will need to be pushed using one of the three following methods:
1. Client services started – the client will communicate with the CommServe® server which will
push out firewall settings.
2. Firewall configurations can manually be pushed to client groups or clients by right-clicking on
the component, selecting all tasks, and then push firewall configuration.
3. When Data Interface Pairs are configured it will automatically push firewall configuration
settings.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 127

Network Control

Network Communication
Communication within a CommCell® environment is based on TCP/IP. It is recommended to use
hostnames for communication between agents and with the CommServe® server. Due to this
recommendation, a properly configured DNS environment with forward and reverse lookup zones
should be used. Host files can also be used in situations where DNS is not available or not reliable. If
using hostnames is not preferred, IP addresses can be used to bypass host name resolution all together.

During agent software installation you will be prompted to choose the hostname of the server you are
installing the software on. The hostname will automatically be populated in the drop down box. If there
are multiple interfaces you can use the drop down box to select the preferred interface. You can also
enter an IP address in place of the hostname though this can lead to communication problems when IP
addresses are changed.

Data Interface Pairs


Data Interface Pairs (DIP) are used to explicitly define the physical IP network path the data will take
from source to target. This is done by specifying source and destination network interfaces using host
name or IP address.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
128 | Data Securit y and Net work Control

Data Interface Pairs can be configured in several different ways:


 Job Configuration Tab of client properties – can be used to configure source and target paths for
a client.
 Data Interface Pairs applet in Control Panel – can be used to configure source and target paths
for clients and MediaAgents.

Robust Network Layer (Network Retries)


During data protection operations, communication is maintained between the agents moving data
(MediaAgent and client agent) and the CommServe® server. This communication is required to keep job
status updates. In situations where communication is lost between the agents and the CommServe
server, the job will continue running based on the maximum number of check intervals and the time
between each interval.

Activity Control
Rules for Activity Control
 If activity is Disabled at any parent level object, all activity in any child levels of the object will
also be disabled. If activity is disabled at the CommCell® level, then the entire CommCell
environment activity is disabled.
 If activity is Enabled at any parent level object, child objects can override that setting and
disable activity at the child object.

Operation Windows
Operation windows allow the CommVault administrator to designate blackout windows in which
designated operations will not run. These rules can be set at the global, client computer group, client,
iDataAgent and subclient levels.

Different operation windows can be defined for data protection jobs, recovery jobs, copy jobs and
administrative jobs. Each defined operation window can have one or more Do not run intervals defined.

Different operation rules can be specified for the same operation type to define specific time intervals
for different days of the week.

Job starts during an operation window blackout period


If a job starts and an operation window is currently preventing jobs from running it will be placed in a
Queued state. This will apply to both indexed and non-indexed jobs. Once the operation window is lifted
and jobs are able to run, the jobs will change to a running state.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 129

Job is running and an operation window blackout period become active


If a job is currently running and an operation window blackout period becomes active indexed and non-
indexed jobs will behave in the following ways:

 Indexed based jobs will finish writing their current chunk then be placed in a waiting state.
When the blackout period is lifted the job will continue from the most successfully written
chunk,
 Non-Indexed jobs will continue writing and will ignore the operation window blackout period.

Allow running jobs to complete past the operation window


The behavior of actively running jobs can be configured in the Job Management applet in the Control
Panel. The setting: Allow running jobs to complete past the operation window, can be selected to allow
all currently running indexed based jobs to continue running even when blackout windows become
active.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
130 | Data Securit y and Net work Control

Network Throttling

The network traffic for Clients and MediaAgents can be throttled based on the network bandwidth in
your environment. This is useful to regulate network traffic and minimize bandwidth congestion.

By default, network throttling is disabled. You can enable the throttling options for an individual client, a
client group consisting of multiple clients, and/or a MediaAgent. Once configured, the throttling options
are applied to all data transfer and control message operations, such as Backup operations including
Laptop Backups, Copy operations including DASH copy, restore operations, etc.

The throttling values set up in the throttling rule regulate the rate at which the data is sent and received.

You can also set up relative bandwidth throttling to ensure performance when the client machine
connects with limited bandwidth. Multiple rules can be created for same client/client group, however
the lowest values set up in different rules takes precedence for each time that intersects.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 131

Configuring Data Encryption

Inline encryption
Data can be encrypted as it is being backed up using inline encryption. Encryption can take place on the
client or on the MediaAgent. There are two steps to implement inline encryption:

 Enable encryption for the client - encryption can be enabled to use specific encryption
algorithms and bit length.
 Encryption is applied at the subclient level - Choose which subclients will be encrypted. This
allows you to specifically define data that will be encrypted.

Offline Copy Based Encryption


Data can be encrypted during auxiliary copy operations. This is useful when the primary location that
data is being backed up to is disk and secondary locations are tapes which will be sent off-site. This
method has the advantage that it will not impact primary backup performance.

Hardware Encryption
Simpana software supports LTO standards for data encryption. For LTO generation 4 drives and above,
the LTO standard includes AES encryption. In this case the drive will perform all encryption and
decryption. The Simpana software can manage encryption keys in the CommServe database and
optionally include the keys on the media for recovery through the Media Explorer tool.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
132 | Data Securit y and Net work Control

Media Password
The Media Password is a CommCell® and/or storage policy level password that is written to all media.
When using Media Explorer or the catalog feature and a Media Password has been set, the
administrator must enter the password before media catalog operations can be conducted. It is strongly
recommended a Media Password is always set.

When using LTO hardware encryption or Simpana offline copy based encryption there is an option to
place the encryption keys on the media. If the keys are placed on the media, a Media Password must be
set or encrypted data will be recoverable without entering any password.

Encryption Best Practices


Use custom subclients to separate data requiring encryption from data that does not require encryption.
This will improve overall performance by only encrypting data that must be secured.

Carefully consider the option to place the encryption keys on the media. Placing the keys on the media
makes direct access using Media Explorer possible. Not placing the keys on the media means that the
CommServe® server must be available to recover data. If keys are not placed on media ensure proper
protection of the CommServe database as that will be the only location where the keys are stored.

Ensure the media password is configured. The CommCell level media password is set in the System
settings in control panel. Optionally, a media password can be configured for specific storage policies.

If the encryption keys will be placed on the media for recovery using Media Explorer, ensure that any
storage policies where data may need to be recovered using Media Explorer have the Enable Erase Data
option deselected. If this option is enabled, Media Explorer can NOT be used to recover data.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 133

Base Folder and Resource Pack Tools

Process Manager
Process Manager can be used for the following information and tasks:
General – Client information
 Instance Name
 Client Host Name
 Software Version
 Installation Path
 Drive Space
 Key Folder

Services - Show all Simpana Services and current status.


 Start a Service
 Stop a Service
 Restart a Service

Processes - Monitor Simpana processes use of resources.


 Dump Process info
 Kill Process
 View Logs for Process

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
134 | Data Securit y and Net work Control

Log Params - Set Log Parameters for troubleshooting


 DBGLevel (0 - 100) verbosity level reported by module. Higher means more info.
 LogFileSize(MB) (0 - 100) Default 5. Max file size of log before creating new log
 LogFileMaxVer (0 -100) Default 2. # of past log copies to keep before overwriting.
 DbgWaitTime (0-600) Seconds to wait before DbgLevel attaches to process/service (0)

Media Explorer
Media Explorer is a standalone utility tool that can directly catalog and restore data without a
CommServe® server or MediaAgent. This is considered a last resort tool and should only be used in
extreme cases. Media explorer has limited capabilities and certain data cannot be restored. If the Erase
Data feature is enabled for a storage policy, no data can be recovered for that policy using the Media
Explorer tool.

DBMaintenance
This DBMaintenance tool is located at <Software Installation Path>\Base directory. From the command
prompt, run dbmaintenance with appropriate parameters from the list of available parameters. Running
the utility without any parameters will give the complete list of supported parameters. The CommCell
services must be stopped before performing database maintenance.

DBMaintenance can be used to perform the following tasks on the CommServe database:
 Check the database for inconsistencies
 Re-index all database tables
 Shrink the database

CommServeDisasterRecoveryGUI
Disaster Recovery Backup data can be restored at any production-site or a hot-site any time using the
CommServe Disaster Recovery Tool. The backup data can be restored from the Export Destination.

The CommServeDisasterRecoveryGUI.exe utility can perform the following:


 CommServe metadata database
 Restore the SRM Server Database
 Change the Name of the CommServe® server
 Change License file
 Perform Post-Recovery Operations

The CommServeDisasterRecoveryGUI.exe utility can be found in the <install path>\Base directory.

TapeToolGui.exe
TapeToolGUI is used for both troubleshooting and performance analysis of tape devices. The latest
version of the TapeToolGUI.exe utility incorporates the DiskRead.exe utility also found in the CommVault
Resource Pack. With this integration you can measure throughput performance from disk to tape – or
tape to tape – without the CommVault software involved.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Data Secur it y and Net work Control | 135

Selecting the Disk Read option, run the tool on the MediaAgent and specify a UNC path to a source
folder on the Client. This will yield full end-to-end throughput performance. The tool gives you the
option to use the Windows ReadFile API to give you an environmental benchmark for real achievable
performance levels.

GxTail
GxTail is a troubleshooting utility used to monitor and view active log files when troubleshooting
problems in real time. It can be used for offline viewing and online monitoring of log files during a
backup or recovery operation. It also provides filters and markers for viewing log files and comes
preconfigured with markers for common failure messages. GxTail can be used to analyze a log file for a
particular job failure and also stream log files to remote Client Agents.

GxTail is installed in the <Simpana Installation>\Base folder and can be accessed by double-clicking the
GxTail.exe file. Log files can be dropped into the GxTail or opened by using the File | Open menu option.
Log files are organized in tabs providing simplified navigation.

Network tools
CommVault provides various network tools used to troubleshoot and test network connectivity. For
information on these tools, go to:

http://documentation.commvault.com/commvault/v10/article?p=features/network/network_tools.htm

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 137

Module 4 – Performance Tuning

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
138 | Perf ormance Tuning

Topics

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 139

Performance

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
140 | Perf ormance Tuning

Establishing Benchmarks

Benchmarks can be divided into two kinds, component and system. Component benchmarks measure
the performance of specific parts of a process, such as a the network, tape or hard disk drive, while
system benchmarks typically measure the performance of the entire process end-to-end.

Establishing a Benchmark focuses your performance tuning and quantifies the effects of your
efforts. Building a benchmark is made up of the following 5 steps:

1. Understand the process


You can’t document or improve something if you don’t know what’s going on. More importantly, you
need to understand what phases a job goes through and how much each phase affects the overall
outcome.

For example: a backup job over a network to a tape library takes 2 hours to complete. You think it
should take a lot less and you spend time, effort, and money to improve your network and tape drives
and parallel the movement of data. The job now takes 1.8 hours to complete. You gained a 10%
improvement.

Looking at the job in more detail we find that the scan phase of the job is taking 1.5 hours and the rest is
the actual data movement. Switching the scan method reduces the scan phase time to 12 minutes. The
job now takes .4 hours. You gained a 78% improvement.

Knowing what phases a job goes through and how much each phase impacts the overall performance
can help you focus your time, effort, and money on the real problems.
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 141

2. Identify the resources involved


Each hardware component is going to have a theoretical performance limit and a practical one.
Attempting to get improvement beyond these limits without changing the resources involved is a waste
of time. Consider using an LTO 6 vs SDLT.

3. Minimize outside influence


Large data movements are usually done during non-production hours for two reasons – one, they can
degrade production work, and two, production work can degrade the movement of data. You want to
minimize competition for resources in order to get a fair benchmark of what performance is actually
achievable. In those cases where competition cannot be eliminated, you must accept the impact to
performance or invest in more resources.

4. Sample, Sample, Sample


A single measurement is not a benchmark. Tape devices have burst speeds that are not sustainable over
the long run. Networks have various degrees of bandwidth availability over a period of time. A single
snapshot check of bandwidth will not give you a realistic expectation. Do periodic testing over the
actual usage of a resource to determine its average performance. Try to level out the peaks and valleys -
or at least try to identify what causes these variations.

5. Write it down
The hardest lessons are the ones you have to learn twice. Once you’ve established your acceptable
and/or expected performance levels for each resource and end-to-end, write them down and use them
as the baseline for comparing future performance.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
142 | Perf ormance Tuning

Storage Performance

Storage Connections

TCP/IP
TCP/IP is the most common network transmission protocol and the least efficient of the three. Factors
that can degrade TCP/IP performance are:

Latency - Packet retransmissions over distance take longer and negatively impact overall throughput for
a transmission path.

Concurrency - TCP/IP was intended to provide multiple users with a shared transmission media. For a
single user, it is an extremely inefficient means to move data.

Line Quality - Transmission packet sizes are negotiated between sender/receiver based on line
quality. A poor line connection can degrade a single link’s performance.

Duplex setting - Automatic detection of connection speed and duplex setting can result in a half-duplex
connection. Full duplex is needed for best performance.

Switches - Each switch in the data path is a potential performance degrader if not properly configured.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 143

SCSI/RAID
SCSI is the most common device protocol used and provides the highest direct connection speed. An
individual SCSI drive’s speed is determined by spindle speed, access time, latency, and buffer. Overall
SCSI throughput is also dependent on how many devices are on the controller and in what type of
configuration. The limitation of SCSI is the distance between devices and the number of devices per
controller.

RAID arrays extend the single addressable capacity and random access performance of a set of disks.
The fundamental difference between reading and writing under RAID is this: when you write data in a
redundant environment, you must access every place where that data is stored; when you read the data
back, you only need to read the minimum amount of data necessary to retrieve the actual data--the
redundant information does not need to be accessed on a read. In a nutshell – writes are slower than
reads.

RAID 0 (striping) or RAID 1 (mirror) or RAID 1+0 with narrow striping are the fastest configuration when
it comes to sequential write performance. Wider striping is better for concurrent use. A RAID 5
configured array regardless of the striping has the worst write performance. It’s even worse than single
disks. Of course the tradeoff is redundancy should a disk fail.

Note that fine tuning a RAID controller for sequential read/write may be counterproductive to
concurrent read/write. A compromise needs to be worked out if backup/archive performance is an
issue.

iSCSI/Fibre Channel
iSCSI or Fibre Channel protocol (FCP) is essentially serial SCSI with increased distance and device
support. SCSI commands and data are assembled into packets and transmitted to devices where the
SCSI command is assembled and executed. Both protocols are more efficient than TCP/IP. FCP has
better statistics than iSCSI for moving data, but not by much. Performance tuning is usually setting the
correct Host Bust Adapter configuration (as recommended by the vendor for sequential I/O) or
hardware mismatch. Best performance is achieved when hardware involved is from the same vendor.
Given that configuration and hardware is optimum, then for both iSCSI and FCP, performance is
inhibited only by available server CPU resources

Disk I/O
Defragment source volume
Performing I/O to disks is a slow process because disks are physical devices that require time to move
the heads to the correct position on the disk before reading or writing. This re-positioning of the head is
exacerbated by having a large number of files or having fragmented files. You can significantly improve
read performance of the source data by de-fragmenting the data on a regular basis.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
144 | Perf ormance Tuning

Use 64 KB block size for target volume


The standard buffer size used by the CommVault software and most transmission methods is 64 KB. The
standard block size used by Windows NTFS File System is 4 KB. UNIX and Netware Files systems use a
standard of 8 KB. The low block size is for backward compatibility with smaller disks (< 2GB). When
writing a 64 KB buffer to smaller disk blocks, there is overhead involved is acquiring and writing each
block. If you use 64 KB block sizes there is no additional overhead. CPU loading is reduced with
potential for improved I/O.

Optimize Readers/Writers
Parallel reads and writes are possible on disk devices. Adding additional readers and writers can
improve throughput. At some point, there are diminishing returns from adding more reads or
writes. For a single disk, the best speed is obtained by having no more than 2 readers/writers.
Additional throughput may be achieved by adding additional readers/writers, but the speed of each data
stream and the amount of improvement will diminish.

For example: Two readers @ 24Mb/sec can move approximately 20GB/hour. Adding an additional
reader may drop the speed of each stream to 20Mb/sec and move 27GB/hour. The throughput is
higher, but the speed of each stream is lower.

Tape
Limit drives per controller
With today’s fast tape drives it is often difficult for a single controller to keep a single tape drive running
at optimum speed. The more tape drives you have on a single controller, the less performance you’ll get
from each tape drive. Ensure there are adequate MediaAgent and network resources to get the most
from the tape library.

Limit drives per MediaAgent


Reading or writing concurrent data streams to tape devices puts stress on the CPU, memory, and bus
capacity of the MediaAgent host. Factor in possible multiplexing of job streams and you’re keeping a
MediaAgent very busy. Over utilization of the host resources can have a dramatic effect on
performance. Have you ever had the experience of slow performance when your CPU utilization rate
approaches 100%? For optimal performance of all data streams to tape devices you should limit the
number of concurrent data streams to tape to no more than 6 per MediaAgent. If more than 6 drives
are accessible to the MediaAgent, you can set this lower limit on the Master Drive Pool properties page.

Keep buffers filled


Tape drives are dependent on the write buffers to feed a continuous stream of data. As the write buffer
low watermark is approached, the tape drive must take action to prevent gaps of no data on the tape.
For static speed tape devices (e.g. DLT), the drive will stop, re-position, and wait for the write buffer to
fill before continuing. This can reduce a tape device throughput performance by as much as 50%. A
variable speed tape device (e.g. LTO) will slow as the write buffer gets lower and uses its speed variation
to maintain the buffer level within a certain range. You still get reduced performance, but the impact is
a lot less than that of a static drive.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 145

So how do you keep the tape’s write buffer filled? Match the buffer input of data to the output of
data. This can be done by providing sufficient fast individual job streams or by multiplexing slower job
streams together. Note: excessive input to a tape write buffer impacts the previous buffer and
operation. Do not multiplex for the sake of just having more parallel data streams and then be
concerned about poor performance.

CommVault’s TapeTool.exe or TapeToolGUI.exe utility located in the <Install home>\Base directory of


the MediaAgent can be used to determine the achievable read/write throughput for a tape device.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
146 | Perf ormance Tuning

Performance Parameters

Chunk Size
A chunk is the unit of data that the MediaAgent software uses to store data on media. For sequential
access media (tape), chunk is defined as data between two file markers. The default chunk size for
indexed data on Tape is 4 GB and for non-indexed data (databases) is 16 GB. For disk libraries the
default chunk size for this type of media is 2 GB. NDMP Libraries use 4 GB chunks.

A higher chunk size will give you better data throughput for backups, but granular restores (e.g., single
file restore) will be slower. On the other hand large restores, like a full machine rebuild will be a bit
faster. Recommended values are: 4 GB, 8 GB, 16 GB or 32 GB.

A lower value is recommended for frequent checks of slower data protection operations, especially
when data is moving across a WAN link. A chunk size can be set as low as 32KB, but we do not
recommend a chunk size below 512KB.

Block Size
CommVault software uses a default block size of 64 KB for tape devices, 32 KB for Centera devices, and
whatever the formatted block size used on disk. MediaAgents can write to media using different block
sizes if the MediaAgent host Operating System and media device support that block size.

With tape devices, the higher the block size, the better the performance. If the block sizes are larger
than 512 KB, read operations from the media will fail and such media will be over written and re-used if
the When Content Verification Failed option is enabled in the Library Properties (Media) dialog box.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 147

Ensure that the device hardware and MediaAgent host Operating System supports higher block sizes. If
block sizes are not supported the data cannot be restored.

Setting Concurrency Parameters


Disk Writers
The previous section on disk I/O contains more detailed discussion on configuring the number of disk
writers. Setting the number of disk writers is a balance between MediaAgent capacity, drive
configuration, data stream speed, and overall throughput.

Multiplexing factor
The previous section on tape I/O contains a more detail discussion on setting a multiplexing factor. Note
that multiplexing factor for disk libraries is unnecessary and not recommended.

Network Agents
Network Agents are parallel processes that read/write buffers to the transmission path. If not fully
used, they consume resources that might be used elsewhere. Each MediaAgent has a setting in its
Properties: Optimize for concurrent LAN backups. This is enabled by default. If this setting is enabled,
the number of Network Agents is forced to 1 and changing the value has no effect.

If LAN Optimization is not checked we recommend:


 For subclients writing over a WAN, set the number of Network Agents to 1.
 For all other conditions, if memory allows and all other tuning actions have been taken, then
increase the number to the max possible for the shared memory available.

Pipeline buffers
The Data Pipe Buffers setting controls the amount of shared memory allocated for the data pipes on
each client of MediaAgent computer. By default, the system allocates 30 pipeline buffers of 64 KB size
on each client computer (e.g., 30 buffers of 64 KB—each allocates about 2 MB of shared memory). You
can configure additional buffers for transferring data between the client and the MediaAgent. Having
more pipeline buffers may translate into faster dedicated (non-multiplexed) backups to tape devices,
and therefore improve data transfer performance.

Of course, additional buffers consume more shared memory. You can set the number of data pipe
buffers on the MediaAgent or the client. If the value is set both on the client and the MediaAgent, the
setting on the client will be used. Refer to the CommVault Online Documentation, Additional Setting
nNumPipelineBuffers, for more information.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
148 | Perf ormance Tuning

Stream Management

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 149

Data Streams

Data Streams are what CommVault software uses to move data from source to destination. The source
can be production data or CommVault protected data. A destination stream will always be to
CommVault protected storage. Understanding the data stream concept will allow a CommCell®
environment to be optimally configured to meet protection and recovery windows. This concept will be
discussed in great detail in the following sections.

Job Streams
Content requiring protection is defined within a subclient. Each subclient will contain one or more
streams for data protection jobs. For most iDataAgents, it is possible to multi-stream subclient
operations. Depending on performance requirements and how the data is organized in the production
environment, multi-streaming source data can be done by adding more subclients or increasing the
streams for an individual subclient.

Multiple Subclients
There are many advantages to use multiple subclients in a CommCell® environment. These advantages
are discussed throughout this book. This section will focus only on the performance aspects of using
multiple subclients.

Running multiple subclients concurrently allows multi-stream read and data movement during
protection operations. This can be used to improve data protection performance and when using multi-
stream restore methods, it can also improve recovery times. Using multiple subclients to define content
is useful in the following situations:

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
150 | Perf ormance Tuning

Using multiple subclients to define data on different physical drives – This method can be used to
optimize read performance by isolating subclient contents to specific physical drives. By running
multiple subclients concurrently each will read content from a specific drive which can improve read
performance.

Using multiple subclients for iDataAgents that don’t support multi-stream operations – This method
can be used for agents such as the Exchange mailbox agent to improve performance by running data
protection jobs on multiple subclients concurrently.

Using multiple subclients to define different backup patterns – This method can be used when the
amount of data requiring protection is too large to fit into a single operation window. Different
subclients can be scheduled to run during different protection periods making use of multiple operation
windows to meet protection needs.

Multi-Stream Subclients
For iDataAgents that support multi-streaming individual subclients can be set to use multiple read
streams for data protection operations. Depending on the iDataAgent being used this can be done
through the Data Readers setting or the Data Streams setting.

Data Readers
Data Readers determine the number of concurrent read operations that will be performed when
protecting a subclient. By default, the number of readers permitted for concurrent read operations is
based on the number of physical disks available. The limit is one reader per physical disk. If there is one
physical disk with two logical partitions, setting the readers to 2 will have no effect. Having too many
simultaneous read operations on a single disk could potentially cause the disk heads to thrash slowing
down read operations and potentially decreasing the life of the disk. The Data Readers setting is
configured in the General tab of the subclient and defaults to two readers.

Allow multiple readers within a drive or Mount Point


When a disk array containing several physical disks is addressed logically by the OS as a single drive
letter, the Allow multiple readers within a drive or mount point can be used as an override. This will
allow a backup job to take advantage of the fast read access of a RAID array. If this option is not selected
the CommVault software will use only use one read operation during data protection jobs.

Data Streams
Some iDataAgents will be configured using data streams and not data readers. For example, Microsoft
SQL and Oracle subclients use data streams to determine the number of job streams that will be used
for data protection operations. Data Streams are configured in the Storage Device tab of the subclient.
Although they will be configured differently in the subclient, they still serve the same purpose of multi-
streaming data protection operations.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 151

Device Streams
As Job Streams are received by the MediaAgent, data is put into chunk format and is written to media as
Device Streams. The number of device streams that can be used will depend on the library type, library
configuration and storage policy configuration.

Storage Policy Device Streams


Device streams are configured in the properties of the storage policy. The general rule of thumb is that
the number of device streams configured in a storage policy should always equal the number of drives
or writers of all libraries defined in the storage policy primary copy. Having fewer number of streams
may be used to throttle parallel throughput, but that doesn’t make maximum efficient use of the
devices and there are other means to restrict allocation of devices. If the number of device streams is
greater than the total number of resources available no benefit will be gained. The CommVault software
uses a throttling mechanism to always use the lowest stream value throughout the data movement
process.

Disk Library Device Streams


When using Simpana deduplication

The default setting for disk library and mount path streams is set to allow maximum. This means the
storage policy device streams setting will govern the total number of streams being written to the
library which defaults to 50. There is also a media management parameter ‘maximum number of data
transfer operations for a deduplication database’ which is also set to 50. It generally recommended not
to go beyond 50 streams without consulting with CommVault engineers first.

When not using Simpana deduplication

For disk libraries the number of device streams is based on the total number of mount path writers for
all mount paths within the library. If a disk library has two mount paths with ten writers each, a total of
twenty device streams can write to the library. It is important to note that since disk libraries allow
multiple write operations multiplexing is not recommended. By increasing the number of mount path
writers, more job streams can be written to device streams on a one-to-one ratio. If network,
MediaAgent and disk resources are adequate increasing the number of writers for a mount path will
have a positive effect on data protection performance.

Tape Library Device Streams


For tape libraries one sequential write operation can be performed to each drive. If there are eight
drives in the library then no more than eight device streams will be used. By default each job stream will
write to a device stream. To allow multiple job streams to be written to a single tape drive, multiplexing
can be enabled. The multiplexing factor will determine how many job streams can be written to a single
device stream. If a multiplexing factor of four is set and there are eight drives a total of thirty two job
streams can be written to eight device streams.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
152 | Perf ormance Tuning

Deduplication and Stream Management

Stream management using Simpana deduplication is important to achieve best data protection
performance. Considering most data after the first full backup will be deduplicated, library performance
becomes less critical and more streams can be written to storage. Using client side deduplication also
significantly reduces network traffic. Considering both of these advantages, considerably more streams
can be used during data protection jobs allowing for more concurrent operations to be conducted.

When using deduplication, the primary bottleneck in most environments will be the deduplication
database. If the deduplication database is properly designed, 50 streams should be configured to
maximize data protection throughput. Stream settings will be configured in two primary areas: the
storage policy and the disk library.

Storage Policy Stream Settings


In the General properties tab of the storage policy, the Device Streams setting should be set to 50. If
global deduplication storage policies are being used, this option should be set in the global
deduplication policy. For best performance, the storage policy device stream setting should NOT be set
higher than 50.

Disk Library and Mount Path Stream Settings


Library and mount path stream settings should be configured to allow up to 50 streams in a spill and fill
load balanced configuration. This will be configured in two primary locations: the library properties and
mount path properties.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Perf ormance Tuning | 153

Library Property Settings


The default stream setting for a disk library is set in the Mount Path Allocation Policy option in the
Mount Paths tab and is configured by default to allow maximum streams. Setting the allowable
maximum streams is suitable since the storage policy maximum streams will be set to 50. However, it is
critical to note that when following standard building block guidelines where one storage policy writes
to one MediaAgent and library, this configuration is adequate. The problem is when CommVault best
practices are not followed and many storage policies write to the same MediaAgent and library. The
result could be a significant degradation in performance. Because of this, a recommendation of setting
the library streams to 50 could be used to avoid potential bottleneck problems from future
misconfigurations.

Most disk libraries will be comprised of multiple mount paths. In order to load balance between mount
paths, it is recommended to set the Mount Path Usage option in the Mount Paths tab to use Spill and
Fill. By setting the number of streams to 50, storage policy streams to 50 and the mount path usage to
spill and fill, maximum performance can be achieved.

Mount Path Settings


In the Allocation Policy tab of the mount path properties, the Mount Path Allocation Policy option can
be used to determine how many write streams will be allowed for the mount path. This option is set to
Maximum Allowed Writers by default and for most situations, should be left at the default setting. By
setting the mount path to maximum allowed writers, the throttling is taken care of by the storage policy
stream settings. In certain cases where a mount path may be under performing, allocating a specific
number of writers can be used to throttle down the number of write operations to the mount path.

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
154 | Perf ormance Tuning

Please Give us your Feedback!

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Index | 155

Index

Activity Control, 128 Meeting Restore Requirements, 76


Agent, 63 Data protection planning
Agents Data availability, 59
Custom Package Install, 99 Service level agreement, 59
Decoupled install, 99 Storage Requirements, 61
Interactive installation, 95 DBMaintenance, 134
Push Install, 97 Deduplication, 34
Restore Only Agent, 100 Best Practices, 45
Archiving, 55 building block, 37
Ask the Educator, 16 Client Side, 35
Best Practices Client Side Disk Cache, 35, 43
Encryption Best Practices, 132 Compression, 43
Business Continuity, 52, 60 Deduplication and Stream Management, 152
Certification, 13, 14 Deduplication Database, 35
Chunk Size, 146 Deduplication Database Scaling, 110
Clients, 63 Deduplication Storage Options, 38
CommCell Architecture, 24 Deduplication Store, 36
CommCell deployment Global Deduplication, 110
Existing CommCell Upgrade Process, 81 Global Deduplication for Secondary Copies,
New CommCell Deployment Process, 80 110
Common Technology Engine, 19 Partitioned Deduplication Database, 40
Best Practices, 28 Signature Generation, 34
CommServe server, 24, 25 Signature Hash Comparison, 35
Cluster, 26 SILO storage, 42
CommServe DR IP Address, 26 Storage Side, 35
CommServeDisasterRecoveryGUI, 134 Variable Content Alignment, 44
DBMaintenance, 134 Direct Attached Storage, 31
DR backup process, 26 Disaster recovery, 48
DR Metadata, 28 Disaster Recovery, 60
Installation best practices, 101 CommCell Disaster Recovery Process, 82
virtualized CommServe, 25 CommServe server, 82
CommVault® Advantage, 10 MediaAgent Recovery, 83
compression, 43 Edge data protection, 55
Content Indexing, 114 Education Advantage, 9
Copy Precedence, 113 Education Services Forum, 11
CVLab, 15 Encryption, 131
Data Interface Pairs, 127 Copy Based Encryption, 131
Data Protection Hardware Encryption, 131
Best Practices, 73 Inline encryption, 131
Database applications, 74 Erase Data, 114
File Backup recommendations, 74 Firewall, 124
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
156 | Index

Configuring Incoming Connections, 125 Chunk Size, 146


Configuring Outgoing Routes, 126 Device Streams, 151
Direct connection, 124 Disk I/O, 143
Gateway, 124 Disk Writers, 147
Options, 126 Establishing Benchmarks, 140
proxy, 124 Multiplexing, 147
Pushing Firewall Settings, 126 Network Agents, 147
Gap Analysis, 49 Optimize Readers/Writers, 144
GridStor, 112 Pipeline buffers, 147
GxTail, 135 Storage connections, 142
Image Level Backups, 56 Stream Management, 148
Index Cache, 27, 29 Streams, 149
Configuration best practices, 102 Subclients, 149
Index Cache Configuration, 90 Tape, 144
Maintaining the Index Cache, 90 Pipeline buffers, 147
Sizing the Index Cache, 90 Process Manager, 133
Indexing, 27 Protecting Applications, 68
Installation Virtualized Applications, 68
Custom Packages, 99 VSS aware applications, 68
Decoupled Install, 99 Recovery Point Objective, 49
Installation Methods - Interactive, 95 Improving Recovery Point Objectives, 76
Push Install, 97 Recovery Time Objective, 48
Silent Install, 100 Improving Recovery Time Objectives, 76
IntelliSnap, 70 Replication, 55
How IntelliSnap Technology Works, 71 Retention, 118
IntelliSnap Architecture, 72 Return on Investment, 50
Learning Bytes, 16 Risk Assessment, 49
Legal Hold, 115 Robust Network Layer, 128
Libraries Service Level Agreement, 48, 59
Configuration best practices, 103 SILO Storage, 42
library Snapshots, 55
Disk library, 30 Storage Area Network, 32
Tape library, 30 Storage policy
Library Business Based Design Strategy, 108
Cloud library, 31 Deduplication’s Impact on Policy Design, 109
Exhaustive Detection, 93 Storage-Based Design Strategy, 107
Library & Drive Configuration Tool, 92 Storage Policy
Plug & Play Library, 31 Hide Storage Policy, 113
Media Explorer, 134 How Many Storage Policies, 111
Media Password, 132 Incremental Storage Policy, 113
MediaAgents, 24 Legal Hold Storage Policy, 115
Installation, 88 Storage Policy Design Best Practices, 116
Installation best practices, 102 Subclient, 74
Multiplexing, 147 Subclients, 63
Network Attached Storage, 32 Allow multiple readers within a drive or
Network Throttling, 130 Mount Point, 150
Operation Windows, 128 Data Readers, 150
Performance tuning Subclients Associations, 115
www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.
Index | 157

TapeToolGui, 134 Virtual Machine Protection best practices, 75


Total Cost of Ownership, 50 VSA Backup Process, 66
Training Units Calculator, 16 VSS, 67
virtual clients, 64 VSS, 67
Virtual Server Agent, 64

www.commvault.com

The Information contained in this document is subject to change without notice. No unauthorized use, copy or distribution.

You might also like