Veeam Backup Replication Best Practices
Veeam Backup Replication Best Practices
Veeam Backup Replication Best Practices
of Contents
Veeam Backup & Replication Best Practices 1.1
Introduction 1.2
Contacting Veeam Software 1.3
Architecture Overview
DNS Resolution 2.1
Veeam Backup Server 2.2
Deployment Method 2.3
Backup Server Placement 2.4
Sizing and System Requirements 2.5
Veeam Backup & Replication Database 2.6
Protecting Veeam Backup & Replication Configuration 2.7
Veeam Enterprise Manager 2.8
vCloud Director Self Service Portal 2.8.1
Search Server and Indexing 2.9
Proxy Servers 2.10
Transport Modes 2.10.1
Direct Storage Access 2.10.2
Virtual Appliance Mode 2.10.3
Network Mode 2.10.4
Backup from Storage Snapshots 2.10.5
NetApp Data ONTAP integration 2.10.5.1
Nimble Storage integration 2.10.5.2
Selecting a Transport Mode 2.11
Sizing a Backup Proxy 2.12
Backup Repository 2.13
Repository Types 2.13.1
SMB 2.13.2
Deduplication Appliances 2.13.3
2
Integration specifics 2.13.4
Windows Server 2012 Deduplication 2.13.5
Repository Planning 2.14
Sizing 2.14.1
Per VM Backup Files 2.14.2
Scale-out Backup Repository 2.14.3
WAN Acceleration 2.15
Anaysing Wan Acceleration Workload 2.15.1
Comparing WAN Acceleration Modes 2.15.2
Sizing For WAN Acceleration 2.15.3
Sizing Targets for WAN Accereration Relationship 2.15.4
Deployments For WAN Acceleration 2.15.5
Is WAN Acceleration Right For me 2.15.6
Tape Support 2.16
Tape Support Deployments 2.16.1
Tape Support Media Information 2.16.2
Tape Support Config Requirements 2.16.3
Tape Support Parallel Processing 2.16.4
Tape Support Virtual Full 2.16.5
Tape Support Writing to Tape 2.16.6
Tape Support Restores 2.16.7
Veeam Explorers 2.17
Interaction with vSphere 2.18
Hyper-V Concerns 2.19
Operational Guidelines
Job Configuration 3.1
Backup Methods 3.2
Encryption 3.3
Deduplication and Compression 3.4
Backup Job 3.5
Backup Copy Job 3.6
Replication Job 3.7
3
Application-Aware Image Processing 3.8
Data Verification Using Virtual Labs 3.9
Applications
Overview of Applications Support 4.1
Active Directory 4.2
Microsoft Exchange 4.3
Microsoft SQL Server 4.4
Microsoft SharePoint Server 4.5
Oracle Database 4.6
MySQL 4.7
IBM Notes/Domino 4.8
SAP HANA 4.9
Appendices
4
Sizing Summary 6.1
Networking Diagrams 6.2
Backup Server 6.3
Proxy Server 6.4
Repository Server 6.5
Storage Integration 6.6
Data Validation 6.7
Application-aware Image Processing 6.8
Enterprise Manager 6.9
5
Veeam Backup & Replication Best Practices
Build 9.5.0.823.
All rights reserved. All trademarks are the property of their respective owners.
Important! Please read the End User Software License Agreement before using the
accompanying software program(s). Using any part of the software indicates that you accept
the terms of the End User Software License Agreement.
6
Introduction
Introduction
Welcome to the Best Practices guide for Veeam Backup & Replication.
As you possess a downloaded version of this e-book, you will notice many references to
external resources for additional information.
The e-book is optimized for digital consumption, and the most recent copy is always
available at:
bp.veeam.expert
Intended Audience
This guide is intended for backup administrators or consultants managing Veeam Backup &
Replication on a daily basis.
Most sections of this guide assume you already have hands on experience with Backup &
Replication, and will serve as an "advanced user guide", meaning that more basic usage
information, system requirements and the like must be found in User Guide in Veeam
Helpcenter.
Service providers delivering BaaS and DRaaS with Veeam Cloud Connect should refer to
the corresponding Veeam Cloud Connect Reference Architecture.
Authors
Preben Berg (@poulpreben)
Andreas Neufert (@AndyandtheVMs)
Tom Sightler
Pascal di Marco
7
Introduction
8
Contacting Veeam Software
Online Support
If you have any questions about Veeam solutions, you may use the following resources:
Customer Support
Should you have any technical concerns, suggestions or questions, please visit the Veeam
Customer Portal at cp.veeam.com to open a case, search our knowledge base, reference
documentation, manage your licenses or obtain the latest product release.
Company Contacts
For the most up-to-date information about company contacts and office locations, please
visit www.veeam.com/contacts.html.
9
DNS Resolution
DNS Resolution
Domain Name System (DNS) resolution is critical for Veeam Backup & Replication
deployment (VBR) and configuration. All infrastructure components should be resolvable
through a fully qualified domain name (FQDN). This is especially important for
vSphere/Hyper-V hosts and clusters. resolvable means that components are accessible
through both forward (A) and reverse (PTR) lookups.
Ensure that the Veeam Backup & Replication server is installed on a machine that has a
resolvable fully qualified domain name (FQDN). To check that the FQDN is resolvable, type
nslookup your-vbr-server-fqdn.domain.local at a command line prompt. If the FQDN is
resolvable, the nslookup command returns the IP and name of the Veeam Backup &
replication server.
Only if DNS resolution is not available you may add the infrastructure components like e.g.
VMware vCenter, ESXi and managed Veeam servers to the local hosts file on all managed
Veeam servers. When using this workaround it is recommended to add both short name and
fully qualified domain name in the hosts file.
When ESXi hosts are added to vCenter it is recommended to use FQDN. When backing up
through the network with the Network Block Device (NBD) transport mode, the FQDN is
returned via VMware API for Data Protection (VADP) so the backup proxy server must be
able to resolve the FQDN via DNS. Using the hosts file the data transport path can be
altered for NBD transfers.
To explicitly alter the data transport path, the hosts file must be deployed on all backup
proxy servers. For easier management, please see the Carbon module and Set-HostsEntry
by Aaron Jensen.
10
DNS Resolution
11
Veeam Backup Server
Backup Server
Veeam Backup & Replication is a modular solution that lets you build a scalable availability
infrastructure for environments of different sizes and configurations. The Backup Server is
the core component. Features & component requirements will affect your decision how you
install the backup server e.g. one datacenter or multiple locations. It could mean that you
choose to install additional backup servers or services in remote locations to optimize the
data streams.
Before installing the Veeam Backup & Replication server it is important to understand the
different data streams generated by the Veeam Backup Server (VBR) Services.
12
Deployment Method
Deployment Method
You may deploy the Veeam Backup & Replication server as either a physical or virtual
server. It will run on any server with Windows Server 2008 R2 or higher installed (64-bit
only). Install Veeam Backup & Replication and its components on dedicated machines.
Backup infrastructure component roles can be co-installed. The following guidelines may
help in deciding which deployment type is the best fit for your environment.
Virtual deployment
For most cases, virtual is the recommended deployment. As it provides high availability for
the backup server component via features like vSphere High Availability, vSphere Fault
Tolerance or Hyper-V Failover Clustering. It also provides great flexibility in sizing and
scaling as the environment grows.
The VM can also be replicated to a secondary location such as a DR site. If the virtual
machine itself should fail or in the event of a datacenter/infrastructure failure, the replicated
VM can be powered on. Best practice in a two-site environment is to install the Backup
server in the DR site, in the event of a disaster it is already available to start the recovery.
Physical deployment
In small-medium environments (up to 500 VMs) it is common to see an all-in-one physical
server running the Backup & Replication server, backup proxy and backup repository
components. This is also referred to as an "Appliance Model" deployment.
In large environments (over 2,500 VMs) installing Backup & Replication services on
separate servers either virtual or physical will provide better performance. When running
many jobs simultaneously, consuming large amounts of CPU and RAM, scaling up the
virtual Backup & Replication server to satisfy the system requirements may become
impractical.
An advantage of running the Veeam Backup & Replication server on a physical server is that
it runs independently from the virtual platform. This might be an ideal situation when
recovering the virtual platform from a disaster. Should the physical server itself fail, there are
additional steps to take before reestablishing operations:
13
Deployment Method
Tip: It is recommended to store the configuration backup, using a file copy job, in a location
that is always available to this standby Backup & Replication server.
14
Backup Server Placement
By evaluating the roles and understanding the data flow between the services it is possible
to optimize overall backup performance and restore throughput significantly.
The amount of collected information is typically very small however the Host discovery
process may take longer or even exceed the default schedule in highly distributed
environments1. If hosts or clusters are connected to vCenter over a high-latency link you
may consider deploying a Backup server locally on the ROBO, then you can create a
vCenter service account with a limited scope to that particular location in order to reduce the
window of the Host discovery process. If the ROBO uses a stand-alone host it is possible to
add the host as a managed server directly instead of through vCenter.
Note: Avoid adding individual hosts to the backup infrastructure if using shared storage in a
vSphere cluster.
15
Backup Server Placement
If storage with advanced integration (HPE, NetApp, EMC, Nimble) are added to the Storage
Integration tab there will additionally be a Storage discovery process periodically
rescanning storage hourly. This process checks all snapshots for virtual machine restore
points for usage within Veeam Explorer for Storage Snapshots. The Veeam Backup &
Replication server itself will not perform the actual scanning of volumes but it will use the
management API's of the storage controller to read information about present snapshots.
Only proxy servers with required storage paths available will be used for the actual storage
rescanning process2.
16
Backup Server Placement
9. Update configuration
database
The scan of a storage controller performs, depending on the protocol, several tasks on the
storage operating system. Therefore it is recommended to have some performance
headroom on the controller. If your controller is already running on >90% CPU utilization,
keep in mind that the scan might take significant time to complete.
The scanning interval of 10 minutes and 7 days can be changed with the following registry
keys.
17
Backup Server Placement
Key: SanRescan_Periodically_Days
Type: REG_DWORD
Default value: 7
Defines in days how frequent we should initiate periodic full rescan after Veeam Backup
service rescan
Per default Veeam will scan all volumes and LUNs on the storage subsystem. During
rescan, each present snapshot produces a snapshot clone, mounts to a proxy server, scans
the filesystem, lookup for discovered VMs and unmounts. This is repeated for every present
snapshot.
Example: A storage system with 50 volumes or LUNs with 10 snapshots for each. Scanning
the entire system means 500 (50x10) mounts and clones are performed. Depending on the
performance of the storage system and the proxy server, this can take significant time.
To minimize the scan time it is recommended to select the volumes used by VMware within
the setup wizard to avoid the overhead of scanning unused data volumes.
When file-level recovery is performed from the Veeam backup console, two mounts are
initiated:
Note: For VMs not running a Windows operating system, a Linux based FLR helper
appliance mounts the backup file for reading the file system.
18
Backup Server Placement
Between 50-400 MB of data is transferred between the console and backup repository. If the
first file mount is performed over a slow connection it may take considerable time to load the
file-level recovery wizard. If there is significant latency between the backup repository and
console, it is recommended to deploy an instance of the console on or closer to the
repository server.
It is possible to avoid the first mount entirely by using "guest file system indexing"3. When
guest file system indexing is enabled, the content of the guest VM is stored in the Veeam
Catalog and presented through Veeam Enterprise Manager. Veeam Enterprise Manager will
initiate the file-level restore with the mount server without requiring the first mount.
Note: If guest file system indexing is disabled restores may still be initiated through
Enterprise Manager however they will still require the first mount to be performed with similar
performance implications as previously described.
Veeam Explorers
Veeam Explorers are installed as part of the backup server and backup console when
installed remotely. When performing item-level recoveries the file-level recovery engine is
leveraged. Please see the previous section for deployment considerations.
The Veeam Explorer for SQL Server, SharePoint and Oracle all use a staging server to allow
selecting a specific point in time for point-in-time restore. This introduces an additional
connection as illustrated below.
19
Backup Server Placement
Proper planning dictates that to get 1-Click Failover working it requires that the vSphere
clusters in each location are connected to separate vCenter servers. In the event of an
outage in the primary datacenter it is only possible for the Backup & Replication server in the
DR site to initiate failover if the vCenter server itself is available.
In cases when it is impossible to have multiple vCenter instances across sites (e.g. Metro
Cluster or similar active-active configurations), the recommended solution is to use vCenter
Server and following these steps in event of a disaster:
1. Replicate vCenter from primary site to secondary site with low RPO
2. Configure VMware DRS affinity rules4 for pinning replica vCenter VM to a specific host
3. Connect to specified host and manually power on replicated vCenter VM
4. Verify vCenter availability through Veeam Backup & Replication
5. Initiate Failover Plans
20
Backup Server Placement
Examples
In this section we will outline two examples based on two enterprises with 50 remote/branch
offices (ROBO). They have the following common characteristics:
2. Add the vCenter Server via the Veeam Backup & Replication console
3. Add the ROBO backup server as Managed Server in the Backup Infrastructure tab
4. Configure the HQ backup server with the roles Backup Repository and optionally WAN
accelerator
5. Configure the ROBO backup server with the roles Backup Proxy, Backup Repository
and optionally as WAN accelerator5
6. Configure one or more Backup Jobs for each ROBO pointing to its local backup
repository
7. At HQ configure one or more Backup Copy Jobs for each ROBO pointing to the backup
repository
8. Install Veeam Backup Console on the ROBO backup server for faster restore via the
local Mount Server
Note: The remote console installation files are on the same installation media as Veeam
Backup & Replication ( \Backup\Shell.x64.msi )
Constraints
Please consider the following constraint:
If a WAN link between HQ and a ROBOs fails, no backup jobs will run, as the backup
server will not be able to communicate with the remote ESXi hosts via the centralized
vCenter Server
21
Backup Server Placement
When performing file-level restore for non-indexed virtual machines at the ROBO via
Veeam Enterprise Manager the restore point will be mounted over the WAN link to HQ
for displaying the contents of the restore point. Thus it is recommended to use indexing
for such virtual machines
3. On vCenter Server, create separate service accounts per ROBO with a limited scope for
displaying only relevant hosts or clusters
4. At the ROBO, add vCenter Server via the Backup Infrastructure tab using the scoped
service account
5. Optional: At the ROBO, configure a local WAN accelerator and create or re-use an
existing WAN accelerator at HQ (please note many-to-one configurations are
supported)
6. At the ROBO, add and configure the Repository Server at HQ (please note many-to-one
configurations are supported)
7. Configure one or more Backup Jobs at each ROBO pointing to its local backup
repository
8. Configure one or more Backup Copy Jobs at each ROBO pointing to the centralized
backup repository at HQ (use WAN acceleration as needed)
9. Install Veeam Backup & Replication Console at HQ. When using the remote console for
connecting to remote instances, it is possible to leverage faster file-level or item-level
restores at HQ via the console's built-in Mount Server
Note: As components are managed by multiple backup servers, always ensure that the
same patch/update/version level is used for the entire Veeam backup infrastructure.
1
. In very large or extremely distributed environments, it is possible to extend the
schedule frequency by altering registry key VolumesDiscover_Periodically_Hours
(REG_DWORD, default: 4) ↩
2
22
Backup Server Placement
3. More information about guest file system indexing in Veeam Helpcenter > Guest file
system indexing ↩
4. VMware Distributed Resource Scheduler > VM-Host Affinity Rules ↩
5. Remember to add sufficient resources if all three roles can run on the remote backup
server. ↩
23
Sizing and System Requirements
Sizing with Veeam is cumulative in respect to configurations, if you want to create an all-in-
one appliance (Appliance Model) add all the resource requirements together (CPU +
Memory) to understand what in total you will need, the same goes if you only wish to have
proxy and repository in one host.
Compute requirements
Recommended Veeam backup server configuration is 1 CPU core (physical or virtual) and
5 GB RAM per 10 concurrently running jobs. Concurrent jobs include any running backup
or replication jobs as well as any job with a continuous schedule such as backup copy jobs
and tape jobs. (Concurrent job – a job that processes a single VM with a single virtual disk)
It is recommended to group multiple virtual machines into a single job for better efficiency
and resource usage. With default configuration it is recommended to configure at minimal a
VM with a single disk up to around 30 VMs per job. The recommendation can be increased
by over 10x (300+ VMs) by leveraging additional features such as per VM backup files.
Please refer to the Job Configuration section of this guide to learn more about job design.
All configuration and session information is stored in the configuration database. In larger
environments the load on the SQL Server hosting the configuration database may be
significant and is highly dependent on the amount of concurrently running jobs. For more
information please see the Backup Server Database section of this guide.
Operating system
The Veeam backup server requires Microsoft Windows 2008 R2 or later (64-bit only). The
latest supported version of Windows OS is always recommended (currently Microsoft
Windows 2016) as it will also support restoring from virtual machines with ReFS file systems
or Windows Server Deduplication enabled.
For the full list of supported operating systems, please refer to the corresponding System
Requirements section of the Veeam User Guide.
24
Sizing and System Requirements
Disk space
This section explains what folders you should plan for when preparing for installation of the
Veeam backup server.
Installation folder
Default location is C:\Program Files\Veeam\Backup and Replication
Plan for 40 GB. If installing in a virtual machine, thin disks may be used. By default the
installer will choose the drive with most available free space for the built in backup
repository.
Log files
Default location is C:\ProgramData\Veeam\Backup
Log file growth will depend on the number and frequency of jobs and the VM count.
Consider that the logging level may also affect the log size, if you need to change the
logging level or log file location refer to this Veeam Knowledge Base article:
http://www.veeam.com/kb1825.
Plan for 3 GB log files generated per 100 virtual machines, with a 24 hour RPO. For
environments with more than 500 VMs it is recommended to change the default location to a
different fast access disk. Many concurrently running jobs may produce a lot of write streams
to log files, than can slow down operations for the Veeam Backup Service and Backup
Manager processes.
This folder is used if VM guest indexing in backup jobs is enabled. For more information,
refer to the Search Server and Indexing section of this guide. To change the default location,
refer to this Veeam Knowledge Base article: http://www.veeam.com/kb1453
25
Sizing and System Requirements
When booting VMs with Instant VM Recovery or SureBackup, this folder is used by default
to store all configuration files and redo logs of the running VM. To offload the changes to a
specific production datastore refer to the corresponding page of the Instant VM Recovery
wizard.
The vPower NFS server is bound to backup repositories and the folder location is defined
per server. To achieve best performance for VMs running off of vPower NFS please
configure the fastest possible storage on the backup server or backup repository. To change
the folder location please see the following steps.
It is recommended to reserve at least 10 GB space for this folder. If you plan to start a
significant number of VMs or run VMs over a longer period increase the space accordingly to
fit the produced/estimated amount of changes generated by the running VMs (conservative
average change rate can be defined as 100 GB per 1 TB VM per 24 hours - or 10%).
Additional disk space is consumed when using Quick Migration. See more information here
> Veeam Help Center > Performing Instant VM Recovery > Before You Begin.
Important! Make sure vPower NFS is configured correctly on the Veeam backup server
itself as it will be used when deploying Virtual Lab for SureBackup or when performing file-
level recovery for Linux-based VMs.
For information on folders required for Enterprise Manager, backup proxy and repository
servers (backup targets) and WAN accelerators, as well as for recommendations on their
sizing please refer to the corresponding sections of this guide.
Other software
26
Sizing and System Requirements
If it is not possible to connect to a remote SQL staging server for Veeam Explorers you can
install Standard or Enterprise versions of SQL (depending on your licensing) locally for
staging databases for item-level restores on the backup server. This installation can also be
used to store the Veeam backup database if required as long as sufficient resources are
assigned to the host machine, however do not run any instances in production from this
installation that may affect the operation of the backups or restore processes. SQL express
is included in the distribution but is limited to a 10GB database.
Other software such as Microsoft Outlook (64-bit) for mail export to PST files via Veeam
Explorer for Exchange, or a PDF viewer for reading Veeam documentation are considered
non-disruptive.
Note: Veeam Backup Enterprise Manager must be updated before updating Veeam backup
servers.
After installing updates open the Veeam Backup & Replication management console. The
Update screen will be displayed and will guide you through updating distributed components
on other Veeam managed servers (like proxy and repository servers, vPower NFS servers,
WAN accelerators and tape servers).
Note: As Veeam deploys no agents on the virtual machines, you do not need to update any
software (agents) on the VMs.
27
Veeam Backup & Replication Database
When planning the Veeam Backup & Replication deployment you must choose the
placement of the configuration database. It may be either a local or remote SQL Server with
several licensing options available. Please see the following recommendations to ensure
your Backup & Replication setup will scale to the size of your infrastructure.
If none of the above apply it is recommended to use Microsoft SQL Server Express Edition
for the sake of simplicity.
28
Veeam Backup & Replication Database
Tip: Veeam Backup & Replication supports Microsoft SQL Server 2008 or higher. To
leverage Microsoft SQL Server 2014 enhancements (cardinality estimator has proved to
show significant improvements for large queries), it is highly recommended to update the
database server to Microsoft SQL Server (Express) 2014 or higher.
Database Placement
It is possible to leverage a remote SQL Server as staging server during restores in Veeam
Explorer products. There are no specific edition requirements for neither SQL Express,
Standard or Enterprise instance of SQL Server installed locally on the backup server. It is
still recommended to run the SQL Server locally (when resource and planning allow) on the
backup server for lowest latency and highest performance.
There may still be scenarios where a remote SQL Server is the better choice:
High Availability - SQL Clustering and AlwaysOn Availability Group on external SQL
Servers can be used for configuration database high availability
Licensing - Some enterprises have dedicated virtual clusters for SQL Server due to
licensing constraints. In such cases, you may place the Veeam configuration database
on existing instances to lower the overall TCO
Sizing
Veeam Backup & Replication may consume high amounts of CPU and RAM while
processing backup or replication jobs. To achieve better performance and load balancing it is
necessary to provide sufficient RAM and CPU resources to Veeam components. Remember
to add additional resources, if the backup server is responsible for multiple roles, such as
repository server or backup proxy.
Up to 50 4 8 GB
Up to 100 8 16 GB
29
Veeam Backup & Replication Database
Note: Concurrently running jobs include any job type with a continuous schedule such as
Backup Copy Jobs.
When running more than 100 jobs concurrently increase compute resources in line with the
table above to meet the resource need of the workload.
Tip:
If you plan to restore encrypted databases with Veeam Explorer for Microsoft SQL Server or
SharePoint you will need a valid encryption certificate on the staging Microsoft SQL Server3.
Follow Microsoft general recommendations for optimal SQL performance, for example, place
the SQL tempdb on the fastest disks for best performance5.
30
Veeam Backup & Replication Database
If using SQL authentication consider that all Veeam UI and Veeam PowerShell changes are
communicated using this authentication.
http://www.veeam.com/kb1250
http://www.veeam.com/kb1448
1. Features Supported by the Editions of SQL Server 2012
https://msdn.microsoft.com/en-us/library/cc645993(v=SQL.110).aspx#CrossBoxScale
↩
2. Generic requirements for SQL Server can be found here:
https://helpcenter.veeam.com/docs/backup/vsphere/system_requirements.html?ver=95
↩
3. For restoring encrypted databases, please see: http://www.veeam.com/kb2006 ↩
http://blogs.msdn.com/b/cindygross/archive/2009/11/20/compilation-of-sql-server-
tempdb-io-best-practices.aspx ↩
31
Protecting Veeam Backup & Replication Configuration
These configuration backups contain all the information about Veeam Backup & Replication
such as Backup Infrastructure components plus objects, Backup jobs (passwords are not
stored by default), Sessions and Tape setup. The configuration backup can be used to
automatically rebuild the Veeam Backup & Replication server with all objects, sessions and
jobs.
To restore all jobs and their metadata (you will be asked for all required passwords during
the restore process if they are not stored in the configuration backup encrypted). Please
refer to the Veeam Backup & Replication User Guide for further details:
https://helpcenter.veeam.com/docs/backup/vsphere/vbr_config.html?ver=95
Tip: If encryption is enabled for configuration backup the passwords are also stored in the
configuration backup files.
1. Make sure the daily configuration backup is not placed in the default location on the
backup server itself
2. Modify the backup configuration backup settings to point to a secure backup repository
on a different location/site
3. Schedule the configuration backup to run when the backup server is least occupied;
4. Make sure the configuration backup is encrypted to protect the configuration details.
Also all passwords are than stored in the configuration backup files
5. Check that you receive notifications about the status of the configuration backup job
32
Protecting Veeam Backup & Replication Configuration
results
6. Think about placement of the backup server, configuration backup and database. This is
highly depended on the overall infrastructure design and DR strategy of your
organization
By default, Veeam Backup & Replication is configured to create a daily configuration backup.
The resulting configuration backup file is stored in the \VeeamConfigBackup\%BackupServer%
folder on the default backup repository. However, for security’s sake, it is recommended that
you do not store configuration backups on the default backup repository or in any other
folder on the backup server. In this case, if the backup server fails, its configuration data will
remain, and you will be able to recover the failed backup server.
Also check the location of the configuration database, when the database is external ensure
this server is replicated to the secondary site. If the server is rebuilt in the event of a disaster,
you may start the secondary location Veeam Backup & Replication server and connect to
the configuration backup. This will help to lower overall Recovery Time Objective (RTO).
Tip Use Veeam's File Copy Job to place a copy of the configuration backup at the DR site.
You can configure another repository for that purpose.
Note All data required for a restore is directly placed within the backup file (which VMs are in
the backup file as well as deduplication and encryption information), even in the event that
configuration database is lost or damaged you can set up a new Veeam backup server and
import the backup files there, or even use the stand-alone “Extract” utility (both a command
line and a graphical version are provided). Then you will be able to restore VMs, files and
application data without restoring the configuration database.
Note: Backup copy jobs do not process configuration backups. Remember that configuration
backups are not processed with backup to tape jobs; if you want to store configuration
backups on tape use file to tape jobs instead.
33
Protecting Veeam Backup & Replication Configuration
C:\Program Files\Veeam
C:\Program Files(x86)\Veeam
C:\VeeamFLR\*
C:\Windows\Veeam
C:\Windows\VeeamLogShipper
C:\Windows\Veeam
C:\Windows\Veeam
C:\Windows\Veeam
Files
34
Protecting Veeam Backup & Replication Configuration
VeeamAgent.exe
VeeamAgent64.exe
Tip: Due to the complex nature of antivirus software some additional exclusions may be
needed. If the antivirus has a logging or history system you can review its logs to detect
whether it has taken any actions that might affected Veeam Backup & Replication
operations.
Consider that other services or process may be using ports configured for the Veeam
vPower NFS Service. To avoid possible issues it is recommended to stop the Veeam vPower
NFS Service if you do not plan to use it. Make sure that none of the NFS ports are used by
other software (including antivirus systems). For more information please refer to this Veeam
Knowledge Base article: http://www.veeam.com/kb1055.
35
Veeam Enterprise Manager
Whether to Deploy?
Enterprise Manager is intended for centralized reporting and management of multiple
backup servers. It provides delegated restore and self-service capabilities as well as the
ability for users to request Virtual Labs from backup administrators. It provides a central
management point for multiple backup servers from a single interface. Enterprise Manager is
also a part of the data encryption and decryption processes implemented in the Veeam
solution and best practice recommend deploying Enterprise Manager in the following
scenarios:
It is recommended to deploy Enterprise Manager if you are using encryption for backup
or backup copy jobs. If you have enabled password loss protection
(https://helpcenter.veeam.com/docs/backup/em/em_manage_keys.html?ver=95) for the
connected backup servers backup files will be encrypted with an additional private key
which is unique for each instance of Enterprise Manager. This will allow Enterprise
Manager administrators to unlock backup files using a challenge/response mechanism
effectively acting as a Public Key Infrastructure (PKI).
In enterprise deployments delegation capabilities can be used to elevate the 1st line
support to perform in-place restores without administrative access.
Searching the Indexes can also be used to find files that have been backed up and the
indexes stored in the Enterprise Manager database.
If the environment includes a single instance of Backup & Replication you may not need to
deploy Enterprise Manager, especially if you want to avoid additional SQL Server database
activity and server resource consumption (which can be especially important if using SQL
Server Express Edition).
36
Veeam Enterprise Manager
Note: If Enterprise Manager is not deployed, password loss protection will be unavailable.
Note: It is possible to restore VM guest files even when application-aware image processing
or file indexing is disabled. If both are disabled, the restore operator must type in guest OS
credentials during a file-level restore.
The backup catalog on the Enterprise Manager server will be used to store indexing data
replicated from the backup catalog on Veeam backup server(s). For more information about
the process, refer to the Enterprise Manager User Guide. To learn more about Veeam
Backup Catalog sizing refer to the “Search Server and Indexing” section of this document.
37
Veeam Enterprise Manager
You have two options to restore through Enterprise Manager: 1-Click Restore to Original
Location or Restore with Custom Settings. When restoring with custom settings make sure
that the restore operator is enabled to also restore Oracle Databases. For more information
see providing access rights
Note: Database restore from storage snapshots via Enterprise Manager is not supported.
Tip: This feature is available only for the Windows-based VMs and requires Veeam Backup
& Replication Enterprise Plus license. The VM needs to be in the same domain with the
Enterprise Manager or in a trusted one (for SID resolution)
1. During the backup of a VM with guest processing enabled, Veeam detects users who
have local administrator access rights to that machine and stores this information in the
Enterprise Manager database.
2. User enters the self-service web portal URL in the web browser and enters the account
name and password to access the necessary VM guest OS.
38
Veeam Enterprise Manager
3. After logging in the user is presented with the most recent restore point for that VM (the
one this user authenticated to) on the Files tab of the web portal.
Note: This feature also works for backups from Veeam Agents for Windows stored on a
Veeam Backup & Replication repository.
For more information on using this feature refer to the Self-Restore of VM Guest Files
section of the Veeam Backup Enterprise Manager User Guide.
39
vCloud Director Self Service Portal
40
vCloud Director Self Service Portal
In order to harden the installation of the vCloud Portal, administrators can work on the IIS
(Internet Information Server) website created by Veeam installer, and leverage all the
security features available in IIS itself.
NOTE: Because the vCloud Portal is a subfolder of the Enterprise Manager installation, in
order to modify its settings, the same settings need to be edited for the entire installation.
The mount operation of the index is instantaneous, and a tenant can browse the content of
the backup set to look for the file(s) he needs. Once the file has been identified, there are
three different options:
tenant can download the file locally into his own workstation from the Self-Service
Backup Portal
tenant can restore the file in its original location inside the guest VM, overwriting the
previous version
41
vCloud Director Self Service Portal
tenant can restore the file in its original location inside the guest VM with a new name,
so that both the new and the previous versions are kept
Option 2 and 3 use the same restore mechanism: Veeam first tries to connect to the Guest
VM via the network, but since this is usually an isolated network inside vCloud Director and
there is no direct connectivity between the vCloud Organization Network and the
management network where Veeam (actually, the mount server) is deployed, VMware VIX
API (up to vSphere 6.0) or VMware vSphere Guest Interaction API (starting from vSphere
6.5) are used to complete a networkless restore.
The file is restored in the original location, with the “RESTORED-“ prefix:
NOTE: vSphere API used for these operations are mainly designed for executing commands
inside the Guest OS, not for file transfers. For this reason, performance of large file restore
operations may not be optimal. Please consider the "Download" option for such activities.
Veeam Backup & Replication uses a Multi-OS FLR Helper Appliance virtual appliance to run
file level restores for non-Microsoft file systems. This appliance is configured by a Veeam
administrator before it can be used for any file restore. Otherwise, the first time a tenant tries
42
vCloud Director Self Service Portal
to restore a file for one if his Linux VMs, he will receive this error in the Self-Service Backup
Portal:
A Veeam administrator needs to configure the appliance from the Veeam Console. This can
be achieved by initiating a file lever restore for any Linux VM:
The restore wizards asks to configure the Helper Appliance. The wizard suggests that the
appliance should be connected to the same network where the guest VM is located, but it
misses the other important information, that the FLR appliance needs to connect first of all to
the Veeam mount server via port TCP/6170.
43
vCloud Director Self Service Portal
44
vCloud Director Self Service Portal
The tenant has the three different options to restore one or more files from the backup set.
While the Download option is immediately consumable by the tenant, the two Restore
options require even more networking configurations, as the Veeam Backup Server would
try to connect to the Guest VM to start the restore process from within the guest, but since
there’s no network connectivity between the two, it will fail:
45
vCloud Director Self Service Portal
For this reason, when Veeam Backup & Replication is used in completely fenced
environments, we suggest to leverage the download options of the vCloud Self-service
portal, and let tenant consume this portal to retrieve the files they need. To avoid a double
operation of downloading the file to their workstations and then uploading them again to the
linux machine, we suggest as a best practice to access the portal from a virtual machine
already running inside the vCloud virtual datacenter. If the machine used to retrieve the files
is not the final destination of the restored files, a tenant will just need a tool like WinSCP to
transfer the file to the linux machine, but both the download and the scp copy will happen in
a local network, with the files not even leaving the service provider datacenter.
Customer1 owns a single linux virtual machine called linux, inside the linux_vapp vcloud
app. He wants to restore a file from the latest backup, so he starts the procedure from the
self-service portal as described before; the customer selects the restore point and asks the
software to initiate the mount operation.
The customer can browse the content of the backup, do searches, and download any file he
may need. In the backend, Veeam Backup & Replication is using the FLR Appliance to
mount the backup and read the linux filesystem used by the linux virtual machine.
The machine is automatically disposed (powered off and deleted from the vSphere
environment):
For the entire duration of the restore process, the FLR will be powered on and used by the
tenant.
The configuration of the FLR appliance can be done in two ways, by assigning a fixed IP
address or by leveraging a DHCP server. As the appliance is often managed as a regular
server, and to be sure it always have an IP address to start and execute the restores, many
administrators configure it with a static IP address. The IP 10.2.50.126 in our example is a
static IP address as you can see from the previous screenshot.
During a file restore from the portal, Veeam Backup & Replication uses the existing
configuration of the FLR appliance, since there is no possibility to change the configuration
from the portal itself. This works perfectly for one single restore operation, but if another
46
vCloud Director Self Service Portal
tenant tries to do a file restore for one of his linux machines after the first customer is already
performing a restore, an error will be returned:
Customer2 has to wait until Customer1 has no restore operation running anymore, before
he can start his own restore. This is done on purpose to avoid multiple FLR appliances to be
spin up using multiple times the same IP address, thus leading to unexpected results.
To allows multiple concurrent restores, the solution is to configure the FLR appliance with a
dynamic IP address, once a service provider has verified that a DHCP server is available in
the port group where the appliance will be connected:
47
vCloud Director Self Service Portal
48
Search Server and Indexing
Indexing
Indexing and Search Overview
Veeam Backup & Replication performs backups at the image-level using APIs available from
the underlying hypervisor. It has no direct visibility of the file structure after backup is
finished. It is possible to Use File Level Recovery (FLR) wizard or Enterprise Manager to
mount VMs from within a backup file and access/restore VM guest files. If a user wants to
perform file restore from the central Enterprise Manager it is not possible within an
acceptable timeframe to mount all backup files and VMs in it to find a file that the Enterprise
Manager user wants to restore. To support advanced file-level restore scenarios Veeam
offers the capability to index files on VMs being backed up. Indexing is available for both
Windows & Linux VMs allowing users of Enterprise Manager to browse and search for the
necessary files and to perform one-click file restores.
The sections below will outline some specific use cases for indexing and describe best
practices and guidelines for sizing.
During the VM backup job run the following operations are performed If configured:
49
Search Server and Indexing
1. Veeam accesses the guest OS (using credentials specified in the job settings) and
injects a small run-time process to collect the list of files.
For Microsoft Windows-based VMs the process gathers file metadata by reading
the MFT data of the supported file system (NTFS and ReFS).
For Linux-based VMs the process leverages the existing “locate” database that is
commonly installed on most Linux distributions. Veeam uses the following software
packages for it: mlocate, gzip and tar
These operations take place in parallel with the backup and do not increase the duration of
the process. For more details on the indexing process refer to the Veeam Backup Enterprise
Manager User Guide.
1. Veeam Backup & Replication creates a catalog (index) of the VM guest OS files and
stores index files on the Veeam backup server in the C:\VBRCatalog\Index\Machines\
{vm_name} folder. Creation of the index is extremely fast and has minimal impact on
2. Once the index is created and stored on Veeam backup servers, the indexing service on
Veeam Backup Enterprise Manager performs index copy — it aggregates index data for
all VM image backups from multiple backup servers to the Enterprise Manager
database while the original index files in the backup servers are deleted to conserve
space. The consolidated indexes are stored on the Enterprise Manager server in the
C:\VBRCatalog\Index\Catalog and are used for search queries .
Important To Note!
To search within the index catalog it is necessary to deploy Veeam Backup Enterprise
Manager, this component is in charge of catalog data replication and retention (see this
section of the User Guide for more details).
If you enable indexing without configuring Enterprise Manager the indexes in the
VBRCatalog folder of the backup server will never be collected or deleted and will
eventually fill up the disk drive.
Windows VM
Temporary space required on the first drive in the VM ( С:\ drive):
50
Search Server and Indexing
This was tested with one million files with 20 characters long filenames in one directory.
Depending on the saved metadata and folder structure of the files, the value can be lower or
higher.
Linux VM
Temporary space required in /tmp :
Linux indexes require around 50% less space because mlocate does not index metadata
such as timestamps and ownership.
When using a Standard license, Enterprise Manager will only keep index data for
restore points still in repositories.
For Enterprise and Enterprise Plus licenses, you can configure Enterprise Manager to
keep indexes even longer, with the default being 3 months. This can significantly
increase the amount of space required for the catalog.
Estimated used space of the final index file after compression is approximately 2 MB per
1,000,000 files for a single VM restore point on the Enterprise Manager server. The indexes
are also stored in the backup files and temporary folders on the backup server.
Example
Below is an example that summarizes the information above. The example is given per
indexed VM containing 10,000,000 files.
2 MB * 10 million files * 60 restore points per month * 3 months index retention = 3.5 GB
Recommended Settings
51
Search Server and Indexing
Place the catalog on a dedicated volume of high performance disk. To change the
default Veeam Catalog folder location refer to this Veeam Knowledge Base article:
http://www.veeam.com/kb1453.
You can enable NTFS compression on the catalog folder. This can reduce the space
requirements by well over 50%. For very large catalogs (with 100s of VMs and 10's of
millions of files) it can be more beneficial to use a Windows 2012 R2 volume with Data
Deduplication enabled. This volume should be dedicated to index files and configured to
run deduplication functions outside of the normal backup window.
It is possible to lower the retention of indexes for offline media (e.g. tape archives) in
order to reduce space requirements. The retention is configurable in Veeam Backup
Enterprise Manager, and it is recommended to configure it for the minimum necessary
to meet the IT policy requirements. Index retention setting is available in the Enterprise
Manager web console under Configuration > Settings > Guest File System Catalog.
To enhance search performance, SSDs can be used. If you plan to index a very large
number of VMs it is recommended to limit the search scope at restore to a single VM
before you click the search button, this will bring faster results.
Notes:
To take advantage of indexing on SUSE Linux Enterprise Server (SLES) you must be
running version 12 or above. In lower versions that do not contain by default the
mlocate package you may try this OpenSUSE package
http://software.opensuse.org/package/mlocate
52
Search Server and Indexing
If you need to use that Veeam Backup Search component (and Microsoft Search Server) for
indexing consider the following notes:
Microsoft Search Server Express Edition can be used as it has no limitations for the
number of indexed files.
Other editions of Microsoft Search Server deliver higher scalability because Search
Server components can be separately installed on multiple servers. If you are using
Enterprise Manager consider that it can spread the load between multiple Microsoft
Search Servers Express automatically.
Microsoft Search Server functionality is used to scan content in the shared VBRCatalog
folder on the Veeam Backup Enterprise Manager server and to create a content index
on the Search Server; this content index is used to process search queries. For more
details, refer to the Veeam Backup Search section of the User Guide.
Note: Though using content index streamlines the search process the index itself can
require significant space on disk in C:\VBRCatalog\Journal\[YYYY_MM]\[search-server] .
Search Server requires an SQL database for its operation. Consider that Microsoft SQL
Server Express Edition leverages only one CPU which limits the Search Server
performance. The database size supported by this edition is also limited (in particular,
10 GB for Microsoft SQL Server 2008 R2 Express Edition or later).
53
Proxy Servers
Proxy Server
With backup proxies you can easily scale Veeam backup infrastructure based on the
organization demands:
In a simple deployment scenario for smaller environments or POC, the backup proxy
is automatically installed on the Veeam backup server as part of the Veeam Backup &
Replication installation.
In advanced deployments, the backup proxy role is manually assigned to one or more
Windows servers. This approach allows for offloading the Veeam backup server,
achieving better performance and reducing the backup window.
Backup proxies can be deployed both in the primary site, where the backup server is
located, or in a remote site where additional infrastructure needs being backed up. A proxy
server is installed on any managed Microsoft Windows server added to the backup
infrastructure. Depending on whether the proxy server is installed on a physical or virtual
machine, different transport modes are available.
A backup proxy handles data traffic between the vSphere or Hyper-V infrastructure and
Backup & Replication during backup, replication (at source and target), VM copy, VM
migration jobs or VM restore. They are also used to detect and scan snapshots to enable
Veeam Explorer for Storage Snapshots features when any supported primary storage is
added to the backup server.
In-line source side data deduplication to skip whitespace and redundant blocks reported
by vSphere Change Block Tracking (CBT) or Veeam File Change Tracking (FCT) for
Hyper-V.
BitLooker: Applies to VMs running Windows OS and using NTFS. For more information,
see the corresponding section of this guide > Deduplication and Compression -
BitLooker
54
Proxy Servers
Technically a backup proxy runs a light-weight transport service that takes a few seconds to
deploy. When you add a Windows-based server to Veeam backup management console
assigning the proxy role to it, Backup & Replication installs the necessary components, and
starts the required services on that server. Any host in a Hyper-V cluster is automatically
enabled as proxy server, when it is added to the infrastructure. When a job is started the
backup server manages dispatch of tasks to proxy servers using its built-in Intelligent Load
Balancer (ILB).
Like any backup vendor using VMware vStorage API for Data Protection (VADP), Backup &
Replication integrates VMware Virtual Disk Development Kit (VDDK) in the Veeam Transport
Service. This is necessary for management interaction with vCenter and ESXi hosts, while in
some scenarios, VDDK is bypassed in favor of Veeam Advanced Data Fetcher for
performance reasons.
Storage optimizations
Stock VDDK transport modes have some limitations, such as being unable to process
multiple disks in parallel, when using virtual appliance transport mode (hot-add), introducing
excessive VMFS metadata updates, when performing replication, or being unable to backup
from NFS based datastores. To overcome these limitations, Veeam introduced logic to
bypass VDDK, when it is more optimal to do so.
Veeam Advanced Data Fetcher (ADF) adds increased queue depth for >2x read
performance on enterprise storage arrays. ADF is supported for Backup from Storage
Snapshots, Direct NFS and virtual appliance mode.
55
Proxy Servers
The default proxy server is configured for 2 simultaneous tasks at installation, whereas
subsequently added proxy servers analyze the CPU configuration. The proxy server
automatically proposes configuring 1 task per CPU core. During deployment, it is determined
which datastores the proxy can access. This information is stored in the configuration
database, and is used at backup time to automatically select the best transport mode
depending on the type of connection between the backup proxy and datastore.
First Backup & Replication checks if data processing can be assigned to a backup proxy
with the following preference:
1. Direct Storage Access (which includes VDDK based Direct SAN or Veeam proprietary
Direct NFS).
2. Virtual appliance mode (hot-add)
3. Network Block Device (NBD)
For more details, see the Transport Modes section of this guide.
After the algorithm identifies all existing backup proxies it distributes tasks via the built-in
Real-time Scheduler (RTS):
1. It discovers the number of tasks being processed at the moment by each proxy and
looks for the server with the lowest load and the best connection.
2. All tasks are added to a "VMs to process" queue. When a proxy task slot becomes
available, RTS will automatically assign the next VM disk backup task to it.
3. Priority goes to the disk that belongs to an already processed VM, after that VMs of
already running jobs have next higher priority.
Tip: At the repository, which writes the backup data, only one thread is writing to the backup
storage per running job. If few jobs with a high number of VMs are processed
simultaneously, you may experience that these threads are cannot fully utilize the available
backup storage performance. If throughput per I/O stream is a bottleneck, consider enabling
per VM backup files.
Tip: Default recommended value is 1 task per core/vCPU, with at least 2 CPUs. To optimize
the backup window, you can cautiously oversubscribe the Max concurrent tasks count, but
monitor CPU and RAM usage carefully.
Parallel Processing
Veeam Backup & Replication supports parallel processing of VMs/VM disks:
It can process multiple VMs within a job simultaneously, increasing data processing
rates.
56
Proxy Servers
If a VM was created with multiple disks, Veeam will process these disks simultaneously
to reduce backup time and minimize VMware snapshot lifetime.
RTS gives priority to currently running parallel processes for VM disk backups.
To achieve the best backup window it is recommended to slightly oversubscribe the tasks
slots, and start more jobs simultaneously. This allow Veeam to leverage the maximum of the
task slots and lead into an optimal backup window.
Note: Parallel processing is a global setting that is turned on by default. If you had upgraded
from older versions please check and enable this setting.
Veeam Installer Service - A service that is installed and started on the Windows server
once it is added to the list of managed servers in the Veeam Backup & Replication
console. This service analyses the system, installs and upgrades necessary
components and services.
VeeamAgent.exe process - a data mover which can be started multiple times (on
demand) for each data stream on the proxy. These processes can operate in either read
or write mode. When used on a proxy server for backup, they are only performing read
operations, while "write" mode is used for writing data on a target backup proxy
(replication). Veeam agents in write mode are also used on all repository types, but will
not be discussed in this chapter.
57
Transport Modes
Transport Modes
Job efficiency and time required for its completion are highly dependent on the data
transport mode. Transport mode is a method used by the Veeam proxy to retrieve VM data
from the source host and write VM data to the target destination.
Both options can be used for Backup from Storage Snapshots. When used with NFS
datastores or Backup from Storage Snapshots, Direct Storage Access mode will also utilize
the Advanced Data Fetcher.
To work in this mode the backup proxy must be deployed as a VM. For smaller deployments
(e.g., several branch offices with a single ESXi host per each office) you can deploy a virtual
backup proxy on a ESXi host that has access to all required datastores. When backup or
replication takes place and a VM snapshot is processed the snapshotted disks are mapped
to the proxy to read data (at backup) and write data (at restore/replication); later they are
unmapped.
Network mode
You may find network mode referred to as nbd in documentation and logs.
58
Transport Modes
The most widespread backup method is network mode, which transports VM data through
the VMkernel interfaces of the VMware ESXi host on which the VM resides.
The benefit of using NBD is the fact that it requires no additional configuration, and is
supported regardless of physical or virtual proxy deployments, or storage protocols used
(including local storage, VMware Virtual Volumes or VMware vSAN). This is also the reason
NBD is used as the fallback method, in case Backup from Storage Snapshots, Direct
Storage Access or Virtual Appliance backup modes fail.
The only requirement is for the proxy to be able to access ESXi hosts on port 902/tcp. NBD
backup throughput is typically limited to using up to 40% of the bandwidth available on the
corresponding VMkernel interfaces. If NBD-SSL is enabled, the throughput is typically 10%
slower than regular NBD. NBD-SSL is enforced for ESXi 6.5 hosts. Read more about this in
Virtual Appliance Mode section - vSphere 6.5 and encryption.
59
Direct Storage Access
The Direct SAN mode uses a direct data path (Fibre Channel or iSCSI) between the VMFS
datastore and the backup proxy for data transfer. The proxy requires read access to the
datastores so Fibre Channel zoning or iSCSI initiator configuration and LUN masking on the
storage array must reflect this. In most cases, the Veeam backup proxies are added to the
same "host group" on the storage as the existing ESXi hosts, in order to ensure all LUNs are
masked correctly.
To use Direct NFS backup mode, the proxies need access to the NFS network and must be
a configured in the NFS server's "exports" for read and/or write access. As NFS based
storage uses IP, the real-time scheduler (RTS) will ensure to always use the backup proxy
with fewest network "hops". This is especially useful, if the NFS network happens to be
routable.
If write access is provided, Veeam will automatically perform full VM restore via Direct
Storage Access for thick provisioned VMs.
Pros
Direct Storage Access mode provides very fast and the most reliable predictable
backup performance (typically, using 8 Gb Fibre Channel or 10 GbE for iSCSI and
NFS).
Produces zero impact on vSphere hosts and VM production networks for backup data
transport.
It is possible to perform full VM restore using Direct Storage Access. This mode will be
used automatically if eligible backup proxies are available in the backup infrastructure,
and the VM disks are thick provisioned.
Direct Storage Access is the fastest backup and restore mode for NFS datastores. It
uses multiple concurrent read and write streams with an increased queue depth via
ADF.
Direct Storage Access for NFS datastores will mitigate the "VM stun" issues that may be
caused by Virtual Appliance Mode (hot-add).
60
Direct Storage Access
Direct Storage Access for FC and iSCSI can be used for replication at the target for the
initial replication (with thick provisioned disks) only. For NFS datastores, Direct Storage
Access can be used for initial and incremental replication passes. There are no
differences on the source replication proxy.
Cons
Typically, Direct Storage Access requires a physical server for Fibre Channel, iSCSI or
NFS connection. For virtual only deployments, Direct Storage Access for iSCSI and
NFS is possible, but would transport the data through networks of the ESXi hosts,
typically making hot-add the more efficient choice.
Restore via Direct Storage Access using Fibre Channel or iSCSI is possible only for
thick-provisioned VM disks. At restore the data stream needs to be coordinated in the
background with vCenter or an ESXi host which can slow down the restore speed.
Consider adding additional hot-add proxy servers for restore (FC/iSCSI only).
Direct SAN mode (FC/iSCSI only) is the most difficult backup mode to configure as it
involves reconfiguring not only the storage but also the SAN, (Fibre Channel zoning,
LUN masking, or reconfiguration of iSCSI targets) to provide the physical proxy
server(s) with direct access to the production VMFS datastores. When such
configuration has been implemented it is extremely important to ensure that HBAs, NIC
drivers and firmwares are up-to-date and that multi path driver software (e.g. MPIO) is
properly configured.
For more information about configuring Direct Storage Access refer to FAQ at Veeam
Community Forums: Direct Storage Access Mode
Example
If datastores or virtual raw device mapping (vRDM) LUNs are connected via shared storage
using Fibre Channel, FCoE or iSCSI, you may add a backup proxy as a member to that
shared storage using LUN masking. This will allow for accessing the storage system for
backup and restore.
Ensure that a connection between the storage and backup proxy can be established. Verify
FC HBAs, zoning, multipath, driver software and iSCSI configurations including any network
changes. To test the connection, you may review volumes visible in Windows Disk
Management, adding one disk per storage system at a time. Once the initial connection has
been verified, add the remaining volumes for that storage array.
61
Direct Storage Access
Recommendations
Use the multipath driver software of the storage vendors choice (preferred integration
into Microsoft MPIO) to avoid disk or cluster failovers at storage level. This will also
prevent the whole storage system from being affected by possible failovers if wrong
data paths are used. It is highly recommended to contact the storage vendor for optimal
settings.
If you attach a large number of volumes to the backup proxy, consider that logging for
the process of searching for the correct volume during the job run can require extra
processing time per VM disk (as well as for overall volume count). To avoid Veeam
logging becoming a bottleneck you can disable logging for this particular task this with
the following registry setting:
To achieve performance/cost optimum, consider using fewer proxies with more CPU
cores available. This will help to fully utilize the HBA or NIC capacity of each proxy
server. A 2 CPU System with 2x 12 cores is considered a good configuration balanced
between throughput and costs.
Changes the Windows SAN Policy to "Offline (shared)". This prevents Windows from
automatically bringing the attached volumes online and also prevents Windows write
operations to the volumes. During Direct SAN restore, if the disks are offline, the proxy
will attempt bringing the volume online, and verify that it is writeable. In case the
operation fails, restore will failover to using NBD mode through the same proxy.
Veeam deploys VMware VDDK to the backup proxy. In most cases, VDDK coordinates
read and write operations (Direct SAN restore) with VMware vSphere allowing
VMware's Software to control the read and write streams in a reliable manner.
62
Direct Storage Access
Disable automount. Open an elevated command prompt and disable automount using
the following commands:
diskpart
automount disable
Present LUNs as read-only to the backup proxy server. This capability is supported by
most modern storage. When possible, implement read-only LUN masking on the
storage system or read-only zoning on the Fibre Channel switches (possible on most
Brocade variants).
Summary
Use Direct Storage Access whenever possible for fast backups reduced load on the ESXi
hosts. Consider using hot-add proxies, as these typically restore faster than Direct SAN
restores. Direct SAN uses VDDK, which can cause excessive metadata updates while hot-
add restore bypasses VDDK.
For NFS datastores, Direct NFS is the best choice for both backup and restore. It delivers
the highest possible throughput, without any negative side effects. You can use it for virtual
and physical proxy deployments.
63
Virtual Appliance Mode
This mode supports a 100% virtual deployment, and uses the VMware ESXi storage I/O
stack, providing very efficient backups and having very little overhead in terms of throughput.
During backup or replication, while the original VM is running off of a VM snapshot, the
original virtual machine disks (VMDK) are mounted via SCSI hot-add to the backup proxy
server. Once the backup or replication job finishes, the disks are unmounted from the proxy
server, and the VM snapshot is committed.
Note: For more information on how it works, refer to the section "Data Backup and Restore
in Virtual Appliance Mode" in Veeam Help Center.
As an example, virtual appliance mode is a good choice for highly dynamic environments,
where it can be difficult for backup administrators to maintain access to newly created
datastores for Direct Storage Access. Prerequisites for using virtual appliance mode are
described in the following knowledge base article: Appliance Mode (Hotadd) Requirements
and Troubleshooting
When planning for the Virtual Appliance mode for a backup proxy consider the time required
for actual hot-add operations (such as adding and removing VM disks from the source virtual
machine) it can add up to 1-2 minutes per VM. For a backup job containing 100 virtual
machines this could result in more than two hours of adding and removing disks with no
actual data processing. To mitigate the issue enable parallel processing and process
multiple disks from the same virtual machine simultaneously (using this transport mode).
Tip: It is recommended to benchmark how such operations affect the backup window by
monitoring a test job in the vSphere console.
Veeam developed Direct Storage Access for NFS based datastores to overcome the
problems with disk hot-add and release which causes significant stuns for NFS based VMs).
Direct Storage Access should be used for all virtual and physical proxy deployment to
backup and restore NFS datastore based VMs.
Pros
64
Virtual Appliance Mode
Using the Virtual Appliance mode for proxy servers enables a fully virtual deployment.
As the proxy will perform source side data deduplication and compression, this mode
will provide satisfactory performance in environments running 1 GbE configurations.
Virtual appliance mode utilizes Veeam Advanced Data Fetcher (ADF), providing
significant increase in throughput for enterprise class storage.
Cons
If working in this mode the backup proxy will occupy the virtual infrastructure resources
impacting consolidation ratio. This could ultimately require additional physical ESXi
hosts and licensing.
This mode requires additional planning and configuration in the enterprise environments
because of the additional large disk Hot-Add processes in VMware vSphere.
Note: For more information see vCenter Server connection overview in the "Veeam Backup
& Replication Server" section of this guide.
It may occur that VMware API reports that unmap and snapshot commit were done correctly
but a snapshot file still remains on disk. These "orphaned snapshots" will grow over time and
can fill up the datastore leading to downtime. To mitigate the issue, Veeam implemented the
following functionality:
Veeam Snapshot Hunter. This feature automatically initiates disk consolidation for VMs
in the "Virtual machine disks consolidation is needed" state. For more information
please see Snapshot Hunter section
Veeam can back up multiple disks of VM in parallel on same proxy (default number
is 4).
Typical "hot-add I/O bursts" during hot-add operations are mitigated by bypassing
65
Virtual Appliance Mode
Note: For additional tips refer to the “Impact of Snapshot Operations” section of this guide.
Backup modes Direct Storage Access and Backup from Storage Snapshots are unavailable
for encrypted virtual machines, and NBD will not be as performant. vSphere 6.5 also
enforces SSL/TLS encryption for network mode (NBD), rendering virtual appliance mode a
much more performant alternative, and will reduce host CPU usage.
Recommendations
Virtual appliance mode should be used when it is not possible to leverage Direct
Storage Access, for example in the case of local datastores, Virtual Volumes (VVOL) or
vSAN.
You will need at least one type of (virtual) SCSI controller added to Proxy Server VM
that is used somewhere at the VMs in your infrastructure to allow VMware to HotAdd
the VM disks at backup.
Add an extra SCSI controller to allow for more VM disks processing in parallel (check
the corresponding Veeam proxy settings, default value is 4). The limit for a single
controller is the maximum number of devices per SCSI controller (15). Max SCSI
66
Virtual Appliance Mode
controllers per VM is 4 = 60 disks max. Adding one additional SCSI controller is usually
sufficient.
When deploying hot-add backup proxies avoid cloning existing VMs as this may lead to
identical UUIDs and cause hot-add operations to fail.
You may re-use any existing Windows server VM (to save on licensing). The Veeam
data mover process runs with ‘below normal’ priority by default.
Note: Changed block tracking (CBT) will be disabled for these hot-add proxies.
Consider that it may impact the backup window in case the said virtual machines should
be included in backup or replication jobs.
Useful links
Specific client OS limitations for Hot-Add processing are documented in Veeam Backup
& Replication Release Notes, * Appliance Mode (Hotadd) Requirements and
Troubleshooting
How to test hotadd manually
67
Network Mode
Network Mode
Network mode is by far the easiest backup mode to implement as it requires no additional
configuration. Veeam uses the same interface to backup and restore VMware configuration
files and to read Change Block Tracking (CBT) information, and download virtual machine
configuration files.
In this mode, the backup proxy will query vCenter for the name of the ESXi host on which
the VM scheduled for backup resides. Typically, hosts are added to vCenter using FQDN,
which means NBD relies heavily on functioning DNS. Regardless if the ESXi hosts are
connected to vCenter using a VMkernel interface on an isolated management network,
VADP backup solutions will attempt to connect to this same interface. Please see the section
on DNS Resolution for more information on how to override the default interface used for
NBD backups.
As the only prerequisite, the backup server and proxy server requires ports 443/tcp and
902/tcp being open to the ESXi hosts.
Note: It is highly recommended to maintain a good network connection between the VMware
ESXi VMKernel port and Veeam Backup & Replication as it will be used by many other
features like Instant VM Recovery, Virtual Lab and SureBackup, Linux FLR appliance, config
files backups etc.
For load balancing Veeam uses a selection of proxy servers based on the network subnet:
Backup proxies in the same subnets as the VMKernel interfaces are selected if you
have the Automatic Selection proxy setting configured in the backup jobs.
68
Network Mode
If no proxy servers are available within same subnet as the VMKernel interface of the
ESXi host, you may have to manually select the proxies that are most suitable to
process the backup job. If Automatic selection is still used, proxies from going through
many network hops, even in other sites may be used to transport data. You can
manually select all eligible proxies to enable load balancing.
69
Network Mode
Pros
Network mode can be used for both backup and restore with same speed.
Being the most mature of all transport modes it supports all types of storage.
Is recommended for NFS based storage in cases where Direct NFS is unavailable.
Using NBD will minimize VM stunning. See also the "Considerations for NFS
Datastores" section of this guide.
As data transfers initiate very quickly, network mode is preferable for processing
incremental backups on relatively static virtual machines (VMs generating a small
amount of change).
70
Network Mode
It can be helpful when dealing with many clusters with individual storage configurations
(e.g. hosting providers). In such deployments, using network mode for data transfer can
help reducing Veeam footprint and costs as well as to increase security (if compared to
other modes and storage configuration).
Cons
Typically, network mode uses only up to 40% of the available bandwidth of the external
VMKernel interface due to throttling mechanisms implemented on the management
interfaces.
Tip: Please see the section on DNS Resolution for information on how to override the
network interface used for NBD backups e.g. when both 1 GbE and 10 GbE VMkernel
interfaces are available, it is preferred to force usage of 10 GbE for highest possible
throughput.
Recommendations
When you choose network mode (NBD), you entirely avoid dealing with hot-add vCenter and
ESXi overhead or physical SAN configuration. NBD is a very fast and reliable way to perform
backups. In emergency situations when you need fast restore the following tips can be
helpful:
Consider setting up at least one virtual backup proxy for hot-add based restores. Then it
will be possible to achieve higher throughput and thus lower RTO.
You can also restore to a thin disk format and later use standard VMware methods to
change the disk format to thick disk if needed. Thin disk restores have to transport less
data.
Another way to overcome this limitation is to use Instant VM Recovery with Storage
vMotion (if license is available) as it is not affected by the same throughput limitations
as the VMkernel interfaces.
As there is no overhead (like SCSI hot-add, or search for the right volumes in Direct
Storage Access) on backup proxies, network mode can be recommended for scenarios
with high-frequency backups or replication jobs, as well as for environments with very
71
Network Mode
To protect VMware, Veeam reduces the number of permitted NBD connections to 28.
Please see the corresponding section in Interaction with vSphere for more information
on how to alter the configuration using registry keys.
72
Backup from Storage Snapshots
Licensing and system requirements are described in the Veeam User Guide: Backup from
Storage Snapshots.
The storage integration covered in this section is VMware only and does not apply for Hyper-
V. Any protocol supported by Backup from Storage Snapshots will utilize the Advanced Data
Fetcher to optimize for retrieving data on enterprise grade storage.
Backup from Storage Snapshots (BfSS) is a feature included in the deep storage array
integrations and a way to optimize and enhance VM backups in a very easy way. The main
objective for implementing BfSS is to minimize the lifetime of a VM snapshot, which reduces
the time for VM snapshot commit and I/O the vSphere environment.
For regular VADP based backups, the VM snapshot is created and remains open (VM snap
lifetime) until the VM backup is completed. Especially with large or highly transactional VMs,
that can lead to large snapshot delta files being created during the backup followed by hours
of snapshot commit tasks within vSphere producing high I/O on the production storage.
73
Backup from Storage Snapshots
Ultimately, these long snapshot commits may lead to unresponsive VMs. For more
information about the impact of VM snapshots please see the "Interaction with vSphere"
section of this book.
How it works
By using BfSS, the VM snapshot lifetime will be significantly reduced. In this section, we will
go through the steps performed.
VM processing limit
74
Backup from Storage Snapshots
When adding a large number of virtual machines to a job, by default steps 1 and 2 (above)
are repeated until all virtual machines within the job have successfully completed. Only then
will BfSS proceed to step 3 and issue the storage snapshot. If adding 100s of jobs to a
backup or replication job, this could cause a very high VM snapshot lifetime for the first VMs
in the job list.
When configuring such large jobs, it is advised to configure the maximum number of VMs
within one storage snapshot. The setting is available in the advanced job settings under the
Integration tab.
Example: When creating a job with 100 VMs, and setting the limit to 10, BfSS will instruct the
job manager to process the first 10 VMs (step 1 and 2), issue the storage snapshot and
proceed with the backup (step 3-7). When step 7 has successfully completed for the first 10
VMs, the job will repeat the above for the following 10 VMs in the job.
As seen below, when ensuring proper configuration of BfSS, minimal VM snapshot lifetime is
achieved, and reduces overall I/O penalty on the production storage for highly transactional
VMs.
75
Backup from Storage Snapshots
Configuration
Enabling BfSS requires minimal configuration, but understanding the tasks and
responsibilities of involved components are key when troubleshooting and optimizing for
high performance and low RTPO.
The backup server is responsible for all API requests towards vSphere and storage arrays
for determining present volumes, snapshots and all necessary details such as initiator
groups, LUN mappings and which protocols are available.
The proxy server(s) are used for reading data from the storage snapshot and sending it to
the backup repository. To leverage Backup from Storage Snapshots, the following
configuration requirements must be met:
76
Backup from Storage Snapshots
Backup server must have access to the management interfaces of the storage array. All
additional prerequisites such as LUN mappings, creation of initiator groups for iSCSI,
altering NFS exports and snapshot management are subsequently handled via this
connection.
Backup proxy servers must be able to directly access the storage array via the same
protocol used for connecting the production datastore (FibreChannel, iSCSI or NFS). As
opposed to using Direct Storage Access, it is not a requirement for the proxy server to have
access to the production datastore itself, as it reads data blocks directly from the cloned
storage snapshot.
As described in previous sections, the backup server and proxy server can be deployed on
one single server or scaled out on different servers. In most environments, where BfSS is
applicable, the components are usually separated for scalability considerations.
When to use
When using Backup from Storage Snapshots, overall jobs processing may take longer, as
additional steps are performed such as mapping vSphere Changed Block Tracking (CBT) to
offsets of the storage snapshot, and the snapshot must be cloned and mounted on the
backup proxy server. The mount overhead can take several seconds on block protocols as
HBAs or initiators must be rescanned. It mostly affect FC deployments.
With this in mind, using BfSS on small VMs or VMs with a very low change rate is not
advised. As the VM snapshot lifetime on such VMs is very short, the benefits of using BfSS
are minimal.
In most environments, large VMs or highly transactional VMs producing large amounts of
changed data benefit most from using BfSS. Using the VM Change Rate Estimation report in
Veeam Availability Suite, you may quickly identify such VMs.
VMs with either virtual or physical Raw Device Mapping (RDM) are not supported with BfSS.
Such VMs will failover to backing up via standard methods if allowed in the job settings.
1
. EMC Unity is supported starting Veeam Backup & Replication 9.0 Update 2
(KB2147) ↩
2. Nimble Storage is supported starting Veeam Backup & Replication 9.5 ↩
3. Cisco HyperFlex is supported starting Veeam Backup & Replication 9.5 Update 2.
Cisco HX utilizes VAAI offloaded storage snapshots, so restores using Veeam Explorer
for Storage Snapshots are not supported. ↩
77
Backup from Storage Snapshots
78
NetApp Data ONTAP integration
Backup from Secondary Snapshots. In case you use NetApp SnapVault or SnapMirror,
Veeam can create a primary snapshot, update the secondary (SV/SM) Snapshot and
backup the CBT changes to the backup file. It is configured with a job setting in the
"Advanced" section if Veeam should allow fallback to the primary snapshot for backup. You
can find the setting within the secondary destination window of your backup job and enable
“Use as the data source”.
Snapshot Orchestration
For NetApp ONTAP storage systems Veeam offers a SnapShot Orchestration only feature.
SnapShot orchestration means to use storage SnapShots as backup target. The feature can
be used without any need to run a real backup to an external repository. Veeam is taking
care of all required storage related tasks like data retention, SnapShot management and
SnapMirror/SnapVault updates to secondary sides.
79
NetApp Data ONTAP integration
To configure a “SnapShot only” job set the Repository to "NetApp SnapShot only"
80
NetApp Data ONTAP integration
The retention policy defines the number of storage snapshots to keep. To store 5 snapshots
a day for 1 week, configure the retention to 35 restore points with a daily schedule. If the job
is configured with a high or lower schedule frequency, adjust the number of restore points
accordingly.
If you use a secondary NetApp ONTAP system with SnapMirror and/or SnapVault you can
set the box for a secondary destination and set the retention.
When using Snapshot Orchestration please take care of the retry scheduler setting.
81
NetApp Data ONTAP integration
If you have for example 100 VMs in one job and 10 of these VMs are failing in the first run
Veeam will rerun the job based on the retry settings. If the setting is set to 3 (default) Veeam
will try 3 more time to process the failed VMs. For every successful retry Veeam will create a
new Snapshot. If all retries are needed to proceed the failed VMs that ends in 3 Snapshots
for one run. It is recommended to not set the value higher than 3 or disable the automatic
retry to avoid a high number of Snapshots being created during every run.
One of the big benefits is that you are still able to use all Veeam restore capabilities from
storage snapshots. For more please refer to the Veeam Explorer for Storage Snapshots
section.
82
Nimble Storage integration
Nimble Storage
This section contains integration specific information for configuring orchestration of
snapshot creation and replication between Nimble Storage arrays.
Veeam will orchestrate the creation of a volume snapshot, and initiate replication to the
secondary Nimble array before the backup job starts.
Note that Veeam Backup & Replication uses its own engine to initiate the creation and
replication of snapshots.
Nimble configuration will not allow empty scheduling. Therefore you can choose Weeks
or Repeat Every Week and Replicate to set to "2" as the minimum — or any desired
configuration, as these configurations will not be used by Veeam.
83
Nimble Storage integration
It is not recommended to rely on storage snapshots as backups, as it violates the 3-2-1 rule.
It is however a great complement to traditional backups to achieve lower RPO, in case the
primary storage array is still available, when a restore is needed.
Note.
Snapshot replication
When configuring backups using the "snapshot only" repository, or regular repositories, it is
possible to configure orchestration of replication to a secondary Nimble Storage array by
checking the Configure secondary destinations for this job.
84
Nimble Storage integration
By clicking Add -- Nimble Snapshot Replicated Copy, it is possible to configure how many
snapshots should be retained at the target Nimble Storage array. During the job run, Veeam
will search for replication settings configured on the Volume Collection for the source volume
being snapshotted. Please see the initial paragraph of this chapter for details on configuring
Volume Collections.
Note.
When configuring replication between Nimble Arrays for integration with Veeam for
Snapshot Only Jobs or for backups to Secondary Array, it is recommended by Nimble to
have each volume have its own volume collection group.
85
Nimble Storage integration
By clicking Add -- Nimble Snapshot Replicated Copy, it is possible to configure how many
snapshots should be retained at the target Nimble Storage array, and furthermore use the
checkbox "Use as the data source". This will instruct the backup proxy to using the
secondary Nimble Storage array as the data source for backups.
86
Selecting a Transport Mode
Name Definition
Very Single host with local disks as primary datastores. Typical ROBO
small configuration.
2-4 hosts with shared storage. Typical ROBO configuration or small
Small
datacenter
Medium 4-20 hosts with shared storage
Large 20-100 hosts with shared storage
Enterprise Over 100 hosts
Keep in mind that within larger datacenters, multiple definitions may apply. As an example, it
is possible that a separate management or DMZ cluster without shared storage could benefit
from using the "Very small" or "Small" recommendations, while the main production
environment is leveraging recommendations based on "Medium" to "Enterprise" datacenter
size.
Very small
Virtual Appliance (Hot-Add) mode is the recommended option, as it gives you the best
performance.
NBD over 10GbE VMKernel interfaces link will provide a very stable and good
performing solution without any special configuration needed.
Direct Storage Access mode or Backup from Storage Snapshots modes are typically
unavailable, as the disks of the host are local and thus cannot be mounted to an
external proxy server.
87
Selecting a Transport Mode
For shared storage connected via FC or iSCSI, you can choose one of the following two
modes:
Physical proxy: Direct Storage Access will provide the best backup performance.
For example, you can configure a physical server with access to FC datastores on
the local site and perform backups to a local repository. If you use thin-provisioned
disks for the VMs, configuring a dedicated backup proxy for restoring via Virtual
Appliance (hot-add) mode can help to increasing restore performance.
Virtual proxy: The Virtual Appliance (hot-add) mode is a good an fast backup
mode. Avoid to backing up VMs on NFS datastores using hot-add. Use Direct
Storage Access or NBD backup modes instead.
NBD over 10 GbE VMKernel Interfaces link will provide a very stable and good
performing solution.
NBD over 1 GbE VMKernel Interfaces can be used for failover and for situations where
you do not have to transport much data.
When using NBD, check the Network Mode chapter for tuning tips.
Large
In addition to the above considerations for Small and Medium, please see the following
guidelines:
When Direct Storage Access, or Backup from Storage Snapshots are unavailable, and
when virtual proxy servers are disallowed, Network Mode (NBD) is the only choice. In
such cases, 10GbE interfaces are a must.
For virtual only deployments (virtual proxies only) in environments with many isolated
clusters, using network mode (NBD) may be ideal. As hot-add requires at least one
proxy within each cluster, it may require many more proxy servers compared to using
network mode.
A combination of hot-add mode for large clusters and NBD mode for smaller clusters
may be ideal.
Enterprise
In addition to the above considerations for Large, please see the following guidelines:
88
Selecting a Transport Mode
When running a central backup server and with multiple branches connected to it, a
dedicated backup server is recommended for at least every 200 branches. Consider
using Veeam Enterprise Manager for federation.
1. In case storage integration is used with Backup from Storage Snapshots (BfSS), the
overhead of mapping blocks from VMware CBT and the storage snapshot can increase
processing time and lead to longer backup windows. To mitigate, consider the majority
if the VMs can be backed up with one of the other transport modes and use BfSS only
for the largest VMs or high change rates (typically 10% of VMs). Veeam ONE Change
Rate Estimation report can help to identify such VMs. ↩
89
Sizing a Backup Proxy
Processing Resources
As described above, you may define the max concurrent tasks value in the backup proxy
settings. It is best practices to plan for 1 physical core or 1 vCPU and 2 GB of RAM for each
of the tasks. A task processes 1 VM disk at a time and CPU/RAM resources are used for
inline data deduplication, compression, encryption and other features that are running on the
proxy itself.
In the User Guide it is stated that proxy servers require 2 GB RAM + 500 MB per task.
Please consider these values as minimum requirements. Using the above mentioned
recommendations allow for growth and additional inline processing features or other special
job settings that increase RAM consumption.
If the proxy is used for other roles like Gateway Server for SMB shares, EMC DataDomain
DDBoost, HPE StoreOnce Catalyst or if you run the backup repository on the server,
remember stacking system requirements for all the different components. Please see related
chapters for each components for further details.
Tip: Doubling the proxy server task count will - in general - reduce the backup window by 2x.
D = Source data in MB
D
T = Throughput = W
CR = Change rate
90
Sizing a Backup Proxy
Example
Our sample infrastructure has the following characteristics:
1,000 VMs
100 TB of consumed storage
8 hours backup window
10% change rate
By inserting these numbers into the equations above, we get the following results.
We use the average throughput to predict how many cores are required to meet the defined
SLA.
T
CF = 100 ≈ 36 cores
The equation is modified to account for decreased performance for incremental backups in
the following result:
T ⋅CR
CI = 25 ≈ 14 cores
As seen above, incremental backups typically have lower compute requirements, on the
proxy servers.
For a physical server, it is recommended to install dual CPUs with 10 cores each. 2
physical servers are required.
For virtual proxy servers, it is recommended to configure multiple proxies with maximum
8 vCPUs to avoid co-stop scheduling issues. 5 virtual proxy servers are required.
If we instead size only for incremental backups rather than full backups, we can predict
alternative full backup window with less compute:
91
Sizing a Backup Proxy
WS = 104857600
14⋅100
W = WS ≈ 21 hours
3600
If the business can accept this increased backup window for periodical full backups, it is
possible to lower the compute requirement by more than 2x and get the following result:
For a physical server, it is recommended to install dual CPUs with 10 cores each. 1
physical server is required.
For virtual proxy servers, it is recommended to configure multiple proxies with maximum
8 vCPUs to avoid co-stop scheduling issues. 2 virtual proxy servers are required.
If you need to achieve a 2x smaller backup window (4 hours), then you may double the
resources - 2x the amount of compute power (split across multiple servers).
The same rule applies if the change rate is 2x higher (20% change rate). To process a 2x
increase in amount of changed data, it is also required to double the proxy resources.
Note: Performance largely depends on the underlying storage and network infrastructure.
Required processing resources may seem too high if compared with traditional agent-based
solutions. However, consider that instead of using all VMs as processing power for all
backup operations (including data transport, source deduplication and compression), Veeam
Backup & Replication uses its proxy and repository resources to offload the virtual
infrastructure. Overall, required CPU and RAM resources utilized by backup and replication
jobs are typically below 5% (and in many cases below 3%) of all virtualization resources.
Consider that some tasks within a job are still sequential processes. For example, a merge
process that write the oldest incremental file into the full file is started after the last VM
finishes backup processing. If you split the VMs into multiple jobs these background
processes are parallelized and overall backup window can be lower. Be as well careful with
big jobs when you use Storage Snapshots at Backup from Storage Snapshots. Guest
processing and Scheduling of jobs that contain multiple snapshots can lead into difficult
scheduling situation and Jobs that spend time waiting for (free) resources. A good size for
jobs that write to per VM chain enabled repositories is 50-200 VMs per Job.
92
Sizing a Backup Proxy
Also, remember that the number of running backup jobs should not exceed 100 jobs
concurrently running (not overall). Veeam can handle more, but a “sweet spot” for database
load, load balancing and overall processing is about 80-100 concurrently running jobs.
Note: Parallel processing may also be limited by max concurrent tasks at the repository
level.
So, in a virtual-only environment you will have slightly more proxies with less proxy task slot
count, while in physical infrastructure with good storage connection you will have a very high
parallel proxy task count per proxy.
The “sweet spot” in a physical environment is about 20 processing tasks 2x10 Core CPU
with 48GB RAM and 2x 16 Gbps FC cards for read + 1-2 10GbE Network cards.
Depending on the primary storage system and backup target storage system, any of the
following methods can be recommended to reach the best backup performance:
Running fewer proxy tasks with a higher throughput per current proxy task
Running higher proxy task count with less throughput per task
As performance depends on multiple factors like storage load, connection, firmware level,
raid configuration, access methods and others, it is recommended to do a Proof of Concept
to define optimal configuration and the best possible processing mode.
93
Sizing a Backup Proxy
Block Size: the smaller the blocks size is, the more RAM is needed for deduplication.
For example, you will see a RAM increase when using LAN mode if compared to Local
target, and even greater (2-4 times) when using WAN. Best practice for most
environments is to use default job settings (Local for backup jobs and LAN for
replication jobs) where another is not mentioned in the documentation or this guide for
specific cases.
94
Backup Repository
Backup Repository
Before you start planning for the repository, go through Veeam Backup & Replication online
documentation at https://www.veeam.com/documentation-guides-datasheets.html to get
basic understanding of repositories.
A backup repository is a storage location used by Veeam Backup & Replication jobs to store
backup files, copies of VMs and metadata for replicated VMs. Technically, a backup
repository is a server that runs the Veeam Transport Service and provides a destination
folder on the backup storage. Each job can use only one repository as its destination
storage, but one repository can be used by multiple jobs.
You can balance the load across the backup infrastructure by setting up several repositories
in the environment and limiting the number of concurrent jobs for each repository, or if you
have a proper license you can leverage Scale-out Backup Repository as explained later on
in this section.
The 3-2-1 rule however is a data protection strategy, whereas availability requires the
different storage implemented in this strategy to support additional capabilities like:
Instant VM recovery
File transforms
Distant copies
Item restoration
Sure Backup
This is the reason why v9.0 introduced two major new features for Veeam backup
repositories : Scale-out Backup Repository and Per-VM Backup chains.
95
Repository Types
Repository Type
Being storage-agnostic, Veeam Backup & Replication supports a wide range of repository
types, each offering its own set of specific capabilities. So when deciding on repository
storage, you might consider the following:
Capacity
Write performance
Read performance
Data density
Security
Backup file utilization
Organization policies may require different storage types for backups with different retention.
In such scenarios, you may configure two backup repositories:
You can consume both layers by setting up a backup copy job from the first to the second
repository, or leverage Scale-out Backup Repository, if licensed.
Since a DAS storage can be fully dedicated to backup operations, this type of repository
is considered to offer a good balance between “performance” and “cost” factors.
A strong benefit of a DAS repository is that it supports the features offered by Veeam
Backup & Replication in a very flexible way. In particular, it provides good read and write
performance, sufficient for Veeam vPower-based features (such as Instant VM
96
Repository Types
However, consider that though DAS is a valuable option in many cases, its scalability may
not meet an organization’s requirements.
Tip: To achieve optimal performance, it is often required to install a battery module to the
server’s controller card in order to enable write-back mode for the internal cache. A DAS is a
shelf with disks, and all the intelligence of the solution is delegated to the controller installed
in the connected server.
Pros Cons
Cost Manageability
Performance Single point of failure
Simplicity Monolithic
SAN Storage
This is a more advanced and manageable solution that offers the same advantages as DAS,
and adds more advantages like higher availability and resiliency.
The volume size and quantity are easily adjustable over time, thus offering a truly scalable
capacity.
Tip: You can configure multiple backup repositories on the SAN storage to increase
repository throughput to the storage system.
Pros Cons
Reliability Complexity
Performance Cost
Technical capabilities
Windows or Linux?
The main difference between Windows and Linux in regards to Veeam repositories is the
way they handle NAS shares – this can be summarized as a choice between NFS and SMB.
Generally, a Linux-based repository can handle a higher throughput than a Windows-based
repository with same CPU/RAM/Disk resources. However, if you deploy Veeam in a small-
97
Repository Types
sized infrastructure, you may want to keep the configuration "all-in-one" on a single Windows
server, so deploying a Linux server as a repository could add extra complexity to the
solution. Other possible concerns relate to cost and administrative burden.
Physical or Virtual?
You can use a virtual machine as a repository server, however, keep in mind that the storage
and associated transport media will be heavily occupied.
If you are using a SAN storage, it can be accessed through software iSCSI initiators, or
directly (as a VMDK or RDM mounted to the Repository VM).
Best practice is to avoid using the same storage technology that is used for the virtualized
infrastructure, as the loss of this single system would lead to the loss of both copies of the
data, the production ones and their backups.
98
SMB
SMB Repository
While an SMB repository is often considered to provide less performance than direct
attached storage, it still can provide very good results as a repository due to leveraging
Veeam’s load-balancing technology for write operations, as explained in the next sections.
Gateway Server
When you set up an SMB share as a repository, the following options are available:
Automatic selection of the server as the SMB gateway proxy (that is, the server that will
host the target-side transport component and thus perform the role of “data writer”
towards the SMB share itself).
Specify a specific server (among the available managed Windows servers in Veeam
Backup & Replication) as a SMB gateway proxy.
The second option is very helpful in situations where the SMB share is located on a remote
location, since it avoids that the automatic selection uses a server that is not local to the
SMB share, thus having all synthetic operations or backup copy jobs occurring over the
WAN link (which is usually slower than the local link). It is always recommended to use an
SMB gateway server as close as possible to the SMB storage. By specifying the SMB
gateway you have a better chance of keeping the data flow under control and avoid data
crossing the WAN links unnecessarily.
As single stream performance for SMB repositories may be suboptimal, you can potentially
increase performance of your SMB storage by configuring several repositories pointing to
the same folder using different gateway servers. With multiple proxies, the automatic SMB
gateway may be a good option and can be configured by selecting Automatic from the
drop-down list.
Tip: Gateway servers must be properly sized as regular Windows repositories. If you are
using Automatic mode, remember that the same machine could be elected backup proxy
and gateway server simultaneously. Apply sizing it accordingly.
Another option for increasing the number of streams is using per VM backup files. Please
see the corresponding section of this guide for more information > Per VM backup files
99
SMB
Even when multiple proxies are used to process a given backup job, only one* Windows
server (called “gateway server") per backup chain will be used to write data to the SMB
share. In Automatic mode the first selected proxy in the running job will become the
gateway server. If per-vm backup files are enabled, this applies to each per-vm chain, thus
multiple gateway servers may be started concurrently.
Here are some recommendations for managing backup proxies used as gateway servers:
The networking between the multiple proxies should be sized correctly to allow data to
flow from each proxy to the gateway server.
As the first backup proxy of a job is used as the gateway server, it may happen that all
the gateway server instances of different jobs (or per-vm backup file chains) are started
on the same proxy. This requires proper sizing of CPU and RAM; ensure resource
monitoring is in place.
Note: Consider that increasing the number of jobs also increases the number of threads to
the NAS storage.
Scaling out using this approach will allow for processing larger amounts of data and optimize
the throughput of the SMB shares. Best practice for large scale environments is to use at
least a mid range or enterprise NAS storage system that provides good I/O performance.
Low end NAS devices often have non official implementations of the SMB protocol that may
improve performance test results, but may also corrupt backup files. For these devices it is
discouraged to use SMB.
100
SMB
101
Deduplication Appliances
Deduplication Appliances
Overview
Deduplication applied to storage is a technique aimed at reducing the storage space
consumption.
Deduplicated storage systems are often optimized for write operations and can offer rather
high ingest rates. However, any random read I/O may suffer from re-hydration processes
required during restores. For this reason we recommend to use these devices mainly as
secondary targets, where parameters like price per GB are more important than restore
performance.
Note: "Re-hydration" is the act of restoring the original blocks in a non-deduplicated form.
During backup files transformation the same blocks are read and then written back to the
appliance where they are de-hydrated (deduplicated) again. This two-step process can
generate significant load on the appliance, slowing down operations.
Also, consider that Instant VM Recovery might not be as fast as expected – unless the
deduplication appliance offers a fast non deduplicated area for the most recent restore
points (such as ExaGrid).
102
Deduplication Appliances
The downside of active fulls is the need to transport the entire amount of virtual machines on
a weekly/monthly basis. This can lead to long snapshot commit, so this mode needs to be
planned carefully. It is recommended to limit the use for primary backup jobs to the
integrated deduplication appliances, where synthetic operations can be used.
If one of the integrated appliance is used, synthetic operations will be performed on the
appliance itself, so they will require minimal additional time and lower I/O.
The repository used for staging should be configured with "Decompress before storing"
advanced option enabled, which ensures previously applied compression at the job level is
ignored.
Also, ensure that the appliance meets Veeam tape requirements described in the User
Guide.
103
Deduplication Appliances
To further reduce restore time, it is recommended to enable file-level indexing for backup
jobs located on deduplication repositories. Indexing VMs will remove the waiting time for
mounting a restore point when browsing contents via Enterprise Manager.
Best Practices
In this section, we will distinguish between integrated and non-integrated deduplication
appliances. Integration is available for:
Limitations:
If the mentioned integration API is unavailable due to licensing restrictions, or if any other
deduplication appliance is used, the appliance should be considered non-integrated.
In order to optimize throughput for deduplication appliances, please use the following
configuration guidelines:
Job configuration
The following settings are configured in the backup job "Edit" wizard under Storage >
Advanced. Options not defined in this table are optional and not related to backup
repositories using deduplication storage.
104
Deduplication Appliances
Configuration
Setting Value
tab
Hardware assisted encryption is available for EMC DataDomain via DDBoost, but must be
configured in the integration specific repository configuration. If enabled on the job level data
reduction efficiency will be significantly degraded.
Repository configuration
The following settings are configured in the "Edit Repository" wizard under Repository >
Advanced.
Setting Value
105
Deduplication Appliances
106
Integration specifics
EMC DataDomain
Selecting DataDomain as a repository will automatically recommend job and repository
settings according to best practices. For more information, refer to vendor guidelines.
Source side deduplication between the Veeam gateway server and DataDomain
appliance. This will reduce the amount of data sent over the network to the appliance
Better LAN parallelization, since DDBoost manages its own network load balancing
algorithms which are considered more efficient than standard network links aggregation
Seamless Veeam files transformations like synthetic full or forever forward incremental
DDBoost can be used through Fibre Channel SAN, providing a totally LAN-free backup
solution
For more details, refer to the DDBoost configuration guide by Rick Vanover: Configuring
EMC Data Domain Boost with Veeam Availability Suite (still applicable for version 9).
ExaGrid
ExaGrid appliances run an integrated Veeam data mover similar to a Linux based backup
repository. With ExaGrid, there is no requirement for a Windows based gateway server.
See Using Veeam Backup and Replication Software with an ExaGrid System for more
information.
ExaGrid recommends configuring 1 job per repository. Thus, if you want to achieve parallel
processing, create several repositories and setup 1 job per repository.
As a rule of thumb, the "landing zone" (which is the zone that will hold most recent set of
data waiting to be deduplicated) should have sufficient capacity for an uncompressed full
backup so that each backup can fully be written there and processed. This ensures
107
Integration specifics
SureBackup, Instant VM Recovery and item-level restores will be usable for the latest
restore point without rehydration overhead.
HPE StoreOnce
Selecting StoreOnce appliance as a repository will automatically recommend job and
repository settings according to best practices. For more information, refer to vendor
guidelines.
If the Catalyst Store is configured as High Bandwidth on the appliance, Low Bandwidth
mode can be forced using the following registry value (ideally, work around the issue by
configuring both Primary and Secondary modes to "Low"):
and
108
Integration specifics
HPE StoreOnce model in: Limitations for HPE StoreOnce to find the maximum limit.
109
Windows Server 2012 Deduplication
1. Use Windows 2012 R2 and apply all patches (some roll-ups contain improvements to
deduplication).
2. Format the disk using the command line "/L" option (for "large size file records") and
64KB cluster size (use parameters /Q /L /A:64K )
3. Follow compression and deduplication guidelines for non-integrated deduplication
storage in previous chapter.
4. Modify garbage collection schedule to run daily rather than weekly.
5. Use backup jobs configured to perform Active full with Incrementals.
6. If possible, spread active full backups over the entire week.
7. Try to keep the .VBK files below 1TB in size (there is no official support from Microsoft
for files bigger than this; see https://msdn.microsoft.com/en-
us/library/hh769303(v=vs.85).aspx). Large files take a long time to deduplicate and will
have to be fully reprocessed if the process is interrupted.
8. Where possible, use multiple volumes. Windows deduplication can process multiple
volumes using multi-core CPU – one CPU core per volume; see
http://blogs.technet.com/b/filecab/archive/2014/12/04/sizing-volumes-for-data-
deduplication-in-windows-server.aspx for details.)
9. Configure deduplication process to run once a day, and for as long as possible.
110
Repository Planning
Configuration Guidelines
Parallel Processing
A repository can be configured to limit the amount of parallel tasks it can process at a time;
with parallel processing enabled (by default) a task is one VMDK handled by the proxy
during a backup job, or by a repository during a backup copy job. If there are many parallel
tasks on the proxy side for only few tasks on the backup repository, this will lead the Veeam
scheduler service to wait for available resources on the repository. To prevent such situation,
you can figure out on which side the bottleneck will be (proxy or repository) and then set the
overall amount of parallel tasks on the proxies equal to the total amount of parallel tasks on
the repositories.
Note: Consider tasks for read operations on backup repositories (like backup copy jobs).
Blocks sizes
During the backup process data blocks are processed in chunks and stored inside backup
files in the backup repository. You can customize the block size during the Job Configuration
using the Storage Optimization setting of the backup job.
By default block size is set to Local target, which is 1 MB before compression. Since
compression ratio is very often around 2x, with this block size Veeam will write around 512
KB or less to the repository per each block.
This value can be used to better configure storage arrays; especially low-end storage
systems can greatly benefit from an optimized stripe size.
There are three layers where the block size can be configured: Veeam block size for the
backup files, the Filesystem, and the Storage volumes.
111
Repository Planning
The Veeam block size of 512KB is going to be written in the underlying filesytem, which has
a block size of 64k. It means that one block will consume 8 blocks at the filesytem lavel, but
no block will be wasted, as the two are aligned. If possible, set the block size at the filesytem
layer as close as possible to the expected Veeam block size.
Then, below the filesytem there is the storage array. Even on some low-end storage
systems, the block size (also called stripe size) can be configured. If possible, again, set the
stripe size as close as possible to the expected Veeam block size. It's important that each
layer is aligned with the others, either by using the same value (if possible) or a value that is
a division of the bigger one. This limits to a minimum the so called write overhead: with a
128KB block size at the storage layer, a Veeam block requires 4 I/O operations to be written.
This is a 2x improvement compared for example with a 64KB stripe size.
Tip: As can be seen from the field, optimal value for the stripe size is often between 256 KB
and 512 KB; however. It is highly recommended to test this prior to deployment whenever
possible.
It is also recommended to use a journaling file systems (this makes exFAT a less reliable
option than NTFS).
112
Repository Planning
A file size limitation can be occasionally reached on NTFS, especially on Windows 2012 R2
with deduplication enabled. This happens due to a hard limit reached on the file records size
because of the high level of file fragmentation. To mitigate the issue, we recommend to
format Windows NTFS repositories with the "/L" (large files) option.
113
Sizing
Repository Sizing
In mid-sized or enterprise environments, the recommended amount of CPU for a repository
is 1 core per concurrent job that processes data on a repository server. At least 2 cores
allow for the Operating System to be more responsive.
It is recommended to configure 4 GB RAM per core. The same amount of resources are
needed for SMB gateway servers. Also, consider that VM recovery processes (Instant
Recovery, FLR and others) require sufficient resources (as described here.
Also, when testing is not possible beforehand, you should make assumptions on
compression and deduplication ratios, change rates, and other factors. The following figures
are typical for most deployments; however, it is important to understand the specific
environment to figure out possible exceptions:
Data reduction thanks to Compression and Deduplication is usually 2:1 or more; it's
common to see 3:1 or better, but you should always be conservative when estimating
required space.
Typical daily change rate is between 2 and 5% in a mid-size or enterprise environment;
this can greatly vary among servers; some servers show much higher values. If
possible, run monitoring tools like Veeam ONE to have a better understanding of the
real change rate values.
Include additional space for one-off full backups.
Include additional space for backup chain transformation (forward forever incremental,
reverse incremental) – at least the size of a full backup multiplied by 1.25x.
Note: When using deduplication appliances, please contact the vendor for sizing guidelines.
Using the numbers above, you can estimate required disk space for any job. Besides,
always leave plenty of extra headroom for future growth, additional full backups, moving
VMs, restoring VMs from tape.
114
Sizing
A repository sizing tool that can be used for estimation is available at http://vee.am/rps. Note
that this tool is not officially supported by Veeam, and it should be used "as is", but it's
nonetheless heavily used by Veeam Architects and regularly updated.
Tip: With Veeam Availability Suite, you can use Veeam ONE together with Veeam Backup &
Replication. Among the many reports, Veeam ONE has the VM Change Rate Estimation
report from the “Infrastructure Assessment” report pack; this can be used as an indicative
pre-deployment assessment of the potential amount of space that should be available on the
backup repositories. This report is built measuring the number of VM virtual disk write
operations supplied by VMware vSphere while additional compression and deduplication
(usually 2 to 3 times) ratio should be assumed.
It is also recommended to periodically run the “Capacity Planning for Backup Repositories”
report from the “Veeam Backup & Replication Reports” pack to analyze the amount of free
space on backup repositories and estimate the projected growth and consequent space
consumption. The report provides recommendations for adjusting the allocated storage
resources in order to meet the future demand for backup storage. Furthermore, it calculates
the amount of additional space that needs to be provisioned to accommodate the necessary
restore points.
For more information on Veeam Availability Suite, please refer to its Reviewer's Guide at
https://www.veeam.com/documentation-guides-datasheets.html
Examples
The examples below explain the impact of backup method and retention policy on the
estimated repository size, assuming the environment is the same in all three cases.
Example 1
Estimated Full Backup Size: 10 * 80GB (Used space) * 50% (2:1 Compression) =
400GB
Estimated Reverse Incremental Size: 10 * 80GB * 50% (2:1 Comp) * 5% (Change Rate)
* 29 (reverse incremental restore points) = 580GB
Spare : 500 GB
Estimated total Backup Size: 400GB + 580GB + 500 = 1480 GB
Example 2
115
Sizing
Estimated Full Backup Size: 10 * 80GB (Used space) * 50% (2:1 Compression) =
400GB
Estimated space for 6 Weekly Fulls (Max required for 30 Day Retention): 400GB * 6 =
2400GB
Estimated Forward Incremental Size Max: 10 * 80GB * 50% * 5% * 32 = 640GB
Estimated total Backup Size: 2400GB + 640GB = 3,040GB (~3TB)
Example 3
Estimated Full Backup Size: 10 * 80GB (Used space) * 50% (2:1 Compression) =
400GB
Estimated space for 3 Monthly Fulls (Max req for 30 Day Retention): 400GB * 3 =
1200GB
Estimated Forward Incremental Size Max: 10 * 80GB * 50% * 5% * 60 = 1200GB
Estimated total Backup Size: 1200GB + 1200GB = 2,400GB (~2.4TB)
To summarize, when estimating the size of the repositories, use the following best practices:
116
Per VM Backup Files
In addition to optimizing write performance with additional streams to multiple files, there are
other positive side effects as well. When using the forward incremental forever backup
mode, you may experience improved merge performance. When backup file compacting is
enabled, per VM backup files require less free space: instead of requiring sufficient space to
temporarily accommodate an additional entire full backup file, only free space equivalent to
the largest backup file in the job is required. Parallel processing to tape will also have
increased performance, as multiple files can be written to separate tape devices
simultaneously.
Per VM backup files is an advanced option available for backup repositories, and it is
disabled by default for new backup repositories. If enabled on an existing repository, an
active full backup is required after the option has been enabled.
117
Per VM Backup Files
When designing your jobs, keep in mind that several operations such as synthetic
operations, health checks and Backup Copy Jobs will be pending until all VMs in the job
have completed successfully. For those reasons, extremely large jobs may be impractical.
Performance
To avoid counter productive effects, attention should be paid on not having too many write
threads towards a storage used as a repository. For example, a low range NAS storage will
probably not react very well to a high amount of parallel processes created by per VM
backup files. To limit this effects, refer to Repository configuration options, especially the
Concurrent tasks limit.
Deduplication
Using Per VM backup file will negatively impact repository space usage since Veeam
deduplication is file based. If backup jobs have been created while grouping similar guests to
optimize deduplication and if Active Full is used, per VM Backup chain might require
additional repository space.
118
Scale-out Backup Repository
Scale-out Backup Repository is an extremely easy way for both medium and large
customers to extend repositories when they run out of space. Instead of facing the long and
cumbersome relocation of backup chains, users will be able to add a new extent (that is any
of the “simple” backup repositories supported by Veeam Backup & Replication) to the
existing Scale-out Repository — or group multiple repositories to create a new one.
The only requirement is the ownership of a proper license, and that at least two simple
repositories have been added to Veeam Backup & Replication already. As per default
settings, it is recommended to enable "per VM backup files" on the Scale-out Backup
Repository for optimal balancing of disk usage.
NOTE: the default backup repository created during the installation cannot be used in a
Scale-out Backup Repository as long as it’s the target of Configuration Backup, as this type
of job is not supported by Scale-out Backup Repository. If the default repository needs to be
added to a Scale-out Backup Repository, consider first to change the target of Configuration
Backup.
For additional technical information, the online documentation is available here : Helpcenter
SoBR.
Data Locality
This is the default policy, and it works by placing all the dependent files of a backup chain
into the same extent. Every extent grouped with this policy has the same chances of
receiving a backup chain as the algorithm treats them equally, and the major parameter for
the initial placement is the free space value.
119
Scale-out Backup Repository
The failure domain is a single extent, as the loss of a given extent impacts only the backup
chains stored into that extent. Policy can be violated by Veeam itself if, for example, one of
the extents has no free space left, and the additional incremental is stored in a different
extent. This because the priority is always to complete a backup or backup copy.
Performance
Performance policy places dependent incremental backup files on a different extent from the
corresponding fulls. In order to choose which extent will hold the different files when using
the performance policy, for each extent users are able to assign it a “role”.
120
Scale-out Backup Repository
Important: When using integrated deduplication devices, virtual synthetic operations may
not work, if the full and incremental backup files are placed on separate extents. Please use
Data Locality mode instead.
Users can configure each repository of the group to accept full backups, incremental
backups or both. As soon as a new backup chain is stored into a performance Scale-out
Backup Repository, the different files are place in accordance to the policy itself.
Note: in order to leverage the performance policy correctly users need to use at least two
different repositories. Even if it’s possible to assign both roles to the same repository, this
configuration makes little sense and the best results can be obtained by splitting full backup
files and incremental backup files over different physical extents.
Performance policy increases the failure domain — a backup chain is split over at least two
repositories, thus the loss of one of the two corrupts the entire backup chain. This is a
consideration that Veeam architects need to evaluate carefully. There is a trade-off between
the increased performance guaranteed by the performance placement policy, and the
increased failure domain.
121
Scale-out Backup Repository
Scale-out Backup Repository is, as the name implies, a scale out architecture, based on
multiple datamovers, with a notion of master and slave repository datamovers.
During backups, the master datamover is always started where the write is happening.
During restore, the master is always started where the VBK is located, as most blocks are
likely retrieved from this location.
A master datamover is the only repository datamover receiving data from a source
datamover (a proxy in a backup job or a source repository in a backup copy job). A master
datamover is able to communicate if needed with other slave datamovers to retrieve their
data.
122
WAN Acceleration
WAN Acceleration
By combining multiple technologies such as network compression, multi-threading, dynamic
TCP window size, variable block size deduplication and global caching, WAN acceleration
provides sufficient capability when the network bandwidth is low or dramatically reduced
when performing Backup Copy and Replication jobs. This technology is specifically designed
to accelerate Veeam job. Any other WAN acceleration technology should be disabled for
Veeam traffic.
The WAN accelerator uses its own digests based on the hashes of the blocks inside a VM
disk, which means that it reads data from the backup files and re-hydrating them on the fly,
or it reads directly from the source VM in case of replication. The WAN accelerator
component will then process those data blocks with much more efficient data deduplication
and compression algorithms. This is the reason why the WAN accelerator consumes
significant amounts of CPU and RAM resources.
To determine how much data has to be transferred over the WAN link with and without WAN
acceleration enabled in a backup copy job, you can compare the daily changes of the
primary backup job statistics (as the same data is transported in a standard backup copy job
without WAN acceleration) with the WAN accelerated backup copy job log and statistics.
123
Anaysing Wan Acceleration Workload
Read — amount of data read from the production storage prior to applying any
compression and deduplication. This is the amount of data that will be optimized by the
WAN accelerator.
124
Anaysing Wan Acceleration Workload
all VMs from the backup job are included in the backup copy job.
The amount of Processed blocks in the backup copy job session is equal to the amount
of Read blocks in the backup job session. This is the most important metric, as it is the
amount of data that has to be processed by the WAN accelerator.
The number of Read blocks for the backup copy job is typically higher than the
amount of Processed - this is due to the backup copy job using a differing fingerprinting
algorithm that works with a different block size compared to the fingerprinting algorithm
125
Anaysing Wan Acceleration Workload
and block size used by backup jobs that created the original backup file. For this
reason, this metric can be ignored.
The amount of Transferred data is the amount of data actually transferred over the
WAN link.
126
Comparing WAN Acceleration Modes
To calculate possible savings and needed bandwidth you may use the following calculator
Bandwidth Calculator.
For example, forward incremental and forever forward incremental methods will make
backup copy jobs work much faster, as read operations will be sequential rather than
random. To avoid similar fragmentation and random I/O on forward incremental modes, keep
backup storage maintenance enabled when possible.
Though a workload penalty may not be significant, it can be a good idea to monitor the
storage latency on the backup repository, especially if the reported bottleneck is Source. If
the storage latency on the backup repository is high, it is recommended that you change the
127
Comparing WAN Acceleration Modes
backup mode in order to increase the throughput of one pair of WAN accelerators.
Configuration
When configuring the WAN accelerator, not all configuration parameters affect both source
and target WAN accelerators. In this section we will highlight what settings should be
considered on each side.
If the link has low latency and high bandwidth, the default setting (5 streams) may be
enough to fully saturate it. If the link is still not saturated, the number of streams may be
increased accordingly.
128
Comparing WAN Acceleration Modes
Testing shows that with high latency links, link speed x 1.5 is a good best practice for
estimating the number of streams required. Below is an example benchmark on a 10 Mbit/s
WAN link with 100 milliseconds of latency.
10 100 0 20 10
Increasing the number of streams to more than required for fully saturating the link will cause
initialization of data transfers to slow down, as the data transfer will wait for all streams to
initialize and stabilize before beginning transferring any data.
Tip: To test different scenarios in the lab before deploying WAN acceleration, you can use a
WAN emulator (such as WANem).
When configuring the cache location for the source WAN accelerator, consider that the
actual cache size on the source is irrelevant, as it is used only for digest files (where block
hashes are stored). However, if a WAN accelerator will be used for bi-directional
129
Comparing WAN Acceleration Modes
acceleration (act as both source and target), follow the guidelines provided in the "Target
WAN Accelerator" section below.
130
Sizing For WAN Acceleration
Hardware
The source WAN accelerator will consume a high amount of CPU and memory whilst re-
applying the WAN optimized compression algorithm. Recommended system configuration is
4 CPU cores and 8 GB RAM. When using an existing Veeam Managed Server for Wan
Acceleration which already has a role such as Veeam Backup & Replication Server, Proxy or
windows Repository ensure you have not overcommitted the CPUs on that host and there is
resource for each source and Target Wan Accelerator. If there is not enough CPU cores free
the job will wait for a free cpu to continue.
The I/O requirements for the source WAN accelerator spikes every time a new VM disk
starts processing. Thus, it is recommended to deploy WAN accelerators on disk
configurations with decent I/O performance.
The typical I/O pattern is made of many small blocks, so using high latency spinning disks is
not recommended.
Disk Size
131
Sizing For WAN Acceleration
Each digest file consumes up to 2% of its source VM disk size. This means, for example,
that a 2 TB VM disk file can produce a digests file up to 40 GB in size.
Additionally, plan for 10 GB of working space for payloads and other temporary files.
For understanding how disk space is consumed, please see the following sections.
Note: As the cache size on the source WAN accelerator will always be ignored, the digests
file will be produced regardless of cache setting been configured. They may consume
considerable disk space.
VeeamWAN\GlobalCache\src
Only a data.veeamdrf file is located in the \VeeamWAN\GlobalCache\src folder. This file will be
synchronized from the target WAN accelerator during the very first job run (or if the cache
was manually cleared) to understand what data blocks are already cached in the target
WAN accelerator. The size of this file is typically up to 2% of the configured target cache
size; thus, it may take some time for the initial data transfer to begin.
VeeamWAN\Digests
On the source WAN accelerator there are the VM disk digests that take up disk space. For
each processed VM disk, a disk digest file is created and placed in \VeeamWAN\Digests\
<JobId>_<VMId>_<DiskId>_<RestorePointID> .
Note: Although the Digest folder is created on the target accelerator no data is stored on the
target normally, however it must be sized into the target in case the digest on the source
becomes corrupt or is missing. In this case the target will calculate its own digests in this
location until the source WAN Accelerator comes back online.
Traffic throttling rules should be created in both directions. See Network Traffic Throttling
and Multithreaded Data Transfer for more information.
The cache size setting configured on the target WAN accelerator will be applied to the
pair of WAN accelerators. This should be taken into account when sizing for many-to-
one scenarios, as configuring 100 GB cache size will result in 100 GB multiplied by the
number of pairs1 configured for each target WAN accelerator.
132
Sizing For WAN Acceleration
2
It is recommended to configure the cache size at 10 GB for each operating system
processed by the WAN accelerator.
Once the target WAN accelerator is deployed, it is recommended to use the cache
population feature (see this section of the User Guide for details). When using this
feature, the WAN accelerator service will scan through selected repositories for
protected operating system types.
It is also possible to seed the initial copy of data to the target repository to further
reduce the amount of data that needs to be transferred during the first run.
Sizing
Hardware
Although a target WAN accelerator will consume less CPU resources than the source, the
I/O requirements for the target side are higher.
For each processed data block, the WAN accelerator will update the cache file (if required),
or it may retrieve the data block from the target repository (if possible). As described in the
user guide, the cache is active on operating system data blocks, while other data blocks are
being processed only with the WAN optimized data reduction algorithm (in-line
compression).
133
Sizing For WAN Acceleration
Tests show that there are no significant performance differences in using spinning disk
drives as storage for the target WAN accelerator cache rather than flash storage. However,
when multiple source WAN accelerators are connected to a single target WAN accelerator
(many-to-one deployment), it is recommended to use SSD or equivalent storage for the
target cache, as the I/O is now the sum of all the difference sources.
Disk Size
Ensure that sufficient space has been allocated for global cache on the target WAN
accelerator.
At least 10 GB per each different OS that is backed up. That is, if you plan to backup VMs
running Windows 8, Windows 2008 R2, Windows 2012 and RHEL 6 (four different operating
systems), you will need at least 10 GB * 4 = 40 GB
Plan for additional 20 GB of working space for cache population, payload and other
temporary files.
Formulas:
Formula for configured cache size (insert this number in configuration wizard):
(Number of operating systems * 10 GB) + 20 GB
Examples:
134
Sizing For WAN Acceleration
Digest space must be built into the equation using the same size for each source target:
For understanding how the disk space is consumed, please see the following sections.
VeeamWAN\GlobalCache\trg
For each pair there will be a subfolder in the trg directory, with a UUID describing which
source WAN accelerator the cache is attached to. In each of those subfolders, the blob.bin
file containing the cache will be located. That file size corresponds to the setting configured
in the management console.
Note: The blob.bin file will exist for all connected source WAN accelerators.
VeeamWAN\GlobalCache\temp
When connecting a new source WAN accelerator, the temp folder will temporarily contain
the data.veeamdrf file that is later transferred to the source containing the cache manifest.
135
Sizing Targets for WAN Accereration Relationship
we can have two types of relationship with our Source and Target accelerators, One to One
and many to One.
One to one is the most simplest form, this is where one Source Accelerator is mapped to
a single Target Accelerator at the other location.
The other type is Many to One where many source accelerators will map to a single target
accelerator in a fan in type design. this is a common configuration and best practice is to
have no more than 4 source accelerators to a single target for resource reasons.
If we assume that we have 3 VMs, each with unique OSes (for instance, Win 2008R2, Win
2012R2, Solaris 10) each OS requires 10GB to be allocated for it.
The Cache itself is wholly independent from the digests required. That is, the Veeam GUI
does not make any determination of how much you can allocate for a digest and so on.
The digest is essentially an index of what cached blocks go where. For digest size, 1TB of
VM disk capacity we are backing up should correspond with 20GB of disk space. That is, for
10VMs we are backing up whose capacity is 2TB, you must account/allocate 40GB for
digest data on the Source WAN Accelerator. This limitation is not applied to the Target WAN
Accelerator.
136
Sizing Targets for WAN Accereration Relationship
For a Many-to-1 setup, the global cache is calculated per 1 Source WAN Accelerator
working with the Target WAN Accelerator.
If we use the same VMs in the previous example, the cache is only required to be 30GB.
However, since we’re using 3 Source WAN Accelerators, the cache size must be 90GB in
response.
On the Target WAN Accelerator, cache size is dictated by the amount of Source WAN
Accelerators plus number of operating systems in use, the digests space on the target end
in this example cannot be excluded from the calculation even though it may never be used.
We may require 120GB of Digest space at source so this needs to be added to the cache
size (90GB) at target resulting in a requirement of 210GB of capacity at a minimum on the
target.
When creating a WAN Accelerator in the user interface it relates to cache sizing only,
digest sizing should be part of the overall design and included as part of the
specification of the WAN Acceleration host.
Note: The target accelerator will use the digest capacity in the event the source digest
becomes unavailable, is rebuilt or becomes corrupt. The target will use the spare capacity to
calculate the digests on the target.
137
Deployments For WAN Acceleration
As the target WAN accelerator can handle multiple incoming streams (as described in the
Many-to-One WAN Acceleration section of the User Guide), it is recommended to maintain a
4:1 ratio between the number of source WAN accelerators per target WAN accelerator.
This guideline is very much dependent on the WAN link speed. Many source sites with low
bandwidth will create little pressure on the target WAN accelerator. So, for instance, in
multiple ROBO configurations a 10:1 ratio can be considered.
If there are sites with very high bandwidth (such as datacenter-to-datacenter replication),
they will produce a much more significant load on both the target WAN accelerator and the
target repository due to the second data block lookup (for more information, refer to the User
Guide).
Note: The secondary data block lookup is used, when a data block is not available in the
WAN accelerator cache. When there is a WAN cache “miss”, the secondary lookup for the
same data block is performed on the target repository. If it is found here, it is read back to
the WAN accelerator instead of re-transmitting over WAN.
Assuming the source and target repositories can deliver the throughput required for the
optimal processing rate, use the guidelines that follow.
Note: The numbers below are processing rates. The WAN link usage is dependent on the
achieved data reduction ratio.
Average throughput per target WAN accelerator: 500 Mbit/s (62.5 MB/s)
Depending on the achieved data reduction rate (typically 10x), the transfer rate over the
WAN link will vary.
If the processing rate is 62.5 MB/s, and the data reduction rate is 10x, then it is
possible to sustain 6.25 MB/s (50 Mbit/s) over the WAN link.
If the WAN link has high bandwidth (above 100Mbps) consider using backup copy
jobs without WAN Acceleration. However, if you use WAN accelerators in that
scenario, it may require deployment of multiple WAN accelerator to fully saturate
the WAN link.
138
Deployments For WAN Acceleration
1. A pair of WAN accelerators means any source WAN accelerator paired with the
sizing. ↩
139
Is WAN Acceleration Right For me
There are a number of ways to determine this based around speed and your available
resources.
Wan Acceleration can be one to one or one to many connections, the first thing you should
consider is the bandwidth available between the locations to see if the cost of optimizing
your traffic is outweighed by the speed of your link.
These numbers are to be considered as a base line , “Your mileage may vary”. The
performance of the underlying storage where the Global Dedupe Cache is located can
greatly impact the performance of the WAN Accelerator function.
Tests show that there are no significant performance differences in using spinning disk
drives as storage for the target WAN accelerator cache rather than flash storage. However,
when multiple source WAN accelerators are connected to a single target WAN accelerator
(many-to-one deployment), it is recommended to use SSD or equivalent storage for the
target cache, as the I/O is now the sum of all the difference sources.
One more point to focus on is the repository used at the target-wan-accelerator, data may be
taken from the repository at the target WAN accelerator if the data is not found in the global
cache but is known to exist in a previous transfer. If slow disks are used it can have an
impact on the speed of the completion of the job and overall processing rate.
140
Is WAN Acceleration Right For me
Other factors are also present such as is this link going to have bi-directional data flow when
using the Wan Accelerators, how many jobs will be using the link at the same time. Measure
your overall saturation of the link prior to using Wan Acceleration to ensure that it meets your
needs.
141
Tape Support
Tape Support
Overview
The diagram below illustrates the main components and processes within the backup
infrastructure when tape support is implemented in Veeam Backup & Replication:
142
Tape Support
Supported
LTO-3 or higher
For VTLs, see the corresponding section under Deduplication Storage
Not supported
IBM "Jaguar" TS11x0 Enterprise tape drives
StorageTek T10000 tape drives
Older Tape drives like DLT or AIT
Drivers
HP drivers: these are not installable with the downloaded install .exe file on a VM (for
143
Tape Support
example, to use with VTL). As a solution, run the install .exe and choose Extract. Use
Device Manager –> Update driver and select the drivers for tape drives and (if you use
HP/HP emulation tape library) for media changer.
Veeam supports medium changers that have no Microsoft Windows drivers available. Make
sure that such device is recognized as an unknown medium changer in the Microsoft Device
Manager list.
It is recommended that you use tape devices with original equipment manufacturer (OEM)
drivers. Limitations VMware does not support tape drives connected directly to ESX(i) 4.x
and later. For more information, see VMware vSphere Release Notes.
For more details and recommendations on configuring vendor-supported tape drives and
media changers on ESX/ESXi, refer to VMware documentation at
http://kb.vmware.com/kb/1016407.
Note: Veeam Backup & Replication uses the MTF (Microsoft Tape Format) industry
format to write data to tape. Veeam Backup & Replication does not support using
WORM (Write Once Read Many) tapes.
144
Tape Support Deployments
Data Movers with Veeam Backup & Replication, the data transfer during archiving and
restore processes is enabled with Veeam Data Mover services. The Data Movers run on
tape servers and other components of backup infrastructure. They receive tasks from the
Veeam backup server and communicate to each other to transfer the data. The Data Movers
are light-weight services that take a few seconds to deploy. Deployment is fully automated:
when you assign a tape server role to a server, Veeam Backup & Replication installs the
necessary components on this server and starts the required services on it.
145
Tape Support Deployments
Replication supports only values divisible by 1024. You can check the resulting range of
block sizes supported by Veeam Backup & Replication for a particular drive in the Drives
properties. For details, see Working with Drives.
Note: If you connect the tape devices via HBA, Veeam Backup & Replication uses the block
size configured for the HBA.
The block size is unified for: All drives in one library (if the drives support different block
sizes)
The tapes must be written with block size that match used for writing such tapes must match
the value currently used for the tape device you are using for restore.
If you have a number of Veeam backup servers, you can easily reconnect a tape server to
another Veeam backup server without reconfiguring the tape device: Veeam backup server
will recognize the library settings automatically. Note that when you reconnect the tape
server, the tape jobs will not run with another Veeam backup server unless you copy the
configuration
146
Tape Support Media Information
Media Management
147
Tape Support Media Information
If you enable the automated drive cleaning option in Veeam Backup & Replication, make
sure that you disabled the drive cleaning tasks on your tape library device.
Veeam Backup & Replication cleans the drives at the beginning of backup to tape jobs or file
to tape job run. The cleaning is not performed during other tape operations such as, for
example, cataloging or export. To clean the drives automatically, Veeam Backup &
Replication performs the following actions:
1. The tape library alerts Veeam Backup & Replication on a drive that requires cleaning.
2. Veeam Backup & Replication waits for a tape job to start.
3. When the tape job locks necessary drives for writing data, Veeam Backup & Replication
checks which of them requires cleaning.
4. Veeam Backup & Replication ejects the tape from the drive, inserts a cleaning tape and
performs the cleaning.
5. Veeam Backup & Replication ejects the cleaning tape and inserts the tape that was
reserved for the tape job.
6. The tape job writes the data on tape.
The cleaning tapes are located in the Unrecognized media pool. The worn-out cleaning
tapes are moved to the Retired media pool automatically.
148
Tape Support Media Information
If a tape job locks multiple drives simultaneously for parallel processing, and one or more
drives require cleaning, all drives wait until the cleaning is finished. After cleaning, all drives
start writing simultaneously.
The automated drive cleaning does not affect creation of media sets.
You cannot enable the automated drive cleaning on standalone tape drives. You cannot start
the drive cleaning manually with Veeam Backup & Replication. The drive cleaning is fully
automated.
For more information about automated drives cleaning, see Automated Drive Cleaning.
Select the Use native SCSI commands instead of Windows driver check box if your library is
an unknown media changer.
Media Information
Veeam Backup Database Veeam Backup & Replication catalogues information about all
archived data and stores this information in the Veeam backup database. The registered
tapes stay in the database until you remove the information about them. You can always
view details for each tape, for example, information about backups written to it, even if the
tape is not inserted in the library. The catalogue lets quickly detect location of the required
items on tape. The catalogue correlates the archived files and the restore points to the
names of the corresponding tapes, both online or offline and the names of the media sets
within which the data was written.
149
Tape Support Media Information
When you start restore, Veeam Backup & Replication prompts for the tapes you need to
bring online. As a result, you can restore data from tape much quicker when necessary.
Veeam Backup & Replication uses the following catalogues for storing the tape-related data:
Tape Catalogue stores information about files/folders archived to tape media with file to
tape jobs, as well as backup files produced by backup to tape jobs. The content of the
Tape catalogue can be examined in the Files view.
Backup catalogue stores information about VMs whose backups are archived to tape
media with backup to tape jobs. The content of the Backup catalogue can be examined
under the Backups > Tape node in the Backup & Replication view
Media Pool
A media pool simply defines a group of tapes managed by Veeam Veeam Backup &
Replication. There are three types of media pools:
Service media pools. Created and managed automatically. It is not possible to modify
their settings. They contains:
Empty media starts out in the Free pool indicating it’s available for use in other
pools.
After inventory or cataloging, media with existing data is placed into the Imported
pool. Review the contents and place such media into the Free pool for overwrite or
leave in Imported pool to keep the data.
Exhausted or broken tapes are placed into the Retired pool and are not used
further.
Media pools are groups of media to which backup data can be written.
Configure each pool settings according to the purpose of the pool, such as the
overwrite protection period that is applied to all media within the pool.
Since v9 a (Custom) Tape Pool can be spanned over multiple tape libraries. The
idea is to use the capacity and drives of multiple tape systems together and to
failover to another tape library in case one library goes offline.
150
Tape Support Media Information
GFS media pools are used to store weekly, monthly, quarterly and yearly backups on
tape.
Configure each pool settings according to the purpose of the pool, such as the
overwrite protection period that is applied to all media within the pool.
Media Set
A media set is a subset of a media pool that contains at least one backup. A new media set
can be created for every backup, or on a time based schedule (i.e. weekly). It is also
possible to reuse the same media set forever. When a media set contains at least one full
backup, it is a self-sufficient restore point. It means that if you have all tapes from the media
set at hand, you can be sure that restore will be successful.
Media Vault
A media vault is used to organize offline media. For example, you have a service
organization that transports the tapes to a safe at a bunker. You can name the vault
accordingly and add some useful information in the description (phone number, place, etc.).
When you need to transport physical tapes to the safe, add these tapes to the vault
manually or set automatic export of offline tapes to a vault in the tape jobs or media pools
properties.
Backup Modes
Backup jobs can create different backup types of backup file chains on disk depending on
the backup mode used. Depending on backup mode, "Backup to Tape" jobs either copies
files to tape or synthesize a full backup. The following rules apply:
When archiving reverse incremental backups, the behavior varies on the type of media
pool used:
Standard Media Pool: The tape job will always copy the full backup and ignore
any rollback files (VRB)
GFS Media Pool: The tape job will create a full backup from VRB files on specified
day(s) as per schedule.
151
Tape Support Media Information
When archiving forward incremental backups, with active or synthetic full scheduled, the
backup chain on tape will be a copy of the backup chain on disk. The virtual full option
in tape job configuration is ignored.
If you archive forward incremental backups without synthetic or active full enabled, or
archive Backup Copy Jobs, the full files are synthesized from existing restore points on
disk. The virtual full backup schedule can be configured on the "Backup to Tape" job.
For more information about virtual full to tape, please see [Veeam Help Center]
(https://helpcenter.veeam.com/docs/backup/vsphere/virtual_full_backup.html?ver=95
If the source backup job contains multiple chains, and the checkbox "Process latest full
backup chain only" in advanced job settings is unchecked, you will be prompted for a
decision, when creating a Backup to Tape job. You may choose to either only the last
backup chain or all existing restore points.
Sizing
For the highest throughput, enabling parallel processing for the Backup to Tape is
recommended. You need to size the servers and storage connection accordingly. It can be
helpful to create multiple partitions with 2-4 tape drives and add these partitions to different
tape servers. Adding these libraries to the media pool and enabling parallel processing will
distribute the load across multiple drives and tape servers.
Note: Parallel processing for tape is unavailable for GFS media pools.
Install Windows 2012 R2 or above on the tape server for best performance. Use the latest
Veeam version and patch level as they often contain tape throughput optimizations.
Perform a POC to test throughput of tape and disk. If you have no opportunity to test speed,
assume that the lowest speed for backup to tape jobs with LTO5/6 is 50MB/s as a
conservative estimate. We highly recommend to do a POC to evaluate real throughput to
avoid additional hardware costs.
The GFS (Grandfather, Father, Son) tape job can help avoid a complex Backup to Tape job
creation by handling weekly, monthly, quarterly and yearly backups in a single job.
For Backup to Tape jobs that use forward incremental (without synthetic or active fulls) jobs
or Backup Copy Jobs as source of the data, it may be required to temporarily disable the job
using pre- and post scripts, as the transform process of forever incremental forever will
terminate the tape job. Another option is to increase the restore points of these jobs
temporarily. By increasing the number of restore points for the source job, the Backups to
Tape job will not be terminated by the merge process. However, please note this will
152
Tape Support Media Information
increase the transform time significantly once the setting is reverted and is highly
discouraged for large jobs. An example of this implementation can be found here: v9 GFS
job - No more continuous?
However, in some situations you may want to use an existing library with non-LTO tapes, or
you need to integrate Veeam Backup & Replication into an existing backup-to-tape software.
Veeam backup files contain all information needed for restore (e.g. deduplication
information, VM metadata, etc.), and you can use the existing backup-to-tape solution to
bring the Veeam backup files on tape. This approach can also support enterprise customer
"Segregation of duty" demands as two complete different teams can handle backups and
tape backups. No single person can delete by mistake or on purpose the primary and tape
chain. Before having two backup solutions co-exist on the same server, please verify they do
not conflict each other.
Tape Encryption
Veeam uses hardware encryption if it is provided by the tape device and enabled in Veeam
Backup & Replication. Tape library should work in the application-managed encryption
mode.
If the hardware based encryption is not supported by the tape device, software based AES-
256 encryption is used. Please note software based encryption may cause significant
performance degradation, if not natively accelerated by the CPU of the tape server.
Hardware based encryption is typically available for LTO-4 or newer libraries, and while a
license is often required, this is usually supported for free by the tape library vendor.
When archiving data, Veeam generates a user key which is stored with data on tape. If you
restore data using another Veeam backup server, provide the password or utilize the
Password Loss Protection in Enterprise Manager. See the User Guide for more information.
If the hardware encryption option is used, and you archive to tape Veeam backups that are
already encrypted on disk, they will be encrypted twice. If you restore such backups with
double encryption on the same Veeam backup server they will be decrypted automatically.
153
Tape Support Media Information
To decrypt on another Veeam backup server, you will need to enter the two passwords
accordingly.
For additional details on tape encryption, see the corresponding section of this guide >
Encryption
Tips
"Short Erase" all tapes before use with Veeam to avoid any problems cause by data
from other backup software
Install latest Windows Updates
Install latest firmware on library, drives, HBA (verify interoperability)
Install separate HBAs for tape is recommended, but not required
A staging area for backup files is required when restoring from tape. Keep this in mind
when sizing backup repositories.
Tape compression should be disabled for tape jobs, when backup files are already
compressed at the backup repository
"File to Tape" engine is optimized for smaller amount of big files (e.g. backup files) only
154
Tape Support Config Requirements
You must have Veeam Backup & Replication Enterprise license or higher is installed on
the Veeam backup server.
You must pre-configure backup job(s) that produce the backup for archiving.
The primary backup job must have at least 2 restore points on disk.
The primary backup copy job must have at least 4 restore points on disk.
You must configure one or more simple media pool with the necessary media set and
retention settings.
You must load tapes to the tape device and configure the target media pool so that it
has access to them. If the media pool has no available tape, the tape job will wait for 72
hours and then terminate.
The backup to tape job processes only VBK (full backups) and VIB files (forward
incremental backups).
If you back up to tape a reverse incremental chain, the tape job will always copy the full
backup.
Microsoft SQL Server log files (VLB) are skipped from processing.
155
Tape Support Parallel Processing
Note: You cannot enable parallel processing for GFS media pools.
To process the tape data in parallel, you can split the data across drives in 2 ways:
The media pool can use the predefined maximum number of drives and process the equal
number of tape jobs simultaneously.
For example, if you set 3 drives as the maximum, you can process up to 3 tape jobs at the
same time. If you have more jobs running at the same time, they are queued. When one of
the jobs finishes and releases its drive, the first queued job takes the drive.
This option is available for backup to tape and file to tape jobs. For example:
When you select processing backup chains in parallel, the media pool processes several
primary jobs simultaneously. If the primary jobs produce per-VM backups, the media pool
processes several per-VM backup chains simultaneously. This option is available for backup
to tape jobs only. For example:
156
Tape Support Parallel Processing
across libraries.
157
Tape Support Virtual Full
Unlike disk backups, tape archives are static: tape jobs cannot rebuild backups once they
are written to tape. Also, the standard backup to tape scheme (archiving new restore points
during each tape session) cannot be used: the tape archive would have one full backup and
an endless chain of increments all of which would be required for restore.
To adapt the forever forward incremental chains to tapes, Veeam Backup & Replication uses
the virtual full. The virtual full mechanism creates a periodic synthesized full backup on tape.
The periodic fulls split the forever incremental backup chain into shorter series of files that
can be effectively stored to tapes. Each series contains one synthesized full backup and a
set of increments. Such series are convenient for restore: you will need to load to the tape
device only those tapes that are part of one series.
The virtual full does not require additional repository disk space: it is synthesized directly on
tape on the fly, when the tape job runs. To build such full backup, Veeam Backup &
Replication uses backup files that are already stored on the backup repository. If the primary
job produces a forever incremental backup chain or is a backup copy job, Veeam Backup &
Replication will periodically create a virtual full backup. You can configure the full backup
with the scheduler .
The virtual full cannot be switched off; however, it is disabled automatically if the primary job
periodically creates active full or synthetic full backups. The virtual full does not depend on
the job settings for incremental backups. If you enable the virtual full for the job, it will be
created in any case, no matter whether you enable or do not enable incremental backups.
Sometimes, the primary job may start when the tape job is still running. By default, the
primary job has priority. In this case, the tape job terminates with error and no data is written
to tape. Select the Prevent this job from being interrupted by primary backup jobs option if
you want to give the tape job a higher priority. If this option is selected, the primary job will
wait until the tape job finishes. Note that the primary job may start with a significant delay.
158
Tape Support Virtual Full
159
Tape Support Writing to Tape
Note: If the file to tape job fails to complete in 3 weeks, it is terminated by timeout.
VM Backup to Tape
To back up data to tape, you need to create and run tape jobs dedicated to archive Veeam
backups that were produced by Veeam backup jobs to tapes. When a backup to tape job
runs, it does not create new backups: it locates already existing backups and copies them
from backup repository to tape. You need to set the source of the tape job: jobs and/or
backup repositories. Jobs as Source The following jobs can be primary for tape jobs:
When the tape job starts on it's schedule, it picks the restore points that were produced by
the primary jobs in period since the last tape job run. If you change the configuration of the
primary jobs, the tape job is updated automatically: it adds new VMs to the list of VMs to
archive or stops archiving VMs that were removed from primary jobs. The primary jobs may
use any backup method:
160
Tape Support Writing to Tape
Forward incremental backup method: When the tape job backs up the forward
incremental chain to tape, it creates a copy of the disk backup chain. The source
backup chain must contain 2 or more restore points.
Reverse incremental backup method: The last backup in the reverse incremental
backup chain is always the full backup. If the source backup chain is reverse
incremental, the tape job will copy the full backup each time the tape job runs. The
increments are skipped. The source backup chain may contain any number of restore
points.
When you add a repository as source to tape job, the tape job constantly scans the selected
repository (or repositories) and writes the newly created backups to tape. The tape job
monitors the selected repository in a background mode. You can set explicit backup
windows for the tape job. In this case, the tape job will start on the set time and archive all
new restore points that were created in period since the last job run. If you create or remove
backup jobs that use this repository, or if you change the configuration of such backup jobs,
you do not need to reconfigure the tape job that archives the repository. Mixed Jobs To one
tape job, you can link an unlimited number of sources. You can mix primary jobs of different
type: backup and backup copy, and of different platform (VMware, Hyper-V, Windows Agent
or Linux Agent). You can add jobs and repositories as source to the same tape job.
Important! The tape job looks only for the Veeam backups that are produced by backup jobs
running on your console. Other files will be skipped. Note that to back up files, you need to
configure file to tape job.
You can add primary jobs to tape jobs at any moment: when you create a tape job, or later.
Adding primary jobs is not obligatory when you create a tape job: you can create an "empty"
job and use it as a secondary destination target. When you link jobs, the tape job processes
161
Tape Support Writing to Tape
them in the same way as the jobs added with the Tape Job Wizard. For more information,
see Linking Backup Jobs to Backup to Tape Jobs.
162
Tape Support Restores
Restores
VM Restore from Tape to Infrastructure
Restoring a VM from tape with Veeam Backup & Replication is a lot like restoring a VM from
disk. For example, you can choose a desired restore point, select the target location or
change the configuration of the restored VM. To restore a VM from tape, you can choose
between the following options:
To choose the needed option, select Restore directly to the infrastructure or Restore through
the staging repository at the Backup Repository step of the Full VM Restore wizard.
When you restore VMs from tape directly to the infrastructure, the restore process publishes
the VMs to the virtual infrastructure copying the VM data directly from tape. This option is
recommended if you want to restore one VM or a small number of VMs from a large backup
that contains a lot of VMs. In this case, you do not need to provide a staging repository for a
large amount of data most of which is not needed to you at the moment. This option is slow
if you restore many VMs. The VMs are restored one by one: this requires a lot of rewinding
of tape as tapes do not provide random access to data.
When you restore VMs from tape through a staging repository, the restore process
temporarily copies the whole restore point to a backup repository or a folder on disk. After
that Veeam starts a regular VM restore. This option is recommended if you want to restore a
lot of VMs from a backup as the disk provides a much faster access to random data blocks
than tape.
This option allows you to copy VM backups from tape to repository. This is helpful if you
need some backups on disk for later use, or also for VM guest OS files restore. You can
restore full backups or incremental backups to a repository or any location of your choice.
The restored backup is registered in the Veeam Backup & Replication console as an
imported disk backup so that you can use it for any restore from disk scenario later on. For
one restore session at a time, you can choose one restore point available on tape.
163
Tape Support Restores
You can restore files and folders that were previously archived with file to tape jobs.
Restoring capabilities allows you to restore files to their original location or another server,
preserving ownership and access permissions. The file restore process allows you to restore
files to any restore point available on tape.
164
Veeam Explorers
Veeam Explorers
Veeam Explorers are tools included in all editions for item-level recovery from several
application. As of v9, following Explorers are available:
Each Explorer has a corresponding user guide available in Helpcenter: Veeam Backup
Explorers User Guide. For specifics of performing granular restore of certain applications
refer to the Applications section of this guide.
VESS is a very easy way to perform item-level recovery directly from storage snapshots.
Veeam is able to use discover and mount any storage snapshot for restores. By combining
the Veeam application consistent with crash consistent snapshots, the RPO for certain
applications can be significantly reduced.
165
Veeam Explorers
2. In case of Block access (iSCSI, FC, FCoE) mount the new LUN to a proxy ESXi and
register a temporary datastore, in case of NFS access the existing NFS datastore and
look for the cloned VM
After restoring and exiting VESS, the temporary datastore, VM and LUN clones will be rolled
back and cleaned up.
166
Interaction with vSphere
While it is possible to connect a Veeam Backup & Replication server directly to ESX(i) hosts,
this section assumes a vSphere environment with at least one vCenter Server, and that the
backup server is integrated at the vCenter Server level, as this is the best practice
configuration in almost all use cases.
vCenter Server
One of the most critical components of any vSphere environment is the vCenter Server. This
server provides a single view of the entire virtual environment, and a central point of
management. Veeam Backup & Replication communicates with the vCenter Server in many
operations. For this reason, fast and stable communication between Veeam Backup &
Replication and the vCenter Server is critical to achieving a stable backup environment.
Problems with connectivity to the vCenter Server is one of the top reasons for failed
Veeam jobs. Having a well-performing vCenter Server with reliable connectivity will
mitigate this issue and provide a strong backbone for a reliable backup infrastructure.
The vCenter Server must be reliable and always available when backup jobs are
running. It must be able to answer queries and perform actions in a reasonable amount
of time. If the vCenter Server performs poorly during normal operations, this should be
corrected prior to implementing Veeam Backup & Replication.
For larger environments, with many concurrent jobs, especially jobs that run at short
intervals, the load on the vCenter Server can be significant. The vCenter Server must
be able to handle increased transactional workload to prevent random job failures due
to command timeouts.
The backup server must have reliable network connectivity to the vCenter Server. It is
generally suggested that the backup server is placed in close logical proximity to the
vCenter Server, but this is not always the best deployment option. In cases where the
167
Interaction with vSphere
backup server and vCenter Server must be deployed across a distance, the only real
requirement is that this connection is consistent and reliable.
When maintenance is being performed on the vCenter Server, best practice would
dictate that all Veeam Backup & Replication jobs must be idle, and the Veeam Backup
Service should be stopped. This includes applying Windows updates (if using the
vCenter Server installable version), vCenter Server patches and upgrades, or any
maintenance that would require the vCenter service to be restarted or the system
rebooted.
Since the original VMDK file is only being used for reads, it provides a consistent view of the
blocks that made up the VM at the time the snapshot was taken. This allows Veeam Backup
& Replication to read this base disk as a consistent image for backup and replication
functions. When the snapshot is removed, the blocks that were written to the delta file are
read and written back into the original VMDK, and finally the delta file is discarded.
As Veeam Backup & Replication leverages the snapshot technology for performing backups,
you should ensure it is possible to snapshot the virtual machine disks, since there are
certain configurations that do not support snapshots. To identify VMs that do not support
snapshots, see VMware KB article 1025279 ; you can also use Veeam ONE assessment
reports to automatically detect them before starting Veeam Availability project.
As with many things in technology, although the concept is simple, the actual implementation
is a little more complex. The following section is a quick look at the impact of various
operations on the VM and underlying infrastructure.
168
Interaction with vSphere
Snapshot Creation
The actual operation of creating a snapshot generally has only a minor impact: the snapshot
file has to be created, and there is a very short “stun” of the VM. This “stun” is generally
short enough (typically, less than 1 sec), so it is rarely an issue except for the most time-
sensitive applications.
Note: Veeam Backup & Replication leverages a standard VM snapshot for the backup
process. These VMware snapshots have a single file size limitations. Keep in mind, that the
maximum file size include all snapshot files and the data disk in total. For example if you
have an old VMFS version 3 the maximum file size (including snapshots) is 2TB and so your
data disk should not be sized over 1.98TB to still be able to create snapshots. For details,
see VMware KB article 1012384.
The default number of concurrently open snapshots per datastore in Veeam Backup &
Replication is 4. This behavior can be changed by creating the following registry key:
Snapshot Open
Simply having a snapshot open for a running VM involves some performance penalty on the
VM, the ESX(i) host and the underlying storage. The host has to track the I/O, split writes to
the snapshot file and update the snapshot file metadata. This overhead, in turn, impacts the
guest (primarily, with slower I/O).
This is generally most notable for VMs with significant write load, and has less impact on
read performance.
From the storage perspective, VMs running with an open snapshot require additional space
to store the snapshot data, and additional I/O load on the datastore. This is generally more
noted on systems with significant write I/O load.
Snapshot Removal
169
Interaction with vSphere
Snapshot removal is the step with the highest impact from the performance perspective. I/O
load increases significantly, due to the extra R/W operations required to commit the
snapshot blocks back into the original VMDK. This eventually leads to the VM “stun”
required to commit the final bits of the snapshot. The “stun” is typically a short pause usually
only a few seconds or less, when the VM is unresponsive ("lost ping"), while the very last
bits of the snapshot file are committed.
VMware vSphere uses the "rolling snapshot" for older versions and the same method as
storage vMotion uses starting from vSphere 6.0u1 to minimize the impact and duration of the
stun, as described below:
For vSphere 6u1 and newer: The host leverages the Storage vMotion Mirror driver to copy
all needed data to the original data disks. When completed, a "Fast Suspend" and "Fast
Resume" is performed (comparable with vMotion) to bring the original data files online.
This “stun” period can be less than one second for small VMs with light load, or several
seconds for larger VMs with significant load. To external clients, this small stun looks like the
server is busy and thus might delay a response for a few seconds. However, applications
that are very sensitive to delays may experience issues with this short period of
unresponsiveness.
How to Mitigate?
To mitigate the impact of snapshots, consider the following recommendations:
Upgrade to vSphere 6u1 or newer to use the new Storage vMotion based Snapshot
commit processing.
Minimize the number of open snapshots per datastore. Multiple open snapshots on
the same datastore are sometimes unavoidable, but the cumulative effect can be bad.
Keep this in mind when designing datastores, deploying VMs and creating backup and
170
Interaction with vSphere
Consider snapshot impact during job scheduling. When possible, schedule backups
and replication job during periods of low activity. Leveraging the Backup Window
functionality can keep long-running jobs from running during production. See the
corresponding setting on the Schedule tab of the job wizard
Use the vStorage APIs for Array Integration (VAAI) where available. VAAI can offer
significant benefits:
Hardware Lock Assist improves the granularity of locking required during snapshot
growth operations, as well as other metadata operations, thus lowering the overall
SAN overhead when snapshots are open.
VAAI in vSphere 5.x offers native snapshot offload support and should provide
significant benefits once vendors release full support.
VAAI is sometimes also available as an ESXi plugin from the NFS storage vendor.
Design datastores with enough IOPS to support snapshots. Snapshots create
additional I/O load and thus require enough I/O headroom to support the added load.
This is especially important for VMs with moderate to heavy transactional workloads.
Creating snapshots in VMware vSphere will cause the snapshot files to be placed on
the same VMFS volumes as the individual VM disks. This means that a large VM, with
multiple VMDKs on multiple datastores, will spread the snapshot I/O load across those
datastores. However, it actually limits the ability to design and size a dedicated
datastore for snapshots, so this has to be factored in the overall design.
Note: This is the default behavior that can be changed, as explained in the VMware
Knowledge Base: http://kb.vmware.com/selfservice/microsites/search.do?
language=en_US&cmd=displayKC&externalId=1002929
Allocate enough space for snapshots. VMware vSphere 5.x puts the snapshot VMDK
on the same datastore with the parent VMDK. If a VM has virtual disks on multiple
datastores, each datastore must have enough space to hold the snapshots for their
volume. Take into consideration the possibility of running multiple snapshots on a single
datastore. According to the best practices, it is strongly recommended to have 10% free
space within a datastore for a general use VM, and at least 20% free space within a
datastore for a VM with high change rate (SQL server, Exchange server, and others).
Note: This is the default behavior that can be changed, as explained in the VMware
Knowledge Base: http://kb.vmware.com/selfservice/microsites/search.do?
language=en_US&cmd=displayKC&externalId=1002929
Watch for low disk space warnings. Veeam Backup & Replication warns you when
there is not enough space for snapshots. The default threshold value for production
datastores is 10 GB. Keep in mind that you must increase this value significantly if using
171
Interaction with vSphere
very large datastores (up to 62 TB). You can increase the warning threshold in the
backup server options, of the Veeam Backup & Replication UI. You can also create a
registry key to prevent Veeam Backup & Replication from taking additional snapshots if
the threshold is breached:
Enable parallel processing. Parallel processing tries to backup multiple VM disks that
belong to a single VM at the same time. This reduces snapshot lifetime to the minimum.
This option is enabled by default. Please note if you upgraded from v6.5 or earlier
versions, you have to enable this option explicitly in the backup server options.
Datastores formatted with the VMFS file system have native capabilities to determine which
cluster node is the owner of a particular VM, while VMs running on NFS datastores rely on
the LCK file that resides within the VM folder.
During hot-add operations, the host on which the hot-add proxy resides will temporarily take
ownership of the VM by changing the contents of the LCK file. This may cause significant
additional "stuns" to the VM. Under certain circumstances, the VM may even end up being
unresponsive. The issue is recognized by VMware and documented in
http://kb.vmware.com/kb/2010953.
Note: This issue does not affect Veeam Direct NFS as part of Veeam Direct Storage Access
processing modes and Veeam Backup from Storage Snapshots on NetApp NFS datastores.
We highly recommend you to use one of these 2 backup modes to avoid problems.
172
Interaction with vSphere
0 (default) – Disabled
1 – “Preferred same Host” If DirectNFS proxy on same host exist it will wait for a free
task slot there. If a proxy on same host do not exist it will use another DirectNFS proxy
(on another host or physical server) or fallback to Virtual Appliance(HotAdd) and then to
Network(NBD) mode.
2 – If there is no DirectNFS proxy on same host as VM, it will fallback to Network mode
(NBD)
Overall HCI solutions should use 1 or 2 (recommended) “1” should be used with HCI
solutions only if the “EnableSameHostHotAddMode” was set to “2”.
This reg key is not used for the Veeam Cisco HyperFlex integration.
If for what ever reason Direct NFS processing can not be used and HotAdd is configured,
ensure that proxies running in the Virtual Appliance mode (Hot-Add) are on the same host
as the protected VMs.
To give preference to a backup proxy located on the same host as the VMs, you can create
the following registry key:
Value = 1 – when proxy A is available on the same host, Veeam Backup & Replication
will leverage it. If proxy A is busy, Veeam Backup & Replication will wait for its
availability; if it becomes unavailable, another Hot-Add proxy (proxy B) will be used.
Value = 2 - when proxy A is available on the same host, Veeam Backup & Replication
will leverage it. If proxy A is busy, Veeam Backup & Replication will wait for its
availability; if it becomes unavailable, Veeam Backup & Replication will switch to NBD
mode.
173
Interaction with vSphere
This solution will typically result in deploying a significant number of proxy servers, and may
not be preferred in some environments. For such environments, it is recommended
switching to Network mode (NBD) if Direct NFS backup mode can not be used.
Snapshot Hunter
At Veeam Support, one of the most commonly raised support cases was for orphaned
snapshots. Orphaned snapshots were caused by VMware’s own failed snapshot commit
operations due to unreleased VMDK file locks during VDDK operations. Veeam uses the
VMware standard VM snapshot processing for backup and replication processes, so
although Veeam was not the origin of the orphaned snapshots, as Veeam uses VMware
snapshots, Veeam is often seen as a root cause as this issue was only discovered when a
backup failed.
If not monitored appropriately, VMware orphaned snapshots can cause many unexpected
problems. The most common problems are overfilled VM datastores, or snapshots growing
so large they are impossible to commit. This is a well-known VMware vSphere issue
described in VMware KB article 1007814. The only way to manually remediate this issue is
cloning the VM and performing a new full VM backup.
Veeam Snapshot Hunter automatically detects any VM with the configuration issue “Virtual
machine disks consolidation needed”. Prior to performing backup of such VMs, Veeam
Backup & Replication will trigger disk consolidation (provided that the datastore performance
threshold specified in the Storage Latency Control settings is not exceeded).
Snapshot Hunter will attempt consolidation eight (8) times. If consolidation fails after all
retries, Veeam Backup & Replication will send an e-mail with a warning.
You can view information on the Snapshot Hunter sessions on the History > System view in
Veeam Backup & Replication console.
Note: Currently, the default behavior of Snapshot Hunter cannot be changed. As Snapshot
Hunter will automatically retry consolidation up to eight times, it may be inappropriate for
some VMs that require planned downtime to consolidate the snapshot manually. Such VMs
should be excluded from backup or replication jobs until the orphaned snapshots are
manually removed.
174
Interaction with vSphere
If you are evaluating Veeam Backup & Replication, use the Infrastructure Assessment
Reports included in Veeam Availability Suite to identify VMs with snapshots that can be
affected by automatic snapshot consolidation.
Modern CPUs have many physical cores and can run many tasks simultaneously. The
impact of having many proxy servers reading data blocks from the production storage at a
very high throughput may be negative. With this in mind, many businesses avoided running
backup or replication jobs during business hours to ensure good response time for their end
users. Storage Latency Control was implemented to help avoid this issue.
When Storage Latency Control is enabled, it monitors the storage read latency on the
production datastores using real-time metrics from the hypervisor. By default, metrics from
the hypervisor are collected every 20 seconds. These settings are inherited from vSphere.
The first Storage Latency Control threshold Stop assigning new tasks to datastore at puts
a limitation on assigning new tasks (one task equals one VM disk). If the latency for a
particular datastore is exceeded, no more proxy tasks will be assigned to it, until the latency
drops below the threshold.
If limiting the number of tasks assigned to the datastore is not sufficient, Storage Latency
Control will throttle the throughput for existing tasks according to the second threshold
Throttle I/O of existing tasks at.
175
Interaction with vSphere
The results of enabling Storage Latency Control are very easy to review using the vSphere
Client.
When to Use?
Storage Latency Control provides a smart way to extend backup windows or even eliminate
backup windows, and run data protection operations during production hours.
176
Interaction with vSphere
When Storage Latency Control is enabled, Veeam Backup & Replication measures the
storage latency before processing each VM disk (and also during processing, if Throttle I/O
of existing tasks at setting is enabled). Furthermore, if the storage latency for a given
datastore is already above the threshold, committing VM snapshots can be delayed. In some
environments, enabling Storage Latency Control will reduce the overall throughput, as
latency increases during the backup window.
However, in most environments having this feature enabled will provide better availability to
production workloads during backup and replication. Thus, if you observe performance
issues during backup and replication, it is recommended to enable Storage Latency Control.
Storage Latency Control is available in Enterprise and Enterprise Plus editions. The
Enterprise Plus customers are offered better granularity, as they can adjust latency
thresholds individually for each datastore. This can be really helpful in infrastructures where
some datastores contain VMs with latency-sensitive applications, while latency thresholds
for datastores containing non-critical systems can be increased to avoid throttling.
All backup vendors that use VMware VADP implement the VMware VDDK kit in their
solutions. This kit provides standard API calls for the backup vendor, and helps to read and
write data. During backup operations, all vendors have to deal with two types of connections:
the VDDK connections to vCenter Server and ESXi, and vendor’s own connections. The
number of VDDK connections may vary for different VDDK versions.
If you try to back up thousands of VMs in a very short time frame, you can run into the SOAP
session count limitation. For example, in vSphere 5.1 the default maximum number of
sessions is 500. If you hit this limitation, you can increase the vCenter Server SOAP
connection limit from 500 to 1000. For details, see http://kb.vmware.com/kb/2004663.
Veeam’s scheduling component does not keep track of the connection count. For this
reason, it is recommended to periodically check the number of vCenter Server connections
within the main backup window to see if you can possibly run into a bottleneck in future, and
increase the limit values on demand only.
177
Interaction with vSphere
You can also optimize the ESXi network (NBD) performance by increasing the NFC buffer
size from 16384 to 32768 MB (or conservatively higher) and reducing the cache flush
interval from 30s to 20s. For details how to do this, see VMware KB article 2052302. After
increaing NFC buffer setting, you can increase the following Veeam Registry setting to add
addition Veeam NBD connections:
Be careful with this setting. If the buffer vs. NFC Connection ratio is too aggressive, jobs
may fail.
This cache is stored into memory, so at each restart of the Veeam services its content is lost;
this is not a problem as the initial retrieval of data is done as soon as the Veeam server is
restarted. From here on, Veeam "subscribed" to a specific API available in vSphere, so that
it can receive in "push" mode any change to the environment, without the need anymore to
do a full search on the vCenter hierarchy during every operation.
178
Interaction with vSphere
The load against vCenter SOAP connection is heavily reduced, as we have now one
single connection per Veeam server instead of each job running a new queiry against
vCenter;
Every navigation operation of the vSphere hierarchy is instantaneous;
The initilisation of every job is almost immediate, as now the Infrastructure Cache
service creates a copy in memory of its cache dedicated to each job, instead of the
Veeam Manager service completing a full search against vCenter:
No special memory consideration needs to be done for the Infrastructure Cache, as its
requirements are really low: as an example, the cache for an environment with 12 hosts and
250 VMs is only 120MB, and this number does not grow linearly since most of the size is
fixed even for smaller environments.
179
Interaction with vSphere
Security
When connecting Veeam Backup & Replication to the vCenter Server infrastructure, you
must supply credentials that the backup server will use to communicate with the vCenter
Server.
The features that Veeam provides, such as backup, restore, replication, and SureBackup,
interact with vSphere at the fundamental level. Certain permissions are required to take
snapshots, create VMs, datastores, and resource groups. Because of this level of
interaction, it is generally recommended that Veeam Backup & Replication uses a restricted
account with the permissions that are required to complete the job.
You can also leverage security to restrict the part of the environment that the backup server
can “see”. This can have multiple benefits beyond security in that it lowers the time required
to parse the vCenter Server hierarchy and reduces the memory footprint required to cache
this information. However, care must be taken when attempting to use this level of
restriction, as some permissions must be provided at the very top of the vCenter Server tree.
Specifically if you access the vCenter over a WAN link such scoping can reduce the
(management background) WAN traffic.
For a detailed description of accounts, rights and permissions required for Veeam Backup &
Replication operations, see the "Required Permissions" document (not changed since V9.0).
180
Hyper-V Concerns
On-Host backup mode, for which backup data processing is on the Hyper-V node
hosting the VM, leveraging non transportable shadow copies by using software VSS
provider.
Off-Host backup mode, for which backup data processing is offloaded to another non
clustered participating Hyper-V node, leveraging transportable shadow copies using
Hardware VSS provider provided by the SAN storage vendor.
Performance wise, since both backup modes are using the exact same Veeam transport
services, the only differentiating factors will be the additional time requested to manage
transportable snapshots (in favor of On-Host mode) and the balance between compute and
backup resources consumption during backup windows (in favor of Off-Host mode).
PRO CON
Simplifies
management
Does not depend on
Requires additional resources from the
third party VSS
hypervisors during the backup window, for IO
On- provider
processing and optimization
Host Does not require
Does not depend on third party VSS provider
additional hardware
Does not require additional hardware
Can be used on any
Hyper-V
infrastructures
No impact on the
compute resources
Adds additional delay for snapshots
on the hosting hyper-
Off- transportation
v
Host Available only for virtualization
Requires third party
infrastructures based on SAN storage
VSS hardware
provider
181
Hyper-V Concerns
Spreading load across hypervisors. It should be kept in mind that the backup load,
instead of being carried by a limited number of dedicated proxies, will be spread through
all the hypervisors. Default Veeam setting is to limit backup to 4 parallel tasks per
hypervisor, which will use a maximum of four cores and 8 GB of RAM. This can be
modified in the “Managed server” section of the Veeam Console, through the “Task limit”
setting. For instance, if the sizing guidelines (please refer to the resource planning
section of this document) results in a total amount of 24 cores and 48 GB of RAM
needed for Veeam transport services, and the infrastructure comprises 12 Hyper-V
servers, each server task limit can be set to 2.
Leveraging storage latency control. This feature allows to protect the volumes
(globally for enterprise edition, and individually for enterprise plus edition) from high
latency, by monitoring and adjusting backup load accordingly. Please refer to user guide
proper section for further information.
182
Hyper-V Concerns
Before version 2016, Microsoft did not offer native change block tracking leading Veeam to
develop a private CBT engine to drastically optimize backups process. In version 2016 of
Hyper-V, Microsoft has implemented its own changed block tracking mechanism, named
“Resilient Change Tracking”. To benefit RCT, the following prerequisites must be met,
otherwise incremental backups or replication will have to read entirely the source virtual
disks to recalculate differences:
183
Hyper-V Concerns
A good rule of thumb to avoid such potential bottleneck is to create a number of volumes
equal or greater than the number of nodes composing the cluster, spreading IOs servicing
across all nodes.
Guest interaction
PowerShell Direct
Introduced by Microsoft in Hyper-V 2016, PowerShell Direct is a new method allowing to
interact with the guest even if no direct network connection is available between the Veeam
guest interaction proxy and the guest itself.
PowerShell Direct can be easily tested on the host, using the following command.
Please refer to the following Technet blogpost for further explanations on where to find and
how to install LIS.
Guest restoration
184
Hyper-V Concerns
Note : this pre-allocation is performed only for Instant VM Recovery Usage. Sure Backup
processing will use a thin provisioning mechanism instead, preserving resources on the
infrastructure.
185
Job Configuration
Job Configuration
In the following section, you will learn more about configuration guidelines for different job
types, and how to optimize both the user experience of using Backup & Replication, and the
backend operations to get the most of the available infrastructure.
186
Backup Methods
Backup Methods
Veeam Backup & Replication stores backups on disk using a simple, self-contained file
based approach. However, there are several methods to create and store those files on the
file system. This section will provide an overview of these methods, their pros and cons, as
well as recommendations on use cases for each one.
Backup mode directly influences disk I/O on both production storage and backup repository,
and backups size; for these reasons it is recommended to carefully review capabilities of the
destination storage when selecting one. Take a look at Deduplication Appliances section of
this guide for important details on using dedicated deduplicating hardware appliances for
storing backups.
For a graphical representation of the mentioned backup modes in this section, please see
Veeam KB1799.
As a generic overview for I/O impact of the backup modes, please see this table:
While changing backup mode is one way of reducing amount of I/O on backup repository it
is also possible to leverage features of the filesystem to avoid extra I/O. Currently Veeam
Backup and Replication supports advanced features of one filesystem, Microsoft ReFS 3.1
(available in Windows Server 2016), to completely eliminate unnecessary read/write
operations in certain configurations. For more details refer to the corresponding section of
this guide. [ReFS chapter is working in progress]
Forward Incremental
187
Backup Methods
The forward incremental backup method is the simplest and easiest to understand; it
generally works well with all storage devices although it requires more storage space than
other backup methods due to the fact that it requires the creation of periodic full backups
(either using active or synthetic backups), typically scheduled weekly. This is necessary
because the incremental backups are dependent on the initial full backup; thus, older full
backups cannot be removed from the retention chain until a newer backup chain is created.
When a new full backup is created, a new chain is started, and the old backups can be
removed once the new chain meets the retention requirements.
Each time an active full is performed (either on schedule or by manually triggering the Active
Full command), a new .VBK file is created by reading all data from the production storage.
Following incremental backups are stored in incremental backup files (.VIB).
When performing active full backups, all blocks are re-read from the source storage.
188
Backup Methods
When to use
Forward incremental backup provides good performance with almost any storage and the
highest level of backup chain consistency since each new chain is populated by re-reading
VM source data. Incremental backups are still processed using Changed Block Tracking
(CBT) thus data reduction is still possible. Active Full can be used in any case where plenty
of repository space is available, the backup window allows enough time and network
bandwidth is sufficient to support reading the source data in full.
Synthetic Full
Synthetic full reads the data already stored in the most recent backup chain (full and its
dependent incrementals) to create a new full backup directly into the destination storage.
If a synthetic full is scheduled, when the job runs, it first creates a normal incremental
backup to collect the most recent changes.
After the job completes the incremental backup, the synthetic full generation is started. It
reads the most recent version of each block for every VM in the job from the backup chain,
and writes those blocks into a new VBK file. This is how a new full backup is "synthetically"
created.
189
Backup Methods
Synthetic full I/O patterns need to be split into two different operation: the creation of the
additional incremental is exactly like any other incremental job. However, the synthetic
creation of the full backup is an I/O intensive process, all in charge of the Veeam repository.
Since the process reads individual blocks from the various files in the chain and writes those
blocks to the full backup file, the I/O pattern is roughly 50%-50% read/write mix. The
processing speed is limited by the IOPS and latency capabilities of the repository storage, so
it may take a significant amount of time. However, there is no impact on the source storage
or production networks during this time as I/O occurs only inside the repository.
NOTE: if an SMB share type of repository is used, the Veeam repository role is executed in
the Gateway Server there is going to be network traffic between the gateway server itself
and the SMB share.
Recommendations on Usage
Due to the way synthetic full works, having many smaller backups jobs with fewer VMs will
allow for faster synthetic full operations. Keep this in mind when setting up jobs that will use
this method or choose to use Per VM Backup Files.
190
Backup Methods
Recommendations on Usage
The primary advantage of using forever forward incremental backup method is space
savings. However, the tradeoff is the required resources for the merge process. The merge
process may take a considerable amount of time, depending on the amount of incremental
changes that the job has to process. The advantage is that the merge process impacts only
the target storage.
Like with synthetic full, it is recommended to have many smaller jobs with a limited number
of VMs, as this can significantly increase the performance of synthetic merge process. Very
large jobs can experience significant increase in time due to extra metadata processing. This
may be remediated by combining forever forward incremental mode with per VM backup
files.
Ideal for VMs with low Jobs with significant change rate may take a long time to
change rate merge
Reverse Incremental
191
Backup Methods
As every other backup method, during its first run reverse incremental backup creates a full
backup file (VBK). All subsequent backups are incremental, that is, only changed data
blocks are copied. During the incremental backup, updated blocks are written directly into
the full backup file, while replaced blocks are taken out and written into a rollback file (.VRB).
This method provides space-efficient backup, as there is only one full backup to store. It also
facilitates retention, since removing old restore points is simply a matter of deleting old VRB
files.
The disadvantage is that creation of rollback files occurs during the backup process itself,
which results in higher I/O load on the target storage and can slow down the backup
process.
Also, over time the in-place injection of new blocks into the full file causes fragmentation of
the VBK file. This can be partially fixed by using compact operations.
This can be especially noticeable for VMs with a high change rate, or when running multiple
concurrent jobs.
Recommendations on Usage
192
Backup Methods
When repository storage uses fast disk with Small NAS boxes with limited I/O
caching RAID controllers and large stripe sizes performance
Deduplication appliances due to
VMs with low change rate
random I/O pattern
High change rate VMs, as VM
snapshot may be open for a long
time
193
Encryption
Encryption
Overview
The encryption technology in Veeam Backup & Replication allows you to protect data both
while it is in transfer between backup components and at rest, when it is stored at its final
destination. This can be disk, tape or a cloud repository. Customers can use one of the
encryption methods or a combination of both to protect against unauthorized access to
important data through all the steps in the data protection process.
Veeam Backup Enterprise Manager additionally provides Password Loss Protection option
that allows authorized Veeam users to recover data from the backup even if the encryption
password is lost. If the password gets lost, the backup server will provide a challenge key for
Enterprise Manager. Using asymmetric encryption with a public/private key pair, Enterprise
Manager generates a response which the backup server can use for unlocking the backup
file without having the password available. For more details on this feature refer to the
corresponding section of the User Guide.
The encryption algorithms used are industry standard in all cases, leveraging AES-256 and
public key encryption methods. Data Encryption section of the User Guide provides detailed
information on the encryption algorithms and standards used by the product.
The following sections describe encryption options available in the product, what they
protect, when they should be used and best practices for their use.
An example is the use of rotated drives for an offsite repository. Because these drives are
rotated offsite, they are at a higher risk of falling into the hands of unauthorized users.
Without encryption enabled, these unauthorized users could install their own copy of Veeam
Backup & Replication and gain access to the stored backups easily.
194
Encryption
On the other hand, if the backup files are encrypted, unauthorized users cannot access any
data in the backups or even learn any critical information about the backup infrastructure as
even backup metadata is encrypted. Without the key used for encryption or access to the
original Veeam Backup & Replication console itself, the backup files remain secure.
This approach provides a method for encrypting backups without compromising backup
performance.
Active full backup is required for enabling encryption to take effect if it was disabled for the
job previously.
Best Practices
Enable encryption if you plan to store backups in locations outside of your security
domain.
While CPU usage for encryption is minimal for most modern processors, some amount
of resources will still be consumed. If Veeam backup proxies are already highly loaded,
take it into account prior to enabling job-level encryption.
Use strong passwords for job encryption and develop a policy for changing them
195
Encryption
regularly. Veeam Backup & Replication helps with this, as it tracks passwords’ age.
Store passwords in a secure location.
Obtain Enterprise or a higher level license for Veeam Backup & Replication, configure
Veeam Backup Enterprise Manager and connect backup servers to it to enable
Password Loss Protection.
Export a copy of the active keyset from Enterprise Manager (see User Guide for more
information).
Back up the Veeam Backup Enterprise Manager configuration database and create an
image-level backup of the Veeam Backup Enterprise Manager server. If these backups
are also encrypted, make sure that passwords are not lost as there will be no Password
Loss Protection for these backups.
Typical use case is to protect data on tapes when media is shipped to an offsite location or
rd
to a 3 party. Without encryption enabled, a lost tape could easily be accessed, and data
stored on tapes could be compromised.
196
Encryption
Tape job encryption should be used any time you want to protect the data stored on tape
from unauthorized access by a 3rd party. Tapes are commonly transported offsite and thus
have a higher chance of being lost and turning up in unexpected places. Encrypting tapes
can provide an added layer of protection if tapes are lost.
If tape jobs are writing already encrypted data to tape (for example, Veeam data from
backup jobs that already have encryption enabled), you may find it acceptable to not use
tape-level encryption. However, be aware that a user who gets access to the tape will be
able to restore the backup files. Although this user will not be able to access the backup data
in those files, some valuable information, for example, job names used for backup files, may
leak.
Best Practices
Enable encryption if you plan to store tapes in locations outside of your security domain.
Consider the risks/benefits of enabling tape job encryption even if the source data is
already encrypted and evaluate appropriately the acceptable level of risk.
Use strong passwords for tape job encryption and develop a policy for changing them
regularly (you can use Veeam Backup & Replication password age tracking capability).
Store passwords in a secure location.
Obtain Enterprise or a higher level license for Veeam Backup & Replication, configure
Veeam Backup Enterprise Manager and connect backup servers to it to enable
Password Loss Protection.
Back up the Veeam Backup Enterprise Manager configuration database and create an
image-level backup of the Veeam Backup Enterprise Manager server. If these backups
are also encrypted, make sure that passwords are not lost as there will be no Password
Loss Protection for these backups.
197
Encryption
Network encryption in Veeam Backup & Replication is controlled via the global Network
Traffic options.
Whenever two backup infrastructure components need to communicate with each other over
the IP network, a dynamic key is generated by the backup server and communicated to each
node over a secure channel. The two components then establish an encrypted connection
between each other using this key, and all communications between these two components
for that session are then encrypted with this key. The key has a one-time use and it's
discarded once the session is completed.
By default, Veeam Backup & Replication automatically encrypts communication between two
nodes if either one or both has an interface configured (if used or not) that is not within the
RFC1918 private address space (10.0.0.0/8, 172.16.0.0/12, 192.168.0.0/16,
169.254.0.0/16). Veeam also automatically uses network-level encryption for any connection
with Veeam Cloud Connect service providers, however Cloud Connect establishes a TLS
1.2 encrypted tunnel to the service provider in a different way. To learn more about specific
Cloud Connect encryption mechanism, watch this YouTube video: How Veeam Cloud
Connect Encryption works.
Best Practices
Enable encryption if network-level attacks are a security concern.
Network-level encryption can use significant CPU resources, especially on the
encrypting side (source) of the connection. Make sure that component nodes have
enough resources. Modern CPU's can offload encryption and reduce the amount of
198
Encryption
CPU resources required. For Intel CPU's specifically, you may check your CPU model
on Intel ARK and look for the AES-NI capability.
199
Deduplication and Compression
Deduplication
What does it do?
The primary purpose of deduplication is to reduce the amount of data that has to be stored
on disk by detecting redundant data within the backup and storing it only once. Veeam
deduplication is based on identifying duplicate blocks inside a single VM disk or across
multiple VMs inside the same job. This is primarily beneficial when VMs are deployed from
the same template since the base image is identical, but it's less useful for incremental data.
Veeam offers 4 different storage optimization settings that impact the size of read blocks and
hash calculations for deduplication:
Local – this is the default setting and is recommended when using a disk-based
repository. When this setting is selected, Veeam reads data and calculates hashes in 1
MB chunks.
LAN – this value is recommended when using a file-based repository such as SMB
200
Deduplication and Compression
shares. When this setting is selected, Veeam reads data and calculates hashes in 512
KB chunks.
WAN – this value is recommended when backing up directly over a slow link or for
replication as it creates the smallest backups files at the cost of memory and backup
performance. When this setting is selected, Veeam reads data and calculates hashes in
256 KB chunks.
Local (>16 TB) – this setting is recommended for large backup jobs with more than 16
TB of source data in the job. When this setting is selected, Veeam reads data hashes
and calculates data on 4 MB blocks.
The smaller the block size, the more CPU will be consumed for hash calculations and the
more RAM will be used to store those hashes.
Note: Local (>16TB) underlying block size has changed in v9.0 from 8 MB to 4 MB. If you
upgrade to Veeam Backup & Replication v9.0 from a previous version, this option will be
displayed as "Local Target (legacy 8MB block size)" in the list and will still use the 8 MB
blocks size. It is recommended that you switch to an option that uses a smaller block size
and create an active full backup to apply the new setting.
However, there are a few special cases where a user might consider disabling this option:
201
Deduplication and Compression
For example, transactional servers like Microsoft Exchange and Microsoft SQL commonly
make small changes across the disk. If you use the 1 MB blocks setting, this can lead to a
great amount of incremental changes each day. The WAN optimization with its smaller block
size of 256 KB may significantly decrease the size of incremental backups. However, this
can have a very significant impact on the backup speed and the amount of memory needed
during the backup process on the repository, especially for large backup jobs.
A 2 TB Microsoft Exchange server may need only 2 GB of RAM on the repository during
backup when using default settings of Local (1 MB) blocks, but would potentially need 8 GB
of RAM on the repository with WAN (256 KB) blocks. Also, transform operations such as
synthetic full backups, forever forward merge and reverse incremental rollback will require
four time the I/O operations compared to the 1 MB block, and this can significantly increase
total backup time. All of this must be taken into consideration prior to changing the defaults.
Best practices
Unless you have a really good understanding of the impact that can cause block size
changing, stick to the defaults.
If you want to change the default block size, be sure to test it well and make sure you
have planned appropriately for the extra I/O and memory requirements on the
repository.
When using a block size smaller than the default one for a large server, it is
recommended to use a backup mode that does not perform synthetic processing (like
forward incremental with scheduled active full).
Note: Block size changes will only become effective after an active full is created.
Compression
What does it do?
The purpose of compression is to reduce the amount of data that has to be transferred
across the wire and stored on disk. Veeam Backup & Replication leverages several different
compression algorithms that provide various balances between compression ratios,
202
Deduplication and Compression
throughput and the amount of CPU use on the backup proxy. Compression provides
maximum effect on space savings in a backup job, so understanding the tradeoffs in these
settings can be very important.
None – this option disables compression for the job. The proxy reads blocks and sends
them uncompressed to the repository where they are written to disk as is.
Dedupe-friendly – this option uses the very simple RLE compression algorithm that
needs very little CPU. It creates somewhat predictable data patterns, which is useful if
users want to leverage 3rd party WAN accelerators with Veeam and/or a deduplication
appliance (without the "decompress before storing" setting). This allows the network
stream to be moderately compressed while still being effectively cached.
Optimal – this is the default compression used on Veeam jobs that leverages LZ4
compression. It provides typical compression ratios around 2:1 with fairly light CPU
overhead. This light CPU overhead allows for excellent throughput with rates up to 150
MB/s per core and even faster decompression rates. This is a most commonly used
practice that allows achieving excellent balance between performance and compression
savings.
High – this option uses zlib compression tuned for low to moderate CPU overhead.
This setting provides for around 10% higher compression ratios compared to optimal,
but uses over 50% more CPU horsepower with rates up to 100 MB/core. If proxies are
not CPU bound, this extra savings may still be very much worth it, especially for larger
repositories or if the bandwidth available is less than the 100 MB/s limit (i.e., 1 Gb links
or less).
Extreme – this option uses zlib compression tuned for high CPU overhead. This
setting uses even more CPU and lowers throughput even further to around 50 MB/core,
while only typically giving around 3-5% additional savings. It is quite rarely used,
however, in cases where bandwidth between the proxy and repository is limited, for
example, when you run primary backups directly through WAN links.
203
Deduplication and Compression
Enabling compression at the job level, and decompressing once sent to the repository will
reduce the traffic between proxy server and backup repository by approximately 50% on
average. If proxy and repository runs on the same server, the compression engine is
automatically bypassed to prevent spending CPU for applying compression. The
uncompressed traffic is sent between local data movers using shared memory instead.
Compression settings can be changed on the job at any time and any new backup sessions
will write new blocks with the new compression mode. Old blocks already stored in backups
will remain in their existing compression level.
Best Practices
Defaults are good, don’t change values without understanding the impact.
Use compression levels above optimal only if you have plenty of CPU and understand
that maximum throughput, especially during full backups, will likely be significantly
lower, especially if the backup proxy CPUs can’t take more load.
Test various compression levels and see how they impact the environment, but always
remember the balance. A single backup job with a few concurrent streams may seem
fine with Extreme compression, but may overload all available proxy CPUs during
production run of all jobs.
Remember that higher compression ratios may also negatively impact restore speeds.
BitLooker
The option "Exclude deleted file blocks" is the third configurable option in job settings. In
several places you will see references to this feature under the name "BitLooker".
204
Deduplication and Compression
When enabled, the proxy server will perform inline analysis of the Master File Table (MFT) of
NTFS file systems and automatically skip blocks that have been marked as deleted.
When upgrading from versions prior to v9.0, this setting is disabled for existing backup jobs.
To enable it for existing jobs, use the following PowerShell commands.
Add-PSSnapIn VeeamPSSnapin;
It is always recommended to leave BitLooker enabled, as it will reduce the amount of backup
storage space required.
205
Backup Job
Backup Job
This screen also provides an advanced object exclusion tool that allows you to select a
parent object and then exclude child objects, or even individual disks within a VM.
Note When selecting very high level containers with many virtual machines, such as
datacenters, clusters or large folders, it is important to keep in mind that tape archive jobs, or
SureBackup jobs with linked jobs cannot exclude certain objects from being processed
Important: Veeam Backup and Replication supports encrypted VMs (in vSphere 6.5) but the
resulting backups will contain unencrypted data. Thus it is strongly recommended to enable
in transit and at rest job level encryption to ensure safety of the data. For more details on
requirements and limitations of the backup of encrypted VMs refer to the corresponding
section of the User Guide.
Grouping VMs running the same operating system or deployed from similar templates into a
single job will increase deduplication rate. Job sizing guidelines still apply, and it is
recommended to monitor the backup window and the size of the job for manageability.
206
Backup Job
Adding resource pools, folders, datastores, or vSphere Tags (vSphere 5.5 and higher) to
backup jobs makes backup management easier. New machines that are member of such
constructs or containers are automatically included in the backup job, and machines
removed from the container are immediately removed from job processing.
When creating jobs based on groups or constructs, ensure that the configured constructs do
not overlap. Overlapping constructs may cause undesired results. For instance, when
creating jobs based on datastores, VMs with disks residing on multiple datastores included
in more than one backup job will cause the VM to be backed up in each job.
Tags
Tags are very convenient for a policy driven approach to data protection. However, it is
recommended to follow these guidelines:
Monitor the number of VMs automatically added to the job to avoid too many VMs being
backed up within a single job
Only one tag can be used to include a VM in a job
Using tags, you can classify VMs by service levels, using different backup jobs for
different service levels
Veeam ONE Business View (OBV) is a very convenient tool for managing vSphere
Tags. OBV allows for creating classification rules and update corresponding tags in
vCenter. Classifications can be defined from CPU, RAM, VM naming convention, folder,
resource pool, datastore etc. OBV can also import VM/host/datastore descriptions from
a CSV file. This feature can be useful when refreshing VMware tags, for example, to
update a CMDB.
Exclusions
It is recommended to limit the number of exclusions in backup jobs. While exclusions can be
very useful, the virtual infrastructure is dynamic and changes rapidly. It is quite possible that
a VM gets moved to a folder or resource pool that is excluded which makes it unprotected.
Monitoring Protected VMs with Veeam ONE is highly recommended.
Also remember that exclusions have higher priority over inclusions in Veeam Backup &
Replication.
207
Backup Job
Replication uses source side deduplication which decreases the amount of data that must
be transferred to the target repository.
When using a deduplication appliance for storing backups, please see the Deduplication
Appliances section of this guide for a detailed description of compression and storage
optimization settings.
Encryption
A detailed description of encryption settings and its influence on the backup infrastructure is
provided in the Encryption section above in this document.
For general guidelines about encryption, refer to the Veeam User Guide: Encryption keys.
Storage maintenance
While data amount is growing and backup window is decreasing, forward incremental
forever backups have become increasingly important in any backup solution. Backup jobs
with no scheduled synthetic or active full backups are becoming more widely adopted.
Forward incremental with weekly synthetic full backups is however still the default setting.
The two main objections towards using a forever forward incremental backup mode are the
following:
The first one is full backup file fragmentation, leading to undesired VBK file growth over
time, and degradation of performance due to fragmentation. Previously it was recommended
to perform periodical active full backups in order to create a new VBK file and backup chain.
This would mitigate issues of fragmentation and remove whitespace left by deleted data
blocks.
The second objection is silent storage corruption. If ever a file or block in the chain got
corrupted by a storage related issue, all subsequent consolidations or restores from this
could be affected.
To address both objections, following features are available under the "Maintenance" tab, in
the Advanced settings of a backup job.
208
Backup Job
Full backup file maintenance will address two issues: VBK file fragmentation caused by
transforms (forward incremental forever, or reverse incremental), and left over whitespace
from deleted data blocks. These issues are mitigated by synthesizing a new full backup file
on the backup repository i.e. copy blocks from the existing VBK file into a new VBK file, and
subsequently deleting the original file. This process may also be referred to as "compacting".
How does it work? During VBK compacting, a new VBK file is created. Existing blocks are
copied from the previous VBK, requiring free space equivalent to the size of an additional full
backup in the repository. In the Restore Point Simulator, this space is part of the "Work
space" parameter. When using Scale-out Backup Repository in Performance Mode, the
compacting process may utilize multiple extents and significantly speed up the compacting
process.
When to use? For every backup job with full transforms. Defragmentation will benefit the
most jobs that are configured to generate a single chain per job, keeping files smaller and
restore speed optimal over time.
How does it work? When a job has finished, storage-level corruption guard will perform a
CRC verification for the most recent restore point. It will validate whether the content of the
backup chain blocks match the content described within the backup file metadata. If a
mismatch is discovered, it will attempt to repair the data block from production storage,
assuming the block still exists and has not been overwritten. If it exists, the backup file will
be repaired. If not, storage-level corruption guard will fail and make the user aware that a
new full backup is required, and that the backup chain must be recovered from a secondary
copy of the backup.
When to use? It is recommended to use storage-level corruption guard for any backup job
with no active full backups scheduled. Synthetic full backups are still "incremental forever"
and may suffer from corruption over time.
209
Backup Job
For more information, please see Veeam Helpcenter: Health Check for Backup Files.
Job Chaining
Chaining backup jobs is convenient in certain circumstances, but should be used with
caution. For example, if a job in such chain fails or stops responding, the entire job chain
delivers poor backup success rate.
A common way to handle multiple jobs is to let the built-in Intelligent Load Balancing (ILB)
handle the proxy/repository resources by starting multiple jobs in parallel by using all
available proxy/repository resources. This allows optimal task scheduling and provides the
shortest backup window.
Load Balancing
When planning jobs schedule, you should consider balancing the load on source and target
disks. Too many jobs accessing the same disk will load the storage significantly; this makes
the job run slower or may have a negative impact on the VMs performance. To mitigate this
problem, you can utilize Storage Latency Control (or Backup I/O Control) settings.
Veeam has a load balancing method that automatically allocates proxy resources making a
choice between all proxies managed by Veeam Backup & Replication that are available at
the moment.
For more details on load balancing, refer to the Veeam Backup & Replication User Guide at
Resource scheduling.
While configuring a backup job, you can disable the automatic proxy selection. Instead, you
can select particular proxies from the list of proxies managed by Veeam backup server, and
appoint them to the job. This is a very good way to manage distributed infrastructures; also it
helps you to keep performance under control.
For example, you can back up a cluster residing on multiple blade chassis. In this case, if
you use virtual proxies, keep the proxies load well-balanced and optimize the network traffic.
210
Backup Job
Dedicated proxies can be also very helpful if you use a stretched cluster and do not want
proxy traffic to go across inter-switch link.
See the illustration below as a good starting point to reach and keep control on high backup
throughput. In this example, administrator wants to keep network traffic as much as possible
inside the chassis; only the proxy-to-repository traffic goes via an external link.
You can use Proxy Affinity to allow only specific proxies to interact with a given repository.
Tip: To optimize load balancing in a distributed environment where backup proxies are rolled
out to multiple sites, it is recommended to select all proxies from the same site in the job.
211
Backup Copy Job
Every backup copy job creates its own folder on the target backup repository and stores its
data in this location. The folder has the same name as the backup copy job.
Once created, a backup copy job will immediately start processing the latest existing restore
point for all VMs included in the job, as long as it has been created less than one
synchronization interval before the start of the backup copy job.
By default, Veeam Backup & Replication keeps 7 restore points on the target backup
repository in case of simple retention policy (see the “Simple Retention Policy” section of the
User Guide for details). If you plan to use Grandfather-Father-Son (GFS) retention, refer to
the “GFS Retention Policy” section for details.
Backup copy jobs file chains layout will depend on the repository option: "Per VM backup
files" will generate one file chain per each VM, otherwise a chain will be generated per each
job.
If a backup copy job cannot process all requested VMs before the end of an incremental
execution interval (by default 24 hours), the job will still create a backup file on the target
backup repository (or multiple files if per-vm chains have been enabled), but some VMs will
be left inconsistent or unprotected. This might be caused by precedence of the backup task
over the backup copy task. The backup copy process will resume from the last full data
transaction during the next synchronization interval.
Limitations of backup copy jobs are described in Veeam Backup & Replication User Guide at
https://helpcenter.veeam.com/backup/vsphere/backup_copy_select_point.html.
Important Note: Jobs with WAN acceleration enabled will process VMs sequentially, while
jobs using direct mode will process included VMs in parallel according to free task slots
availability on backup repositories.
By design, a backup copy job is a process that runs continuously. This process includes
several stages.
A copy job restarts every time at the defined Copy every interval setting (default is 12:00
AM daily) and monitors for new restore points of the selected VMs to appear in the specified
sources. On the Schedule tab it is possible to define time period when data transfers are
allowed. This is especially helpful, when transferring multiple times per day (e.g. hourly
synchronization interval), or again when the bandwidth used to transfer the backup copy
jobs can only be used during the night.
The concept of the "interval" is used to define two parameters: how often the job should be
looking for new points, and for daily intervals at what time it should start looking for points. If
you set an interval of 1 day, that equals to instruct the backup copy job that once a day,
starting at the selected time, it should begin looking for new restore points. When the restore
point is found, the copy job will copy it. However, once a single point is copied, another point
for that VM will not be copied until the next interval starts.
Another reason for this design is that you may run local backups more often (for example,
hourly), but you may only want to copy data offsite only daily or weekly, thus you can set the
backup copy "interval" independently of the schedule of the backup jobs it is using as
source.
1. Pre-job activity — if enabled, the pre-job scripts are executed at the very beginning of
a copy interval.
2. Health check — if scheduled, backup file integrity is verified before the next copy is
initiated.
3. Data transfer (synchronization) phase — during this phase, the backup copy job
checks for a new restore point in the source, creates a file for a new restore point at the
target and starts copying the state of the latest restore point of each processed VM to
the target repository. The data transfer (synchronization) phase starts at specific time
configured in the job properties (see Synchronization Intervals). You can define any
interval needed in minutes, hours or days. Moreover, you can specify the time slot
during which data can and cannot be transferred over the network, thus regulating
network usage (see Backup Copy Window).
213
Backup Copy Job
4. Transform phase — copy jobs are by nature running in "forever forward incremental"
mode, and perform transform operations on the target backup repository accordingly.
Additionally, it is possible to schedule health checks or backup file compacting as
described in the Backup Job section. The transform phase begins when all VMs are
successfully copied to the target, or if the synchronization interval expires.
Note: the transform process itself puts additional pressure on the target repository. In
large environments with deduplication storage appliances used as backup repositories
or with backup copy jobs processing a large number of VMs or big VMs, the transform
process can take a significant amount of time. For non-integrated deduplication
appliances, it is recommended to use the "Read entire restore point..." option. This
forces the Backup Copy Job to running forward incremental with periodical full backups
copied entirely from the source backup repository rather than being synthesized from
existing data.
5. Compact full backups — if enabled, the recent full backup file is re-created on the
same repository, writing all the blocks close to each other as much as possible to
reduce fragmentation.
6. Post-job activity — if enabled, several post-job activities are executed before the job
enters the idle phase, such as post-job scripts and sending e-mail reports.
7. Idle phase — for the most time, the backup copy job remains in the Idle state, waiting
for a new restore point to appear on the source backup repository. When the
synchronization interval expires, a new interval starts at step 1.
For more information, refer to the corresponding section of the User Guide > Backup Copy
Job.
214
Backup Copy Job
job you want to copy. The job container will protect all the VMs in the selected source
job(s).
Select from backup: this method is equivalent to the Select from infrastructure
method, but allows for selecting specific VMs inside specific backups. This is helpful,
when only certain critical VMs should be copied offsite.
Initial synchronization
When creating the initial copy to the secondary repository, it is recommended to use backup
seeding (see Creating Seed for Backup Copy Job) whenever possible. Especially when
transferring large amounts of data over less performant WAN links, the seeding approach
can help mitigating initial synchronization issues.
While Backup Copy Jobs were designed for WAN resiliency, the initial copy is more error
prone, as it is typically transferring data outside the datacenter over less reliable links (high
latency, or packet loss). Another issue that can be solved by seeding is when the full backup
is larger than the amount of data that can be transferred in an interval. Even if the interval
can be extended to accomodate the initial transfer, this may lead to upload times of even
multiple days. Seeding can speed up the initial sync by removing the need for the sync.
The most frequent synchronization issues are described in the User Guide > Handling
Backup Copy Job Issues.
Additional Options
Restore Point Lookup
By default, after a restart of the job interval (the Copy every setting), a backup copy job
analyzes the VM list it has to protect, and searches backwards in time for newer restore
point states. If the state of the restore point in the target repository is older than the state in
the source repository, the new state is transferred.
215
Backup Copy Job
For example, if the backup job is scheduled to run at 10:20 PM, and the backup copy job
uses the default schedule of copying the latest restore point state every day at 10:00 PM,
the state copied by the backup copy job is typically one day behind. In the image below, you
can see some VMs affected by this behavior.
The following forum thread provides a very good explanation of the backup copy scheduler
and the LookForward registry key > Veeam Community Forums - Backup Copy Intervals
Job Seeding
Usually, a backup copy is used to send data remotely. If it is necessary to send data over a
slow link, you can seed the backup copy job by taking the following steps:
1. Create a "local" backup copy job and target it at a removable device used as a backup
216
Backup Copy Job
repository, or copy the backup files afterwards. Run the created backup copy job to
create a full backup set on this device. Note that also the .vbm file has to be moved.
2. Once the backup copy job is over, delete the local backup copy job from the Veeam
console.
3. Transport the removable device with the created backup files to the destination site.
4. Copy backup file to the target backup repository.
5. Import the backup on the target. If already imported, perform a rescan.
6. Create the final backup copy job on the Veeam console. On the Target step of the
Backup copy job wizard, use the Map backup link and select the transported backup
— this backup will be used as a “seed”.
If you are using a WAN accelerated transfer, refer to the WAN Accelerator section for proper
cache population procedure:
https://helpcenter.veeam.com/backup/vsphere/wan_populate_cache.html.
Note: Only the initial first run of a reverse incremental chain can be used with seeding (but
any forward incremental chain can be used). See kb1856 for more information.
217
Replication Job
Replication Job
Note: This section focuses on replicating VMs to your own virtual infrastructure. When
implementing Cloud Connect replication for DRaaS only source configuration details of
this section are relevant to end user side of the deployment. For more information on
implementing DRaaS on Cloud Connect provider side refer to Cloud Connect
Reference Architecture document.
Replication jobs are used to replicate VMs to another or the same virtual environment
(instead of creating deduplicated and compressed backup files at backup run). Veeam can
store up to 28 restore points (on VMware platforms).
Like backup, replication is a job-driven process. In many ways, it works similarly to forward
incremental backup:
During the first run of a replication job, Veeam Backup & Replication copies a whole VM
image and registers the replicated VM on the target ESXi host. In case of replicating to
a cluster a host with least VMs registered at the moment will be used.
During subsequent runs, the replication job copies only incremental changes, and
creates restore points for the VM replica — so the VM can be recovered to the selected
state. Every restore point is in fact a regular VMware snapshot.
When you perform incremental replication, data blocks that have changed since the last
replication cycle are written to the snapshot delta file next to the full VM replica. The
number of restore points in the chain depends on the retention policy settings.
Replication infrastructure and process are very similar to those used for backup. They
include a source host, a target host with associated datastores, one or two proxy servers
and a repository. The source host and the target host are the two terminal points between
which the replicated data is moved.
Replicated data is collected, elaborated and transferred with the help of Veeam data movers.
The data movers involved in replication are the source proxy, the target proxy and the
repository. The data mover hosted on the repository processes replica metadata files.
Important! Although the replica data is written to the target datastore, certain replica
metadata must be located on a backup repository. This metadata is used by the source
proxy and thus should be deployed closer to the source host and therefore no
compression/uncompression processing is used.
1. When a new replication session is started, the source-side data mover (proxy task)
218
Replication Job
performs the same operations as in backup process. In addition, in cases when VMware
CBT mechanism cannot be used, the source-side data mover interacts with the
repository data mover to obtain replica metadata — in order to detect which blocks have
changed since the previous job run.
2. The source-side data mover compresses the copied blocks of data and transfers them
to the target data mover. Note: In on-site replication scenarios, the source-side
transport service and the target-side transport service may run on the same backup
proxy.
3. The target-side data mover uncompresses replica data and writes it to the destination
datastore.
Veeam Backup & Replication supports a number of replication scenarios that depend on the
location of the target host and will be discussed later in this section.
During replication cycles, Veeam Backup & Replication creates the following files for a VM
replica:
During the first replication cycle, Veeam Backup & Replication copies these files to the
selected datastore to the <ReplicaName> folder, and registers a VM replica on the target
host.
Replica restore points (snapshot delta files). During incremental runs, the replication job
creates a snapshot delta file in the same folder, next to a full VM replica.
Replica metadata where replica checksums are stored. Veeam Backup & Replication
uses this file to quickly detect changed blocks of data between two replica states.
Metadata files are stored on the backup repository.
During the first run of a replication job, Veeam Backup & Replication creates a replica with
empty virtual disks on the target datastore. Disks are then populated with data copied from
the source side.
To streamline the replication process, you can deploy the backup proxy on a virtual machine.
The virtual backup proxy must be registered on an ESXi host with direct connection to the
target datastore. In this case, the backup proxy will be able to use the Virtual Appliance
(hotadd) transport mode for writing replica data to target. In case of NFS datastore at target,
you can as well use Direct Storage access mode (Direct NFS) to write the data.
If the Virtual Appliance mode is applicable, replica virtual disks are mounted to the backup
proxy and populated through the ESX I/O stack. This results in increased writing speed and
fail-safe replication to ESXi targets. For information on Virtual Appliance mode, see
https://helpcenter.veeam.com/docs/backup/vsphere/virtual_appliance.html?ver=95.
219
Replication Job
If the backup proxy is deployed on a physical server, or the Virtual Appliance or Direct NFS
mode cannot be used for other reasons, Veeam Backup & Replication will use the Network
transport mode to populate replica disk files. For information on the Network mode, see
https://helpcenter.veeam.com/docs/backup/vsphere/network_mode.html?ver=95.
The Direct SAN mode (as part of Direct Storage Access) can only be used together with
replication targets in case of transferring thick-provisioned VM disks at the first replication
run. As replication restore points are based on VMware snapshots, that are thin provisioned
by definition, Veeam will failback to Virtual Appliance (HotAdd) mode or Network mode, if
configured at proxy transport settings. Direct SAN mode or backup from storage snapshots
can be used on the source side in any scenario.
Note: Veeam Backup and Replication supports replicating VMs residing on VVOLs but
VVOLs are not supported as replication target datastore. Replication of encrypted VMs is
supported but comes with requirements and limitations outlined in the corresponding section
of the User Guide. Replication of encrypted VMs is NOT supported when the target is
Veeam Cloud Connect.
Onsite Replication
If the source and target hosts are located in the same site, you can use one backup proxy
for data processing and a backup repository for storing replica metadata. The backup proxy
must have access to both source host and target host. In this scenario, the source-side data
mover and the target-side data mover will be started on the same backup proxy. Replication
data will be transferred between these two data movers and will not be compressed.
220
Replication Job
Offsite Replication
The common requirement for offsite replication is that one Veeam data mover runs in the
production site (closer to the source host), and another data mover runs in a remote site
(closer to the target host). During backup, the data movers maintain a stable connection,
which allows for uninterrupted operation over WAN or slow links.
Thus, to replicate across remote sites, deploy at least one local backup proxy in each site:
The backup repository must be deployed in the production site, closer to the source backup
proxy.
Tip: It is recommended to place a Veeam backup server on the replica target side so that it
can perform a failover when the source side is down. When planning off-site replication,
consider advanced possibilities — replica seeding, replica mapping and WAN acceleration.
These mechanisms reduce the amount of replication traffic while network mapping and re-IP
streamline replica configuration.
For offsite replication, open the connections between the Veeam backup components:
The Veeam backup server must have access to the vCenter Server, the ESXi hosts, the
source backup proxy and the target backup proxy.
The source backup proxy must have access to the Veeam backup server, the source
ESXi host, backup repository holding the replica metadata, the target proxy, and the
source vCenter Server.
The target backup proxy must have access to the Veeam backup server, the source
proxy, the target ESXi host, and the target vCenter Server.
221
Replication Job
The source proxy compresses data and sends it via the WAN to the target proxy, where the
data is uncompressed. Note that you also can seed the replica by sending the backup files
offsite (using some external media, for example) and then only synchronize it with
incremental job runs.
In this scenario:
The Veeam backup server in the production site will be responsible for backup jobs
(and/or local replication).
The Veeam backup server in the DR site will control replication from the production site
to the DR site.
Thus, in disaster situation, all recovery operations (failover, failback and other) will be
performed by the Veeam backup server in the DR site. Additionally, it may be worth installing
the Veeam Backup Enterprise Manager to have visibility across the two Veeam backup
servers so that you only have to license the source virtual environment once (used from both
backup servers)
Tip: Plan for possible failover carefully. DNS and possibly authentication services (Active
Directory, for example, or DHCP server if some replicated VMs do not use static addresses)
should be implemented redundant across both sides. vCenter Server (and vCD)
infrastructure should be as well considered for the failover scenario. In most cases, Veeam
do not need a vCenter Server for replica target processing. It can be best practice to add the
ESXi hosts from the replica target side (only) directly to Veeam Backup & Replication as
managed servers and to perform replication without vCenter Server on the target side. In
this scenario a failover can be performed from the Veeam console without an working
vCenter Server itself (for example to failover the vCenter Server virtual machine).
222
Replication Job
Also, when replicating VMs to a remote DR site, you can manage network traffic by applying
traffic throttling rules or limiting the number of data transfer connections. See Veeam Backup
& Replication User Guide for more information:
https://helpcenter.veeam.com/docs/backup/vsphere/setting_network_traffic_throttling.html?
ver=95.
Tip: Replication can leverage WAN acceleration allowing a more effective use of the link
between the source and remote sites. For more information, see the User Guide
https://helpcenter.veeam.com/docs/backup/vsphere/wan_acceleration.html?ver=95 or the
present document (the “WAN Acceleration“ section above).
In some circumstances, you can get a better RTO with an RPO greater or equal to 24 hours,
using replicas from backup. A common example beside the usage of proactive VM restores,
is a remote office infrastructure, where the link between the remote site and the
headquarters provides limited capacity.
In this case, the data communication link should be mostly used for the critical VM replicas
synchronization with a challenging RPO. Now, assuming that a backup copy job runs for all
VMs every night, some non-critical VMs can be replicated from the daily backup file. This
requires only one VM snapshot and only one data transfer.
223
Replication Job
You can find additional information about replica from backup in the appropriate section of
the Veeam Backup & Replication User Guide:
https://helpcenter.veeam.com/backup/vsphere/replica_from_backup.html?ver=95
Tip: This feature is sometimes named and used as proactive restore. Together with
SureReplica, it is a powerful feature for availability.
224
Application-Aware Image Processing
225
Application-Aware Image Processing
6. Veeam Backup & Replication runs post-thaw script (if any) for the Microsoft
Windows/Linux guest OS.
7. Backup data transfer and snapshot commit is performed.
8. Finally, log file truncation is performed with VSS (for Microsoft SQL Server and
Exchange Server) or using native Oracle commands (for Oracle databases on Linux).
Note that you can use pre- and post-job scripting to automate job global settings from the
Veeam Backup & Replication server itself. It is recommended to use the VM guest
processing options for interaction with VMs.
226
Application-Aware Image Processing
Windows
and VMware
Linux (with VSS-
applications supported Windows Windows
VM guest and known applications with VSS- (no VSS- Linux with
OS type user for (without aware aware applications
Guest OS known user applications applications)
processing) for Guest
OS
processing)
Guest OS
processing
Y Y Y Y Y
is
applicable
Use
VMware
N Y N N N
Tools
quiescence
VMware
Tools
quiescence
with Y N N N N
VMware
Script
processing
Enable
Veeam
Application-
N N Y N N
Aware
Image
Processing
Enable
Veeam
Application-
Aware
Image N N N Y N
Processing
and
InGuest
Scripts
Disable
Veeam
Application-
N N N N Y
Aware
Image
Processing
227
Application-Aware Image Processing
To coordinate proper VSS and indexing activities, Veeam Backup & Replication deploys a
small executable component inside a VM. It is installed only during VSS quiescence
procedure and removed immediately after the processing is finished, producing very low
impact on VM performance and stability. As for connection method for accessing VM guest
OS, Veeam first tries to connect to the VM over network using RPC and then by VMware
VIX channel through VMware Tools (for Windows guest only).
Any Windows server managed by Veeam Backup and Replication can be selected to act as
guest interaction proxy but the preference would be given to the server that has IP address
in the same subnet as subject VM. This functionality allows for having only small limited
range of ports to allow through the firewalls in restricted environments and for that reason it
is recommended to have guest interaction proxies in all VM subnets that are not supposed
to be directly accessible from the network where Veeam backup server resides.
For details on network configuration refer to the section "Required ports" below.
Tip: If the backup server has no network connection to the VMs and deploying additional
guest interaction proxies is not practical/possible (for example, service provider
environments), order in which backup server or guest interaction proxy tries to communicate
to a VM can be changed using the following registry key:
RPC connections means injecting the file via the "ADMIN$" share on the target VM. See
Veeam Knowledge Base article at http://www.veeam.com/kb1230 for more information.
Consider that this is a global setting that will be applied on the Veeam backup server level
228
Application-Aware Image Processing
Tip: To verify the credentials you supplied on the Guest Processing step of the job wizard,
click Test Now button.
Windows OS
Linux OS
Certificate-based
No Yes No
authentication available
VMware Tools must be
Yes Yes No
installed and up to date
Required Ports
229
Application-Aware Image Processing
The following ports should be open between the Veeam backup server and VM for guest OS
processing:
For Windows VMs - remote RPC ports, including Dynamic Port Range (TCP ports 1025
to 5000 - for Microsoft Windows 2003, 49152-65535 - for Microsoft Windows 2008 and
newer); TCP\UDP ports 135, 137-139, 445.
For Linux VMs – SSH port (default is TCP port 22)
For details, refer to the Veeam Backup & Replication User Guide
(https://helpcenter.veeam.com/docs/backup/vsphere/used_ports.html?ver=95).
Sizing
Since guest processing produces very low impact on VM performance, no special
considerations on sizing are required. If you use VSS processing with VMware Tools
quiescence or Veeam in-guest processing, you need free space on each drive of the VM for
the software VSS snapshot. Please check Microsoft requirements for more information.
File exclusions
Another operation Veeam Backup can do on guest OS level (NTFS only) is excluding certain
files or folders from the backup. Alternatively the job can be configured to include only
specified files or folders in the backup.
This functionality operates very similarly and shares a lot of characteristics with excluding
Windows page file and deleted file blocks. It may help reduce size of the backup files or
implement additional data protection strategies for specific data. Backups for which this
option was enabled remain image-level and hypervisor APIs are used to retrieve VM data.
File exclusion feature uses a combination of NTFS MFT data and guest file system indexes
collected by in-guest coordination process to determine which virtual disk blocks belong to
the excluded files and thus should not be included in the backup.
Full file/folder paths, environment variables or file masks can be used to define exclusions.
For more details on configuring exclusions and its limitations refer to the corresponding User
Guide section.
Note: Generic file exclusions (defined for high level folders) are most effective. File masks
exclusions require guest file system indexes and generating indexes may put additional
stress on guest VM and will increase backup time. For this reason it is recommended to
230
Application-Aware Image Processing
avoid using file system masks especially on fileservers with large number (thousands) of
small files and use high level folder exclusions instead. When using include filters, file
exclusions are created for everything else and can take significant time.
1. Virtual machine NTFS MFT is read into the memory cache on the backup proxy, data
blocks that store excluded files are marked as deleted.
2. When sending data blocks to target repository data is read both from the VM snapshot
and memory cache on the backup proxy. Target repository reconstructs VM disks
without excluded VM blocks.
3. Virtual machine NTFS is modified using the data in the cache on the proxy and
information about excluded data blocks is saved in the backup file or replica metadata.
This information is necessary as CBT is not aware of which blocks were excluded and is
used to determine which blocks should be processed during the next backup session.
1 Only this account is able to bypass the UAC prompt for launching processes with
entirely disabled, unless the user account is the local administrator account (SID S-...-500).
231
Data Verification Using Virtual Labs
232
Data Verification Using Virtual Labs
When a SureBackup job is executed the static routes to reach the masqueraded networks
are temporarily added to the routing table on the Veeam backup server. To review the
routing table, you can open a command prompt on the Veeam backup server and execute:
route print -4
You may run this command before and after starting the SureBackup job to compare the
differences.
The routes are added just after the Virtual Lab appliance has booted and has been correctly
initialized by the Veeam backup server. As static routes are added, this will ensure the
Virtual Lab appliance is the gateway for all packets destined to the masquerade networks.
To avoid network reconfiguration of physical components, place the backup server and the
Virtual Lab appliance in the same network subnet.
2. Virtual Lab appliance network interfaces are reconfigured for appropriate isolated
networks.
4. The SureBackup job waits for IP configuration to be published and stabilized through
VMware Tools.
5. A static route for the configured masqueraded networks is added dynamically to the
routing table of the Veeam backup server. Those static routes define the IP address of
the Virtual Lab appliance as the gateway towards the masquerated networks.
233
Data Verification Using Virtual Labs
2. Veeam reconfigures the VMs and connects them to the isolated port groups of the
Virtual Lab. If a network connection is configured to be connected to a port group that is
not available in the Virtual Lab, those network are disconnected automatically.
3. Veeam creates a snapshot for the VMs in order to redirect write operations to a
production datastore selected during the Virtual Lab configuration.
4. If the domain controller role is selected, registry settings are injected in the VM to
ensure the NETLOGON service will not shutdown due to missing peer communication.1
6. During boot VMware Tools announce IP configuration of VMs. The SureBackup job
waits for this information to stabilize.
Note: If VMware Tools are not installed on the virtual machine the job will wait for the
duration of Maximum allowed boot time configured for the VMs. This will slow down
SureBackup jobs significantly. Therefore, it is always recommended to install VMware Tools
on a verified VM.
2. PING tests are initiated according to the masqueraded network configuration. The ping
is sent from the Veeam backup server using the static routes added during the job
execution. Since the masquerade network is not part of the Veeam backup server's own
subnet, the packet is sent to the gateway matching the Virtual Lab network (usually the
virtual lab appliance).
3. Application-specific testing uses scripts and is enabled based on the roles assigned
to a VM in the application group configuration. The built-in roles will check
corresponding TCP ports for a given service. The built-in role for SQL Server provides
additional testing (see next section), and custom scripts may be used for third party
applications. Requests are sent from the Veeam backup server, and the routing to the
virtual machine is handled by the Virtual Lab proxy appliance.
234
Data Verification Using Virtual Labs
Note: This feature reads the entire backup file, and requires significant time to
complete.
If Linked Jobs are configured for the SureBackup job, linked VMs will start booting once all
virtual machines explicitly defined within the Application Group have been successfully
booted and verified. Remember that by default 3 VMs are tested at the same time in a
Linked Job. There may be more than 3 VMs linked, but the following ones will stay in the
testing queue. The limit can be adjusted in the SureBackup job configuration wizard, and
may be increased if the backup repository can handle the load accordingly.
235
Data Verification Using Virtual Labs
Default startup
Role Default test script
options
600s maximum
boot time
DNS Server Connection test on port 53
120s application
timeout
1800s maximum
Domain Controller
boot time
(authoritative or non Connection test on port 389
120s application
authoritative)
timeout
1800s maximum
boot time
Global Catalog Connection test on port 3268
120s application
timeout
1800s maximum
boot time
Mail Server 120s application test on port 25
timeout
Connection
1800s maximum
boot time Run “USE” SQL command against
SQL server
120s application all defined databases on the server
timeout
1800s maximum
Veeam Backup for Office boot time
Connection test on port 9191
365 120s application
timeout
600s maximum
boot time
Web Server 120s application test on port 80
timeout
Connection
Note : You will notice that the Domain Controller startup mode (authoritative or not) can now
be choosen. Veeam will mark the server accordingly so it boots in the selected mode. This is
especially useful if many DC needs to be tested in a single SureBackup job. Please remind
that if a single (or the first) Domain Controller is booted, it might use the authoritative mode.
Subsequent Domain controllers must then use non-authoritative mode and will then
synchronize from the authoritative one.
236
Data Verification Using Virtual Labs
A dedicated Visual Basic script is included to allow for testing whether all databases on a
given instance are available. This script is available in the Veeam installation folder as the
Veeam.Backup.SqlChecker.vbs file.
By default, the script tries to retrieve and check all instances; you can optionally configure
one or more specific instances to be tested. The script enumerates all databases and
checks if these databases are available, using the USE <db> statement.
When running scripts that require authentication, when executed the script will impersonate
the service account under which the Veeam Backup Service is running (default is SYSTEM).
To specify different credentials configure them in the 'Credentials' tab in the Application
Group settings.
Important! To ensure successful authentication it is required for the specified user to have
public access to all databases.
The SqlChecker.vbs script also accepts two additional parameters to use SQL
authentication instead of Windows based authentication. In order to use SQL authentication
you need to add a custom test script instead of the built-in SQL Server role, and specify the
following path and arguments:
237
Data Verification Using Virtual Labs
1. Open the Veeam installation folder and look in the SbRoles folder. All roles are defined
in the XML files available in this folder.
2. To create custom roles, duplicate one of the above mentioned files and modify the
<Id> tag using a UUID generator (such as https://www.uuidgenerator.net). Use this
When creating custom roles for Linux-based applications you may need to execute the
generated code locally within the VM. To do so, use \Putty\plink.exe shipped with the
product and located in the Veeam Backup & Replication installation directory.
When executing bash scripts locally on a Linux virtual machine using plink.exe , the exit
codes are passed to the SureBackup job, enabling correct error reporting. If using
plink.exe in combination with a SSH private key, you should connect manually (one time)
to the VM via SSH using putty.exe from the Veeam backup server in order to accept the
target VM SSH fingerprint; otherwise, the SureBackup job will wait for this input and
ultimately timeout.
238
Data Verification Using Virtual Labs
Common Issues
When performing SureBackup, there are few common issues you may come across. Most of
these issues are described in Veeam knowledge base articles:
When restoring Windows 2008 R2 virtual machines with the VMXNET3 network
adapter, the resulting virtual machine obtains a new NIC, and all network settings have
to be adjusted manually. The solution is explained in Veeam KB1570
When using DHCP with leases bound to MAC addresses, ensure that the vNIC MAC
address is configured as static . Otherwise the VM will boot with a MAC in the Virtual
Lab, and the VM may get a different IP address > Setting a static MAC address for a
virtual NIC
Some Linux distributions use udev for assigning names to NICs. If the MAC address
changes during replication or Instant VM Recovery, the NIC's configuration file may not
be applied. For more information, please see RHEL6 SureBackup
Troubleshooting Mode
If you need to troubleshoot Virtual Lab, it is recommended to start sessions in the
Troubleshooting Mode. To do so:
3. Select Start.
The SureBackup lab will now start in troubleshooting mode, which means that errors will not
cause the Virtual Lab to shut down immediately.
If the selected VM is in an application group, this VM and previous ones are started. If the
VM is part of a linked job, the entire Application Group and the selected VM is started.
This mode is especially helpful during an implementation phase while measuring application
boot times via vPower NFS, or implementing custom verification scripts. When you have
finished troubleshooting, you can stop the SureBackup session manually.
Tip: On the Virtual Lab appliance, ICMP traffic is blocked on all network interfaces
connected to isolated networks, unless you check the "Allow proxy appliance to act as
internet proxy for virtual machines in this lab". This may lead to undesired behavior of some
239
Data Verification Using Virtual Labs
When Distributed vSwitch (dvSwitch) is available, port groups can span multiple ESXi hosts
("Advanced Multi-Host" configuration). Distributed vSwitches are typically required when
using Virtual Lab for replicas (SureReplica) as replicas will often span multiple hosts.
vSphere Distributed Resource Scheduler (DRS) may also distribute VMs across multiple
hosts within a cluster once they are started.
Important! Please check the following help article and the links at the bottom of the
webpage before you configure Virtual Labs for Distributed vSwitch: Advanced Multi-Host
Virtual Labs.
Even in environments where Distributed vSwitch is available, make sure that the Veeam
backup server and the Virtual Lab proxy appliance are placed in the same VLAN to prevent
network packets (sent to the masquerading IP subnets) from being routed.
240
Data Verification Using Virtual Labs
Most DR datacenters are configured with different IP networks from production to allow for
“active-active” configurations. In such cases, layer 3 (L3) is used for networking
configuration and routing is in place to establish communications between the production
site and the DR site.
For more information, please see the Backup Server Placement section of this guide.
241
Data Verification Using Virtual Labs
1. For more information about Domain Controller restore, please see the corresponding
thread in Veeam ↩
242
Overview of Applications Support
It is possible to ensure data safety and transactional consistency for applications not
covered in this guide using pre-freeze and post-thaw scripts that will execute inside of the
virtual machine. Subject application has to provide the way to prepare itself appropriately.
Generally speaking pre-freeze and post-thaw scripts have to (depending on the capabilities
of the application):
Certain applications do not require these steps as they include self-healing mechanics or
maintain transactional consistency by other means, application documentation has to be
checked and/or application vendor has to be contacted for specifics on achieving this.
Note that in addition to configuring application consistency for such applications, restore
process has to be properly planned as additional steps would have to be followed to restore
them as well. Using U-AIR (Universal Application Item Recovery) functionality allows for
performing restores of any applications including custom in-house built provided the native
application management tools are used.
243
Active Directory
Active Directory
Veeam Backup and Replication natively supports backup of Microsoft Active Directory
controllers and allows for image level and granular AD items restore.
Preparation
For Microsoft Active Directory, check the tombstone lifetime settings, as described in Veeam
Explorers User Guide at Veeam Help Center
(https://helpcenter.veeam.com/docs/backup/explorers/vead_recommendations.html?
ver=95).
Job configuration
For backup and restore of domain controllers to work properly application aware image
processing opption has to be enabled in the job properties. For more details refer to the
corresponding section of the User Guide.
Recovery verification
There are two Domain Controller roles available in application group configuration - for
authoritative and non-authoritative restore. When testing recovery of one domain controller
only choosing role with authoritative restore will speed up verification process.
244
Active Directory
245
Microsoft Exchange
Microsoft Exchange
Veeam Backup and Replication supports variety of Exchange configuration including DAG
deployments. For more details refer to the corresponding section of the User Guide.
Preparation
DAG clustered configurations may require adjusting cluster timeouts to avoid failovers during
backup as per KB1744.
Job configuration
For backup and restore of Exchange servers to work properly application aware image
processing opption has to be enabled in the job properties. For more details refer to the
corresponding section of the User Guide.
246
Microsoft SQL Server
Preparation
Transaction logs are processed periodically and stored in temporary folder inside of the VM
before shipping to repository/shipping server. Default location of the temporary folder is
%allusersprofile%\Veeam\Backup. To change temporary folder use SqlTempLogPath
(STRING) registry value as described at How It Works: SQL Server and Transaction Log
Backup:
For the list of all registry keys responsible to fine-tuning MS SQL server backup (for example
excluding certain databases from processing) refer to KB2182.
As restore is integral part of SQL Server protection, special attention should be paid to
planning Veeam Explorer for SQL configuration, specifically network connectivity between
mount server and staging servers in restricted environments. Ports used for communication
between them are listed in the Used Ports section of the User Guide.
Job configuration
When backing up AlwaysOn availability group make sure all cluster nodes are processed by
the same backup job for transaction logs processing and restores to work properly. Consider
increasing cluster timeouts in case failover occurs during the backup, similar to Exchange
DAG as per KB1744.
247
Microsoft SQL Server
One special case of custom settings that must be configured on staging server is encryption.
When performing restore/export of encrypted database please refer to KB2006 for details on
configuring the staging server.
248
Microsoft SharePoint Server
Job configuration
For backup and restore of SharePoint servers to work properly application aware image
processing opption has to be enabled in the job properties. For more details refer to the
corresponding section of the User Guide. As SharePoint deployments may spread across
several servers make sure to familiarize yourself with the Required Microsoft SharePoint
Backup Job Settings section of the User Guide.
249
Oracle Database
Oracle
Veeam Backup and Replication natively supports backup of Oracle database servers and
allows for image level and granular Oracle databases restore.
Note: 32-bit Oracle instances on 64-bit Linux, and Oracle RAC are not supported.
Preparation
Only databases in ARCHIVELOG mode will be backed up online, databases in
NOARCHIVELOG mode will be shut down which will cause database availability
disruption.
Logs are stored temporarily on the guest filesystem before they are shipped for processing.
This may cause undesired behavior if there is no enough space available in default location
and changing temporary location from default is recommended as per KB 2093.
When backing up Oracle on Linux, the backup server is used for initiating connections,
whereas a Guest Interaction Proxy will be selected for Oracle on Windows.
As restore is integral part of Oracle protection, special attention should be paid to planning
Veeam Explorer for Oracle configuration, specifically network connectivity between mount
server and staging servers in restricted environments. Ports used for communication
between them are listed in the corresponding section of the User Guide
(https://helpcenter.veeam.com/docs/backup/vsphere/used_ports.html?ver=95#explorers).
Permissions
Certain level of access is expected from the user account configured for performing Oracle
backup. Refer to the corresponding section of the User Guide for details
(https://helpcenter.veeam.com/docs/backup/explorers/veo_connection_to_source_server.ht
ml?ver=95).
When processing Linux instances, the same user account specified for application
awareness is used to process the Oracle backup. For Windows instances, you may specify
two separate accounts.
Note: It is not possible to use different accounts to access different Oracle instances running
on the same VM, make sure specified credentials can be used to access all instances on a
VM in those cases.
250
Oracle Database
Windows OS
User account used to connect to a VM should have local administrator privileges on guest
VM and read/write access to database files on filesystem level.
In addition this account or separate Oracle account in case it is different should have
SYSDBA rights, this can be achieved by adding it to ora_dba local group.
Linux OS
Root account or account elevated to root should be used to connect to a VM. Automatic
adding to sudoers can be enabled for the account but note that sudoers file entry will not
be removed automatically. Persistent sudoers file entry with NOPASSWD: ALL option can
be added manually, for example:
1
This account should be included in the oinstall group to access Oracle database files
hierarchy, and to asmadmin group (where applies).
In addition this account or separate Oracle account in case it is different should have
SYSDBA rights, this can be achieved by adding it to dba local group.
Job configuration
Refer to the corresponding section of the User Guide
(https://helpcenter.veeam.com/docs/backup/vsphere/replica_vss_transaction_oracle_vm.htm
l?ver=95) for details on configuring Oracle database backup and transaction logs
processing.
Avoid using aggressive logs truncation settings for databases protected with Data Guard as
it may affect logs synchronization to secondary server. Data Guard should have enough time
to transport logs remotely before they are truncated thus generally having "Delete logs older
than" option less than 24 hours is not recommended.
Job workflow
Oracle on Linux backup workflow
1. Coordination component which will perform all the necessary steps is injected into the
251
Oracle Database
guest VM. This component is the same as the one used for Linux application-aware
image processing in general.
2. Perform application discovery. This is done using native OS methods, coordination
component queries /etc/oraInst.loc and reads inventory.xml which is then compared to
/etc/oratab information.
3. Status and version of instance(s) is fetched.
4. Disk group information is retrieved for ASM instances.
5. Log mode is identified, this information will later be used for decisions on how exactly
the database has to be processed. Database files, CDB (Oracle 12 only) and current
DBID information is retrieved.
6. At this step archive log necessary information was collected and Veeam will start doing
actual backup, modifying database state - current archive log is archived and all archive
log information is retrieved.
7. PFILE backup is created and archived into the backup metadata.
8. Additional information is collected and recorded (current DBID, SCN, Sequence IDs,
database unique name, domain, recovery file destination, basic listener information and
current archive log).
9. Coordination component is shut down and then restarted again to finalize the backup:
database is put into backup mode and database snapshot is created.
VSS Pluggable
VSS enabled
disabled database
Same
Oracle VSS writer is engaged,
Oracle worflow
NOARCHIVELOG databases are shut N/A
11 as for
down and excluded from VSS processing
Linux
Same
Oracle VSS writer is engaged, Same workflow
Oracle worflow
NOARCHIVELOG databases are shut as for Linux, VSS
12 as for
down and excluded from VSS processing writer is skipped
Linux
252
Oracle Database
Ensure the account used to connect to target/staging server has enough permissions on
operating system and database as described in the corresponding section of User Guide or
earlier in this guide.
Note: When restoring to Linux ensure that account used to connect to restore target server
has valid shell.
Restore workflow
When performing restore Veeam Explorer follows the following steps:
253
Oracle Database
254
MySQL
MySQL/MariaDB
Backup and restore of MySQL and MariaDB is covered in this whitepaper:
https://www.veeam.com/wp-consistent-protection-mysql-mariadb.html.
255
IBM Notes/Domino
Domino
Backup and restore of IBM Lotus Domino is covered in this Veeam webinar:
https://www.veeam.com/videos/backing-up-non-vss-aware-applications-ibm-lotus-domino-
4867.html
256
SAP HANA
SAP HANA
Pre-freeze scripts can be used to create HANA snapshot before the backup starts. This
snapshot can be used as transactionally consistent state of database after restoring HANA
VM.
257
POC Guide
POC Guide
Organizations are modernizing their data centers in order to provision IT services faster,
strengthen security and control, and lower operational costs. While building modern data
centers, organizations invest in server virtualization, modern storage applications and cloud-
based services. However, businesses are facing new demands from end users including
access to data and applications 24/7, no patience for downtime or data loss, and exponential
data growth at 30-50% per year.
Veeam bridges this gap by providing customers a new kind of solution -- Availability for the
Modern Data Center, which delivers RTPO of < 15 minutes for all applications and data.
Organizations now can leverage their investments in the modern data center to meet new
demands of the always-on business.
This section of the document will demonstrate how Veeam solution can be used throughout
an entire datacenter availability project, beginning with the first assessment phase to the
project implementation from the technical perspective.
258
POC Guide
Note: While these guidelines focus on enterprise customers with more than 100 hosts or
1,000 virtual machines, Veeam Availability Suite is applicable to any infrastructure size.
259
Assessment
Assessment
Before starting a project, it is very important to understand customers’ needs, vision and the
IT environment. While the first two can be the outcome of an initial project meeting, the IT
environment can be analyzed with Veeam ONE, which is a part of the Veeam Availability
Suite.
The following information is very important and can help to streamline the project and
proactively prevent situations that impact the environment:
260
Assessment
This report will help you identify storage systems that are under heavy pressure or at its
maximum load. Let Veeam ONE run at least 24 hours and check if there are high latency
situations.
261
Assessment
This report will help you identify VMs with a high change rate at the block level (relevant for
incremental backups). You can later configure the backup or replication job to process them
at the beginning of the backup window, to address the longer job runtimes. In general, this
report will give you numbers for backup target storage planning.
VM Configuration Assessment
262
Assessment
This report will help you assess VMs readiness for performing backup with Veeam Backup &
Replication. It analyzes configuration of VMs in the virtual environment and shows potential
issues and possible limitations that can cause the backup process to fail or prevent VMs
from being properly backed up.
Infrastructure Overview
Active Snapshots
263
Assessment
VMware snapshots are often done to save a specific state of the VM for some time. While
they are created very easily, administrators forget to delete them over time. Together with
administrators, you can release all snapshots that are not needed anymore. This will help
prevent datastore downtimes because of snapshots filling up the whole physical storage.
Orphaned Snapshots
264
Assessment
This report detects VM snapshots that are still active on datastores but do not show up in the
VMware Snapshot Manager. Veeam Backup & Replication and its Snapshot Hunter will
correct this situation by consolidating these snapshots, which can bring extra load at the first
backup POC. We strongly recommend that you tune the VMware environment and
consolidate all orphaned snapshots before you start a Backup & Replication project.
In the later POC phase, create a separate account for a VMware user and use this account
for all authentication operations in Veeam Backup & Replication. With the Infrastructure
Changes by User report, you can track and document all changes done by this user.
Inventory
265
Assessment
This report provides the most complete and up-to-date configuration information on all
objects in the virtual environment. It can be used offline at the planning phase to address
any virtual infrastructure-related questions.
There are many additional interesting reports in the Veeam Availability Suite.
Check out the VMware Optimization or Hyper-V Optimization sections of Veeam ONE
Reporter. A good example is the Garbage Files Report that can identify possible wasted
space on datastores. In some cases, it helped to free up 10 TB+ of space on the tier 1
storage.
266
Accelerated Evaluation
Accelerated Evaluation
Many customers decide to do a small scale Proof of Concept (POC) after seeing their first
live demonstration and presentation meetings with partners or Veeam System Engineers.
The idea is to get started with the interface of Veeam Backup & Replication and to test if
everything works as expected/presented within the customer’s environment.
As enterprise environments are sometimes very complicated from the firewall and storage
perspective, in most cases customers decide to do a POC in small test environments.
Typically, a test environment includes:
It is possible to carry out a Veeam Backup & Replication POC in such environment with only
a single Veeam backup server on a VM with 8 cores and 8-16GB of RAM. (Since this test is
focused on the user interface experience, no special preparation is needed from the
performance perspective.)
Customers often drive this POC themselves. To assist customers with this task, Veeam has
published a good Evaluator's Guide that includes configuration screenshots with minimal
required background information.
One thing to remember when running a POC with Veeam is that you want to test something
with meaning, testing a backup because it backs up is important however having a goal is
also important.
Even for a small POC a plan is essential, write up can be as simple as:
How many machines, set a specific number and record their names.
What applications are you testing and why, what is the criteria for success on each
machine.
What types of recovery are you going to test and why (Veeam currently has 57 ways to
recover).
What are your expectations from the testing process.
What functionality do you want to see in action.
We all know Veeam will protect virtual machines, the aim of your POC should be to see how
well it lives up to your expectation at doing specific types of protection and recovery.
267
Accelerated Evaluation
268
Enhanced Evaluation
Enhanced Evaluation
Based on the information gathered during the assessment phase and customer
requirements, you may design a solution on paper and possibly implement it. Most likely
such designs are going to change over multiple revisions during the implementation phase
after communicating with other departments e.g. security, networking and storage teams. It
may also happen that the customer comes up with the new demands based on new findings.
This may delay in the implementation and ultimately lead to increased cost.
This chapter about The Enhanced Evaluation should help you avoiding such situations. We
will explain how the approach used by Veeam architects can help you simplify and
streamline the design phase and steer all project participants towards the same goals. This
will optimize the implementation phase and ultimately cut cost due to less time spent
revising the design and realigning stakeholders.
269
Workshop Example
Infrastructure Discovery
1. Start with the first main customer datacenter. Figure out the following:
i. Is this the same platform as the main datacenter, if not what is it?
ii. Are there any storage replication/mirroring involved?
iii. Is Active/Active cluster used?
For proper backup proxy implementation and backup mode selection, it is important to
know where the data that you want to back up is located, and whether you can access
all data from a single site.
270
Workshop Example
i. Start with the repository, discussing customer demands. In the example, customer
wanted to have the backup data in both datacenters. If so, you could decide to
implement repositories on both sides (half of the data on each side) and use the
backup copy job to move data to the second site.
ii. Discuss proxy implementation. The customer agreed to implement physical proxy
servers connected to their Fibre Channel network. As the customer used thick-
provisioned VMware VM disks, this ensured a fast and reliable backup and restore.
Check out the “Backup Proxy” section of this guide to determine the best proxy
implementation and select a transport mode for the environment.
iii. Plan for the backup server. In this example, it was placed on a VM and replicated to
the second datacenter. (The underlying datastore of the VM was not replicated to
the second site, only the VM.)
iv. Add other required components. The customer was already using two IBM TS3500
libraries for long-term retention with the existing backup software (agents). They
prepared a partition on each library with 4x LTO6 drives for use with Veeam. You
would proceed and connect them to the 2 physical servers (having the proxy and
repository roles assigned), and additionally assign the tape server role to these
servers.
6. Define OS/applications:
In this example, the customer used a third small datacenter with a single storage system
(Quorum) for the storage virtualization. During the discussion the customer identified 50 VMs
that were business-critical and needed full performance even at disaster recovery. Thus, in
the next step, you would add 2 ESXi hosts to that Quorum datacenter and replicate these 50
VMs every hour to that datacenter. The connection speed is to be 10 GbE. So, in case of
disaster recovery the customer could just boot up all VMs with full speed.
271
Workshop Example
Important! It is very important to use all available Veeam possibilities to implement the best
RTO and RPO times in customer’s environment.
For the VM recovery scenario, you can mix classic VM restore (best for small VMs), Instant
VM Recovery (best for huge data servers) and VM replica failover (best for database
systems with extreme I/O requirements). Together with the customer, check the “possible
failure areas” (single storage system/ whole datacenter/ 1 datastore) and decide if the
designed Veeam implementation fits into these needs and is in line with the budget.
Proxy/Repository Systems
Proxy and repository servers should be placed in the VMKernel networks. Veeam Backup &
Replication uses the VMKernel interfaces to read out configuration data and disk data (in
case of NBD), and to map Veeam vPower NFS datastores for fast recovery (Instant VM
Recovery).
272
Workshop Example
Veeam ONE
Veeam ONE components should be placed next to the vCenter Server and should be able to
read from the backup server and ESXi hosts (over the CIM protocol) as well. See Veeam
ONE documentation for more information: Veeam ONE Deployment Guide.
Enterprise Manager
When Veeam Backup Enterprise Manager is used with Self-Restore Services, it should be
placed in the internal DMZ in most cases.
Restore Points
273
Workshop Example
In the sample case, the customer needed daily backup with 14 restore points; the points
were to be held on 2 sites (copied with backup copy job). The customer also wanted to
offload the weekly full backups on tape and hold them for a period slightly longer than one
year in both tape libraries.
The customer also needed to replicate the most critical VMs to the Quorum datacenter
hourly, between 7:00 and 19:00. The number of replication restore points to be maintained
was the maximum possible (here 28 restore points).
In many architecture meetings, planning for the retention policies is the most time-
consuming part as you are likely to engage different administrators and management team
members from different departments in this process. These team members have to translate
their file-based existing restore point policies into a new way (image-level backup). It is a
matter of concern because a longer retention chain will result in expensive storage space
costs.
Important! Remember to agree on backing up Microsoft SQL Server transaction logs with
Veeam Backup & Replication.
If speaking about the storage sizing, the tool at Veeam Restore Points Calculator can help to
illustrate the retention chains on disk and estimate the required capacity.
274
Preparation
Typically, the enhanced POC is carried out under the following conditions:
The environment is close to the production environment, with all firewalls in place.
Involved storage systems are similar to the production storage systems.
Veeam storage integration is used whenever possible.
To demonstrate the good working load balancing and scalability, 100-200 VMs are
backed up/replicated.
All major applications are backed up to test all restore scenarios.
Preparation steps
1. Prepare for the POC planning with the Veeam User Guide and this document.
2. Complete a POC document including all your designs and plans, including chosen
servers for the tests and why they are important. Set a criteria for success on each
machine and what is tested.
3. Check out the necessary firewall ports and help the customer with the internal firewall
change requests. Refer to the corresponding sections in the User Guide and this
document.
Tip: Perform firewall planning very carefully: if something is misconfigured, this may
block the entire POC. In most cases, it is not easy to detect problems and introduce
firewall changes, when the POC is already running. However, it is a good idea to ask
the customer to have the firewall administrator at hand in case you need an urgent
change.
4. Create a separate vCenter Server account for Veeam ONE (read-only + datastore
browsing + CIM) so that you are able to track what users do.
5. If you want to use the storage integration feature, check out the corresponding chapter
in this guide, set up the storage and the SAN network together with the storage
administrators. Limit the scope of storage system rescan to the volumes used in the
POC.
6. If you want to use SureBackup, make sure that a virtualized Domain Controller is
present if needed (e.g. for Microsoft Exchange).
275
Preparation
7. Let the customer prepare all used antivirus systems upfront so that you do not run into
trouble. Check the "Antivirus" section of this guide and Veeam KB1999.
8. Ask the customer to prepare a decent performing storage system for the POC. Avoid
low-end NAS appliances for enhanced evaluations.
9. Let the customer prepare all operating systems and database installations. Set up
Veeam Backup & Replication and backup infrastructure components together with the
customer and place the folders correctly.
10. Ensure that the document relating to all the testing is accurate and up to date including
all success criteria for each machine being tested. This will keep control for POC, each
test and a schedule can be built around the testing avoiding random testing of features.
276
Automation
Automation
The bigger the environment, the more automation is needed to reduce the administration
effort. For example, if you are operating 40 branch offices with independent Veeam
installations, you may want to roll out and configure backup servers with scripts, and
automatically create jobs in the same location. Another example is automatic job creation for
2,000-3,000 VMs with exactly the same configurations, which can limit user-caused
misconfiguration.
Command line
Following operations are managed through the Windows command line:
PowerShell
Operations in Veeam Backup & Replication can be automated with Veeam PowerShell snap-
in in the following areas:
Configuration
Job creation/job editing
Working with external schedulers (UC4/TWS and other) to start Veeam jobs
Restores
Reporting
Datacenter migration (quick migration or replication)
The PowerShell plugin is available with all commercial versions of the product.
Note: PowerShell plugin is also available with Veeam Backup FREE, although limited:
http://www.veeam.com/blog/veeam-backup-free-edition-now-with-powershell.html
Our customers and partners use this functionality to scale out backup infrastructure
environments to nearly 100,000 VMs under a single Veeam Backup Enterprise Manager
instance with multiple backup servers located in different datacenters.
The best starting point to get in touch with the Veeam PowerShell plugin is to read the
Veeam PowerShell User Guide > Veeam Help Center - PowerShell Reference.
277
Automation
You can find help for the scripts in the Veeam Community Forums - PowerShell section. If
you need some examples, refer to the following thread: Getting Started and Code Examples
RESTful API
In the Veeam Enterprise Manager, there is as well RESTful API that allows you to create
workflows in orchestration tools or to integrate Veeam Backup Enterprise Manager (self-
services) in your own “cloud” portal. Specifically, this is an option that comes with Enterprise
Plus Editions and is focused on the hosting business.
278
Infrastructure Hardening
Infrastructure Hardening
Running your Veeam Backup & Replication infrastructure in a secure configuration is a
daunting task even for security professionals. This chapter provides practical advice to help
administrators to harden their infrastructure following security best practices so that they can
confidently deploy their Veeam services and lower their chances of being compromised.
Hardening is about securing the infrastructure against attacks, by reducing its attack surface
and thus eliminating as many risks as possible. One of the main measures in hardening is
removing all non-essential software programs and utilities from the deployed Veeam
components. While these components may offer useful features to the administrator, if they
provide ‘back-door’ access to the system, they must be removed during the hardening
process.
But also, creating visibility in what goes on in the infrastructure is part of hardening your
infrastructure. Making sure you will notice when an attack is/or has taken place and then
making sure logs and traces are saved for law-enforcement and security specialists when
needed.
Protect
Protecting your infrastructure successfully is all about understanding what and whom you
are protecting, your Veeam infrastructure, against. If you know what and whom you are
protecting against, makes it easier to take the correct countermeasures. One of those
countermeasures is hardening.
Looking at the different Veeam Backup & Replication components you have to protect the
following components:
Consider the Veeam Backup & Replication server to be the Number 1 target on your
infrastructure and it should have very restricted access. As a general rule the backup server
is the single greatest target a hacker can claim on your network. Also the backup
repositories which holds the backup files are a primary target.
279
Infrastructure Hardening
Hardening
Within the hardening process of your Veeam infrastructure there are a few steps everyone
should always consider and act upon, namely:
1. Secure by Design
3. Console Access
5. Required Permissions
6. Encryption
8. Segmentation
9. Visibility
Secure by Design
Overly complex designs become harder for the IT team to manage and overlook and it
makes it easier for an attacker to exploit and stay in the shadows. Simpler designs that can
be easily overviewed are in basis more secure. Adding security to an already existing
infrastructure is much harder and costly than thinking about it while designing a new or
refreshing an existing infrastructure. In a virtual infrastructure, it is good use to build up a
Master image which has been hardened from the start. Removing all known attack vectors
and only open up access when Veeam components are added and needs specific (port)
openings or extra software to function properly. This way all builds are consistent and kept
up-to-date which makes it secure in the basis.
Consider the Veeam Backup & Replication server to be the Number 1 target on your
infrastructure and it should have very restricted access. As a general rule the backup server
is the single greatest target a hacker can claim on your network.
280
Infrastructure Hardening
Remove all non-essential software programs and utilities from the deployed Veeam
components. While these programs may offer useful features to the administrator, if they
provide ‘back-door’ access to the system, they must be removed during the hardening
process. Think about additional software like web browsers, java, adobe reader and such.
All parts which do not belong to the operating system or to active Veeam components,
remove it. It will make maintaining an up-to-date patch level much easier.
Remove the Backup & Replication Console from the Veeam Backup & Replication
server. The console is installed locally on the backup server by default.
Switch off the Veeam vPower NFS Service if you do not plan on using the following
Veeam features: SureBackup, Instant Recovery, or Other-OS File Level Recovery (FLR)
operations.
For uninstalling Veeam Backup & Replication Console, first de-install all Veeam Explorers:
Example for uninstalling the Veeam Backup & Replication console is: msiexec /x {D0BCF408-
A05D-45AA-A982-5ACC74ADFD8A}
Enterprise Manager
When Enterprise Manager is not in use de-install it and remove it from your environment.
Console Access
The Veeam Backup & Replication console is a client-side component that provides access to
the backup server. The console lets several backup operators and admins log in to Veeam
Backup & Replication simultaneous and perform all kind of data protection and disaster
281
Infrastructure Hardening
Install the Veeam Backup & Replication console on a central management server that is,
positioned in a DMZ and protected with 2-factor authentication. Do NOT install the console
on the local desktops of backup & recovery admins.
Containment to keep the attackers from moving around too easily. Some standard measures
and policies are:
Do not use user accounts for admin access, reducing incidents and accidents
Give every Veeam admin his own admin account or add their admin account to the
appropriate security group within Veeam, for traceability and easy adding and removal
Only give out access to what is needed for the job
Limit users who can log in using Remote Desktop and/or Veeam Backup Console
Add 2-factor authentication to highly valuable assets
Monitor your accounts for suspicious activity
A role assigned to the user defines the user activity scope: what operations in Veeam
Backup & Replication the user can perform. Role security settings affect the following
operations
For Admin accounts adding 2-factor authentication is also a must to secure the
infrastructure.
282
Infrastructure Hardening
And for service accounts use 25+ characters combined with a password tool for easier
management. An Admin can copy and paste the password when needed, increasing
security of the service accounts.
Lockout policy
Use a Lockout policy that complements a clever password management policy. Accounts will
be locked after a small number of incorrect attempts. This can stop password guessing
attacks dead in the water. But be careful that this can also lock everyone out of the backup &
replication system for a period! For service accounts, sometimes it is better just to raise
alarms fast. Instead of locking the accounts. This way you gain visibility into suspicious
behavior towards your data/infrastructure.
Required Permissions
Use the principle of least privilege. Provide the minimal required permissions needed for
the accounts to run. The accounts used for installing and using Veeam Backup & Replication
must have the following permissions.
If VMware vCenter Server is added to the backup infrastructure, an account that has
administrator permissions is required. Instead of granting administrator permissions to the
account, you can configure more granular permissions. Veeam has identified the minimum
permissions required for the various software functions. Review the "Required Permissions"
document (not changed since V9.0) and configure the accounts used by Veeam Backup &
Replication to meet these requirements.
Particularly, backup proxies must be considered the target for compromise. During backup,
proxies obtain from the backup server credentials required to access virtual infrastructure
servers. A person having administrator privileges on a backup proxy can intercept the
credentials and use them to access the virtual infrastructure.
283
Infrastructure Hardening
Ensure timely guest OS updates on backup infrastructure servers. Install the latest
updates and patches on backup infrastructure servers to minimize the risk of exploiting
guest OS vulnerabilities by attackers.
Choose strong encryption algorithms for SSH. To communicate with Linux servers
deployed as part of the backup infrastructure, Veeam Backup & Replication uses SSH.
Make sure that for the SSH tunnel you use a strong and proven encryption algorithm,
with sufficient key length. Ensure that private keys are kept in a highly secure place, and
cannot be uncovered by a 3rd party.
Encryption
Backup and replica data is a highly potential source of vulnerability. To secure data stored in
backups and replicas, follow these guidelines:
Ensure physical security of target servers. Check that only authorized personnel
have access to the room where your target servers (backup repositories and hosts)
reside.
Restrict user access to backups and replicas. Check that only authorized users have
permissions to access backups and replicas on target servers.
Encrypt data in backups. Use Veeam Backup & Replication inbuilt encryption to
protect data in backups. To guarantee security of data in backups, follow Encryption
Best Practices.
Backup and replica data can be intercepted in-transit, when it is communicated from source
to target over a network. To secure the communication channel for backup traffic, consider
these guidelines:
Isolate backup traffic. Use an isolated network to transport data between backup
infrastructure components — backup server, backup proxies, repositories and so on.
(also see segmentation)
Encrypt network traffic. By default, Veeam Backup & Replication encrypts network
traffic traveling between public networks. To ensure secure communication of sensitive
data within the boundaries of the same network, you can also encrypt backup traffic in
private networks. For details, see Enabling Network Data Encryption.
284
Infrastructure Hardening
The Backup & Replication configuration database stores credentials to connect to virtual
servers and other systems in the backup & replication infrastructure. All passwords stored in
the database are encrypted. However, a user with administrator privileges on the backup
server can decrypt the passwords, which presents a potential threat.
To secure the Backup & Replication configuration database, follow these guidelines:
Restrict user access to the database. Check that only authorized users can access
the backup server and the server that hosts the Veeam Backup & Replication
configuration database (if the database runs on a remote server).
Encrypt data in configuration backups. Enable data encryption for configuration
backup to secure sensitive data stored in the configuration database. For details, see
Creating Encrypted Configuration Backups.
Segmentation
Add local protection mechanics, in addition to the border firewalls, intrusion detection,
patching and such. You can make use of local mechanisms, like up-to-date anti-malware,
firewalls and network segmentation. This way you create different rings-of-defense slowing
an attacker down.
A good practice is to place the backup repositories in a special segment not accessible by
any user. Like for instance the production storage is only available to the virtual
infrastructure components and application servers. Not directly accessible by any user!
To segment your infrastructure and Veeam Backup & Replication components, make sure
the firewalls on the local server installations have the correct Ports opened.
You can also deploy VMware NSX as a counter measure with micro-segmentation to make
sure the attack surface is as narrow as possible without blocking everyone to use the
services. Visibility into the network and all data flows is crucial to help you protect all
different rings/cells within your infrastructure. You can add the Veeam components to NSX
policies to make sure they can communicate with each other without opening it up to any
user.
Ports
Try not to use obscure ports and other tricks to try and hide Veeam ports and protocols in
use, while this may look like a good choice. In practice this often makes the infrastructure
harder to manage which opens other possibilities for attackers. Obscurity is not security!
You can check which ports are in use by which service on a Windows system by using:
285
Infrastructure Hardening
netstat -bona > portlist.txt you can open the text file with for instance notepad
portlist.txt
Visibility
To know when you are under attack or have been breached it is vital to have visibility in the
whole data flow path. You should be able to know what is ‘normal behavior’ and what is
NOT. Monitor your accounts and Veeam infrastructure for suspicious activity. Place virtual
trip-wires, like e.g. creating a non-used admin account with alarms tied to it. When any
activity on that account is observed, it will trigger a red alert instantly. There are several
systems out there that can help you by alerting suspicious behavior so you get aware that
someone is snooping around and is trying to gain access to your infrastructure. Visibility is
Key!
It is important to get alerts as soon as possible while defending against other attacks like
viruses, malware and ransomware. The biggest fear of these attacks is that they may
propagate to other systems fast. Having visibility into for e.g. potential ransomware activity is
a big deal.
A system that detects possible ransomware activity is Veeam ONE 9.5. There is a pre-
defined alarm called “Possible ransomware activity.” This alarm will trigger if there is a
high CPU utilization combined with lots of writes to disk.
VMware vRealize Network Insight can take VMs, objects, groupings and their physical
elements and easily fingerprint the application and determine the internal and external
flows, the client connections, etc. this way you get an analysis of what is ‘normal’
behavior and what is not.
Recovery Strategy
Have a recovery strategy in place, before you find out your infrastructure is breached you
should know what to do when being compromised through attacks. Backup your data and
make sure the backups cannot be accessed by an attacker to wipe them out. An offsite copy
(air-gap) or read-only on any media is highly recommended to survive any attack.
286
Infrastructure Hardening
The 3-2-1 rule is very general and it works for all data types (individual and corporate) and
all environment types (physical and virtual). When backing up VMware or Hyper-V
environments with Veeam, this rule becomes the “3-2-1-0 backup rule” where 0 means “0
errors” during the automatic recoverability verification of every backup with Veeam’s
SureBackup.
Veeam Backup & Replication™ can help you to fulfill all 3-2-1-0 backup rule requirements.
Have at least three copies of data: Setup Backup Jobs to create several backups for
each of your VMware or Hyper-V VMs.
Keep one backup copy offsite: Setup Backup Copy Jobs to transfer your backup
offsite faster with built-in WAN acceleration, or use Veeam Backup Cloud Edition to
store your backups to one of 15 public clouds, including Windows Azure, Amazon
Glacier, Google Cloud Storage and more.
287
Backup & Replication Anatomy
288
Backup
Backup
This section provides a step-by-step description of a VMware virtual machine backup
process implemented in Veeam Backup & Replication.
1. Initialization Phase
A backup job can be started automatically or manually in the Veeam Backup & Replication
console, Veeam Backup Enterprise Manager web console, by means of PowerShell,
RESTful API and other.
In the initialization phase, Veeam Backup & Replication prepares resources necessary for a
backup job. To help you better understand firewall settings and connection initiation flow, the
process is illustrated by the diagram (see below):
1. When a backup job is initialized, the Veeam Backup Manager process is started on the
Veeam backup server.
2. Veeam Backup Manager reads job settings from the Veeam Backup configuration
database and creates a list of VM tasks to process (one task stands for one VM disk).
3. Veeam Backup Manager connects to the Veeam Backup Service. The Veeam Backup
Service includes a resource scheduling component for managing all tasks and
resources in the backup infrastructure. The resource scheduler checks what resources
are available, and assigns backup proxies and repositories to process that job tasks
using Veeam's load balancing.
4. After the necessary backup infrastructure resources have been assigned, Veeam
Backup Manager connects to the Transport Services on the target repository and on the
backup proxy. The Transport Services, in their turn, start the Veeam Data Movers. On
the backup proxy, a new Veeam Data Mover is started for each task that the proxy is
processing.
5. Veeam Backup Manager establishes a connection with Veeam Data Movers on the
backup repository and backup proxy, and sets a number of rules for data transfer (such
as network traffic throttling rules, and so on).
6. Veeam Data Movers on the backup proxy and repository establish a connection with
each other for data transfer.
289
Backup
7. Veeam Backup Manager connects to the vCenter Server or ESXi host and gathers
metadata about VMs and hosts engaged in the backup process. At this step, no
connection between the Veeam backup server and VM guest networks is established.
290
Backup
291
Backup
If pre-freeze and post-thaw scripts are enabled in the backup job properties, Veeam Backup
& Replication obtains information about the guest’s IP address from VMware Tools. Veeam
uses this IP address to connect to the guest network over SSH and perform in-guest
processing tasks. Scripts reside on the backup server and are injected in the guest OS at
the time of backup.
If there is no network connectivity with a Linux-based VM, Veeam Backup & Replication will
not fail over to the VIX communication channel. In such cases, as an alternative method, you
can use VMware Tools quiescence and let VMware Tools run the necessary scripts that will
need to be created inside the guest OS (see location details for Windows / Linux guest at:
https://pubs.vmware.com/vsphere-50/topic/com.vmware.datarecovery.admin.doc_20/GUID-
6F339449-8A9F-48C0-BE70-91A2654A79D2.html.
3. Creating a VM Snapshot
Now, Veeam Backup & Replication requests the vCenter Server or ESXi host to initiate a VM
snapshot creation. A VM snapshot is required to use VMware VADP backup methods and
leverage features like VMware Changed Block Tracking (CBT).
292
Backup
Right after the VM snapshot is taken, all quiesced disk I/O activities in the guest OS are
resumed.
5. VM Data Transport
To read and transfer data from the VM snapshot, Veeam Backup & Replication can use one
of the following transport modes:
Network (NBD)
For more information about each transport mode, see Veeam Backup & Replication User
Guide or a corresponding section below.
293
Backup
In the Virtual Appliance transport mode, Veeam backup proxy connects to the ESXi host
where the VM resides, and reads the necessary VM configuration files (such as *.vmx). VM
disks as of the snapshot state are hot-added to a virtualized Veeam backup proxy. The proxy
reads VM data and unmaps the VM disks when finished.
294
Backup
6. Committing VM Snapshot
After Veeam backup proxy finishes reading VM data, Veeam backup server requests the
vCenter Server or ESXi host to initiate a VM snapshot commit.
295
VM Restore
VM Restore
This section provides a step-by-step description of a full virtual machine restore process
implemented in Veeam Backup & Replication.
1. Initialization Phase
In the initialization phase, Veeam Backup & Replication prepares the resources necessary
for full VM recovery. It performs the following steps:
2. Checks available backup infrastructure resources and assigns a proxy server for
transferring restored VM data to the target host/datastore.
3. Communicates with Transport Services on the backup proxy and backup repository
where the backup files reside.\ Transport Services, in their turn, start Veeam Data
Movers. Veeam Data Movers on the backup proxy and repository establish a connection
with each other for data transfer.
4. Connects to the vCenter Server or ESXi host where the restored VM will be registered.
2. Restoring VM Configuration
296
VM Restore
Veeam Backup & Replication retrieves VM configuration data from the backup and restores
it on the chosen ESXi host/datastore. Next, it instructs VMware vSphere to register the
restored VM on the host. If a user selects to change VM configuration (for example, disk
format or network settings) during restore, Veeam makes the necessary amendments.
3. Creating VM Snapshot
Veeam Backup & Replication requests the vCenter Server or ESXi host to initiate a VM
snapshot creation on the restored VM.
297
VM Restore
4. VM Data Transport
Veeam Backup Manager instructs VMware vSphere to create virtual disks for the VM.
To write VM disk data to the target datastore, Veeam Backup & Replication can use one of
the 3 transport modes:
Network (NBD)
For more information about each transport mode, see Veeam Backup & Replication User
Guide and the corresponding sections of this document.
In the Direct SAN Access mode, Veeam Backup & Replication connects to the ESXi host
where the restored VM is registered. The ESXi host locates the VM disks, retrieves
metadata about the disk layout on the storage, and sends this metadata to the backup proxy.
The backup proxy uses this metadata to copy VM data blocks to the datastore via SAN.
298
VM Restore
299
VM Restore
5. Committing VM Snapshot
After the proxy finishes writing VM disk data, Veeam Backup & Replication requests the
vCenter Server or ESXi host to initiate a snapshot commit for the restored VM.
300
VM Restore
301
Instant VM Recovery
Instant VM Recovery
This section provides a step-by-step description of the Instant VM Recovery process
implemented in Veeam Backup & Replication.
1. Initialization Phase
In the initialization phase, Veeam Backup & Replication prepares resources necessary for
Instant VM Recovery. It performs the following steps:
1. Starts the Veeam Backup Manager process on the Veeam backup server.
2. Checks with the Veeam Backup Service whether the necessary backup infrastructure
resources are available for instant VM Recovery.
3. Communicates with the Transport Service on the backup repository to start Veeam Data
Mover.
2. NFS Mapping
When backup infrastructure resources are prepared, Veeam Backup & Replication maps an
empty NFS datastore to the selected ESXi host. It uses the Veeam vPower NFS Service for
this purpose.
302
Instant VM Recovery
Next, Veeam Backup & Replication creates in the Veeam NFS datastore VM configuration
files and links to virtual disk files. Virtual disk files remain in the backup on the repository,
while all changes to these files are written to the cache file.
To migrate VM disk data to a production datastore, use VMware Storage vMotion or Veeam
Quick Migration. For details, see Veeam Backup & Replication User Guide.
303
Instant VM Recovery
304
Windows File-Level Restore
1. Initialization Phase
In the initialization phase, Veeam Backup & Replication prepares resources necessary for
Microsoft Windows file-level restore. It performs the following steps:
1. Checks with the Veeam Backup Service whether the necessary backup infrastructure
resources are available for Microsoft Windows file-level restore.
2. Starts Veeam Data Movers on the Veeam backup server and backup repository.
3. Mounts the content of backup files to the backup server with the help of Veeam’s
proprietary driver.
The backup files remain on the backup repository. Guest files inside the backup can be
accessed in Veeam Backup browser or Microsoft Windows File explorer on the backup
server, mapped by default in the C:\VeeamFLR folder (can be changed via registry key).
305
Windows File-Level Restore
306
Windows File-Level Restore
After all restore activities are completed and the user closes the Veeam Backup browser (or
the browser is closed by timeout), the content of the backup files is dismounted from the
backup server.
307
Replication
Replication
This section provides a step-by-step description of a VMware virtual machine replication
process implemented in Veeam Backup & Replication.
In many aspects, the replication initialization phase is similar to the initialization phase of the
backup process. Veeam Backup & Replication starts the necessary processes, builds the list
of VMs to replicate, assigns backup infrastructure resources for the job and starts Veeam
Data Movers on two backup proxies (source and target) and the backup repository that is
used for storing replica metadata.
Next, Veeam Backup & Replication performs in-guest processing tasks, triggers VM
snapshot creation, registers a replica VM on the target host and performs data transfer from
the source host and datastore to the target host and datastore. The source and target
proxies can use one of 3 available data transport modes for reading data from source and
writing data to target.
This diagram illustrates the replication process with the NBD transport mode used for
reading and writing VM data. For examples of the Direct SAN/NFS Access and HotAdd
transport modes, see the “Backup Anatomy” section above in this Appendix.
The following diagram illustrates a possible placement of the Veeam Backup & Replication
components in a distributed environment, with a WAN link between the production and DR
sites.
308
Replication
309
Sizing Summary
Sizing with Veeam is cumulative in respect to configurations, if you want to create an all-in-
one appliance (Appliance Model) add all the resource requirements together (CPU +
Memory) to understand what in total you will need, the same goes if you only wish to have
proxy and repository in one host.
Please also bear in mind that these figures reflect Veeam’s resource requirement, you must
take the hosts system requirements into your calculation, this will depend on what you are
using which is why we have not detailed it here.
Minimum Resources
310
Sizing Summary
Storage space for Guest indexing before Enterprise manager flush: 2MB per 1 million
files (compressed)
This is based on a rounded figure offering approximately 30 VMs in a single backup job
which will finish around an 8 hours backup window if in a per job backup, if a per VM
repository is used more can be added. Please read the sizing and repository section for a
full detailed description of parallelization of workloads in a Proxy.
When sizing a repository server remember, the ability to execute a task on the repository will
be affected by the proxy’s ability to process all the tasks from the proxy’s. If a repository has
20 cores, then the maximum processed tasks will be no more than 20 tasks from any proxy
or group of proxies in the backend fabric of Veeam to that repository.
311
Sizing Summary
Veeam Backup & Replication may consume high amounts of CPU and RAM while
processing backup or replication jobs. To achieve better performance and load balancing it is
necessary to provide sufficient RAM and CPU resources. If possible, follow these guidelines:
Note: Concurrently running jobs include any job type with a continuous schedule such as
Backup Copy Jobs. When running more than 100 jobs concurrently increase compute
resources in line with the table above to meet the resource need of the workload. Veeam
installation package includes SQL server 2012 Express Edition, the basic limitations of this
software are as follows:
If any of the below apply consider using SQL standard or Enterprise editions
312
Networking Diagrams
Networking Diagrams
There is a detailed list of ports used by Veeam Backup & Replication available in the User
Guide, but sometimes a more visual approach is helpful – you can use the diagrams below
for that purpose.
313
Backup Server
Backup Server
314
Proxy Server
Proxy Server
The following ports are required for the proxy server.
315
Repository Server
Repository Server
The following ports are required for the repository server.
316
Repository Server
317
Storage Integration
Storage Integrations
The following ports are required for integrated storage.
318
Data Validation
Data Validation
The following ports are required, when using SureBackup, SureReplica, or On-demand
Sandbox from Storage Snapshots.
319
Application-aware Image Processing
320
Enterprise Manager
Enterprise Manager
The following ports are required for Enterprise Manager
321