NetBackup™ Backup Planning and Performance Tuning Guide
- NetBackup capacity planning
- Primary server configuration guidelines
- Size guidance for the NetBackup primary server and domain
- Factors that limit job scheduling
- More than one backup job per second
- Stagger the submission of jobs for better load distribution
- NetBackup job delays
- Selection of storage units: performance considerations
- About file system capacity and NetBackup performance
- About the primary server NetBackup catalog
- Guidelines for managing the primary server NetBackup catalog
- Adjusting the batch size for sending metadata to the NetBackup catalog
- Methods for managing the catalog size
- Performance guidelines for NetBackup policies
- Legacy error log fields
- Media server configuration guidelines
- NetBackup hardware design and tuning considerations
- About NetBackup Media Server Deduplication (MSDP)
- Data segmentation
- Fingerprint lookup for deduplication
- Predictive and sampling cache scheme
- Data store
- Space reclamation
- System resource usage and tuning considerations
- Memory considerations
- I/O considerations
- Network considerations
- CPU considerations
- OS tuning considerations
- MSDP tuning considerations
- MSDP sizing considerations
- Cloud tier sizing and performance
- Accelerator performance considerations
- Media configuration guidelines
- About dedicated versus shared backup environments
- Suggestions for NetBackup media pools
- Disk versus tape: performance considerations
- NetBackup media not available
- About the threshold for media errors
- Adjusting the media_error_threshold
- About tape I/O error handling
- About NetBackup media manager tape drive selection
- How to identify performance bottlenecks
- Best practices
- Best practices: NetBackup SAN Client
- Best practices: NetBackup AdvancedDisk
- Best practices: Disk pool configuration - setting concurrent jobs and maximum I/O streams
- Best practices: About disk staging and NetBackup performance
- Best practices: Supported tape drive technologies for NetBackup
- Best practices: NetBackup tape drive cleaning
- Best practices: NetBackup data recovery methods
- Best practices: Suggestions for disaster recovery planning
- Best practices: NetBackup naming conventions
- Best practices: NetBackup duplication
- Best practices: NetBackup deduplication
- Best practices: Universal shares
- NetBackup for VMware sizing and best practices
- Best practices: Storage lifecycle policies (SLPs)
- Best practices: NetBackup NAS-Data-Protection (D-NAS)
- Best practices: NetBackup for Nutanix AHV
- Best practices: NetBackup Sybase database
- Best practices: Avoiding media server resource bottlenecks with Oracle VLDB backups
- Best practices: Avoiding media server resource bottlenecks with MSDPLB+ prefix policy
- Best practices: Cloud deployment considerations
- Measuring Performance
- Measuring NetBackup performance: overview
- How to control system variables for consistent testing conditions
- Running a performance test without interference from other jobs
- About evaluating NetBackup performance
- Evaluating NetBackup performance through the Activity Monitor
- Evaluating NetBackup performance through the All Log Entries report
- Table of NetBackup All Log Entries report
- Evaluating system components
- About measuring performance independent of tape or disk output
- Measuring performance with bpbkar
- Bypassing disk performance with the SKIP_DISK_WRITES touch file
- Measuring performance with the GEN_DATA directive (Linux/UNIX)
- Monitoring Linux/UNIX CPU load
- Monitoring Linux/UNIX memory use
- Monitoring Linux/UNIX disk load
- Monitoring Linux/UNIX network traffic
- Monitoring Linux/Unix system resource usage with dstat
- About the Windows Performance Monitor
- Monitoring Windows CPU load
- Monitoring Windows memory use
- Monitoring Windows disk load
- Increasing disk performance
- Tuning the NetBackup data transfer path
- About the NetBackup data transfer path
- About tuning the data transfer path
- Tuning suggestions for the NetBackup data transfer path
- NetBackup client performance in the data transfer path
- NetBackup network performance in the data transfer path
- NetBackup server performance in the data transfer path
- About shared memory (number and size of data buffers)
- Default number of shared data buffers
- Default size of shared data buffers
- Amount of shared memory required by NetBackup
- How to change the number of shared data buffers
- Notes on number data buffers files
- How to change the size of shared data buffers
- Notes on size data buffer files
- Size values for shared data buffers
- Note on shared memory and NetBackup for NDMP
- Recommended shared memory settings
- Recommended number of data buffers for SAN Client and FT media server
- Testing changes made to shared memory
- About NetBackup wait and delay counters
- Changing parent and child delay values for NetBackup
- About the communication between NetBackup client and media server
- Processes used in NetBackup client-server communication
- Roles of processes during backup and restore
- Finding wait and delay counter values
- Note on log file creation
- About tunable parameters reported in the bptm log
- Example of using wait and delay counter values
- Issues uncovered by wait and delay counter values
- Estimating the effect of multiple copies on backup performance
- Effect of fragment size on NetBackup restores
- Other NetBackup restore performance issues
- About shared memory (number and size of data buffers)
- NetBackup storage device performance in the data transfer path
- Tuning other NetBackup components
- When to use multiplexing and multiple data streams
- Effects of multiplexing and multistreaming on backup and restore
- How to improve NetBackup resource allocation
- Encryption and NetBackup performance
- Compression and NetBackup performance
- How to enable NetBackup compression
- Effect of encryption plus compression on NetBackup performance
- Information on NetBackup Java performance improvements
- Information on NetBackup Vault
- Fast recovery with Bare Metal Restore
- How to improve performance when backing up many small files
- How to improve FlashBackup performance
- Veritas NetBackup OpsCenter
- Tuning disk I/O performance
Tuning universal shares
Two key methods for improving performance of universal shares include tuning the number of vpfsd instances and setting the ingest mode.
A universal share uses one vpfsd instance by default. In most cases, one instance is adequate. Increasing the number of vpfsd instances might improve universal share performance, although it also requires more CPU and memory. You can increase the number of vpfsd instances from 1 to up to 16 and distribute the shares cross all the vpfsd instances.
In extensive testing with a maximum recommended 50 concurrently active shares, setting the number of vpfsd instances to 2 resulted in favorable results without overconsuming CPU and memory resources. Increasing the number of vpfsd instances to more than 2 should be done carefully and in conjunction with extensive benchmark testing.
There is no requirement to balance the number of universal shares per vpfsd instance because it is done automatically.
To make this change, set the numOfInstance value in the vpfsd_config.json file on the MSDP server.
Ingest mode was natively introduced in NetBackup 10.0 to give DBAs a way to increase performance of dumps to the universal share. When ingest mode of a universal share is enabled, the share will persist all the data from memory to disk on the client side at the end of the dump. The ingest mode is faster than normal mode as it does not guarantee all the ingested data is persisted to disk until the ingest mode is turn off. Therefore, turning ingest mode off is critical for data dump integrity.
You can enable and disable the ingest mode for a specific share on the NFS/SMB client side.
To leverage ingest mode, the .vpfs_special_control_config file must be updated to turn it on at the beginning of the dump, and then to turn it off at the end of the dump. Ideally, this action would be included in a script that the DBA calls to execute a database dump to the target universal share.
For more information, consult the NetBackup Deduplication Guide.
Improving performance with emergency engineering binaries (EEBs)
When dumping MSSQL data to universal shares, performance can be improved in some deployments by applying an EEB. The following table lists the relevant EEBs that are needed for each release of NetBackup.
NetBackup Version
Required EEB
9.1.0.1
4047040 v24
10.0
4070421 v8
10.0.0.1
4078688 v6
10.1
4091734 v6
10.1.1
4102406 v1
Universal share as a replacement for Oracle incremental forever backup (Co-pilot).
Beginning in NetBackup 10.1.1, universal share is becoming the replacement for Oracle Co-pilot feature. Internal tests have shown performance can be better with the initial dump, however, the merge part is still bottlenecked with the small I/O updates from Oracle RMAN. This blog describes the Oracle incremental merge feature really well:
https://blogs.oracle.com/maa/post/using-oracle-incremental-merge
The merge elapsed time is proportionally increased with the amount of changed data. Internal tests showed that for database change rate greater than 5%, the incremental merge performance can be slower than the full backup. For this reason, we don't recommend the Co-pilot feature if the database change rate is substantially greater than 5%. However, the merge operation has very low overhead on the Oracle client. In other words, normal operation can resume while the merge is in progress with very little impact to Oracle server performance. If this is acceptable, then a universal share can still be a good alternative to stream-based backup even if the change rate is larger.
Another advantage of the incremental forever backup feature is faster restore, as there is always a latest full image of the database available for access, while with stream-based backups, merging of multiple incremental backup images may be necessary to form the final backup image. Additionally, the Oracle incremental forever backups enables instant access capability that allows you to access the database much faster for certain database operations.