NetBackup™ for Hadoop Administrator's Guide
- Introduction
- Prerequisites and best practices for the Hadoop plug-in for NetBackup
- Configuring NetBackup for Hadoop
- About configuring NetBackup for Hadoop
- Managing backup hosts
- Adding Hadoop credentials in NetBackup
- Configuring the Hadoop plug-in using the Hadoop configuration file
- Configuring NetBackup for a highly-available Hadoop cluster
- Configuring a custom port for the Hadoop cluster
- Configuring number of threads for backup hosts
- Configuring number of streams for backup hosts
- Configuring distribution algorithm and golden ratio for backup hosts
- Configuring communication between NetBackup and Hadoop clusters that are SSL-enabled (HTTPS)
- Configuration for a Hadoop cluster that uses Kerberos
- Hadoop.conf configuration for parallel restore
- Create a BigData policy for Hadoop clusters
- Disaster recovery of a Hadoop cluster
- Performing backups and restores of Hadoop
- Troubleshooting
- About troubleshooting NetBackup for Hadoop issues
- About NetBackup for Hadoop debug logging
- Troubleshooting backup issues for Hadoop data
- Backup operation fails with error 6609
- Backup operation failed with error 6618
- Backup operation fails with error 6647
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Backup operation fails with error 6654
- Backup operation fails with bpbrm error 8857
- Backup operation fails with error 6617
- Backup operation fails with error 6616
- Backup operation fails with error 84
- NetBackup configuration and certificate files do not persist after the container-based NetBackup appliance restarts
- Unable to see incremental backup images during restore even though the images are seen in the backup image selection
- One of the child backup jobs goes in a queued state
- Troubleshooting restore issues for Hadoop data
- Restore fails with error code 2850
- NetBackup restore job for Hadoop completes partially
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Restore operation fails when Hadoop plug-in files are missing on the backup host
- Restore fails with bpbrm error 54932
- Restore operation fails with bpbrm error 21296
- Hadoop with Kerberos restore job fails with error 2850
- Configuration file is not recovered after a disaster recovery
- Index
Create a BigData policy for Hadoop clusters
Backup policies provide the instructions that follows to back up clients. To configure backup policies for the Hadoop plug-in for , use the type as the .
Note:
The host name and port of the NameNode must be the same as the values that you specified with the HTTP address parameter in the core-site.xml of the Hadoop cluster.
To create a BigData policy for Hadoop clusters
- Open the web UI.
- On the left, click Protection > Policies.
- On the Policies tab, click Add.
- On the Attributes tab, for the Policy type select BigData.
- On the Schedules tab, click Add to create a new schedule.
You can create a schedule for a Full backup, Differential incremental backup, or Cumulative incremental backup for your BigData policy. After you set the schedule, Hadoop data is backed up automatically as per the set schedule without any further user intervention.
- On the Clients tab, enter the IP address or the host name of the
NameNode. - On the Backup selections tab, enter the following parameters and their values as shown:
Application_Type=hadoop
The parameter values are case-sensitive.
Backup_Host=IP_address or hostname
The backup host must be a Linux computer. The backup host can be a NetBackup client or a media server.
You can specify multiple backup hosts.
File path or the directory to back up.
You can specify multiple file paths.
Note:
The directory or folder that is specified for the backup selection when you define a BigData Policy with Application_Type=hadoop must not contain a space or a comma in their names.
- Click Create.
For more information on using NetBackup for BigData applications, refer to the Veritas NetBackup documentation page.