Please enter search query.
Search <book_title>...
NetBackup™ for Hadoop Administrator's Guide
Last Published:
2022-09-08
Product(s):
NetBackup (10.1)
- Introduction
- Verifying the pre-requisites and best practices for the Hadoop plug-in for NetBackup
- Configuring NetBackup for Hadoop
- About configuring NetBackup for Hadoop
- Managing backup hosts
- Adding Hadoop credentials in NetBackup
- Configuring the Hadoop plug-in using the Hadoop configuration file
- Configuration for a Hadoop cluster that uses Kerberos
- Configuring NetBackup policies for Hadoop plug-in
- Disaster recovery of a Hadoop cluster
- Performing backups and restores of Hadoop
- Troubleshooting
- About troubleshooting NetBackup for Hadoop issues
- About NetBackup for Hadoop debug logging
- Troubleshooting backup issues for Hadoop data
- Backup operation fails with error 6609
- Backup operation failed with error 6618
- Backup operation fails with error 6647
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Backup operation fails with error 6654
- Backup operation fails with bpbrm error 8857
- Backup operation fails with error 6617
- Backup operation fails with error 6616
- Backup operation fails with error 84
- NetBackup configuration and certificate files do not persist after the container-based NetBackup appliance restarts
- Unable to see incremental backup images during restore even though the images are seen in the backup image selection
- One of the child backup jobs goes in a queued state
- Troubleshooting restore issues for Hadoop data
- Restore fails with error code 2850
- NetBackup restore job for Hadoop completes partially
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Restore operation fails when Hadoop plug-in files are missing on the backup host
- Restore fails with bpbrm error 54932
- Restore operation fails with bpbrm error 21296
- Configuration file is not recovered after a disaster recovery
- Index
Configuring the Hadoop plug-in using the Hadoop configuration file
The backup hosts use the hadoop.conf file to save the configuration settings of the Hadoop plug-in. You need to create a separate file for each backup host and copy it to the /usr/openv/var/global/. You need to manually create the hadoop.conf file in JSON format. This file is not available by default with the installer.
Note:
You must not provide a blank value for any of the parameters, or the backup job fails.
Ensure that you configure all the required parameters to run the backup and restore operations successfully.
With this release, the following plug-in settings can be configured:
Following is an example of the hadoop.conf file.
Note:
For non-HA environment, the fail-over parameters are not required.
{
"application_servers":
{
"hostname_of_the_primary_namenode":
{
"failover_namenodes":
[
{
"hostname":"hostname_of_failover_namenode",
"port":port_of_the_failover_namenode
}
],
"port":port_of_the_primary_namenode
}
},
"number_of_threads":number_of_threads
}