Veritas NetBackup for Hadoop Administrator's Guide

Last Published:
Product(s): NetBackup (8.1)
  1. Introduction
    1.  
      Protecting Hadoop data using NetBackup
    2.  
      Backing up Hadoop data
    3.  
      Restoring Hadoop data
    4.  
      Deploying the Hadoop plug-in
    5.  
      NetBackup for Hadoop terminologies
    6.  
      Limitations
  2. Installing and deploying Hadoop plug-in for NetBackup
    1.  
      About installing and deploying the Hadoop plug-in
    2. Pre-requisites for installing the Hadoop plug-in
      1.  
        Operating system and platform compatibility
      2.  
        License for Hadoop plug-in for NetBackup
    3.  
      Best practices for deploying the Hadoop plug-in
    4.  
      Preparing the Hadoop cluster
    5.  
      Downloading the Hadoop plug-in
    6.  
      Installing the Hadoop plug-in
    7.  
      Verifying the installation of the Hadoop plug-in
  3. Configuring NetBackup for Hadoop
    1.  
      About configuring NetBackup for Hadoop
    2. Managing backup hosts
      1.  
        Whitelisting a NetBackup client on NetBackup master server
      2.  
        Configure a NetBackup Appliance as a backup host
    3.  
      Adding Hadoop credentials in NetBackup
    4. Configuring the Hadoop plug-in using the Hadoop configuration file
      1.  
        Configuring NetBackup for a highly-available Hadoop cluster
      2.  
        Configuring a custom port for the Hadoop cluster
      3.  
        Configuring number of threads for backup hosts
    5.  
      Configuration for a Hadoop cluster that uses Kerberos
    6. Configuring NetBackup policies for Hadoop plug-in
      1. Creating a BigData backup policy
        1. Creating BigData policy using the NetBackup Administration Console
          1.  
            Using the Policy Configuration Wizard to create a BigData policy for Hadoop clusters
          2.  
            Using the NetBackup Policies utility to create a BigData policy for Hadoop clusters
        2.  
          Using NetBackup Command Line Interface (CLI) to create a BigData policy for Hadoop clusters
    7.  
      Disaster recovery of a Hadoop cluster
  4. Performing backups and restores of Hadoop
    1. About backing up a Hadoop cluster
      1.  
        Pre-requisite for running backup and restore operations for a Hadoop cluster with Kerberos authentication
      2.  
        Backing up a Hadoop cluster
      3.  
        Best practices for backing up a Hadoop cluster
    2. About restoring a Hadoop cluster
      1. Restoring Hadoop data on the same Hadoop cluster
        1.  
          Using the Restore Wizard to restore Hadoop data on the same Hadoop cluster
        2.  
          Using the bprestore command to restore Hadoop data on the same Hadoop cluster
      2.  
        Restoring Hadoop data on an alternate Hadoop cluster
      3.  
        Best practices for restoring a Hadoop cluster
  5. Troubleshooting
    1.  
      About troubleshooting NetBackup for Hadoop issues
    2.  
      About NetBackup for Hadoop debug logging
    3. Troubleshooting backup issues for Hadoop data
      1.  
        Backup operation for Hadoop fails with error code 6599
      2.  
        Backup operation fails with error 6609
      3.  
        Backup operation failed with error 6618
      4.  
        Backup operation fails with error 6647
      5.  
        Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
      6.  
        Backup operation fails with error 6654
      7.  
        Backup operation fails with bpbrm error 8857
      8.  
        Backup operation fails with error 6617
      9.  
        Backup operation fails with error 6616
    4. Troubleshooting restore issues for Hadoop data
      1.  
        Restore fails with error code 2850
      2.  
        NetBackup restore job for Hadoop completes partially
      3.  
        Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
      4.  
        Restore operation fails when Hadoop plug-in files are missing on the backup host
      5.  
        Restore fails with bpbrm error 54932
      6.  
        Restore operation fails with bpbrm error 21296

Backing up Hadoop data

Hadoop data is backed up in parallel streams wherein Hadoop DataNodes stream data blocks simultaneously to multiple backup hosts.

Note:

All the directories specified in Hadoop backup selection must be snapshot-enabled before the backup.

The following diagram provides an overview of the backup flow:

Figure: Backup flow

Backup flow

As illustrated in the following diagram:

  1. A scheduled backup job is triggered from the master server.

  2. Backup job for Hadoop data is a compound job. When the backup job is triggered, first a discovery job is run.

  3. During discovery, the first backup host connects with the NameNode and performs a discovery to get details of data that needs to be backed up.

  4. A workload discovery file is created on the backup host. The workload discovery file contains the details of the data that needs to be backed up from the different DataNodes.

  5. The backup host uses the workload discovery file and decides how the workload is distributed amongst the backup hosts. Workload distribution files are created for each backup host.

  6. Individual child jobs are executed for each backup host. As specified in the workload distribution files, data is backed up.

  7. Data blocks are streamed simultaneously from different DataNodes to multiple backup hosts.

The compound backup job is not completed untill all the child jobs are completed. After the child jobs are completed, NetBackup cleans all the snapshots from the NameNode. Only after the cleanup activity is completed, the compound backup job is completed.

See About backing up a Hadoop cluster.