NetBackup™ Web UI Cloud Object Store Administrator's Guide

Last Published:
Product(s): NetBackup & Alta Data Protection (10.3.0.1, 10.3)
  1. Introduction
    1.  
      Overview of NetBackup protection for Cloud object store
    2.  
      Features of NetBackup Cloud object store workload support
  2. Managing Cloud object store assets
    1.  
      Prerequisites for adding Cloud object store accounts
    2.  
      Permissions required for Amazon S3 cloud provider user
    3.  
      Permissions required for Azure
    4.  
      Limitations and considerations
    5. Adding Cloud object store accounts
      1.  
        Creating cross-account access in AWS
      2.  
        Check certificate for revocation
      3.  
        Managing Certification Authorities (CA) for NetBackup Cloud
      4.  
        Adding a new region
    6.  
      Manage Cloud object store accounts
  3. Protecting Cloud object store assets
    1. About accelerator support
      1.  
        How NetBackup accelerator works with Cloud object store
      2.  
        Accelerator notes and requirements
      3.  
        Accelerator force rescan for Cloud object store (schedule attribute)
      4.  
        Accelerator backup and NetBackup catalog
    2.  
      About incremental backup
    3.  
      About policies for Cloud object store assets
    4.  
      Planning for policies
    5.  
      Prerequisites for Cloud object store policies
    6.  
      Creating a backup policy
    7.  
      Setting up attributes
    8.  
      Creating schedule attributes for policies
    9.  
      Configuring the Start window
    10.  
      Configuring exclude dates
    11.  
      Configuring include dates
    12.  
      Configuring the Cloud objects tab
    13.  
      Adding conditions
    14.  
      Adding tag conditions
    15.  
      Example of conditions and tag conditions
    16. Managing Cloud object store policies
      1.  
        Copy a policy
      2.  
        Deactivating or deleting a policy
      3.  
        Manually backup assets
  4. Recovering Cloud object store assets
    1.  
      Prerequisites for recovering Cloud object store objects
    2.  
      Configuring Cloud object retention properties
    3.  
      Recovering Cloud object store assets
  5. Troubleshooting
    1.  
      Recovery for Cloud object store using web UI for original bucket recovery option starts but job fails with error 3601
    2.  
      Recovery Job does not start
    3.  
      Restore fails: "Error bpbrm (PID=3899) client restore EXIT STATUS 40: network connection broken"
    4.  
      Access tier property not restored after overwrite existing to original location
    5.  
      Reduced accelerator optimization in Azure for OR query with multiple tags
    6.  
      Backup is failed and shows a certificate error with Amazon S3 bucket names containing dots (.)
    7.  
      Azure backup job fails when space is provided in a tag query for either tag key name or value.
    8.  
      The Cloud object store account has encountered an error
    9.  
      Bucket list empty when selecting it in policy selection
    10.  
      Creating second account on Cloudian fails by selecting existing region
    11.  
      Restore failed with 2825 incomplete restore operation
    12.  
      Bucket listing of cloud provider fails when adding bucket in Cloud objects tab
    13.  
      AIR import image restore fails on the target domain if the Cloud store account is not added in the target domain.
    14.  
      Backup for Azure Data Lake fails when a back-level media server is used with backup host or storage server version 10.3
    15.  
      Backup fails partially in Azure Data Lake: "Error nbpem (pid=16018) backup of client
    16.  
      Recovery for Azure Data Lake fails: "This operation is not permitted as the path is too deep"
    17.  
      Empty directories are not backed up in Azure Data Lake
    18.  
      Recovery error: "Invalid alternate directory location. You must specify a string with length less than 1025 valid characters"
    19.  
      Recovery error: "Invalid parameter specified"
    20.  
      Restore fails: "Cannot perform the COSP operation, skipping the object: [/testdata/FxtZMidEdTK]"
    21.  
      Cloud store account creation fails with incorrect credentials
    22.  
      Discovery failures due to improper permissions
    23.  
      Restore failures due to object lock

Prerequisites for Cloud object store policies

Before you begin creating a policy for a Cloud object store account, consider the following prerequisites.

  • A valid Cloud object store account to access the bucket(s) and objects.

  • Keep handy information about the bucket(s) and the criteria that you want to use for selecting objects from them, in the Cloud objects tab.

  • In the Cloud objects tab, you must have permission to view and select the Cloud object store account(s), and the access host to specify the backup host or scale-out server for a policy.

  • Evaluate the requirement for NetBackup accelerators in your environment. If you want to use accelerators, you need to specify this while creating the policy.

  • If you plan to use any other backup host or scale-out server, other than the one used for Cloud object store account validation: Make sure that required ports are opened and configurations are done for communication from the backup host or scale-out server to the cloud provider endpoint using REST API calls.

    You can use a scale-out server if you have a large number of buckets in your Cloud object store. NetBackup Snapshot Manager can scale out as many data mover containers as needed at run time, and then scale them down when the data protection jobs are completed. You do not need to worry about configuring multiple backup hosts, and creating multiple policies to distribute the load across these backup hosts.

  • Evaluate the requirement for NetBackup multistreaming in your environment. For a given bucket, NetBackup creates one stream per query defined for the bucket in the policy. If you want to use multistreaming, you can specify this while creating the policy. To use multistream, you also need to configure the number of jobs for the buckets as client in the Client attributes section, under primary server Host properties. Add the client name and set the Maximum data streams as required.