NetBackup™ for Hadoop Administrator's Guide
- Introduction
- Verifying the pre-requisites and best practices for the Hadoop plug-in for NetBackup
- Configuring NetBackup for Hadoop
- About configuring NetBackup for Hadoop
- Managing backup hosts
- Adding Hadoop credentials in NetBackup
- Configuring the Hadoop plug-in using the Hadoop configuration file
- Configuration for a Hadoop cluster that uses Kerberos
- Configuring NetBackup policies for Hadoop plug-in
- Disaster recovery of a Hadoop cluster
- Performing backups and restores of Hadoop
- Troubleshooting
- About troubleshooting NetBackup for Hadoop issues
- About NetBackup for Hadoop debug logging
- Troubleshooting backup issues for Hadoop data
- Backup operation fails with error 6609
- Backup operation failed with error 6618
- Backup operation fails with error 6647
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Backup operation fails with error 6654
- Backup operation fails with bpbrm error 8857
- Backup operation fails with error 6617
- Backup operation fails with error 6616
- Backup operation fails with error 84
- NetBackup configuration and certificate files do not persist after the container-based NetBackup appliance restarts
- Unable to see incremental backup images during restore even though the images are seen in the backup image selection
- One of the child backup jobs goes in a queued state
- Troubleshooting restore issues for Hadoop data
- Restore fails with error code 2850
- NetBackup restore job for Hadoop completes partially
- Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed up or restored for Hadoop
- Restore operation fails when Hadoop plug-in files are missing on the backup host
- Restore fails with bpbrm error 54932
- Restore operation fails with bpbrm error 21296
- Configuration file is not recovered after a disaster recovery
- Index
Configuring NetBackup for a highly-available Hadoop cluster
To protect a highly-available Hadoop cluster, when you configure NetBackup for Hadoop cluster:
Specify one of the NameNodes (primary) as the client in the BigData policy.
Specify the same NameNode (primary and fail-over) as application server when you execute the tpconfig command.
Create a
hadoop.conffile, update it with the details of the NameNodes (primary and fail-over), and copy it to all the backup hosts. Thehadoop.conffile is in JSON format.Hostname and port of the NameNode must be same as you have specified with the http address parameter in the
core-site.xmlof the Hadoop cluster.User name of the primary and fail-over NameNode must be same.
Do not provide a blank value for any of the parameters, or the backup job fails.
To update the hadoop.conf file for highly-available Hadoop cluster
- Update the
hadoop.conffile with the following parameters:{ "application_servers": { "hostname_of_primary_namenode1": { "failover_namenodes": [ { "hostname": "hostname_of_failover_namenode1", "port": port_of_failover_namenode1 } ], "port":port_of_primary_namenode1 } } } - If you have multiple Hadoop clusters, use the same
hadoop.conffile to update the details. For example,{ "application_servers": { "hostname_of_primary_namenode1": { "failover_namenodes": [ { "hostname": "hostname_of_failover_namenode1", "port": port_of_failover_namenode1 } ], "port"::port_of_primary_namenode1 }, "hostname_of_primary_namenode2": { "failover_namenodes": [ { "hostname": "hostname_of_failover_namenode2", "port": port_of_failover_namenode2 } ], "port":port_of_primary_namenode2 } } } - Copy this file to the following location on all the backup hosts:
/usr/openv/var/global/