Configuration
EnterpriseBacula Enterprise Only
This solution is only available for Bacula Enterprise. For subscription inquiries, please reach out to sales@baculasystems.com.
The following chapter presents the information on HDFS Plugin parameters, estimation and backup parameters, and restore parameters.
Plugin Parameters
The following parameters affect any type of HDFS Plugin Job (Backup, Estimation or Restore).
url=<string>specifies the URL of the HDFS instance. This parameter is mandatory.user=<string>specifies the User who owns theroot path "/"in the HDFS instance. Bacula needs to know this user in order to create snapshots in the system. This parameter is mandatory.
Plugin Estimation and Backup Parameters
include=<string>specifies which files should be backed up from the HDFS System. This parameter is optional. There may be more than oneincludeparameter.regexinclude=<regex>specifies, using a Regular Expression, which files should be backed up from the HDFS System. This parameter is optional. There may be more than oneregexincludeparameter.exclude=<string>specifies which files should NOT be backed up from the HDFS System. This parameter is optional. There may be more than oneexcludeparameter.regexexclude=<regex>specifies, using a Regular Expression, which files should NOT be backed up from the HDFS System. This parameter is optional. There may be more than oneregexexcludeparameter.
If none of the optional paramaters include, regexinclude,
exclude or regexexclude are specified then all files from the
Hadoop File System to which the user bacula has access will be
backed up.
Plugin Restore Parameters
user=<string>specifies an account where restore will be performed. This parameter is optional. If not set, theuserparameter from the backup Job will be used.url=<string>specifies the URL of the HDFS system during a restore. This parameter is optional. If not set, theurl=<string>parameter from the backup Job will be used.restore_local=<yes or no>specifies that the files should be restored to a local directory based on thewhere=restore job parameter. This parameter is optional and defaults to no.
Fileset Examples
In the example below, all files inside the path btest1 will be
backed up.
Fileset {
Name = FS_Hdfs
Include {
Plugin = "hdfs: user=hadoop URL=hdfs://localhost:9000 include=btest1/*"
}
}
In the example below, all files that do not end with tmp inside the
path btest1 will be backed up.
Fileset {
Name = FS_Hdfs_without_tmp
Include {
Plugin = "hdfs: user=hadoop URL=hdfs://localhost:9000 include=btest1/* exclude=*tmp"
}
}
This example is the same as the exclude one above, but using
regexexclude instead:
Fileset {
Name = FS_Hdfs_without_tmp
Include {
Plugin = "hdfs: user=hadoop URL=hdfs://localhost:9000 include=btest1/* regexclude=.*\\.tmp\\Z(?ms)"
}
}
In the example below, all files that end with .pdf inside the path
path1 will be backed up.
Fileset {
Name = FS_Hdfs_without_tmp
Include {
Plugin = "hdfs: user=hadoop URL=hdfs://localhost:9000 include=btest1/* regexinclude=.*\\.pdf\\Z(?ms)"
}
}
In the example below, all files will be backed up.
Fileset {
Name = FS_Hdfs_everything
Include {
Options {
Compression = LZO
}
Plugin = "hdfs: user=hadoop URL=hdfs://localhost:9000"
}
}
See also
Previous articles:
Next articles:
Go back to: Hadoop Distributed File System Plugin.