public class HBaseSnapshotBinaryInputFormat extends AbstractBinaryInputFormat
1. The snapshot name. This points to a pre-created snapshot of the graph
table on HBase. HBaseStoreManager.HBASE_SNAPSHOT
e.g. janusgraphmr.ioformat.conf.storage.hbase.snapshot-name=janusgraph-snapshot
2. The snapshot restore directory. This is specified as a temporary restore
directory on the same File System as hbase root dir. The restore directory is
used to restore the table and region structure from the snapshot to scan the
table, but with no data coping involved.
HBaseStoreManager.HBASE_SNAPSHOT_RESTORE_DIR
e.g. janusgraphmr.ioformat.conf.storage.hbase.snapshot-restore-dir=/tmp
It is also required that the Hadoop configuration directory, which contains core-site.xml, is in the classpath for access to the hadoop cluster. This requirement is similar to the configuration requirement for hadoop-gremlin
Additionally, the HBase configuration directory, which contains hbase-site.xml,
should be placed in the classpath as well. If it is not, hbase.rootdir property
needs to be set as a pass-through property in the graph property file.
e.g. janusgraphmr.ioformat.conf.storage.hbase.ext.hbase.rootdir=/hbase
hadoopConf, janusgraphConf, mrConf| Constructor and Description |
|---|
HBaseSnapshotBinaryInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.mapreduce.RecordReader<StaticBuffer,Iterable<Entry>> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit,
org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) |
org.apache.hadoop.conf.Configuration |
getConf() |
List<org.apache.hadoop.mapreduce.InputSplit> |
getSplits(org.apache.hadoop.mapreduce.JobContext jobContext) |
void |
setConf(org.apache.hadoop.conf.Configuration config) |
public List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext jobContext) throws IOException, InterruptedException
getSplits in class org.apache.hadoop.mapreduce.InputFormat<StaticBuffer,Iterable<Entry>>IOExceptionInterruptedExceptionpublic org.apache.hadoop.mapreduce.RecordReader<StaticBuffer,Iterable<Entry>> createRecordReader(org.apache.hadoop.mapreduce.InputSplit inputSplit, org.apache.hadoop.mapreduce.TaskAttemptContext taskAttemptContext) throws IOException, InterruptedException
createRecordReader in class org.apache.hadoop.mapreduce.InputFormat<StaticBuffer,Iterable<Entry>>IOExceptionInterruptedExceptionpublic void setConf(org.apache.hadoop.conf.Configuration config)
setConf in interface org.apache.hadoop.conf.ConfigurablesetConf in interface org.apache.tinkerpop.gremlin.hadoop.structure.io.HadoopPoolsConfigurablesetConf in class AbstractBinaryInputFormatpublic org.apache.hadoop.conf.Configuration getConf()
getConf in interface org.apache.hadoop.conf.ConfigurablegetConf in interface org.apache.tinkerpop.gremlin.hadoop.structure.io.HadoopPoolsConfigurablegetConf in class AbstractBinaryInputFormatCopyright © 2012–2020. All rights reserved.