- calculateAutoBalancedSplits(List<InputSplit>, long) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Calculates the number of MapReduce input splits for the map tasks.
- CELL_TTL_COLUMN_SPEC - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- CELL_VISIBILITY_COLUMN_SPEC - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- CellCounter - Class in org.apache.hadoop.hbase.mapreduce
-
A job with a a map and reduce phase to count cells in a table.
- CellCounter() - Constructor for class org.apache.hadoop.hbase.mapreduce.CellCounter
-
- CellCreator - Class in org.apache.hadoop.hbase.mapreduce
-
Facade to create Cells for HFileOutputFormat.
- CellCreator(Configuration) - Constructor for class org.apache.hadoop.hbase.mapreduce.CellCreator
-
- CellDeserializer() - Constructor for class org.apache.hadoop.hbase.mapreduce.CellSerialization.CellDeserializer
-
- CellImporter() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellImporter
-
- CellReducer() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellReducer
-
- CellSerialization - Class in org.apache.hadoop.hbase.mapreduce
-
Use to specify the type of serialization for the mappers and reducers
- CellSerialization() - Constructor for class org.apache.hadoop.hbase.mapreduce.CellSerialization
-
- CellSerialization.CellDeserializer - Class in org.apache.hadoop.hbase.mapreduce
-
- CellSerialization.CellSerializer - Class in org.apache.hadoop.hbase.mapreduce
-
- CellSerializer() - Constructor for class org.apache.hadoop.hbase.mapreduce.CellSerialization.CellSerializer
-
- CellSortImporter() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellSortImporter
-
- CellSortReducer - Class in org.apache.hadoop.hbase.mapreduce
-
Emits sorted Cells.
- CellSortReducer() - Constructor for class org.apache.hadoop.hbase.mapreduce.CellSortReducer
-
- cellVisibilityExpr - Variable in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- CellWritableComparable() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable
-
- CellWritableComparable(Cell) - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable
-
- CellWritableComparablePartitioner() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparablePartitioner
-
- CellWritableComparator() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable.CellWritableComparator
-
- CF_RENAME_PROP - Static variable in class org.apache.hadoop.hbase.mapreduce.Import
-
- checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.hbase.mapred.TableOutputFormat
-
- checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat
-
- checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
Checks if the output table exists and is enabled.
- cleanup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, ImmutableBytesWritable>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.HashMapper
-
- cleanup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication.Verifier
-
- cleanup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper
-
- cleanup(Mapper<WALKey, WALEdit, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.WALPlayer.WALMapper
-
- cleanupJob(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- close(Reporter) - Method in class org.apache.hadoop.hbase.mapred.TableOutputFormat.TableRecordWriter
-
- close() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- close() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- close() - Method in class org.apache.hadoop.hbase.mapreduce.CellSerialization.CellDeserializer
-
- close() - Method in class org.apache.hadoop.hbase.mapreduce.CellSerialization.CellSerializer
-
- close() - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash.Reader
-
- close() - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization.KeyValueDeserializer
-
Deprecated.
- close() - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization.KeyValueSerializer
-
Deprecated.
- close(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat.MultiTableRecordWriter
-
- close(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat.TableRecordWriter
-
Closes the writer, in this case flush table commits.
- close() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Closes the split.
- close() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Closes the split.
- close() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- closeTable() - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
- closeTable() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- COLUMN_LIST - Static variable in class org.apache.hadoop.hbase.mapred.TableInputFormat
-
space delimited list of columns
- columns - Variable in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
- columns - Variable in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
The grouping columns.
- COLUMNS_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- combineTableNameSuffix(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- CompactionTool - Class in org.apache.hadoop.hbase.regionserver
-
- CompactionTool() - Constructor for class org.apache.hadoop.hbase.regionserver.CompactionTool
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable.CellWritableComparator
-
- compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparable.KeyValueWritableComparator
-
- compareTo(TableSplit) - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- compareTo(Import.CellWritableComparable) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable
-
- compareTo(Import.KeyValueWritableComparable) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparable
-
- compareTo(TableSplit) - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Compares this split against the given one.
- conf - Variable in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- CONF_DEST_PREFIX - Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
-
Configuration prefix for overrides for the destination filesystem
- CONF_SKIP_TMP - Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
-
- CONF_SOURCE_PREFIX - Static variable in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
-
Configuration prefix for overrides for the source filesystem
- configure(JobConf) - Method in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
- configure(JobConf) - Method in class org.apache.hadoop.hbase.mapred.HRegionPartitioner
-
- configure(JobConf) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormat
-
- configureCfRenaming(Configuration, Map<String, String>) - Static method in class org.apache.hadoop.hbase.mapreduce.Import
-
Sets a configuration property with key
Import.CF_RENAME_PROP
in conf that tells
the mapper how to rename column families.
- configureIncrementalLoad(Job, Table, RegionLocator) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
Configure a MapReduce Job to perform an incremental load into the given
table.
- configureIncrementalLoad(Job, TableDescriptor, RegionLocator) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
Configure a MapReduce Job to perform an incremental load into the given
table.
- configureIncrementalLoad(Job, List<HFileOutputFormat2.TableInfo>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
- configureIncrementalLoadMap(Job, TableDescriptor) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- configureSplitTable(Job, TableName) - Static method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Sets split table in map-reduce job.
- convertScanToString(Scan) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Writes the given scan into a Base64 encoded string.
- convertStringToScan(String) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Converts the given Base64 string back into a Scan instance.
- CopyTable - Class in org.apache.hadoop.hbase.mapreduce
-
Tool used to copy a table to another one which can be on a different setup.
- CopyTable() - Constructor for class org.apache.hadoop.hbase.mapreduce.CopyTable
-
- create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int) - Method in class org.apache.hadoop.hbase.mapreduce.CellCreator
-
- create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) - Method in class org.apache.hadoop.hbase.mapreduce.CellCreator
-
- create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, List<Tag>) - Method in class org.apache.hadoop.hbase.mapreduce.CellCreator
-
- CREATE_TABLE_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- createCompositeKey(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
Creates a composite key to use as a mapper output key when using
MultiTableHFileOutputFormat.configureIncrementaLoad to set up bulk ingest job
- createCompositeKey(byte[], ImmutableBytesWritable) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
Alternate api which accepts an ImmutableBytesWritable for the suffix
- createCompositeKey(String, ImmutableBytesWritable) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the
suffix
- createGroupKey(byte[][]) - Method in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
Create a key by concatenating multiple column values.
- createGroupKey(byte[][]) - Method in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
Create a key by concatenating multiple column values.
- createKey() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- createKey() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- createNInputSplitsUniform(InputSplit, int) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Create n splits for one InputSplit, For now only support uniform distribution
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.HFileInputFormat
-
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
Builds a TableRecordReader.
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat
-
- createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.WALInputFormat
-
- createRegionSizeCalculator(RegionLocator, Admin) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- createScanFromConfiguration(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Sets up a
Scan
instance, applying settings from the configuration property
constants defined in
TableInputFormat
.
- createSubmittableJob(String[]) - Method in class org.apache.hadoop.hbase.mapred.RowCounter
-
- createSubmittableJob(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.CellCounter
-
Sets up the actual job.
- createSubmittableJob(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.CopyTable
-
Sets up the actual job.
- createSubmittableJob(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.Export
-
Sets up the actual job.
- createSubmittableJob(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable
-
- createSubmittableJob(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.Import
-
Sets up the actual job.
- createSubmittableJob(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
Sets up the actual job.
- createSubmittableJob(Configuration, String[]) - Method in class org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication
-
Sets up the actual job.
- createSubmittableJob(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.RowCounter
-
Sets up the actual job.
- createSubmittableJob(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.SyncTable
-
- createSubmittableJob(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
Sets up the actual job.
- createValue() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- createValue() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- createVisibilityExpTags(String) - Method in class org.apache.hadoop.hbase.mapreduce.DefaultVisibilityExpressionResolver
-
- createVisibilityExpTags(String) - Method in interface org.apache.hadoop.hbase.mapreduce.VisibilityExpressionResolver
-
Convert visibility expression into tags to be serialized.
- CREDENTIALS_LOCATION - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- getAdmin() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Allows subclasses to get the
Admin
.
- getArgumentsFromCommandLine(Configuration, String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.ExportUtils
-
- getAttributesKeyColumnIndex() - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getBadLineCount() - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- getBadLineCount() - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- getBadLineCount() - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
- getBestLocations(Configuration, HDFSBlocksDistribution) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getCellTTLColumnIndex() - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getCellVisibilityColumnIndex() - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.DefaultVisibilityExpressionResolver
-
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
Returns the current configuration.
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.HRegionPartitioner
-
Returns the current configuration.
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
-
Returns the current configuration.
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Returns the current configuration.
- getConf() - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
- getConf() - Method in class org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool
-
- getCurrentHash() - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash.Reader
-
Get the current hash
- getCurrentKey() - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash.Reader
-
Get the current key
- getCurrentKey() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Returns the current key.
- getCurrentKey() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Returns the current key.
- getCurrentKey() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- getCurrentValue() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Returns the current value.
- getCurrentValue() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Returns the current value.
- getCurrentValue() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- getDeserializer(Class<Cell>) - Method in class org.apache.hadoop.hbase.mapreduce.CellSerialization
-
- getDeserializer(Class<KeyValue>) - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization
-
Deprecated.
- getDeserializer(Class<Mutation>) - Method in class org.apache.hadoop.hbase.mapreduce.MutationSerialization
-
- getDeserializer(Class<Result>) - Method in class org.apache.hadoop.hbase.mapreduce.ResultSerialization
-
- getEncodedRegionName() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the region's encoded name.
- getEndRow() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getEndRow() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the end row.
- getFamily(int) - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getFamilyArray() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getFamilyByteBuffer() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getFamilyLength() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getFamilyOffset() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getFamilyPosition() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getHtd() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getJar(Class) - Static method in class org.apache.hadoop.hbase.mapreduce.JarFinder
-
Returns the full path to the Jar containing the class.
- getLength() - Method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getLength() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getLength() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getLength() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getLength() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the length of the split.
- getLocations() - Method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getLocations() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getLocations() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getLocations() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getLocations() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the region's location as an array.
- getMapperClass(JobContext) - Static method in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
Get the application's mapper class.
- getNumberOfThreads(JobContext) - Static method in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
The number of threads in the thread pool that will run the map function.
- getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat
-
- getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
Returns the output committer.
- getPartition(ImmutableBytesWritable, V2, int) - Method in class org.apache.hadoop.hbase.mapred.HRegionPartitioner
-
- getPartition(ImmutableBytesWritable, VALUE, int) - Method in class org.apache.hadoop.hbase.mapreduce.HRegionPartitioner
-
Gets the partition number for a given key (hence record) given the total
number of partitions i.e.
- getPartition(Import.CellWritableComparable, Cell, int) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparablePartitioner
-
- getPartition(Import.KeyValueWritableComparable, KeyValue, int) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparablePartitioner
-
Deprecated.
- getPartition(ImmutableBytesWritable, VALUE, int) - Method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- getPos() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- getPos() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- getPos() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- getProgress() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- getProgress() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- getProgress() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
The current progress of the record reader through its data.
- getProgress() - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
The current progress of the record reader through its data.
- getProgress() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- getQualifier(int) - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getQualifierArray() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getQualifierByteBuffer() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getQualifierLength() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getQualifierOffset() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getQualifierPosition() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.hbase.mapred.MultiTableSnapshotInputFormat
-
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Builds a TableRecordReader.
- getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
-
- getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.hbase.mapred.TableOutputFormat
-
Creates a new record writer.
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat
-
- getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
Creates a new record writer.
- getRegion() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getRegionInfo() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- getRegionInfo() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getRegionInfosFromManifest(SnapshotManifest) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getRegionLocation() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getRegionLocation() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the region location.
- getRegionLocator() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- getRegionSize(byte[]) - Method in class org.apache.hadoop.hbase.mapreduce.RegionSizeCalculator
-
Returns size of given region in bytes.
- getRegionSizeMap() - Method in class org.apache.hadoop.hbase.mapreduce.RegionSizeCalculator
-
- getRestoreDir() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getRowArray() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getRowByteBuffer() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getRowKeyColumnIndex() - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getRowLength() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getRowOffset() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getRowPosition() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getScan() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Gets the scan defining the actual details like columns etc.
- getScan() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getScan() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns a Scan object from the stored string representation.
- getScanner() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- getScans() - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
Allows subclasses to get the list of
Scan
objects.
- getSequenceId() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getSerializedSize(boolean) - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getSerializer(Class<Cell>) - Method in class org.apache.hadoop.hbase.mapreduce.CellSerialization
-
- getSerializer(Class<KeyValue>) - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization
-
Deprecated.
- getSerializer(Class<Mutation>) - Method in class org.apache.hadoop.hbase.mapreduce.MutationSerialization
-
- getSerializer(Class<Result>) - Method in class org.apache.hadoop.hbase.mapreduce.ResultSerialization
-
- getSkipBadLines() - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- getSkipBadLines() - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- getSkipBadLines() - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
- getSnapshotDirs(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- getSnapshotManifest(Configuration, String, Path, FileSystem) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getSnapshotsToScans(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- getSplitAlgo(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.hbase.mapred.MultiTableSnapshotInputFormat
-
- getSplits(JobConf, int) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Calculates the splits that will serve as input for the map tasks.
- getSplits(JobConf, int) - Method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
-
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
Calculates the splits that will serve as input for the map tasks.
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormat
-
- getSplits(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Calculates the splits that will serve as input for the map tasks.
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Calculates the splits that will serve as input for the map tasks.
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat
-
- getSplits(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getSplits(Scan, SnapshotManifest, List<HRegionInfo>, Path, Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getSplits(Scan, SnapshotManifest, List<HRegionInfo>, Path, Configuration, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- getSplits(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.WALInputFormat
-
- getStartEndKeys() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
- getStartEndKeys() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- getStartRow() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getStartRow() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the start row.
- getSuffix(byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
- getTable() - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Allows subclasses to get the
Table
.
- getTable() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getTable() - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Allows subclasses to get the
Table
.
- getTable() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the table name.
- getTableDescriptor() - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- getTableName() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- getTableName(byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
- getTableName() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the table name converted to a byte array.
- getTableNameSuffixedWithFamily(byte[], byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- getTagsArray() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTagsByteBuffer() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTagsLength() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTagsOffset() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTagsPosition() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTimestamp() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getTimestampKeyColumnIndex() - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- getTs() - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- getTs() - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- getTypeByte() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getValueArray() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getValueByteBuffer() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getValueLength() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getValueOffset() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getValuePosition() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- getVisibilityExpressionResolver() - Method in class org.apache.hadoop.hbase.mapreduce.CellCreator
-
- GROUP_COLUMNS - Static variable in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
JobConf parameter to specify the columns used to produce the key passed to
collect from the map phase
- GROUP_COLUMNS - Static variable in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
JobConf parameter to specify the columns used to produce the key passed to
collect from the map phase.
- GroupingTableMap - Class in org.apache.hadoop.hbase.mapred
-
Extract grouping columns from input record
- GroupingTableMap() - Constructor for class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
- GroupingTableMapper - Class in org.apache.hadoop.hbase.mapreduce
-
Extract grouping columns from input record.
- GroupingTableMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
- IdentityTableMap - Class in org.apache.hadoop.hbase.mapred
-
Pass the given key and record as-is to reduce
- IdentityTableMap() - Constructor for class org.apache.hadoop.hbase.mapred.IdentityTableMap
-
constructor
- IdentityTableMapper - Class in org.apache.hadoop.hbase.mapreduce
-
Pass the given key and record as-is to the reduce phase.
- IdentityTableMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.IdentityTableMapper
-
- IdentityTableReduce - Class in org.apache.hadoop.hbase.mapred
-
Write to table each key, record pair
- IdentityTableReduce() - Constructor for class org.apache.hadoop.hbase.mapred.IdentityTableReduce
-
- IdentityTableReducer - Class in org.apache.hadoop.hbase.mapreduce
-
Convenience class that simply writes all values (which must be
Put
or
Delete
instances)
passed to it out to the configured HBase table.
- IdentityTableReducer() - Constructor for class org.apache.hadoop.hbase.mapreduce.IdentityTableReducer
-
- IGNORE_MISSING_FILES - Static variable in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
- Import - Class in org.apache.hadoop.hbase.mapreduce
-
Import data written by
Export
.
- Import() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import
-
- Import.CellImporter - Class in org.apache.hadoop.hbase.mapreduce
-
A mapper that just writes out KeyValues.
- Import.CellReducer - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.CellSortImporter - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.CellWritableComparable - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.CellWritableComparable.CellWritableComparator - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.CellWritableComparablePartitioner - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.Importer - Class in org.apache.hadoop.hbase.mapreduce
-
Write table content out to files in hdfs.
- Import.KeyValueImporter - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.KeyValueReducer - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.KeyValueSortImporter - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.KeyValueWritableComparable - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.KeyValueWritableComparable.KeyValueWritableComparator - Class in org.apache.hadoop.hbase.mapreduce
-
- Import.KeyValueWritableComparablePartitioner - Class in org.apache.hadoop.hbase.mapreduce
-
- IMPORT_FORMAT_VER - Static variable in class org.apache.hadoop.hbase.mapreduce.ResultSerialization
-
- Importer() - Constructor for class org.apache.hadoop.hbase.mapreduce.Import.Importer
-
- ImportTsv - Class in org.apache.hadoop.hbase.mapreduce
-
Tool to import data from a TSV file.
- ImportTsv() - Constructor for class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- ImportTsv.TsvParser - Class in org.apache.hadoop.hbase.mapreduce
-
- ImportTsv.TsvParser.BadTsvLineException - Exception in org.apache.hadoop.hbase.mapreduce
-
- includeRegionInSplit(byte[], byte[]) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
Test if the given region is to be included in the InputSplit while
splitting the regions of a table.
- includeRegionInSplit(byte[], byte[]) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Test if the given region is to be included in the InputSplit while splitting
the regions of a table.
- incrementBadLineCount(int) - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- incrementBadLineCount(int) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- incrementBadLineCount(int) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
- init() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
Build the scanner.
- init() - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
Build the scanner.
- init() - Method in class org.apache.hadoop.hbase.mapreduce.DefaultVisibilityExpressionResolver
-
- init() - Method in interface org.apache.hadoop.hbase.mapreduce.VisibilityExpressionResolver
-
Giving a chance for the initialization.
- initCredentials(JobConf) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
- initCredentials(Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
- initCredentialsForCluster(Job, String) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
- initCredentialsForCluster(Job, Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Obtain an authentication token, for the specified cluster, on behalf of the current user
and add it to the credentials for the given map reduce job.
- initialize(JobConf) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormat
-
- initialize(JobConf) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Handle subclass specific set up.
- initialize(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
- initialize(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Handle subclass specific set up.
- initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Initializes the reader.
- initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Build the scanner.
- initialize(TableSnapshotInputFormatImpl.InputSplit, Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- initializeTable(Connection, TableName) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Allows subclasses to initialize the table information.
- initializeTable(Connection, TableName) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Allows subclasses to initialize the table information.
- initJob(String, String, String, Class<? extends TableMap>, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
Use this before submitting a TableMap job.
- initJob(String, String, Class<? extends TableMap>, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.IdentityTableMap
-
Use this before submitting a TableMap job.
- initJob(String, Scan, String, Class<? extends TableMapper>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
Use this before submitting a TableMap job.
- initJob(String, Scan, Class<? extends TableMapper>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.IdentityTableMapper
-
Use this before submitting a TableMap job.
- initMultiTableSnapshotMapperJob(Map<String, Collection<Scan>>, Class<? extends TableMap>, Class<?>, Class<?>, JobConf, boolean, Path) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets up the job for reading from one or more multiple table snapshots, with one or more scans
per snapshot.
- initMultiTableSnapshotMapperJob(Map<String, Collection<Scan>>, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Sets up the job for reading from one or more table snapshots, with one or more scans
per snapshot.
- initTableMapJob(String, String, Class<? extends TableMap>, Class<?>, Class<?>, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapJob(String, String, Class<? extends TableMap>, Class<?>, Class<?>, JobConf, boolean) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
- initTableMapJob(String, String, Class<? extends TableMap>, Class<?>, Class<?>, JobConf, boolean, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(TableName, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(byte[], Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, boolean, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(byte[], Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(byte[], Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableMap job.
- initTableMapperJob(List<Scan>, Class<? extends TableMapper>, Class<?>, Class<?>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a Multi TableMap job.
- initTableMapperJob(List<Scan>, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a Multi TableMap job.
- initTableMapperJob(List<Scan>, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a Multi TableMap job.
- initTableReduceJob(String, Class<? extends TableReduce>, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReduceJob(String, Class<? extends TableReduce>, JobConf, Class) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReduceJob(String, Class<? extends TableReduce>, JobConf, Class, boolean) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReducerJob(String, Class<? extends TableReducer>, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReducerJob(String, Class<? extends TableReducer>, Job, Class) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReducerJob(String, Class<? extends TableReducer>, Job, Class, String, String, String) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableReducerJob(String, Class<? extends TableReducer>, Job, Class, String, String, String, boolean) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Use this before submitting a TableReduce job.
- initTableSnapshotMapJob(String, String, Class<? extends TableMap>, Class<?>, Class<?>, JobConf, boolean, Path) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets up the job for reading from a table snapshot.
- initTableSnapshotMapJob(String, String, Class<? extends TableMap>, Class<?>, Class<?>, JobConf, boolean, Path, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets up the job for reading from a table snapshot.
- initTableSnapshotMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Sets up the job for reading from a table snapshot.
- initTableSnapshotMapperJob(String, Scan, Class<? extends TableMapper>, Class<?>, Class<?>, Job, boolean, Path, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Sets up the job for reading from a table snapshot.
- INPUT_FILES_SEPARATOR_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
- INPUT_TABLE - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Job parameter that specifies the input table.
- InputSplit() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- InputSplit(TableDescriptor, HRegionInfo, List<String>, Scan, Path) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- instantiateFilter(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.Import
-
Create a
Filter
to apply to all incoming keys (
KeyValues
) to
optionally not include in the job output
- isRecoverySupported() - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- isSplitable(JobContext, Path) - Method in class org.apache.hadoop.hbase.mapreduce.HFileInputFormat
-
- isValidArguements(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.ExportUtils
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapred.Driver
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapred.RowCounter
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.CellCounter
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.CopyTable
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.Driver
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.Export
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.HashTable
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.Import
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.RowCounter
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.SyncTable
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
Main entry point.
- main(String[]) - Static method in class org.apache.hadoop.hbase.regionserver.CompactionTool
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.snapshot.ExportSnapshot
-
- main(String[]) - Static method in class org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool
-
- map(ImmutableBytesWritable, Result, OutputCollector<ImmutableBytesWritable, Result>, Reporter) - Method in class org.apache.hadoop.hbase.mapred.GroupingTableMap
-
Extract the grouping columns from value to construct a new key.
- map(ImmutableBytesWritable, Result, OutputCollector<ImmutableBytesWritable, Result>, Reporter) - Method in class org.apache.hadoop.hbase.mapred.IdentityTableMap
-
Pass the key, value to reduce
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
Extract the grouping columns from value to construct a new key.
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, ImmutableBytesWritable>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.HashMapper
-
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.IdentityTableMapper
-
Pass the key, value to reduce.
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellImporter
-
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, Import.CellWritableComparable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellSortImporter
-
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.Importer
-
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueImporter
-
Deprecated.
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, Import.KeyValueWritableComparable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueSortImporter
-
Deprecated.
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication.Verifier
-
Map method that compares every scanned row with the equivalent from
a distant cluster.
- map(ImmutableBytesWritable, Result, Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper
-
- map(LongWritable, Text, Mapper<LongWritable, Text, ImmutableBytesWritable, Put>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
Convert a line of TSV text into an HBase table row.
- map(LongWritable, Text, Mapper<LongWritable, Text, ImmutableBytesWritable, Text>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
Convert a line of TSV text into an HBase table row.
- map(WALKey, WALEdit, Mapper<WALKey, WALEdit, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.WALPlayer.WALMapper
-
- MAPPER_CLASS - Static variable in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
- MAPPER_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- MAPREDUCE_INPUT_AUTOBALANCE - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Specify if we enable auto-balance to set number of mappers in M/R jobs.
- MapreduceDependencyClasspathTool - Class in org.apache.hadoop.hbase.util
-
Generate a classpath string containing any jars required by mapreduce jobs.
- MapreduceDependencyClasspathTool() - Constructor for class org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool
-
- MapReduceExtendedCell - Class in org.apache.hadoop.hbase.util
-
A wrapper for a cell to be used with mapreduce, as the output value class for mappers/reducers.
- MapReduceExtendedCell(Cell) - Constructor for class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- MAX_AVERAGE_REGION_SIZE - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
In auto-balance, we split input by ave region size, if calculated region size is too big, we can set it.
- MultiTableHFileOutputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
Create 3 level tree directory, first level is using table name as parent
directory and then use family name as child directory, and all related HFiles
for one family are under child directory
-tableName1
-columnFamilyName1
-columnFamilyName2
-HFiles
-tableName2
-columnFamilyName1
-HFiles
-columnFamilyName2
- MultiTableHFileOutputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableHFileOutputFormat
-
- MultiTableInputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
Convert HBase tabular data from multiple scanners into a format that
is consumable by Map/Reduce.
- MultiTableInputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
-
- MultiTableInputFormatBase - Class in org.apache.hadoop.hbase.mapreduce
-
- MultiTableInputFormatBase() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
- MultiTableOutputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
Hadoop output format that writes to one or more HBase tables.
- MultiTableOutputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat
-
- MultiTableOutputFormat.MultiTableRecordWriter - Class in org.apache.hadoop.hbase.mapreduce
-
Record writer for outputting to multiple HTables.
- MultiTableRecordWriter(Configuration, boolean) - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableOutputFormat.MultiTableRecordWriter
-
- MultiTableSnapshotInputFormat - Class in org.apache.hadoop.hbase.mapred
-
MultiTableSnapshotInputFormat generalizes
TableSnapshotInputFormat
allowing a MapReduce job to run over one or more table snapshots, with one or more scans
configured for each.
- MultiTableSnapshotInputFormat() - Constructor for class org.apache.hadoop.hbase.mapred.MultiTableSnapshotInputFormat
-
- MultiTableSnapshotInputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
MultiTableSnapshotInputFormat generalizes
TableSnapshotInputFormat
allowing a MapReduce job to run over one or more table snapshots, with one or more scans
configured for each.
- MultiTableSnapshotInputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormat
-
- MultiTableSnapshotInputFormatImpl - Class in org.apache.hadoop.hbase.mapreduce
-
Shared implementation of mapreduce code over multiple table snapshots.
- MultiTableSnapshotInputFormatImpl() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- MultithreadedTableMapper<K2,V2> - Class in org.apache.hadoop.hbase.mapreduce
-
Multithreaded implementation for @link org.apache.hbase.mapreduce.TableMapper
- MultithreadedTableMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
- MutationSerialization - Class in org.apache.hadoop.hbase.mapreduce
-
- MutationSerialization() - Constructor for class org.apache.hadoop.hbase.mapreduce.MutationSerialization
-
- RAW_SCAN - Static variable in class org.apache.hadoop.hbase.mapreduce.ExportUtils
-
- read(Configuration, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellWritableComparable
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueWritableComparable
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit
-
- readFields(DataInput) - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Reads the values of each field.
- RecordReader() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.RecordReader
-
- recoverTask(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- reduce(ImmutableBytesWritable, Iterator<Put>, OutputCollector<ImmutableBytesWritable, Put>, Reporter) - Method in class org.apache.hadoop.hbase.mapred.IdentityTableReduce
-
No aggregation, output pairs of (key, record)
- reduce(ImmutableBytesWritable, Iterable<Cell>, Reducer<ImmutableBytesWritable, Cell, ImmutableBytesWritable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.CellSortReducer
-
- reduce(Writable, Iterable<Mutation>, Reducer<Writable, Mutation, Writable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.IdentityTableReducer
-
Writes each given record, consisting of the row key and the given values,
to the configured OutputFormat
.
- reduce(Import.CellWritableComparable, Iterable<Cell>, Reducer<Import.CellWritableComparable, Cell, ImmutableBytesWritable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellReducer
-
- reduce(Import.KeyValueWritableComparable, Iterable<KeyValue>, Reducer<Import.KeyValueWritableComparable, KeyValue, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueReducer
-
Deprecated.
- reduce(ImmutableBytesWritable, Iterable<KeyValue>, Reducer<ImmutableBytesWritable, KeyValue, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSortReducer
-
Deprecated.
- reduce(K, Iterable<Put>, Reducer<K, Put, K, Put>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.PutCombiner
-
- reduce(ImmutableBytesWritable, Iterable<Put>, Reducer<ImmutableBytesWritable, Put, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.PutSortReducer
-
- reduce(ImmutableBytesWritable, Iterable<Text>, Reducer<ImmutableBytesWritable, Text, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- REGION_SERVER_CLASS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
Optional specification of the rs class name of the peer cluster
- REGION_SERVER_IMPL - Static variable in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
Optional specification of the rs impl name of the peer cluster
- RegionSizeCalculator - Class in org.apache.hadoop.hbase.mapreduce
-
Computes size of each region for given table and given column families.
- RegionSizeCalculator(RegionLocator, Admin) - Constructor for class org.apache.hadoop.hbase.mapreduce.RegionSizeCalculator
-
Computes size of each region for table and given column families.
- resetCacheConfig(Configuration) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Enable a basic on-heap cache for these jobs.
- restart(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
Restart from survivable exceptions by creating a new scanner.
- restart(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
Restart from survivable exceptions by creating a new scanner.
- restart(byte[]) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Restart from survivable exceptions by creating a new scanner.
- restart(byte[]) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Restart from survivable exceptions by creating a new scanner.
- RESTORE_DIR_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- RESTORE_DIRS_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- restoreSnapshots(Configuration, Map<String, Path>, FileSystem) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
Restore each (snapshot name, restore directory) pair in snapshotToDir
- ResultSerialization - Class in org.apache.hadoop.hbase.mapreduce
-
- ResultSerialization() - Constructor for class org.apache.hadoop.hbase.mapreduce.ResultSerialization
-
- retrieveGetCounterWithStringsParams(TaskAttemptContext) - Static method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
In new mapreduce APIs, TaskAttemptContext has two getCounter methods
Check if getCounter(String, String) method is available.
- RowCounter - Class in org.apache.hadoop.hbase.mapred
-
A job with a map to count rows.
- RowCounter() - Constructor for class org.apache.hadoop.hbase.mapred.RowCounter
-
- RowCounter - Class in org.apache.hadoop.hbase.mapreduce
-
A job with a just a map phase to count rows.
- RowCounter() - Constructor for class org.apache.hadoop.hbase.mapreduce.RowCounter
-
- ROWKEY_COLUMN_SPEC - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapred.RowCounter
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.CellCounter
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.CopyTable
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.Export
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.Import
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- run(Mapper<ImmutableBytesWritable, Result, K2, V2>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
Run the application's maps using a thread pool.
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.replication.VerifyReplication
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.RowCounter
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.SyncTable
-
- run(String[]) - Method in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
- run(String[]) - Method in class org.apache.hadoop.hbase.regionserver.CompactionTool
-
- run(String[]) - Method in class org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool
-
- SCAN - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Base-64 encoded scanner.
- SCAN_BATCHSIZE - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Set the maximum number of values to return for each call to next().
- SCAN_CACHEBLOCKS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Set to false to disable server-side caching of blocks for this scan.
- SCAN_CACHEDROWS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
The number of rows for caching that will be passed to scanners.
- SCAN_COLUMN_FAMILY - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Column Family to Scan
- SCAN_COLUMNS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Space delimited list of columns and column families to scan.
- SCAN_MAXVERSIONS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
The maximum number of version to return.
- SCAN_ROW_START - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Scan start row
- SCAN_ROW_STOP - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Scan stop row
- SCAN_TIMERANGE_END - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
The ending timestamp used to filter columns with a specific range of versions.
- SCAN_TIMERANGE_START - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
The starting timestamp used to filter columns with a specific range of versions.
- SCAN_TIMESTAMP - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
The timestamp used to filter columns with a specific timestamp.
- SCANS - Static variable in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
-
Job parameter that specifies the scan list.
- SEPARATOR_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- serialize(Cell) - Method in class org.apache.hadoop.hbase.mapreduce.CellSerialization.CellSerializer
-
- serialize(KeyValue) - Method in class org.apache.hadoop.hbase.mapreduce.KeyValueSerialization.KeyValueSerializer
-
Deprecated.
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.DefaultVisibilityExpressionResolver
-
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.GroupingTableMapper
-
Sets the configuration.
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.HRegionPartitioner
-
Sets the configuration.
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormat
-
Sets the configuration.
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Sets the configuration.
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
- setConf(Configuration) - Method in class org.apache.hadoop.hbase.util.MapreduceDependencyClasspathTool
-
- setEndKey(Configuration, byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- setEndRow(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- setEndRow(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- setHTable(Table) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- setHTable(Table) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- setHTable(Table) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Sets the HBase table.
- setInput(Configuration, Map<String, Collection<Scan>>, Path) - Static method in class org.apache.hadoop.hbase.mapred.MultiTableSnapshotInputFormat
-
- setInput(JobConf, String, Path) - Static method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInput(JobConf, String, Path, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInput(Configuration, Map<String, Collection<Scan>>, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormat
-
- setInput(Configuration, Map<String, Collection<Scan>>, Path) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
Configure conf to read from snapshotScans, with snapshots restored to a subdirectory of
restoreDir.
- setInput(Job, String, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInput(Job, String, Path, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInput(Configuration, String, Path) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInput(Configuration, String, Path, RegionSplitter.SplitAlgorithm, int) - Static method in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
Configures the job to use TableSnapshotInputFormat to read from a snapshot.
- setInputColumns(byte[][]) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
- setInputColumns(byte[][]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- setInputColumns(byte[][]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- setMapperClass(Job, Class<? extends Mapper<ImmutableBytesWritable, Result, K2, V2>>) - Static method in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
Set the application's mapper class.
- setNumberOfThreads(Job, int) - Static method in class org.apache.hadoop.hbase.mapreduce.MultithreadedTableMapper
-
Set the number of threads in the pool for running maps.
- setNumMapTasks(String, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets the number of map tasks for the given job configuration to the
number of regions the given table has.
- setNumReduceTasks(String, JobConf) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets the number of reduce tasks for the given job configuration to the
number of regions the given table has.
- setNumReduceTasks(String, Job) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Sets the number of reduce tasks for the given job configuration to the
number of regions the given table has.
- setRowFilter(Filter) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
Allows subclasses to set the
Filter
to be used.
- setRowFilter(Filter) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- setRowFilter(Filter) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- setScan(Scan) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
Sets the scan defining the actual details like columns etc.
- setScan(Scan) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
Sets the scan defining the actual details like columns etc.
- setScan(Scan) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
Sets the scan defining the actual details like columns etc.
- setScannerCaching(JobConf, int) - Static method in class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
Sets the number of rows to return and cache with each scanner iteration.
- setScannerCaching(Job, int) - Static method in class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
Sets the number of rows to return and cache with each scanner iteration.
- setScans(List<Scan>) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
Allows subclasses to set the list of
Scan
objects.
- setSequenceId(long) - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- setSnapshotDirs(Configuration, Map<String, Path>) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- setSnapshotToScans(Configuration, Map<String, Collection<Scan>>) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- setStartKey(Configuration, byte[]) - Static method in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- setStartRow(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- setStartRow(byte[]) - Method in class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- setTable(Table) - Method in class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
- setTableRecordReader(TableRecordReader) - Method in class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
- setTableRecordReader(TableRecordReader) - Method in class org.apache.hadoop.hbase.mapreduce.MultiTableInputFormatBase
-
- setTableRecordReader(TableRecordReader) - Method in class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- setTimestamp(long) - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- setTimestamp(byte[]) - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- setup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, ImmutableBytesWritable>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.HashMapper
-
- setup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellImporter
-
- setup(Mapper<ImmutableBytesWritable, Result, Import.CellWritableComparable, Cell>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.CellSortImporter
-
- setup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.Importer
-
- setup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueImporter
-
Deprecated.
- setup(Mapper<ImmutableBytesWritable, Result, Import.KeyValueWritableComparable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.Import.KeyValueSortImporter
-
Deprecated.
- setup(Reducer<ImmutableBytesWritable, Put, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.PutSortReducer
-
- setup(Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper
-
- setup(Reducer<ImmutableBytesWritable, Text, ImmutableBytesWritable, KeyValue>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
Handles initializing this class with objects specific to it (i.e., the parser).
- setup(Mapper<LongWritable, Text, ImmutableBytesWritable, Put>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
Handles initializing this class with objects specific to it (i.e., the parser).
- setup(Mapper<LongWritable, Text, ImmutableBytesWritable, Text>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
Handles initializing this class with objects specific to it (i.e., the parser).
- setup(Mapper<WALKey, WALEdit, ImmutableBytesWritable, Mutation>.Context) - Method in class org.apache.hadoop.hbase.mapreduce.WALPlayer.WALMapper
-
- setupJob(JobContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- SHUFFLE_MAPS - Static variable in class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
Specify if we have to shuffle the map tasks.
- SimpleTotalOrderPartitioner<VALUE> - Class in org.apache.hadoop.hbase.mapreduce
-
A partitioner that takes start and end keys and uses bigdecimal to figure
which reduce a key belongs to.
- SimpleTotalOrderPartitioner() - Constructor for class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- SKIP_EMPTY_COLUMNS - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- SKIP_LINES_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- SNAPSHOT_INPUTFORMAT_LOCALITY_ENABLED_DEFAULT - Static variable in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- SNAPSHOT_INPUTFORMAT_LOCALITY_ENABLED_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
Whether to calculate the block location for splits.
- SNAPSHOT_TO_SCANS_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.MultiTableSnapshotInputFormatImpl
-
- SPLIT_ALGO - Static variable in class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
For MapReduce jobs running multiple mappers per region, determines
what split algorithm we should be using to find split points for scanners.
- START - Static variable in class org.apache.hadoop.hbase.mapreduce.SimpleTotalOrderPartitioner
-
- START_TIME_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.WALInputFormat
-
- STORAGE_POLICY_PROPERTY - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- STORAGE_POLICY_PROPERTY_CF_PREFIX - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- SyncMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.SyncTable.SyncMapper
-
- SyncTable - Class in org.apache.hadoop.hbase.mapreduce
-
- SyncTable(Configuration) - Constructor for class org.apache.hadoop.hbase.mapreduce.SyncTable
-
- SyncTable.SyncMapper - Class in org.apache.hadoop.hbase.mapreduce
-
- SyncTable.SyncMapper.Counter - Enum in org.apache.hadoop.hbase.mapreduce
-
- TABLE_MAP_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
- TABLE_NAME - Static variable in class org.apache.hadoop.hbase.mapreduce.Import
-
- TableHash() - Constructor for class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash
-
- TableInputFormat - Class in org.apache.hadoop.hbase.mapred
-
Convert HBase tabular data into a format that is consumable by Map/Reduce.
- TableInputFormat() - Constructor for class org.apache.hadoop.hbase.mapred.TableInputFormat
-
- TableInputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
Convert HBase tabular data into a format that is consumable by Map/Reduce.
- TableInputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableInputFormat
-
- TableInputFormatBase - Class in org.apache.hadoop.hbase.mapred
-
- TableInputFormatBase() - Constructor for class org.apache.hadoop.hbase.mapred.TableInputFormatBase
-
- TableInputFormatBase - Class in org.apache.hadoop.hbase.mapreduce
-
- TableInputFormatBase() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableInputFormatBase
-
- TableMap<K extends org.apache.hadoop.io.WritableComparable<? super K>,V> - Interface in org.apache.hadoop.hbase.mapred
-
Scan an HBase table to sort by a specified sort column.
- TableMapper<KEYOUT,VALUEOUT> - Class in org.apache.hadoop.hbase.mapreduce
-
Extends the base Mapper
class to add the required input key
and value classes.
- TableMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableMapper
-
- TableMapReduceUtil - Class in org.apache.hadoop.hbase.mapred
-
- TableMapReduceUtil() - Constructor for class org.apache.hadoop.hbase.mapred.TableMapReduceUtil
-
- TableMapReduceUtil - Class in org.apache.hadoop.hbase.mapreduce
-
- TableMapReduceUtil() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil
-
- TableOutputCommitter - Class in org.apache.hadoop.hbase.mapreduce
-
Small committer class that does not do anything.
- TableOutputCommitter() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableOutputCommitter
-
- TableOutputFormat - Class in org.apache.hadoop.hbase.mapred
-
Convert Map/Reduce output and write it to an HBase table
- TableOutputFormat() - Constructor for class org.apache.hadoop.hbase.mapred.TableOutputFormat
-
- TableOutputFormat<KEY> - Class in org.apache.hadoop.hbase.mapreduce
-
Convert Map/Reduce output and write it to an HBase table.
- TableOutputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableOutputFormat
-
- TableOutputFormat.TableRecordWriter - Class in org.apache.hadoop.hbase.mapred
-
Convert Reduce output (key, value) to (HStoreKey, KeyedDataArrayWritable)
and write to an HBase table.
- TableOutputFormat.TableRecordWriter - Class in org.apache.hadoop.hbase.mapreduce
-
Writes the reducer output to an HBase table.
- TableRecordReader - Class in org.apache.hadoop.hbase.mapred
-
Iterate over an HBase table data, return (Text, RowResult) pairs
- TableRecordReader() - Constructor for class org.apache.hadoop.hbase.mapred.TableRecordReader
-
- TableRecordReader - Class in org.apache.hadoop.hbase.mapreduce
-
Iterate over an HBase table data, return (ImmutableBytesWritable, Result)
pairs.
- TableRecordReader() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableRecordReader
-
- TableRecordReaderImpl - Class in org.apache.hadoop.hbase.mapred
-
Iterate over an HBase table data, return (Text, RowResult) pairs
- TableRecordReaderImpl() - Constructor for class org.apache.hadoop.hbase.mapred.TableRecordReaderImpl
-
- TableRecordReaderImpl - Class in org.apache.hadoop.hbase.mapreduce
-
Iterate over an HBase table data, return (ImmutableBytesWritable, Result)
pairs.
- TableRecordReaderImpl() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableRecordReaderImpl
-
- TableRecordWriter(BufferedMutator) - Constructor for class org.apache.hadoop.hbase.mapred.TableOutputFormat.TableRecordWriter
-
- TableRecordWriter(JobConf) - Constructor for class org.apache.hadoop.hbase.mapred.TableOutputFormat.TableRecordWriter
-
Instantiate a TableRecordWriter with a BufferedMutator for batch writing.
- TableRecordWriter() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableOutputFormat.TableRecordWriter
-
- TableReduce<K extends org.apache.hadoop.io.WritableComparable,V> - Interface in org.apache.hadoop.hbase.mapred
-
Write a table, sorting by the input key
- TableReducer<KEYIN,VALUEIN,KEYOUT> - Class in org.apache.hadoop.hbase.mapreduce
-
Extends the basic Reducer
class to add the required key and
value input/output classes.
- TableReducer() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableReducer
-
- TABLES_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.WALPlayer
-
- tableSeparator - Static variable in class org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2
-
- TableSnapshotInputFormat - Class in org.apache.hadoop.hbase.mapred
-
TableSnapshotInputFormat allows a MapReduce job to run over a table snapshot.
- TableSnapshotInputFormat() - Constructor for class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
-
- TableSnapshotInputFormat - Class in org.apache.hadoop.hbase.mapreduce
-
TableSnapshotInputFormat allows a MapReduce job to run over a table snapshot.
- TableSnapshotInputFormat() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat
-
- TableSnapshotInputFormat.TableSnapshotRegionSplit - Class in org.apache.hadoop.hbase.mapred
-
- TableSnapshotInputFormat.TableSnapshotRegionSplit - Class in org.apache.hadoop.hbase.mapreduce
-
- TableSnapshotInputFormatImpl - Class in org.apache.hadoop.hbase.mapreduce
-
Hadoop MR API-agnostic implementation for mapreduce over table snapshots.
- TableSnapshotInputFormatImpl() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl
-
- TableSnapshotInputFormatImpl.InputSplit - Class in org.apache.hadoop.hbase.mapreduce
-
Implementation class for InputSplit logic common between mapred and mapreduce.
- TableSnapshotInputFormatImpl.RecordReader - Class in org.apache.hadoop.hbase.mapreduce
-
Implementation class for RecordReader logic common between mapred and mapreduce.
- TableSnapshotRegionSplit() - Constructor for class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSnapshotRegionSplit(TableSnapshotInputFormatImpl.InputSplit) - Constructor for class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSnapshotRegionSplit(HTableDescriptor, HRegionInfo, List<String>, Scan, Path) - Constructor for class org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSnapshotRegionSplit() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSnapshotRegionSplit(TableSnapshotInputFormatImpl.InputSplit) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSnapshotRegionSplit(HTableDescriptor, HRegionInfo, List<String>, Scan, Path) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormat.TableSnapshotRegionSplit
-
- TableSplit - Class in org.apache.hadoop.hbase.mapred
-
A table split corresponds to a key range [low, high)
- TableSplit() - Constructor for class org.apache.hadoop.hbase.mapred.TableSplit
-
default constructor
- TableSplit(TableName, byte[], byte[], String) - Constructor for class org.apache.hadoop.hbase.mapred.TableSplit
-
Constructor
- TableSplit(byte[], byte[], byte[], String) - Constructor for class org.apache.hadoop.hbase.mapred.TableSplit
-
- TableSplit - Class in org.apache.hadoop.hbase.mapreduce
-
A table split corresponds to a key range (low, high) and an optional scanner.
- TableSplit() - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Default constructor.
- TableSplit(TableName, Scan, byte[], byte[], String) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Creates a new instance while assigning all variables.
- TableSplit(TableName, Scan, byte[], byte[], String, long) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Creates a new instance while assigning all variables.
- TableSplit(TableName, Scan, byte[], byte[], String, String, long) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Creates a new instance while assigning all variables.
- TableSplit(TableName, byte[], byte[], String) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Creates a new instance without a scanner.
- TableSplit(TableName, byte[], byte[], String, long) - Constructor for class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Creates a new instance without a scanner.
- TextSortReducer - Class in org.apache.hadoop.hbase.mapreduce
-
Emits Sorted KeyValues.
- TextSortReducer() - Constructor for class org.apache.hadoop.hbase.mapreduce.TextSortReducer
-
- TIMESTAMP_CONF_KEY - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv
-
- TIMESTAMPKEY_COLUMN_SPEC - Static variable in class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- toString() - Method in class org.apache.hadoop.hbase.mapred.TableSplit
-
- toString() - Method in class org.apache.hadoop.hbase.mapreduce.HashTable.TableHash
-
- toString() - Method in class org.apache.hadoop.hbase.mapreduce.TableSplit
-
Returns the details about this instance as a string.
- toString() - Method in class org.apache.hadoop.hbase.util.MapReduceExtendedCell
-
- ts - Variable in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
Timestamp for all inserted rows
- TsvImporterMapper - Class in org.apache.hadoop.hbase.mapreduce
-
Write table content out to files in hdfs.
- TsvImporterMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-
- TsvImporterTextMapper - Class in org.apache.hadoop.hbase.mapreduce
-
Write table content out to map output files.
- TsvImporterTextMapper() - Constructor for class org.apache.hadoop.hbase.mapreduce.TsvImporterTextMapper
-
- TsvParser(String, String) - Constructor for class org.apache.hadoop.hbase.mapreduce.ImportTsv.TsvParser
-
- ttl - Variable in class org.apache.hadoop.hbase.mapreduce.TsvImporterMapper
-