|
||||||||||
| PREV CLASS NEXT CLASS | FRAMES NO FRAMES | |||||||||
| SUMMARY: NESTED | FIELD | CONSTR | METHOD | DETAIL: FIELD | CONSTR | METHOD | |||||||||
java.lang.Objectorg.apache.hadoop.hbase.mapred.TableMapReduceUtil
@Deprecated @InterfaceAudience.Public @InterfaceStability.Stable public class TableMapReduceUtil
Utility for TableMap and TableReduce
| Constructor Summary | |
|---|---|
TableMapReduceUtil()
Deprecated. |
|
| Method Summary | |
|---|---|
static void |
addDependencyJars(org.apache.hadoop.mapred.JobConf job)
Deprecated. |
static void |
initCredentials(org.apache.hadoop.mapred.JobConf job)
Deprecated. |
static void |
initMultiTableSnapshotMapperJob(Map<String,Collection<Scan>> snapshotScans,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
Deprecated. Sets up the job for reading from one or more multiple table snapshots, with one or more scans per snapshot. |
static void |
initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Use this before submitting a TableMap job. |
static void |
initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars)
Deprecated. |
static void |
initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
Class<? extends org.apache.hadoop.mapred.InputFormat> inputFormat)
Deprecated. Use this before submitting a TableMap job. |
static void |
initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Use this before submitting a TableReduce job. |
static void |
initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job,
Class partitioner)
Deprecated. Use this before submitting a TableReduce job. |
static void |
initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job,
Class partitioner,
boolean addDependencyJars)
Deprecated. Use this before submitting a TableReduce job. |
static void |
initTableSnapshotMapJob(String snapshotName,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
Deprecated. Sets up the job for reading from a table snapshot. |
static void |
limitNumMapTasks(String table,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Ensures that the given number of map tasks for the given job configuration does not exceed the number of regions for the given table. |
static void |
limitNumReduceTasks(String table,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Ensures that the given number of reduce tasks for the given job configuration does not exceed the number of regions for the given table. |
static void |
setNumMapTasks(String table,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Sets the number of map tasks for the given job configuration to the number of regions the given table has. |
static void |
setNumReduceTasks(String table,
org.apache.hadoop.mapred.JobConf job)
Deprecated. Sets the number of reduce tasks for the given job configuration to the number of regions the given table has. |
static void |
setScannerCaching(org.apache.hadoop.mapred.JobConf job,
int batchSize)
Deprecated. Sets the number of rows to return and cache with each scanner iteration. |
| Methods inherited from class java.lang.Object |
|---|
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait |
| Constructor Detail |
|---|
public TableMapReduceUtil()
| Method Detail |
|---|
public static void initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job)
table - The table name to read from.columns - The columns to scan.mapper - The mapper class to use.outputKeyClass - The class of the output key.outputValueClass - The class of the output value.job - The current job configuration to adjust.
public static void initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars)
public static void initTableMapJob(String table,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
Class<? extends org.apache.hadoop.mapred.InputFormat> inputFormat)
table - The table name to read from.columns - The columns to scan.mapper - The mapper class to use.outputKeyClass - The class of the output key.outputValueClass - The class of the output value.job - The current job configuration to adjust.addDependencyJars - upload HBase jars and jars for any of the configured
job classes via the distributed cache (tmpjars).
public static void initMultiTableSnapshotMapperJob(Map<String,Collection<Scan>> snapshotScans,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
throws IOException
snapshotScans - map of snapshot name to scans on that snapshot.mapper - The mapper class to use.outputKeyClass - The class of the output key.outputValueClass - The class of the output value.job - The current job to adjust. Make sure the passed job is
carrying all necessary HBase configuration.addDependencyJars - upload HBase jars and jars for any of the configured
job classes via the distributed cache (tmpjars).
IOException
public static void initTableSnapshotMapJob(String snapshotName,
String columns,
Class<? extends TableMap> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapred.JobConf job,
boolean addDependencyJars,
org.apache.hadoop.fs.Path tmpRestoreDir)
throws IOException
snapshotName - The name of the snapshot (of a table) to read from.columns - The columns to scan.mapper - The mapper class to use.outputKeyClass - The class of the output key.outputValueClass - The class of the output value.job - The current job to adjust. Make sure the passed job is
carrying all necessary HBase configuration.addDependencyJars - upload HBase jars and jars for any of the configured
job classes via the distributed cache (tmpjars).tmpRestoreDir - a temporary directory to copy the snapshot files into. Current user should
have write permissions to this directory, and this should not be a subdirectory of rootdir.
After the job is finished, restore directory can be deleted.
IOException - When setting up the details fails.TableSnapshotInputFormat
public static void initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job)
throws IOException
table - The output table.reducer - The reducer class to use.job - The current job configuration to adjust.
IOException - When determining the region count fails.
public static void initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job,
Class partitioner)
throws IOException
table - The output table.reducer - The reducer class to use.job - The current job configuration to adjust.partitioner - Partitioner to use. Pass null to use
default partitioner.
IOException - When determining the region count fails.
public static void initTableReduceJob(String table,
Class<? extends TableReduce> reducer,
org.apache.hadoop.mapred.JobConf job,
Class partitioner,
boolean addDependencyJars)
throws IOException
table - The output table.reducer - The reducer class to use.job - The current job configuration to adjust.partitioner - Partitioner to use. Pass null to use
default partitioner.addDependencyJars - upload HBase jars and jars for any of the configured
job classes via the distributed cache (tmpjars).
IOException - When determining the region count fails.
public static void initCredentials(org.apache.hadoop.mapred.JobConf job)
throws IOException
IOException
public static void limitNumReduceTasks(String table,
org.apache.hadoop.mapred.JobConf job)
throws IOException
table - The table to get the region count for.job - The current job configuration to adjust.
IOException - When retrieving the table details fails.
public static void limitNumMapTasks(String table,
org.apache.hadoop.mapred.JobConf job)
throws IOException
table - The table to get the region count for.job - The current job configuration to adjust.
IOException - When retrieving the table details fails.
public static void setNumReduceTasks(String table,
org.apache.hadoop.mapred.JobConf job)
throws IOException
table - The table to get the region count for.job - The current job configuration to adjust.
IOException - When retrieving the table details fails.
public static void setNumMapTasks(String table,
org.apache.hadoop.mapred.JobConf job)
throws IOException
table - The table to get the region count for.job - The current job configuration to adjust.
IOException - When retrieving the table details fails.
public static void setScannerCaching(org.apache.hadoop.mapred.JobConf job,
int batchSize)
job - The current job configuration to adjust.batchSize - The number of rows to return in batch with each scanner
iteration.
public static void addDependencyJars(org.apache.hadoop.mapred.JobConf job)
throws IOException
IOExceptionTableMapReduceUtil.addDependencyJars(org.apache.hadoop.mapreduce.Job)
|
||||||||||
| PREV CLASS NEXT CLASS | FRAMES NO FRAMES | |||||||||
| SUMMARY: NESTED | FIELD | CONSTR | METHOD | DETAIL: FIELD | CONSTR | METHOD | |||||||||