org.apache.hadoop.hbase.client.Put.add(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], byte[], long, byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], ByteBuffer, long, ByteBuffer)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabel(Configuration, String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabels(Configuration, String[])
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String)
Use addPeer(String, ReplicationPeerConfig, Map) instead.
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String, String) |
org.apache.hadoop.hbase.client.Result.addResults(ClientProtos.RegionLoadStats)
|
org.apache.hadoop.hbase.client.MetaScanner.allTableRegions(Configuration, Connection, TableName, boolean)
|
org.apache.hadoop.hbase.client.Table.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use Table.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.HTable.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use HTable.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.Table.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.HTable.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.clearAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HConnection.clearCaches(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(byte[]) |
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String, String)
|
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, User) |
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitHConnection(Connection, ServerName, AdminProtos.AdminService.BlockingInterface, ClientProtos.ClientService.BlockingInterface) |
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteException)
Use RemoteException.unwrapRemoteException() instead.
In fact we should look into deprecating this whole class - St.Ack 2010929
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections(boolean) |
org.apache.hadoop.hbase.client.HConnection.deleteCachedRegionLocation(HRegionLocation)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteConnection(Configuration) |
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteFamilyVersion(byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteStaleConnection(HConnection) |
org.apache.hadoop.hbase.client.HConnectionManager.execute(HConnectable)
Internal method, do not use thru HConnectionManager.
|
org.apache.hadoop.hbase.client.HTableInterface.exists(List)
|
org.apache.hadoop.hbase.client.HTable.exists(List) |
org.apache.hadoop.hbase.client.HBaseAdmin.flush(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.flush(String)
|
org.apache.hadoop.hbase.client.HTableInterface.flushCommits()
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName, boolean)
You can pass master flag but nothing special is done.
|
org.apache.hadoop.hbase.client.HTable.getAllRegionLocations()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.getAuths(Configuration, String)
|
org.apache.hadoop.hbase.zookeeper.ZKUtil.getChildDataAndWatchForNewChildren(ZooKeeperWatcher, String)
Unused
|
org.apache.hadoop.hbase.client.HConnection.getClient(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Result.getColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(String)
|
org.apache.hadoop.hbase.client.HTable.getConnection()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnection.getCurrentNrHRS()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.HColumnDescriptor.getDataBlockEncodingOnDisk()
|
org.apache.hadoop.hbase.HRegionInfo.getDaughterRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.HTable.getEndKeys()
|
org.apache.hadoop.hbase.client.Mutation.getFamilyMap()
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.MetaScanner.getHRegionInfo(Result)
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result, byte[])
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfoAndServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[]) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
|
org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService() |
org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
Since 0.96.0
|
org.apache.hadoop.hbase.client.HConnection.getMaster()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HRegionInfo.getMergeRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HTableDescriptor.getName()
|
org.apache.hadoop.hbase.filter.Filter.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterWrapper.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.getNonceGenerator()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HTableDescriptor.getOwnerString() |
org.apache.hadoop.hbase.MetaTableAccessor.getRegion(Connection, byte[])
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegionCount(Configuration, String) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[])
|
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[], boolean)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(byte[], byte[], boolean) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(String)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(TableName, byte[], boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HTable.getRegionLocations()
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[])
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[], boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTableInterface.getRowOrBefore(byte[], byte[])
As of version 0.92 this method is deprecated without
replacement. Since version 0.96+, you can use reversed scan.
getRowOrBefore is used internally to find entries in hbase:meta and makes
various assumptions about the table (which are true for hbase:meta but not
in general) to be efficient.
|
org.apache.hadoop.hbase.client.HTable.getRowOrBefore(byte[], byte[])
Use reversed scan instead.
|
org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRowOrBefore(ClientProtos.ClientService.BlockingInterface, byte[], byte[], byte[])
since 0.99 - use reversed scanner instead.
|
org.apache.hadoop.hbase.client.HTable.getScannerCaching()
|
org.apache.hadoop.hbase.HRegionInfo.getSeqNumDuringOpen(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.ClusterStatus.getServerInfo()
|
org.apache.hadoop.hbase.HRegionInfo.getServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.HTable.getStartEndKeys()
|
org.apache.hadoop.hbase.client.HTable.getStartKeys()
|
org.apache.hadoop.hbase.HTableDescriptor.getTableDir(Path, byte[]) |
org.apache.hadoop.hbase.HRegionInfo.getTableName()
|
org.apache.hadoop.hbase.client.HTableInterface.getTableName()
|
org.apache.hadoop.hbase.client.ClientScanner.getTableName()
|
org.apache.hadoop.hbase.HRegionInfo.getTableName(byte[])
|
org.apache.hadoop.hbase.client.HConnection.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(Pattern)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(String)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.getUserPermissions(Configuration, String)
|
org.apache.hadoop.hbase.HRegionInfo.getVersion()
HRI is no longer a VersionedWritable
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.getWriteBuffer()
|
org.apache.hadoop.hbase.client.HTable.getWriteBuffer()
since 0.96. This is an internal buffer that should not be read nor write.
|
org.apache.hadoop.hbase.client.HTableInterface.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Table.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Mutation.getWriteToWAL()
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.client.HTableInterface.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.client.HTable.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.isAccessControllerRunning(Configuration)
|
org.apache.hadoop.hbase.client.HTableInterface.isAutoFlush()
|
org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[], byte[][]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(TableName, byte[][])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isTableDisabled(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, TableName)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(TableName)
|
org.apache.hadoop.hbase.client.Result.list()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.listLabels(Configuration, String)
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.listPeers()
|
org.apache.hadoop.hbase.client.HConnection.listTableNames()
|
org.apache.hadoop.hbase.client.HConnection.listTables()
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[], boolean, boolean) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName, boolean, boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String, String)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainAndCacheToken(Configuration, UserGroupInformation)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainToken(Configuration)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(Configuration, UserGroupInformation, Job)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(JobConf, UserGroupInformation)
|
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
|
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback) |
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], List) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put, int) |
org.apache.hadoop.hbase.client.HTablePool.putTable(HTableInterface) |
org.apache.hadoop.hbase.client.Result.raw()
|
org.apache.hadoop.hbase.HRegionInfo.readFields(DataInput)
Use protobuf deserialization instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.HTableDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.client.HConnection.relocateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.relocateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.client.HBaseAdmin.rollHLogWriter(String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.setAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean)
in 0.96. When called with setAutoFlush(false), this function also
set clearBufferOnFail to true, which is unexpected but kept for historical reasons.
Replace it with setAutoFlush(false, false) if this is exactly what you want, though
this is the method you want for most cases.
|
org.apache.hadoop.hbase.client.HTable.setAutoFlush(boolean) |
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean, boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlushTo(boolean)
|
org.apache.hadoop.hbase.HColumnDescriptor.setEncodeOnDisk(boolean)
As of release 0.98
(HBASE-9870).
This will be removed in HBase 2.0.0. This method does nothing now.
|
org.apache.hadoop.hbase.client.Append.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Mutation.setFamilyMap(NavigableMap>)
|
org.apache.hadoop.hbase.client.Increment.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Put.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Delete.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.HColumnDescriptor.setKeepDeletedCells(boolean)
|
org.apache.hadoop.hbase.HTableDescriptor.setName(byte[]) |
org.apache.hadoop.hbase.HTableDescriptor.setName(TableName) |
org.apache.hadoop.hbase.HTableDescriptor.setOwner(User) |
org.apache.hadoop.hbase.HTableDescriptor.setOwnerString(String) |
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.setPeerTableCFs(String, String)
|
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(TableName, boolean)
does nothing since since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setScannerCaching(int)
|
org.apache.hadoop.hbase.client.HConnectionManager.setServerSideHConnectionRetries(Configuration, String, Log)
Internal method, do not use.
|
org.apache.hadoop.hbase.client.HTableInterface.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Append.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Mutation.setWriteToWAL(boolean)
|
org.apache.hadoop.hbase.client.Increment.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Put.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Delete.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheBloomsOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataInL1()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheIndexesOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCompressTags()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldEvictBlocksOnClose()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldPrefetchBlocksOnOpen()
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String, String)
|
org.apache.hadoop.hbase.filter.Filter.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterWrapper.transform(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(byte[], byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], byte[], Object, ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.zookeeper.ZKUtil.updateExistingNodeData(ZooKeeperWatcher, String, byte[], int)
Unused
|
org.apache.hadoop.hbase.HRegionInfo.write(DataOutput)
|
org.apache.hadoop.hbase.HColumnDescriptor.write(DataOutput)
|
org.apache.hadoop.hbase.HTableDescriptor.write(DataOutput)
Writables are going away.
Use MessageLite.toByteArray() instead.
|