org.apache.hadoop.hbase.client.Put.add(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], byte[], long, byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], ByteBuffer, long, ByteBuffer)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabel(Configuration, String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabels(Configuration, String[])
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String)
Use addPeer(String, ReplicationPeerConfig, Map) instead.
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String, String) |
org.apache.hadoop.hbase.client.Result.addResults(ClientProtos.RegionLoadStats)
|
org.apache.hadoop.hbase.client.MetaScanner.allTableRegions(Configuration, Connection, TableName, boolean)
|
org.apache.hadoop.hbase.client.Table.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use Table.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.HTable.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use HTable.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.Table.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.HTable.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.protobuf.RequestConverter.buildBulkLoadHFileRequest(List>, byte[], boolean)
use buildBulkLoadHFileRequest(final List> familyPaths,
final byte[] regionName, boolean assignSeqNum, List clusterIds)
|
org.apache.hadoop.hbase.quotas.QuotaSettings.buildSetQuotaRequestProto(QuotaSettings)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.protobuf.ProtobufUtil.bulkLoadHFile(ClientProtos.ClientService.BlockingInterface, List>, byte[], boolean)
use bulkLoadHFile(final ClientService.BlockingInterface client,
final List> familyPaths, final byte[] regionName, boolean assignSeqNum,
List clusterIds) instead.
|
org.apache.hadoop.hbase.client.coprocessor.SecureBulkLoadClient.bulkLoadHFiles(List>, Token>, String, byte[]) |
org.apache.hadoop.hbase.security.visibility.VisibilityClient.clearAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HConnection.clearCaches(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(byte[]) |
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String, String)
|
org.apache.hadoop.hbase.HTableDescriptor.convert() |
org.apache.hadoop.hbase.HColumnDescriptor.convert() |
org.apache.hadoop.hbase.ClusterStatus.convert() |
org.apache.hadoop.hbase.ClusterStatus.convert(ClusterStatusProtos.ClusterStatus) |
org.apache.hadoop.hbase.HColumnDescriptor.convert(HBaseProtos.ColumnFamilySchema) |
org.apache.hadoop.hbase.HTableDescriptor.convert(HBaseProtos.TableSchema) |
org.apache.hadoop.hbase.client.Result.createCompleteResult(List) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, User) |
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitHConnection(Connection, ServerName, AdminProtos.AdminService.BlockingInterface, ClientProtos.ClientService.BlockingInterface) |
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteException)
Use RemoteException.unwrapRemoteException() instead.
In fact we should look into deprecating this whole class - St.Ack 2010929
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections(boolean) |
org.apache.hadoop.hbase.client.HConnection.deleteCachedRegionLocation(HRegionLocation)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteConnection(Configuration) |
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteFamilyVersion(byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteStaleConnection(HConnection) |
org.apache.hadoop.hbase.client.HConnectionManager.execute(HConnectable)
Internal method, do not use thru HConnectionManager.
|
org.apache.hadoop.hbase.client.HTableInterface.exists(List)
|
org.apache.hadoop.hbase.client.HTable.exists(List) |
org.apache.hadoop.hbase.client.HBaseAdmin.flush(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.flush(String)
|
org.apache.hadoop.hbase.client.HTableInterface.flushCommits()
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName, boolean)
You can pass master flag but nothing special is done.
|
org.apache.hadoop.hbase.client.HTable.getAllRegionLocations()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.getAuths(Configuration, String)
|
org.apache.hadoop.hbase.zookeeper.ZKUtil.getChildDataAndWatchForNewChildren(ZooKeeperWatcher, String)
Unused
|
org.apache.hadoop.hbase.client.HConnection.getClient(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Result.getColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(String)
|
org.apache.hadoop.hbase.client.Admin.getCompactionStateForRegion(byte[]) |
org.apache.hadoop.hbase.client.HTable.getConnection()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnection.getCurrentNrHRS()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.HColumnDescriptor.getDataBlockEncodingOnDisk()
|
org.apache.hadoop.hbase.HRegionInfo.getDaughterRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.HTable.getEndKeys()
|
org.apache.hadoop.hbase.client.Mutation.getFamilyMap()
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.MetaScanner.getHRegionInfo(Result)
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result, byte[])
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfoAndServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[]) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
|
org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
Since 0.96.0
|
org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService() |
org.apache.hadoop.hbase.client.HConnection.getMaster()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HRegionInfo.getMergeRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HTableDescriptor.getName()
|
org.apache.hadoop.hbase.filter.Filter.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterWrapper.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.getNonceGenerator()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HTableDescriptor.getOwnerString()
since 0.94.1
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegion(Connection, byte[])
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegionCount(Configuration, String) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[])
|
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[], boolean)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(byte[], byte[], boolean) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(String)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(TableName, byte[], boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HTable.getRegionLocations()
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[])
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[], boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTableInterface.getRowOrBefore(byte[], byte[])
As of version 0.92 this method is deprecated without
replacement. Since version 0.96+, you can use reversed scan.
getRowOrBefore is used internally to find entries in hbase:meta and makes
various assumptions about the table (which are true for hbase:meta but not
in general) to be efficient.
|
org.apache.hadoop.hbase.client.HTable.getRowOrBefore(byte[], byte[])
Use reversed scan instead.
|
org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRowOrBefore(ClientProtos.ClientService.BlockingInterface, byte[], byte[], byte[])
since 0.99 - use reversed scanner instead.
|
org.apache.hadoop.hbase.client.Table.getRpcTimeout()
Use getReadRpcTimeout or getWriteRpcTimeout instead
|
org.apache.hadoop.hbase.client.HTable.getRpcTimeout() |
org.apache.hadoop.hbase.client.Scan.getScanMetrics()
|
org.apache.hadoop.hbase.client.HTable.getScannerCaching()
|
org.apache.hadoop.hbase.HRegionInfo.getSeqNumDuringOpen(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.ClusterStatus.getServerInfo()
|
org.apache.hadoop.hbase.HRegionInfo.getServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException.getSnapshotDescription()
since 1.3.0, will be removed in 3.0.0
|
org.apache.hadoop.hbase.client.HTable.getStartEndKeys()
|
org.apache.hadoop.hbase.client.HTable.getStartKeys()
|
org.apache.hadoop.hbase.client.Result.getStats() |
org.apache.hadoop.hbase.HTableDescriptor.getTableDir(Path, byte[]) |
org.apache.hadoop.hbase.HRegionInfo.getTableName()
|
org.apache.hadoop.hbase.client.HTableInterface.getTableName()
|
org.apache.hadoop.hbase.client.ClientScanner.getTableName()
|
org.apache.hadoop.hbase.HRegionInfo.getTableName(byte[])
|
org.apache.hadoop.hbase.client.HConnection.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(Pattern)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(String)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.getUserPermissions(Configuration, String)
|
org.apache.hadoop.hbase.HRegionInfo.getVersion()
HRI is no longer a VersionedWritable
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.getWriteBuffer()
|
org.apache.hadoop.hbase.client.HTable.getWriteBuffer()
since 0.96. This is an internal buffer that should not be read nor write.
|
org.apache.hadoop.hbase.client.HTableInterface.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Table.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Mutation.getWriteToWAL()
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.client.HTableInterface.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.client.HTable.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.isAccessControllerRunning(Configuration)
|
org.apache.hadoop.hbase.client.HTableInterface.isAutoFlush()
|
org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.Result.isPartial()
|
org.apache.hadoop.hbase.client.Admin.isSnapshotFinished(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[], byte[][]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(TableName, byte[][])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isTableDisabled(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, TableName)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(TableName)
|
org.apache.hadoop.hbase.client.Result.list()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.listLabels(Configuration, String)
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.listPeers()
|
org.apache.hadoop.hbase.client.Admin.listSnapshots() |
org.apache.hadoop.hbase.client.Admin.listSnapshots(Pattern) |
org.apache.hadoop.hbase.client.Admin.listSnapshots(String) |
org.apache.hadoop.hbase.client.HConnection.listTableNames()
|
org.apache.hadoop.hbase.client.HConnection.listTables()
|
org.apache.hadoop.hbase.client.Admin.listTableSnapshots(Pattern, Pattern) |
org.apache.hadoop.hbase.client.Admin.listTableSnapshots(String, String) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[], boolean, boolean) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName, boolean, boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String, String)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainAndCacheToken(Configuration, UserGroupInformation)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainToken(Configuration)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(Configuration, UserGroupInformation, Job)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(JobConf, UserGroupInformation)
|
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
|
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback) |
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], List) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put, int) |
org.apache.hadoop.hbase.client.HTablePool.putTable(HTableInterface) |
org.apache.hadoop.hbase.client.Result.raw()
|
org.apache.hadoop.hbase.HRegionInfo.readFields(DataInput)
Use protobuf deserialization instead.
|
org.apache.hadoop.hbase.HTableDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.HColumnDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.client.HConnection.relocateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.relocateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.client.HBaseAdmin.rollHLogWriter(String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.setAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean)
in 0.96. When called with setAutoFlush(false), this function also
set clearBufferOnFail to true, which is unexpected but kept for historical reasons.
Replace it with setAutoFlush(false, false) if this is exactly what you want, though
this is the method you want for most cases.
|
org.apache.hadoop.hbase.client.HTable.setAutoFlush(boolean) |
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean, boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlushTo(boolean)
|
org.apache.hadoop.hbase.HColumnDescriptor.setEncodeOnDisk(boolean)
As of release 0.98
(HBASE-9870).
This will be removed in HBase 2.0.0. This method does nothing now.
|
org.apache.hadoop.hbase.client.Delete.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Put.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Increment.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Append.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Mutation.setFamilyMap(NavigableMap>)
|
org.apache.hadoop.hbase.HColumnDescriptor.setKeepDeletedCells(boolean)
|
org.apache.hadoop.hbase.HTableDescriptor.setName(byte[]) |
org.apache.hadoop.hbase.HTableDescriptor.setName(TableName) |
org.apache.hadoop.hbase.HTableDescriptor.setOwner(User)
since 0.94.1
|
org.apache.hadoop.hbase.HTableDescriptor.setOwnerString(String)
since 0.94.1
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.setPeerTableCFs(String, String)
|
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(TableName, boolean)
does nothing since since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.Table.setRpcTimeout(int)
Use setReadRpcTimeout or setWriteRpcTimeout instead
|
org.apache.hadoop.hbase.client.HTable.setRpcTimeout(int) |
org.apache.hadoop.hbase.client.HTable.setScannerCaching(int)
|
org.apache.hadoop.hbase.client.HConnectionManager.setServerSideHConnectionRetries(Configuration, String, Log)
Internal method, do not use.
|
org.apache.hadoop.hbase.client.Scan.setStartRow(byte[])
use Scan.withStartRow(byte[]) instead. This method may change the inclusive of
the stop row to keep compatible with the old behavior.
|
org.apache.hadoop.hbase.client.Result.setStatistics(ClientProtos.RegionLoadStats) |
org.apache.hadoop.hbase.client.Scan.setStopRow(byte[])
use Scan.withStartRow(byte[]) instead. This method may change the inclusive of
the stop row to keep compatible with the old behavior.
|
org.apache.hadoop.hbase.quotas.QuotaSettings.setupSetQuotaRequest(MasterProtos.SetQuotaRequest.Builder)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.client.HTableInterface.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Delete.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Put.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Increment.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Append.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Mutation.setWriteToWAL(boolean)
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheBloomsOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataInL1()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheIndexesOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCompressTags()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldEvictBlocksOnClose()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldPrefetchBlocksOnOpen()
|
org.apache.hadoop.hbase.client.Admin.snapshot(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.client.Admin.snapshot(String, TableName, HBaseProtos.SnapshotDescription.Type) |
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String, String)
|
org.apache.hadoop.hbase.client.Admin.takeSnapshotAsync(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.protobuf.ProtobufUtil.toBulkLoadDescriptor(TableName, ByteString, Map>, Map, long)
use toBulkLoadDescriptor(TableName tableName, ByteString encodedRegionName,
Map> storeFiles, Map storeFilesSize, long bulkloadSeqId,
List clusterIds) instead.
|
org.apache.hadoop.hbase.filter.FilterList.toString(int) |
org.apache.hadoop.hbase.filter.Filter.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterWrapper.transform(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(byte[], byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], byte[], Object, ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.zookeeper.ZKUtil.updateExistingNodeData(ZooKeeperWatcher, String, byte[], int)
Unused
|
org.apache.hadoop.hbase.HRegionInfo.write(DataOutput)
|
org.apache.hadoop.hbase.HTableDescriptor.write(DataOutput)
Writables are going away.
Use MessageLite.toByteArray() instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.write(DataOutput)
|