Deprecated Methods |
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(byte[])
|
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteException)
Use RemoteException.unwrapRemoteException() instead.
In fact we should look into deprecating this whole class - St.Ack 2010929 |
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983 |
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections(boolean)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteConnection(Configuration)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteStaleConnection(HConnection)
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName, boolean)
You can pass master flag but nothing special is done. |
org.apache.hadoop.hbase.zookeeper.ZKUtil.getChildDataAndWatchForNewChildren(ZooKeeperWatcher, String)
Unused |
org.apache.hadoop.hbase.client.HTable.getConnection()
This method will be changed from public to package protected. |
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(Configuration)
|
org.apache.hadoop.hbase.client.HConnection.getCurrentNrHRS()
This method will be changed from public to package protected. |
org.apache.hadoop.hbase.client.Mutation.getFamilyMap()
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[])
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List)
|
org.apache.hadoop.hbase.HTableDescriptor.getOwnerString()
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(byte[], byte[], boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.getRowOrBefore(byte[], byte[])
As of version 0.92 this method is deprecated without
replacement.
getRowOrBefore is used internally to find entries in .META. and makes
various assumptions about the table (which are true for .META. but not
in general) to be efficient. |
org.apache.hadoop.hbase.client.HTable.getScannerCaching()
Use Scan.setCaching(int) and Scan.getCaching() |
org.apache.hadoop.hbase.ClusterStatus.getServerInfo()
Use ClusterStatus.getServers() |
org.apache.hadoop.hbase.HTableDescriptor.getTableDir(Path, byte[])
|
org.apache.hadoop.hbase.client.HConnection.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(Pattern)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(String)
|
org.apache.hadoop.hbase.HRegionInfo.getVersion()
HRI is no longer a VersionedWritable |
org.apache.hadoop.hbase.client.HTable.getWriteBuffer()
since 0.96. This is an internal buffer that should not be read nor write. |
org.apache.hadoop.hbase.client.Mutation.getWriteToWAL()
|
org.apache.hadoop.hbase.HTableDescriptor.isDeferredLogFlush()
use HTableDescriptor.getDurability() |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[])
|
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[], byte[][])
|
org.apache.hadoop.hbase.client.HConnection.isTableDisabled(byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(byte[])
use HBaseAdmin.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableEnabled(byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, byte[])
use HBaseAdmin.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, String)
use HBaseAdmin.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, TableName)
use HBaseAdmin.isTableEnabled(org.apache.hadoop.hbase.TableName tableName) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(String)
use HBaseAdmin.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(TableName)
use HBaseAdmin.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[], byte[])
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[])
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[], boolean, boolean)
|
org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(Configuration, InputStream)
in 0.96 onwards. HBase will no longer rely on zoo.cfg
availability. |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[])
|
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
since 0.96 - Use HTableInterface.batch(java.util.List extends org.apache.hadoop.hbase.client.Row>, java.lang.Object[]) instead |
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
since 0.96 - Use HTableInterface.batchCallback(java.util.List extends org.apache.hadoop.hbase.client.Row>, java.lang.Object[], org.apache.hadoop.hbase.client.coprocessor.Batch.Callback) instead |
org.apache.hadoop.hbase.client.HTablePool.putTable(HTableInterface)
|
org.apache.hadoop.hbase.HTableDescriptor.readFields(DataInput)
Writables are going away. Use pb HTableDescriptor.parseFrom(byte[]) instead. |
org.apache.hadoop.hbase.HRegionInfo.readFields(DataInput)
Use protobuf deserialization instead. |
org.apache.hadoop.hbase.HColumnDescriptor.readFields(DataInput)
Writables are going away. Use pb HColumnDescriptor.parseFrom(byte[]) instead. |
org.apache.hadoop.hbase.master.RegionState.readFields(DataInput)
Writables are going away |
org.apache.hadoop.hbase.client.HConnection.relocateRegion(byte[], byte[])
|
org.apache.hadoop.hbase.HTableDescriptor.setDeferredLogFlush(boolean)
use HTableDescriptor.setDurability(Durability) |
org.apache.hadoop.hbase.HTableDescriptor.setName(byte[])
|
org.apache.hadoop.hbase.HTableDescriptor.setName(TableName)
|
org.apache.hadoop.hbase.HTableDescriptor.setOwner(User)
|
org.apache.hadoop.hbase.HTableDescriptor.setOwnerString(String)
|
org.apache.hadoop.hbase.client.HTable.setScannerCaching(int)
Use Scan.setCaching(int) |
org.apache.hadoop.hbase.client.Mutation.setWriteToWAL(boolean)
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(byte[], byte[], Object, HRegionLocation)
|
org.apache.hadoop.hbase.zookeeper.ZKUtil.updateExistingNodeData(ZooKeeperWatcher, String, byte[], int)
Unused |
org.apache.hadoop.hbase.catalog.CatalogTracker.waitForMetaServerConnection(long)
Use #getMetaServerConnection(long) |
org.apache.hadoop.hbase.HTableDescriptor.write(DataOutput)
Writables are going away.
Use MessageLite.toByteArray() instead. |
org.apache.hadoop.hbase.HRegionInfo.write(DataOutput)
Use protobuf serialization instead. See HRegionInfo.toByteArray() and
HRegionInfo.toDelimitedByteArray() |
org.apache.hadoop.hbase.HColumnDescriptor.write(DataOutput)
Writables are going away. Use HColumnDescriptor.toByteArray() instead. |
org.apache.hadoop.hbase.master.RegionState.write(DataOutput)
Writables are going away |