@InterfaceAudience.Private public final class ProtobufUtil extends Object
限定符和类型 | 字段和说明 |
---|---|
static byte[] |
PB_MAGIC
Magic we put ahead of a serialized protobuf message.
|
public static final byte[] PB_MAGIC
public static byte[] prependPBMagic(byte[] bytes)
PB_MAGIC
, to flag what
follows as a protobuf in hbase. Prepend these bytes to all content written to znodes, etc.bytes
- Bytes to decoratebytes
with magic prepended (Creates a new
byte array that is bytes.length
plus PB_MAGIC
.length.public static boolean isPBMagicPrefix(byte[] bytes)
bytes
- Bytes to check.bytes
has PB_MAGIC
for a prefix.public static boolean isPBMagicPrefix(byte[] bytes, int offset, int len)
bytes
- Bytes to check.offset
- offset to start atlen
- length to usebytes
has PB_MAGIC
for a prefix.public static void expectPBMagicPrefix(byte[] bytes) throws DeserializationException
bytes
- bytes to checkDeserializationException
- if we are missing the pb magic prefixpublic static int lengthOfPBMagic()
PB_MAGIC
public static IOException getRemoteException(com.google.protobuf.ServiceException se)
se
- ServiceException that wraps IO exception thrown by the serverpublic static IOException getServiceException(com.google.protobuf.ServiceException se)
se
- ServiceException that wraps IO exception thrown by the serverpublic static IOException handleRemoteException(Exception e)
getRemoteException(ServiceException)
but more generic, able to handle more than
just ServiceException
. Prefer this method to
getRemoteException(ServiceException)
because trying to
contain direct protobuf references.e
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName toServerName(ServerName serverName)
serverName
- the ServerName to converttoServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName)
public static ServerName toServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName proto)
proto
- the protocol buffer ServerName to convertpublic static List<ServerName> toServerNameList(List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName> proto)
proto
- protocol buffer ServerNameListpublic static HTableDescriptor[] getHTableDescriptorArray(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.GetTableDescriptorsResponse proto)
proto
- the GetTableDescriptorsResponsepublic static byte[][] getSplitKeysArray(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.CreateTableRequest proto)
proto
- the CreateTableRequestpublic static Durability toDurability(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Durability proto)
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Durability toDurability(Durability d)
public static Get toGet(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get proto) throws IOException
proto
- the protocol buffer Get to convertIOException
public static Consistency toConsistency(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Consistency consistency)
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Consistency toConsistency(Consistency consistency)
public static Put toPut(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto) throws IOException
proto
- The protocol buffer MutationProto to convertIOException
public static Put toPut(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto, CellScanner cellScanner) throws IOException
proto
- The protocol buffer MutationProto to convertcellScanner
- If non-null, the Cell data that goes with this proto.IOException
public static Delete toDelete(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto) throws IOException
proto
- the protocol buffer Mutate to convertIOException
public static Delete toDelete(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto, CellScanner cellScanner) throws IOException
proto
- the protocol buffer Mutate to convertcellScanner
- if non-null, the data that goes with this delete.IOException
public static Append toAppend(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto, CellScanner cellScanner) throws IOException
cellScanner
- proto
- the protocol buffer Mutate to convertIOException
public static Mutation toMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto) throws IOException
proto
- the protocol buffer Mutate to convertIOException
public static Increment toIncrement(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto, CellScanner cellScanner) throws IOException
proto
- the protocol buffer Mutate to convertIOException
public static Get toGet(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto proto, CellScanner cellScanner) throws IOException
proto
- the protocol buffer Mutate to convert.cellScanner
- IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.ReadType toReadType(Scan.ReadType readType)
public static Scan.ReadType toReadType(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.ReadType readType)
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan toScan(Scan scan) throws IOException
scan
- the client Scan to convertIOException
public static Scan toScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan proto) throws IOException
proto
- the protocol buffer Scan to convertIOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor toCursor(Cursor cursor)
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor toCursor(Cell cell)
public static Cursor toCursor(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Cursor cursor)
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get toGet(Get get) throws IOException
get
- the client GetIOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutation(Increment increment, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builder, long nonce)
increment
- public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation, long nonce) throws IOException
type
- mutation
- IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builder) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builder, long nonce) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutationNoData(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builder) throws IOException
type
- mutation
- builder
- IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutationNoData(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation) throws IOException
type
- mutation
- IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto toMutationNoData(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType type, Mutation mutation, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builder, long nonce) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result toResult(Result result)
result
- the client Result to convertpublic static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result toResult(Result result, boolean encodeTags)
result
- the client Result to convertencodeTags
- whether to includeTags in converted protobuf result or not
When @encodeTags is set to true, it will return all the tags in the response.
These tags may contain some sensitive data like acl permissions, etc.
Only the tools like Export, Import which needs to take backup needs to set
it to true so that cell tags are persisted in backup.
Refer to HBASE-25246 for more context.public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result toResult(boolean existence, boolean stale)
existence
- the client existence to sendpublic static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result toResultNoData(Result result)
result
- the client Result to convertpublic static Result toResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result proto)
proto
- the protocol buffer Result to convertpublic static Result toResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result proto, boolean decodeTags)
proto
- the protocol buffer Result to convertdecodeTags
- whether to decode tags into converted client Result
When @decodeTags is set to true, it will decode all the tags from the
response. These tags may contain some sensitive data like acl permissions,
etc. Only the tools like Export, Import which needs to take backup needs to
set it to true so that cell tags are persisted in backup.
Refer to HBASE-25246 for more context.public static Result toResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result proto, CellScanner scanner) throws IOException
proto
- the protocol buffer Result to convertscanner
- Optional cell scanner.IOException
public static org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator toComparator(ByteArrayComparable comparator)
comparator
- the ByteArrayComparable to convertpublic static ByteArrayComparable toComparator(org.apache.hadoop.hbase.protobuf.generated.ComparatorProtos.Comparator proto) throws IOException
proto
- the protocol buffer Comparator to convertIOException
public static Filter toFilter(org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter proto) throws IOException
proto
- the protocol buffer Filter to convertIOException
public static org.apache.hadoop.hbase.protobuf.generated.FilterProtos.Filter toFilter(Filter filter) throws IOException
filter
- the Filter to convertIOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType toDeleteType(KeyValue.Type type) throws IOException
type
- IOException
public static KeyValue.Type fromDeleteType(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.DeleteType type) throws IOException
type
- The DeleteTypeIOException
public static Throwable toException(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair parameter) throws IOException
parameter
- the protocol buffer Parameter to convertIOException
- if failed to deserialize the parameter@Deprecated public static Result getRowOrBefore(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, byte[] regionName, byte[] row, byte[] family) throws IOException
client
- regionName
- row
- family
- IOException
public static boolean bulkLoadHFile(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, List<Pair<byte[],String>> familyPaths, byte[] regionName, boolean assignSeqNum) throws IOException
client
- familyPaths
- regionName
- assignSeqNum
- IOException
public static boolean bulkLoadHFile(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, List<Pair<byte[],String>> familyPaths, byte[] regionName, boolean assignSeqNum, List<String> clusterIds) throws IOException
client
- familyPaths
- regionName
- assignSeqNum
- IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceResponse execService(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall call, byte[] regionName) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceResponse execService(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.MasterProtos.MasterService.BlockingInterface client, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall call) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceResponse execRegionServerService(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CoprocessorServiceCall call) throws IOException
client
- call
- IOException
public static <T extends com.google.protobuf.Service> T newServiceStub(Class<T> service, com.google.protobuf.RpcChannel channel) throws Exception
Exception
public static HRegionInfo getRegionInfo(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, byte[] regionName) throws IOException
admin
- regionName
- IOException
public static void closeRegion(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, ServerName server, byte[] regionName, boolean transitionInZK) throws IOException
admin
- regionName
- transitionInZK
- IOException
public static boolean closeRegion(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, ServerName server, byte[] regionName, int versionOfClosingNode, ServerName destinationServer, boolean transitionInZK) throws IOException
admin
- regionName
- versionOfClosingNode
- IOException
public static void warmupRegion(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, HRegionInfo regionInfo) throws IOException
admin
- regionInfo
- IOException
public static void openRegion(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, ServerName server, HRegionInfo region) throws IOException
admin
- region
- IOException
public static List<HRegionInfo> getOnlineRegions(org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin) throws IOException
admin
- IOException
public static List<HRegionInfo> getOnlineRegions(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin) throws IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo getServerInfo(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin) throws IOException
IOException
public static List<String> getStoreFiles(org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, byte[] regionName, byte[] family) throws IOException
IOException
public static List<String> getStoreFiles(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, byte[] regionName, byte[] family) throws IOException
IOException
public static void split(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, HRegionInfo hri, byte[] splitPoint) throws IOException
admin
- hri
- splitPoint
- IOException
public static void mergeRegions(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin, HRegionInfo region_a, HRegionInfo region_b, boolean forcible, User user) throws IOException
admin
- region_a
- region_b
- forcible
- true if do a compulsory merge, otherwise we will only merge
two adjacent regionsuser
- effective userIOException
public static long getTotalRequestsCount(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionLoad rl)
public static byte[] toDelimitedByteArray(com.google.protobuf.Message m) throws IOException
m
- Message to get delimited pb serialization of (with pb magic prefix)IOException
public static Permission toPermission(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission proto)
proto
- the protobuf Permissionpublic static TablePermission toTablePermission(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission proto)
proto
- the protobuf Permissionpublic static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission toPermission(Permission perm)
perm
- the client Permissionpublic static List<Permission.Action> toPermissionActions(List<org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action> protoActions)
protoActions
- the list of protobuf Actionspublic static Permission.Action toPermissionAction(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action action)
action
- the protobuf Actionpublic static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.Permission.Action toPermissionAction(Permission.Action action)
action
- the client Actionpublic static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission toUserPermission(UserPermission perm)
perm
- the client UserPermissionpublic static UserPermission toUserPermission(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UserPermission proto)
proto
- the protobuf UserPermissionpublic static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions toUserTablePermissions(com.google.common.collect.ListMultimap<String,TablePermission> perm)
perm
- the list of user and table permissionspublic static void grant(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, boolean mergeExistingPermissions, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyuserShortName
- the short name of the user to grant permissionsactions
- the permissions to be grantedcom.google.protobuf.ServiceException
public static void grant(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, TableName tableName, byte[] f, byte[] q, boolean mergeExistingPermissions, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyuserShortName
- the short name of the user to grant permissionstableName
- optional table namef
- optional column familyq
- optional qualifieractions
- the permissions to be grantedcom.google.protobuf.ServiceException
public static void grant(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, String namespace, boolean mergeExistingPermissions, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxynamespace
- the short name of the user to grant permissionsactions
- the permissions to be grantedcom.google.protobuf.ServiceException
public static void revoke(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyuserShortName
- the short name of the user to revoke permissionsactions
- the permissions to be revokedcom.google.protobuf.ServiceException
public static void revoke(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, TableName tableName, byte[] f, byte[] q, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyuserShortName
- the short name of the user to revoke permissionstableName
- optional table namef
- optional column familyq
- optional qualifieractions
- the permissions to be revokedcom.google.protobuf.ServiceException
public static void revoke(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, String userShortName, String namespace, Permission.Action... actions) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyuserShortName
- the short name of the user to revoke permissionsnamespace
- optional table nameactions
- the permissions to be revokedcom.google.protobuf.ServiceException
public static List<UserPermission> getUserPermissions(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxycom.google.protobuf.ServiceException
public static List<UserPermission> getUserPermissions(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, TableName t) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxyt
- optional table namecom.google.protobuf.ServiceException
public static List<UserPermission> getUserPermissions(com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol, byte[] namespace) throws com.google.protobuf.ServiceException
It's also called by the shell, in case you want to find references.
protocol
- the AccessControlService protocol proxynamespace
- name of the namespacecom.google.protobuf.ServiceException
public static com.google.common.collect.ListMultimap<String,TablePermission> toUserTablePermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions proto)
proto
- the protobuf UserPermissionpublic static org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.Token toToken(org.apache.hadoop.security.token.Token<AuthenticationTokenIdentifier> token)
token
- the Token instance to copypublic static org.apache.hadoop.security.token.Token<AuthenticationTokenIdentifier> toToken(org.apache.hadoop.hbase.protobuf.generated.AuthenticationProtos.Token proto)
proto
- the protobuf Token messagepublic static String getRegionEncodedName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier) throws DoNotRetryIOException
regionSpecifier
- the region specifierDoNotRetryIOException
- if the specifier type is unsupportedpublic static ScanMetrics toScanMetrics(byte[] bytes)
public static org.apache.hadoop.hbase.protobuf.generated.MapReduceProtos.ScanMetrics toScanMetrics(ScanMetrics scanMetrics, boolean reset)
public static void toIOException(com.google.protobuf.ServiceException se) throws IOException
se
- the ServiceException
instance to convert into an IOException
IOException
public static org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell toCell(Cell kv, boolean encodeTags)
public static Cell toCell(org.apache.hadoop.hbase.protobuf.generated.CellProtos.Cell cell, boolean decodeTags)
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NamespaceDescriptor toProtoNamespaceDescriptor(NamespaceDescriptor ns)
public static NamespaceDescriptor toNamespaceDescriptor(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NamespaceDescriptor desc) throws IOException
IOException
public static <T extends com.google.protobuf.Message> T getParsedGenericInstance(Class<?> runtimeClass, int position, com.google.protobuf.ByteString b) throws IOException
runtimeClass
- the runtime type of the classposition
- the position of the argument in the class declarationb
- the ByteString which should be parsed to get the instance createdIOException
public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor toCompactionDescriptor(HRegionInfo info, byte[] family, List<org.apache.hadoop.fs.Path> inputPaths, List<org.apache.hadoop.fs.Path> outputPaths, org.apache.hadoop.fs.Path storeDir)
public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.CompactionDescriptor toCompactionDescriptor(HRegionInfo info, byte[] regionName, byte[] family, List<org.apache.hadoop.fs.Path> inputPaths, List<org.apache.hadoop.fs.Path> outputPaths, org.apache.hadoop.fs.Path storeDir)
public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FlushDescriptor toFlushDescriptor(org.apache.hadoop.hbase.protobuf.generated.WALProtos.FlushDescriptor.FlushAction action, HRegionInfo hri, long flushSeqId, Map<byte[],List<org.apache.hadoop.fs.Path>> committedFiles)
public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.RegionEventDescriptor toRegionEventDescriptor(org.apache.hadoop.hbase.protobuf.generated.WALProtos.RegionEventDescriptor.EventType eventType, HRegionInfo hri, long seqId, ServerName server, Map<byte[],List<org.apache.hadoop.fs.Path>> storeFiles)
public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.RegionEventDescriptor toRegionEventDescriptor(org.apache.hadoop.hbase.protobuf.generated.WALProtos.RegionEventDescriptor.EventType eventType, byte[] tableNameAsBytes, byte[] encodedNameAsBytes, byte[] regionNameAsBytes, long seqId, ServerName server, Map<byte[],List<org.apache.hadoop.fs.Path>> storeFiles)
public static String getShortTextFormat(com.google.protobuf.Message m)
m
- m
public static SlowLogParams getSlowLogParams(com.google.protobuf.Message message)
message
- Message object Message
public static TableName toTableName(org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName tableNamePB)
public static org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName toProtoTableName(TableName tableName)
public static TableName[] getTableNameArray(List<org.apache.hadoop.hbase.protobuf.generated.TableProtos.TableName> tableNamesList)
public static CellVisibility toCellVisibility(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CellVisibility proto)
proto
- public static CellVisibility toCellVisibility(byte[] protoBytes) throws DeserializationException
protoBytes
- DeserializationException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CellVisibility toCellVisibility(CellVisibility cellVisibility)
cellVisibility
- public static Authorizations toAuthorizations(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Authorizations proto)
proto
- public static Authorizations toAuthorizations(byte[] protoBytes) throws DeserializationException
protoBytes
- DeserializationException
public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Authorizations toAuthorizations(Authorizations authorizations)
authorizations
- public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions toUsersAndPermissions(String user, Permission perms)
public static org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions toUsersAndPermissions(com.google.common.collect.ListMultimap<String,Permission> perms)
public static com.google.common.collect.ListMultimap<String,Permission> toUsersAndPermissions(org.apache.hadoop.hbase.protobuf.generated.AccessControlProtos.UsersAndPermissions proto)
public static TimeUnit toTimeUnit(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeUnit proto)
proto
- public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TimeUnit toProtoTimeUnit(TimeUnit timeUnit)
timeUnit
- public static ThrottleType toThrottleType(org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.ThrottleType proto)
proto
- public static org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.ThrottleType toProtoThrottleType(ThrottleType type)
type
- public static QuotaScope toQuotaScope(org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.QuotaScope proto)
proto
- public static org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.QuotaScope toProtoQuotaScope(QuotaScope scope)
scope
- public static QuotaType toQuotaScope(org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.QuotaType proto)
proto
- public static org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.QuotaType toProtoQuotaScope(QuotaType type)
type
- public static org.apache.hadoop.hbase.protobuf.generated.QuotaProtos.TimedQuota toTimedQuota(long limit, TimeUnit timeUnit, QuotaScope scope)
limit
- the allowed number of request/data per timeUnittimeUnit
- the limit time unitscope
- the quota scopepublic static org.apache.hadoop.hbase.protobuf.generated.WALProtos.BulkLoadDescriptor toBulkLoadDescriptor(TableName tableName, com.google.protobuf.ByteString encodedRegionName, Map<byte[],List<org.apache.hadoop.fs.Path>> storeFiles, Map<String,Long> storeFilesSize, long bulkloadSeqId)
tableName
- The tableName into which the bulk load is being imported into.encodedRegionName
- Encoded region name of the region which is being bulk loaded.storeFiles
- A set of store files of a column family are bulk loaded.storeFilesSize
- Map of store files and their lengthsbulkloadSeqId
- sequence ID (by a force flush) used to create bulk load hfile
namepublic static org.apache.hadoop.hbase.protobuf.generated.WALProtos.BulkLoadDescriptor toBulkLoadDescriptor(TableName tableName, com.google.protobuf.ByteString encodedRegionName, Map<byte[],List<org.apache.hadoop.fs.Path>> storeFiles, Map<String,Long> storeFilesSize, long bulkloadSeqId, List<String> clusterIds)
tableName
- The tableName into which the bulk load is being imported into.encodedRegionName
- Encoded region name of the region which is being bulk loaded.storeFiles
- A set of store files of a column family are bulk loaded.storeFilesSize
- Map of store files and their lengthsbulkloadSeqId
- sequence ID (by a force flush) used to create bulk load hfile nameclusterIds
- The list of cluster Ids with the clusters where the bulk even had
already been processed.public static void mergeDelimitedFrom(com.google.protobuf.Message.Builder builder, InputStream in) throws IOException
builder
- current message builderin
- Inputsream with delimited protobuf dataIOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, InputStream in, int size) throws IOException
builder
- current message builderin
- InputStream containing protobuf datasize
- known size of protobuf dataIOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, InputStream in) throws IOException
builder
- current message builderin
- InputStream containing protobuf dataIOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, com.google.protobuf.ByteString bs) throws IOException
builder
- current message builderbs
- ByteString containing theIOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, byte[] b) throws IOException
builder
- current message builderb
- byte arrayIOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, byte[] b, int offset, int length) throws IOException
builder
- current message builderb
- byte arrayoffset
- length
- IOException
public static void mergeFrom(com.google.protobuf.Message.Builder builder, com.google.protobuf.CodedInputStream codedInput, int length) throws IOException
IOException
public static ReplicationLoadSink toReplicationLoadSink(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ReplicationLoadSink cls)
public static ReplicationLoadSource toReplicationLoadSource(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ReplicationLoadSource cls)
public static List<ReplicationLoadSource> toReplicationLoadSourceList(List<org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.ReplicationLoadSource> clsList)
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.VersionInfo getVersionInfo()
public static List<SecurityCapability> toSecurityCapabilityList(List<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability> capabilities)
capabilities
- capabilities returned in the SecurityCapabilitiesResponse messagepublic static byte[] toCompactionEventTrackerBytes(Set<String> storeFiles)
public static Set<String> toCompactedStoreFiles(byte[] bytes) throws IOException
IOException
public static List<LogEntry> toSlowLogPayloads(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LogEntry logEntry)
OnlineLogRecord
logEntry
- slowlog response protobuf instancepublic static boolean toClearSlowLogPayload(org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ClearSlowLogResponses clearSlowLogResponses)
AdminProtos.ClearSlowLogResponses
to booleanclearSlowLogResponses
- Clear slowlog response protobuf instancepublic static List<LogEntry> toBalancerDecisionResponse(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LogEntry logEntry)
public static List<LogEntry> getBalancerDecisionEntries(org.apache.hadoop.hbase.protobuf.generated.MasterProtos.BalancerDecisionsResponse response)
public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.LogRequest toBalancerDecisionRequest(int limit)
public static RegionState parseMetaRegionStateFrom(byte[] data, int replicaId) throws DeserializationException
data
- protobuf serialized data with meta server name.replicaId
- replica ID for this regionDeserializationException
- if the data is invalid.public static ServerName parseServerNameFrom(byte[] data) throws DeserializationException
data
- Data with a serialize server name in it; can handle the old style
servername where servername was host and port. Works too with data that
begins w/ the pb 'PBUF' magic and that is then followed by a protobuf that
has a serialized ServerName
in it.data
is null else converts passed data
to a ServerName instance.DeserializationException
- when data cannot be de-serialized as expected.public static org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLocation toRegionLocation(HRegionLocation loc)
public static HRegionLocation toRegionLocation(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionLocation proto)
Copyright © 2007–2022 The Apache Software Foundation. All rights reserved.