Package | Description |
---|---|
org.deeplearning4j.text.tokenization.tokenizer | |
org.deeplearning4j.text.tokenization.tokenizerfactory |
Modifier and Type | Class and Description |
---|---|
class |
BertWordPieceStreamTokenizer
A tokenizer that works with a vocab from a published bert model and tokenizes a token at a time from a stream
|
class |
BertWordPieceTokenizer
A tokenizer that works with a vocab from a published bert model
|
class |
DefaultStreamTokenizer
Tokenizer based on the
StreamTokenizer |
class |
DefaultTokenizer
Default tokenizer
|
class |
NGramTokenizer |
Constructor and Description |
---|
NGramTokenizer(Tokenizer tokenizer,
Integer minN,
Integer maxN) |
Modifier and Type | Method and Description |
---|---|
Tokenizer |
BertWordPieceTokenizerFactory.create(InputStream toTokenize) |
Tokenizer |
DefaultTokenizerFactory.create(InputStream toTokenize) |
Tokenizer |
NGramTokenizerFactory.create(InputStream toTokenize) |
Tokenizer |
TokenizerFactory.create(InputStream toTokenize)
Create a tokenizer based on an input stream
|
Tokenizer |
BertWordPieceTokenizerFactory.create(String toTokenize) |
Tokenizer |
DefaultTokenizerFactory.create(String toTokenize) |
Tokenizer |
NGramTokenizerFactory.create(String toTokenize) |
Tokenizer |
TokenizerFactory.create(String toTokenize)
The tokenizer to createComplex
|
Copyright © 2019. All rights reserved.