Uses of Class
org.apache.lucene.analysis.Tokenizer
-
Packages that use Tokenizer Package Description org.apache.lucene.analysis API and code to convert text into indexable/searchable tokens.org.apache.lucene.analysis.standard Standards-based analyzers implemented with JFlex. -
-
Uses of Tokenizer in org.apache.lucene.analysis
Subclasses of Tokenizer in org.apache.lucene.analysis Modifier and Type Class Description classCharTokenizerAn abstract base class for simple, character-oriented tokenizers.classKeywordTokenizerEmits the entire input as a single token.classLetterTokenizerA LetterTokenizer is a tokenizer that divides text at non-letters.classLowerCaseTokenizerLowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together.classWhitespaceTokenizerA WhitespaceTokenizer is a tokenizer that divides text at whitespace.Fields in org.apache.lucene.analysis declared as Tokenizer Modifier and Type Field Description protected TokenizerReusableAnalyzerBase.TokenStreamComponents. sourceConstructors in org.apache.lucene.analysis with parameters of type Tokenizer Constructor Description TokenStreamComponents(Tokenizer source)Creates a newReusableAnalyzerBase.TokenStreamComponentsinstance.TokenStreamComponents(Tokenizer source, TokenStream result)Creates a newReusableAnalyzerBase.TokenStreamComponentsinstance. -
Uses of Tokenizer in org.apache.lucene.analysis.standard
Subclasses of Tokenizer in org.apache.lucene.analysis.standard Modifier and Type Class Description classClassicTokenizerA grammar-based tokenizer constructed with JFlexclassStandardTokenizerA grammar-based tokenizer constructed with JFlex.classUAX29URLEmailTokenizerThis class implements Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29 URLs and email addresses are also tokenized according to the relevant RFCs.
-