Uses of Class
org.apache.lucene.analysis.Analyzer.TokenStreamComponents

Package
Description
Text analysis.
Analyzer for Arabic.
Analyzer for Bulgarian.
Analyzer for Bengali Language.
Analyzer for Brazilian Portuguese.
Analyzer for Catalan.
Analyzer for Chinese, Japanese, and Korean, which indexes bigrams.
Analyzer for Sorani Kurdish.
Fast, general-purpose grammar-based tokenizers.
Analyzer for Simplified Chinese, which indexes words.
Basic, general-purpose analysis components.
A general-purpose Analyzer that can be created with a builder-style API.
Analyzer for Czech.
Analyzer for Danish.
Analyzer for German.
Analyzer for Greek.
Fast, general-purpose URLs and email addresses tokenizers.
Analyzer for English.
Analyzer for Spanish.
Analyzer for Estonian.
Analyzer for Basque.
Analyzer for Persian.
Analyzer for Finnish.
Analyzer for French.
Analyzer for Irish.
Analyzer for Galician.
Analyzer for Hindi.
Analyzer for Hungarian.
Analyzer for Armenian.
Analysis components based on ICU
Analyzer for Indonesian.
Analyzer for Italian.
Analyzer for Japanese.
Analyzer for Korean.
Analyzer for Lithuanian.
Analyzer for Latvian.
Miscellaneous Tokenstreams.
Analyzer for Nepali.
Analyzer for Dutch.
Analyzer for Norwegian.
Analyzer for Polish.
Analyzer for Portuguese.
Automatically filter high-frequency stopwords.
Analyzer for Romanian.
Analyzer for Russian.
Word n-gram filters.
Analyzer for Serbian.
Fast, general-purpose grammar-based tokenizer StandardTokenizer implements the Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29.
Analyzer for Swedish.
Analyzer for Tamil.
Analyzer for Telugu Language.
Analyzer for Thai.
Analyzer for Turkish.
Unicode collation support.
Support for document suggestion