Uses of Interface
org.apache.lucene.analysis.tokenattributes.OffsetAttribute

Packages that use OffsetAttribute
Package
Description
Text analysis.
Analyzer for Chinese, Japanese, and Korean, which indexes bigrams.
Fast, general-purpose grammar-based tokenizers.
Analyzer for Simplified Chinese, which indexes words.
Construct n-grams for frequently occurring terms and phrases.
A filter that decomposes compound words you find in many Germanic languages into the word parts.
Basic, general-purpose analysis components.
Fast, general-purpose URLs and email addresses tokenizers.
Tokenizer that breaks text into words with the Unicode Text Segmentation algorithm.
Analyzer for Japanese.
Analyzer for Korean.
MinHash filtering (for LSH).
Miscellaneous Tokenstreams.
Character n-gram tokenizers and filters.
Analysis components for path-like strings such as filenames.
Set of components for pattern-based (regex) analysis.
Provides various convenience classes for creating payloads on Tokens.
Word n-gram filters.
Fast, general-purpose grammar-based tokenizer StandardTokenizer implements the Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29.
Analysis components for Synonyms.
Analyzer for Thai.
General-purpose attributes for text analysis.
Utility functions for text analysis.
Tokenizer that is aware of Wikipedia syntax.
The logical representation of a Document for indexing and searching.
Code to maintain and access indices.
Monitoring framework
Highlighting search terms.
Analyzer based autosuggest.
The UnifiedHighlighter -- a flexible highlighter that can get offsets from postings, term vectors, or analysis.