|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use AttributeSource | |
---|---|
org.apache.lucene.analysis | API and code to convert text into indexable/searchable tokens. |
org.apache.lucene.analysis.ar | Analyzer for Arabic. |
org.apache.lucene.analysis.bg | Analyzer for Bulgarian. |
org.apache.lucene.analysis.br | Analyzer for Brazilian Portuguese. |
org.apache.lucene.analysis.cjk | Analyzer for Chinese, Japanese, and Korean, which indexes bigrams (overlapping groups of two adjacent Han characters). |
org.apache.lucene.analysis.cn | Analyzer for Chinese, which indexes unigrams (individual chinese characters). |
org.apache.lucene.analysis.cn.smart |
Analyzer for Simplified Chinese, which indexes words. |
org.apache.lucene.analysis.compound | A filter that decomposes compound words you find in many Germanic languages into the word parts. |
org.apache.lucene.analysis.cz | Analyzer for Czech. |
org.apache.lucene.analysis.de | Analyzer for German. |
org.apache.lucene.analysis.el | Analyzer for Greek. |
org.apache.lucene.analysis.en | Analyzer for English. |
org.apache.lucene.analysis.es | Analyzer for Spanish. |
org.apache.lucene.analysis.fa | Analyzer for Persian. |
org.apache.lucene.analysis.fi | Analyzer for Finnish. |
org.apache.lucene.analysis.fr | Analyzer for French. |
org.apache.lucene.analysis.ga | Analysis for Irish. |
org.apache.lucene.analysis.gl | Analyzer for Galician. |
org.apache.lucene.analysis.hi | Analyzer for Hindi. |
org.apache.lucene.analysis.hu | Analyzer for Hungarian. |
org.apache.lucene.analysis.hunspell | Stemming TokenFilter using a Java implementation of the Hunspell stemming algorithm. |
org.apache.lucene.analysis.icu | Analysis components based on ICU |
org.apache.lucene.analysis.icu.segmentation | Tokenizer that breaks text into words with the Unicode Text Segmentation algorithm. |
org.apache.lucene.analysis.id | Analyzer for Indonesian. |
org.apache.lucene.analysis.in | Analysis components for Indian languages. |
org.apache.lucene.analysis.it | Analyzer for Italian. |
org.apache.lucene.analysis.ja | Analyzer for Japanese. |
org.apache.lucene.analysis.lv | Analyzer for Latvian. |
org.apache.lucene.analysis.miscellaneous | Miscellaneous TokenStreams |
org.apache.lucene.analysis.ngram | Character n-gram tokenizers and filters. |
org.apache.lucene.analysis.nl | Analyzer for Dutch. |
org.apache.lucene.analysis.no | Analyzer for Norwegian. |
org.apache.lucene.analysis.path | Analysis components for path-like strings such as filenames. |
org.apache.lucene.analysis.payloads | Provides various convenience classes for creating payloads on Tokens. |
org.apache.lucene.analysis.phonetic | Analysis components for phonetic search. |
org.apache.lucene.analysis.position | Filter for assigning position increments. |
org.apache.lucene.analysis.pt | Analyzer for Portuguese. |
org.apache.lucene.analysis.reverse | Filter to reverse token text. |
org.apache.lucene.analysis.ru | Analyzer for Russian. |
org.apache.lucene.analysis.shingle | Word n-gram filters |
org.apache.lucene.analysis.sinks | Implementations of the SinkTokenizer that might be useful. |
org.apache.lucene.analysis.snowball | TokenFilter and Analyzer implementations that use Snowball
stemmers. |
org.apache.lucene.analysis.standard | Standards-based analyzers implemented with JFlex. |
org.apache.lucene.analysis.stempel | Stempel: Algorithmic Stemmer |
org.apache.lucene.analysis.sv | Analyzer for Swedish. |
org.apache.lucene.analysis.synonym | Analysis components for Synonyms. |
org.apache.lucene.analysis.th | Analyzer for Thai. |
org.apache.lucene.analysis.tr | Analyzer for Turkish. |
org.apache.lucene.analysis.wikipedia | Tokenizer that is aware of Wikipedia syntax. |
org.apache.lucene.collation |
CollationKeyFilter
converts each token into its binary CollationKey using the
provided Collator , and then encode the CollationKey
as a String using
IndexableBinaryStringTools , to allow it to be
stored as an index term. |
org.apache.lucene.facet.enhancements | Enhanced category features |
org.apache.lucene.facet.enhancements.association | Association category enhancements |
org.apache.lucene.facet.index.streaming | Expert: attributes streaming definition for indexing facets |
org.apache.lucene.index | Code to maintain and access indices. |
org.apache.lucene.queryParser | A simple query parser implemented with JavaCC. |
org.apache.lucene.queryParser.core.config | Contains the base classes used to configure the query processing |
org.apache.lucene.queryParser.standard.config | Standard Lucene Query Configuration |
org.apache.lucene.search.highlight | The highlight package contains classes to provide "keyword in context" features typically used to highlight search terms in the text of results pages. |
org.apache.lucene.util | Some utility classes. |
Uses of AttributeSource in org.apache.lucene.analysis |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis | |
---|---|
class |
ASCIIFoldingFilter
This class converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if one exists. |
class |
CachingTokenFilter
This class can be used if the token attributes of a TokenStream are intended to be consumed more than once. |
class |
CannedTokenStream
TokenStream from a canned list of Tokens. |
class |
CharTokenizer
An abstract base class for simple, character-oriented tokenizers. |
class |
EmptyTokenizer
Emits no tokens |
class |
FilteringTokenFilter
Abstract base class for TokenFilters that may remove tokens. |
class |
ISOLatin1AccentFilter
Deprecated. If you build a new index, use ASCIIFoldingFilter
which covers a superset of Latin 1.
This class is included for use with existing
indexes and will be removed in a future release (possibly Lucene 4.0). |
class |
KeywordMarkerFilter
Marks terms as keywords via the KeywordAttribute . |
class |
KeywordTokenizer
Emits the entire input as a single token. |
class |
LengthFilter
Removes words that are too long or too short from the stream. |
class |
LetterTokenizer
A LetterTokenizer is a tokenizer that divides text at non-letters. |
class |
LimitTokenCountFilter
This TokenFilter limits the number of tokens while indexing. |
class |
LowerCaseFilter
Normalizes token text to lower case. |
class |
LowerCaseTokenizer
LowerCaseTokenizer performs the function of LetterTokenizer and LowerCaseFilter together. |
class |
MockFixedLengthPayloadFilter
TokenFilter that adds random fixed-length payloads. |
class |
MockTokenizer
Tokenizer for testing. |
class |
MockVariableLengthPayloadFilter
TokenFilter that adds random variable-length payloads. |
class |
NumericTokenStream
Expert: This class provides a TokenStream
for indexing numeric values that can be used by NumericRangeQuery or NumericRangeFilter . |
class |
PorterStemFilter
Transforms the token stream as per the Porter stemming algorithm. |
class |
StopFilter
Removes stop words from a token stream. |
class |
TeeSinkTokenFilter
This TokenFilter provides the ability to set aside attribute states that have already been analyzed. |
static class |
TeeSinkTokenFilter.SinkTokenStream
TokenStream output from a tee with optional filtering. |
class |
TokenFilter
A TokenFilter is a TokenStream whose input is another TokenStream. |
class |
Tokenizer
A Tokenizer is a TokenStream whose input is a Reader. |
class |
TokenStream
A TokenStream enumerates the sequence of tokens, either from
Field s of a Document or from query text. |
class |
TypeTokenFilter
Removes tokens whose types appear in a set of blocked types from a token stream. |
class |
WhitespaceTokenizer
A WhitespaceTokenizer is a tokenizer that divides text at whitespace. |
Methods in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
abstract boolean |
TeeSinkTokenFilter.SinkFilter.accept(AttributeSource source)
Returns true, iff the current state of the passed-in AttributeSource shall be stored
in the sink. |
Constructors in org.apache.lucene.analysis with parameters of type AttributeSource | |
---|---|
CharTokenizer(AttributeSource source,
Reader input)
Deprecated. use CharTokenizer.CharTokenizer(Version, AttributeSource, Reader) instead. This will be
removed in Lucene 4.0. |
|
CharTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new CharTokenizer instance |
|
KeywordTokenizer(AttributeSource source,
Reader input,
int bufferSize)
|
|
LetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use LetterTokenizer.LetterTokenizer(Version, AttributeSource, Reader) instead.
This will be removed in Lucene 4.0. |
|
LetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new LetterTokenizer using a given AttributeSource . |
|
LowerCaseTokenizer(AttributeSource source,
Reader in)
Deprecated. use LowerCaseTokenizer.LowerCaseTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
LowerCaseTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new LowerCaseTokenizer using a given AttributeSource . |
|
NumericTokenStream(AttributeSource source,
int precisionStep)
Expert: Creates a token stream for numeric values with the specified precisionStep using the given AttributeSource . |
|
Tokenizer(AttributeSource source)
Deprecated. use Tokenizer.Tokenizer(AttributeSource, Reader) instead. |
|
Tokenizer(AttributeSource source,
Reader input)
Construct a token stream processing the given input using the given AttributeSource. |
|
TokenStream(AttributeSource input)
A TokenStream that uses the same attributes as the supplied one. |
|
WhitespaceTokenizer(AttributeSource source,
Reader in)
Deprecated. use WhitespaceTokenizer.WhitespaceTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
WhitespaceTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Construct a new WhitespaceTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.ar |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ar | |
---|---|
class |
ArabicLetterTokenizer
Deprecated. (3.1) Use StandardTokenizer instead. |
class |
ArabicNormalizationFilter
A TokenFilter that applies ArabicNormalizer to normalize the orthography. |
class |
ArabicStemFilter
A TokenFilter that applies ArabicStemmer to stem Arabic words.. |
Constructors in org.apache.lucene.analysis.ar with parameters of type AttributeSource | |
---|---|
ArabicLetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use ArabicLetterTokenizer.ArabicLetterTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
ArabicLetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Deprecated. Construct a new ArabicLetterTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.bg |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.bg | |
---|---|
class |
BulgarianStemFilter
A TokenFilter that applies BulgarianStemmer to stem Bulgarian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.br |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.br | |
---|---|
class |
BrazilianStemFilter
A TokenFilter that applies BrazilianStemmer . |
Uses of AttributeSource in org.apache.lucene.analysis.cjk |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cjk | |
---|---|
class |
CJKBigramFilter
Forms bigrams of CJK terms that are generated from StandardTokenizer or ICUTokenizer. |
class |
CJKTokenizer
Deprecated. Use StandardTokenizer, CJKWidthFilter, CJKBigramFilter, and LowerCaseFilter instead. |
class |
CJKWidthFilter
A TokenFilter that normalizes CJK width differences:
Folds fullwidth ASCII variants into the equivalent basic latin
Folds halfwidth Katakana variants into the equivalent kana
|
Constructors in org.apache.lucene.analysis.cjk with parameters of type AttributeSource | |
---|---|
CJKTokenizer(AttributeSource source,
Reader in)
Deprecated. |
Uses of AttributeSource in org.apache.lucene.analysis.cn |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn | |
---|---|
class |
ChineseFilter
Deprecated. Use StopFilter instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
class |
ChineseTokenizer
Deprecated. Use StandardTokenizer instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
Constructors in org.apache.lucene.analysis.cn with parameters of type AttributeSource | |
---|---|
ChineseTokenizer(AttributeSource source,
Reader in)
Deprecated. |
Uses of AttributeSource in org.apache.lucene.analysis.cn.smart |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cn.smart | |
---|---|
class |
SentenceTokenizer
Tokenizes input text into sentences. |
class |
WordTokenFilter
A TokenFilter that breaks sentences into words. |
Constructors in org.apache.lucene.analysis.cn.smart with parameters of type AttributeSource | |
---|---|
SentenceTokenizer(AttributeSource source,
Reader reader)
|
Uses of AttributeSource in org.apache.lucene.analysis.compound |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.compound | |
---|---|
class |
CompoundWordTokenFilterBase
Base class for decomposition token filters. |
class |
DictionaryCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
class |
HyphenationCompoundWordTokenFilter
A TokenFilter that decomposes compound words found in many Germanic languages. |
Uses of AttributeSource in org.apache.lucene.analysis.cz |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.cz | |
---|---|
class |
CzechStemFilter
A TokenFilter that applies CzechStemmer to stem Czech words. |
Uses of AttributeSource in org.apache.lucene.analysis.de |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.de | |
---|---|
class |
GermanLightStemFilter
A TokenFilter that applies GermanLightStemmer to stem German
words. |
class |
GermanMinimalStemFilter
A TokenFilter that applies GermanMinimalStemmer to stem German
words. |
class |
GermanNormalizationFilter
Normalizes German characters according to the heuristics of the German2 snowball algorithm. |
class |
GermanStemFilter
A TokenFilter that stems German words. |
Uses of AttributeSource in org.apache.lucene.analysis.el |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.el | |
---|---|
class |
GreekLowerCaseFilter
Normalizes token text to lower case, removes some Greek diacritics, and standardizes final sigma to sigma. |
class |
GreekStemFilter
A TokenFilter that applies GreekStemmer to stem Greek
words. |
Uses of AttributeSource in org.apache.lucene.analysis.en |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.en | |
---|---|
class |
EnglishMinimalStemFilter
A TokenFilter that applies EnglishMinimalStemmer to stem
English words. |
class |
EnglishPossessiveFilter
TokenFilter that removes possessives (trailing 's) from words. |
class |
KStemFilter
A high-performance kstem filter for english. |
Uses of AttributeSource in org.apache.lucene.analysis.es |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.es | |
---|---|
class |
SpanishLightStemFilter
A TokenFilter that applies SpanishLightStemmer to stem Spanish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.fa |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fa | |
---|---|
class |
PersianNormalizationFilter
A TokenFilter that applies PersianNormalizer to normalize the
orthography. |
Uses of AttributeSource in org.apache.lucene.analysis.fi |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fi | |
---|---|
class |
FinnishLightStemFilter
A TokenFilter that applies FinnishLightStemmer to stem Finnish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.fr |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.fr | |
---|---|
class |
ElisionFilter
Removes elisions from a TokenStream . |
class |
FrenchLightStemFilter
A TokenFilter that applies FrenchLightStemmer to stem French
words. |
class |
FrenchMinimalStemFilter
A TokenFilter that applies FrenchMinimalStemmer to stem French
words. |
class |
FrenchStemFilter
Deprecated. Use SnowballFilter with
FrenchStemmer instead, which has the
same functionality. This filter will be removed in Lucene 5.0 |
Uses of AttributeSource in org.apache.lucene.analysis.ga |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ga | |
---|---|
class |
IrishLowerCaseFilter
Normalises token text to lower case, handling t-prothesis and n-eclipsis (i.e., that 'nAthair' should become 'n-athair') |
Uses of AttributeSource in org.apache.lucene.analysis.gl |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.gl | |
---|---|
class |
GalicianMinimalStemFilter
A TokenFilter that applies GalicianMinimalStemmer to stem
Galician words. |
class |
GalicianStemFilter
A TokenFilter that applies GalicianStemmer to stem
Galician words. |
Uses of AttributeSource in org.apache.lucene.analysis.hi |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.hi | |
---|---|
class |
HindiNormalizationFilter
A TokenFilter that applies HindiNormalizer to normalize the
orthography. |
class |
HindiStemFilter
A TokenFilter that applies HindiStemmer to stem Hindi words. |
Uses of AttributeSource in org.apache.lucene.analysis.hu |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.hu | |
---|---|
class |
HungarianLightStemFilter
A TokenFilter that applies HungarianLightStemmer to stem
Hungarian words. |
Uses of AttributeSource in org.apache.lucene.analysis.hunspell |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.hunspell | |
---|---|
class |
HunspellStemFilter
TokenFilter that uses hunspell affix rules and words to stem tokens. |
Uses of AttributeSource in org.apache.lucene.analysis.icu |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.icu | |
---|---|
class |
ICUFoldingFilter
A TokenFilter that applies search term folding to Unicode text, applying foldings from UTR#30 Character Foldings. |
class |
ICUNormalizer2Filter
Normalize token text with ICU's Normalizer2 |
class |
ICUTransformFilter
A TokenFilter that transforms text with ICU. |
Uses of AttributeSource in org.apache.lucene.analysis.icu.segmentation |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.icu.segmentation | |
---|---|
class |
ICUTokenizer
Breaks text into words according to UAX #29: Unicode Text Segmentation (http://www.unicode.org/reports/tr29/) |
Uses of AttributeSource in org.apache.lucene.analysis.id |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.id | |
---|---|
class |
IndonesianStemFilter
A TokenFilter that applies IndonesianStemmer to stem Indonesian words. |
Uses of AttributeSource in org.apache.lucene.analysis.in |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.in | |
---|---|
class |
IndicNormalizationFilter
A TokenFilter that applies IndicNormalizer to normalize text
in Indian Languages. |
class |
IndicTokenizer
Deprecated. (3.6) Use StandardTokenizer instead. |
Constructors in org.apache.lucene.analysis.in with parameters of type AttributeSource | |
---|---|
IndicTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Deprecated. |
Uses of AttributeSource in org.apache.lucene.analysis.it |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.it | |
---|---|
class |
ItalianLightStemFilter
A TokenFilter that applies ItalianLightStemmer to stem Italian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.ja |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ja | |
---|---|
class |
JapaneseBaseFormFilter
Replaces term text with the BaseFormAttribute . |
class |
JapaneseKatakanaStemFilter
A TokenFilter that normalizes common katakana spelling variations
ending in a long sound character by removing this character (U+30FC). |
class |
JapanesePartOfSpeechStopFilter
Removes tokens that match a set of part-of-speech tags. |
class |
JapaneseReadingFormFilter
A TokenFilter that replaces the term
attribute with the reading of a token in either katakana or romaji form. |
class |
JapaneseTokenizer
Tokenizer for Japanese that uses morphological analysis. |
Uses of AttributeSource in org.apache.lucene.analysis.lv |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.lv | |
---|---|
class |
LatvianStemFilter
A TokenFilter that applies LatvianStemmer to stem Latvian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.miscellaneous |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.miscellaneous | |
---|---|
class |
EmptyTokenStream
An always exhausted token stream. |
class |
PrefixAndSuffixAwareTokenFilter
Links two PrefixAwareTokenFilter . |
class |
PrefixAwareTokenFilter
Joins two token streams and leaves the last token of the first stream available to be used when updating the token values in the second stream based on that token. |
class |
SingleTokenTokenStream
A TokenStream containing a single token. |
class |
StemmerOverrideFilter
Provides the ability to override any KeywordAttribute aware stemmer
with custom dictionary-based stemming. |
Uses of AttributeSource in org.apache.lucene.analysis.ngram |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ngram | |
---|---|
class |
EdgeNGramTokenFilter
Tokenizes the given token into n-grams of given size(s). |
class |
EdgeNGramTokenizer
Tokenizes the input from an edge into n-grams of given size(s). |
class |
NGramTokenFilter
Tokenizes the input into n-grams of the given size(s). |
class |
NGramTokenizer
Tokenizes the input into n-grams of the given size(s). |
Constructors in org.apache.lucene.analysis.ngram with parameters of type AttributeSource | |
---|---|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
EdgeNGramTokenizer.Side side,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
EdgeNGramTokenizer(AttributeSource source,
Reader input,
String sideLabel,
int minGram,
int maxGram)
Creates EdgeNGramTokenizer that can generate n-grams in the sizes of the given range |
|
NGramTokenizer(AttributeSource source,
Reader input,
int minGram,
int maxGram)
Creates NGramTokenizer with given min and max n-grams. |
Uses of AttributeSource in org.apache.lucene.analysis.nl |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.nl | |
---|---|
class |
DutchStemFilter
Deprecated. Use SnowballFilter with
DutchStemmer instead, which has the
same functionality. This filter will be removed in Lucene 5.0 |
Uses of AttributeSource in org.apache.lucene.analysis.no |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.no | |
---|---|
class |
NorwegianLightStemFilter
A TokenFilter that applies NorwegianLightStemmer to stem Norwegian
words. |
class |
NorwegianMinimalStemFilter
A TokenFilter that applies NorwegianMinimalStemmer to stem Norwegian
words. |
Uses of AttributeSource in org.apache.lucene.analysis.path |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.path | |
---|---|
class |
PathHierarchyTokenizer
Tokenizer for path-like hierarchies. |
class |
ReversePathHierarchyTokenizer
Tokenizer for domain-like hierarchies. |
Uses of AttributeSource in org.apache.lucene.analysis.payloads |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.payloads | |
---|---|
class |
DelimitedPayloadTokenFilter
Characters before the delimiter are the "token", those after are the payload. |
class |
NumericPayloadTokenFilter
Assigns a payload to a token based on the Token.type() |
class |
TokenOffsetPayloadTokenFilter
Adds the Token.setStartOffset(int)
and Token.setEndOffset(int)
First 4 bytes are the start |
class |
TypeAsPayloadTokenFilter
Makes the Token.type() a payload. |
Uses of AttributeSource in org.apache.lucene.analysis.phonetic |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.phonetic | |
---|---|
class |
BeiderMorseFilter
TokenFilter for Beider-Morse phonetic encoding. |
class |
DoubleMetaphoneFilter
Filter for DoubleMetaphone (supporting secondary codes) |
class |
PhoneticFilter
Create tokens for phonetic matches. |
Uses of AttributeSource in org.apache.lucene.analysis.position |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.position | |
---|---|
class |
PositionFilter
Set the positionIncrement of all tokens to the "positionIncrement", except the first return token which retains its original positionIncrement value. |
Uses of AttributeSource in org.apache.lucene.analysis.pt |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.pt | |
---|---|
class |
PortugueseLightStemFilter
A TokenFilter that applies PortugueseLightStemmer to stem
Portuguese words. |
class |
PortugueseMinimalStemFilter
A TokenFilter that applies PortugueseMinimalStemmer to stem
Portuguese words. |
class |
PortugueseStemFilter
A TokenFilter that applies PortugueseStemmer to stem
Portuguese words. |
Uses of AttributeSource in org.apache.lucene.analysis.reverse |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.reverse | |
---|---|
class |
ReverseStringFilter
Reverse token string, for example "country" => "yrtnuoc". |
Uses of AttributeSource in org.apache.lucene.analysis.ru |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.ru | |
---|---|
class |
RussianLetterTokenizer
Deprecated. Use StandardTokenizer instead, which has the same functionality.
This filter will be removed in Lucene 5.0 |
class |
RussianLightStemFilter
A TokenFilter that applies RussianLightStemmer to stem Russian
words. |
class |
RussianLowerCaseFilter
Deprecated. Use LowerCaseFilter instead, which has the same
functionality. This filter will be removed in Lucene 4.0 |
class |
RussianStemFilter
Deprecated. Use SnowballFilter with
RussianStemmer instead, which has the
same functionality. This filter will be removed in Lucene 4.0 |
Constructors in org.apache.lucene.analysis.ru with parameters of type AttributeSource | |
---|---|
RussianLetterTokenizer(AttributeSource source,
Reader in)
Deprecated. use RussianLetterTokenizer.RussianLetterTokenizer(Version, AttributeSource, Reader)
instead. This will be removed in Lucene 4.0. |
|
RussianLetterTokenizer(Version matchVersion,
AttributeSource source,
Reader in)
Deprecated. Construct a new RussianLetterTokenizer using a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.shingle |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.shingle | |
---|---|
class |
ShingleFilter
A ShingleFilter constructs shingles (token n-grams) from a token stream. |
class |
ShingleMatrixFilter
Deprecated. Will be removed in Lucene 4.0. This filter is unmaintained and might not behave correctly if used with custom Attributes, i.e. Attributes other than the ones located in org.apache.lucene.analysis.tokenattributes . It also uses
hardcoded payload encoders which makes it not easily adaptable to other use-cases. |
Uses of AttributeSource in org.apache.lucene.analysis.sinks |
---|
Methods in org.apache.lucene.analysis.sinks with parameters of type AttributeSource | |
---|---|
boolean |
DateRecognizerSinkFilter.accept(AttributeSource source)
|
boolean |
TokenRangeSinkFilter.accept(AttributeSource source)
|
boolean |
TokenTypeSinkFilter.accept(AttributeSource source)
|
Uses of AttributeSource in org.apache.lucene.analysis.snowball |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.snowball | |
---|---|
class |
SnowballFilter
A filter that stems words using a Snowball-generated stemmer. |
Uses of AttributeSource in org.apache.lucene.analysis.standard |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.standard | |
---|---|
class |
ClassicFilter
Normalizes tokens extracted with ClassicTokenizer . |
class |
ClassicTokenizer
A grammar-based tokenizer constructed with JFlex |
class |
StandardFilter
Normalizes tokens extracted with StandardTokenizer . |
class |
StandardTokenizer
A grammar-based tokenizer constructed with JFlex. |
class |
UAX29URLEmailTokenizer
This class implements Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29 URLs and email addresses are also tokenized according to the relevant RFCs. |
Constructors in org.apache.lucene.analysis.standard with parameters of type AttributeSource | |
---|---|
ClassicTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new ClassicTokenizer with a given AttributeSource . |
|
StandardTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new StandardTokenizer with a given AttributeSource . |
|
UAX29URLEmailTokenizer(AttributeSource source,
Reader input)
Deprecated. use UAX29URLEmailTokenizer.UAX29URLEmailTokenizer(Version, AttributeSource, Reader) instead. |
|
UAX29URLEmailTokenizer(Version matchVersion,
AttributeSource source,
Reader input)
Creates a new UAX29URLEmailTokenizer with a given AttributeSource . |
Uses of AttributeSource in org.apache.lucene.analysis.stempel |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.stempel | |
---|---|
class |
StempelFilter
Transforms the token stream as per the stemming algorithm. |
Uses of AttributeSource in org.apache.lucene.analysis.sv |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.sv | |
---|---|
class |
SwedishLightStemFilter
A TokenFilter that applies SwedishLightStemmer to stem Swedish
words. |
Uses of AttributeSource in org.apache.lucene.analysis.synonym |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.synonym | |
---|---|
class |
SynonymFilter
Matches single or multi word synonyms in a token stream. |
Uses of AttributeSource in org.apache.lucene.analysis.th |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.th | |
---|---|
class |
ThaiWordFilter
TokenFilter that use BreakIterator to break each
Token that is Thai into separate Token(s) for each Thai word. |
Uses of AttributeSource in org.apache.lucene.analysis.tr |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.tr | |
---|---|
class |
TurkishLowerCaseFilter
Normalizes Turkish token text to lower case. |
Uses of AttributeSource in org.apache.lucene.analysis.wikipedia |
---|
Subclasses of AttributeSource in org.apache.lucene.analysis.wikipedia | |
---|---|
class |
WikipediaTokenizer
Extension of StandardTokenizer that is aware of Wikipedia syntax. |
Constructors in org.apache.lucene.analysis.wikipedia with parameters of type AttributeSource | |
---|---|
WikipediaTokenizer(AttributeSource source,
Reader input,
int tokenOutput,
Set<String> untokenizedTypes)
Creates a new instance of the WikipediaTokenizer . |
Uses of AttributeSource in org.apache.lucene.collation |
---|
Subclasses of AttributeSource in org.apache.lucene.collation | |
---|---|
class |
CollationKeyFilter
Converts each token into its CollationKey , and then
encodes the CollationKey with IndexableBinaryStringTools , to allow
it to be stored as an index term. |
class |
ICUCollationKeyFilter
Converts each token into its CollationKey , and
then encodes the CollationKey with IndexableBinaryStringTools , to
allow it to be stored as an index term. |
Uses of AttributeSource in org.apache.lucene.facet.enhancements |
---|
Subclasses of AttributeSource in org.apache.lucene.facet.enhancements | |
---|---|
class |
EnhancementsCategoryTokenizer
A tokenizer which adds to each category token payload according to the CategoryEnhancement s defined in the given
EnhancementsIndexingParams . |
Uses of AttributeSource in org.apache.lucene.facet.enhancements.association |
---|
Subclasses of AttributeSource in org.apache.lucene.facet.enhancements.association | |
---|---|
class |
AssociationListTokenizer
Tokenizer for associations of a category |
Uses of AttributeSource in org.apache.lucene.facet.index.streaming |
---|
Subclasses of AttributeSource in org.apache.lucene.facet.index.streaming | |
---|---|
class |
CategoryAttributesStream
An attribute stream built from an Iterable of
CategoryAttribute . |
class |
CategoryListTokenizer
A base class for category list tokenizers, which add category list tokens to category streams. |
class |
CategoryParentsStream
This class adds parents to a CategoryAttributesStream . |
class |
CategoryTokenizer
Basic class for setting the CharTermAttribute s and
PayloadAttribute s of category tokens. |
class |
CategoryTokenizerBase
A base class for all token filters which add term and payload attributes to tokens and are to be used in CategoryDocumentBuilder . |
class |
CountingListTokenizer
CategoryListTokenizer for facet counting |
Uses of AttributeSource in org.apache.lucene.index |
---|
Methods in org.apache.lucene.index that return AttributeSource | |
---|---|
AttributeSource |
FieldInvertState.getAttributeSource()
|
Uses of AttributeSource in org.apache.lucene.queryParser |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser | |
---|---|
static class |
QueryParserTestBase.QPTestFilter
Filter which discards the token 'stop' and which expands the token 'phrase' into 'phrase1 phrase2' |
Uses of AttributeSource in org.apache.lucene.queryParser.core.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.core.config | |
---|---|
class |
AbstractQueryConfig
This class is the base of QueryConfigHandler and FieldConfig . |
class |
FieldConfig
This class represents a field configuration. |
class |
QueryConfigHandler
This class can be used to hold any query configuration and no field configuration. |
Uses of AttributeSource in org.apache.lucene.queryParser.standard.config |
---|
Subclasses of AttributeSource in org.apache.lucene.queryParser.standard.config | |
---|---|
class |
StandardQueryConfigHandler
This query configuration handler is used for almost every processor defined in the StandardQueryNodeProcessorPipeline processor pipeline. |
Uses of AttributeSource in org.apache.lucene.search.highlight |
---|
Subclasses of AttributeSource in org.apache.lucene.search.highlight | |
---|---|
class |
OffsetLimitTokenFilter
This TokenFilter limits the number of tokens while indexing by adding up the current offset. |
class |
TokenStreamFromTermPositionVector
|
Uses of AttributeSource in org.apache.lucene.util |
---|
Methods in org.apache.lucene.util that return AttributeSource | |
---|---|
AttributeSource |
AttributeSource.cloneAttributes()
Performs a clone of all AttributeImpl instances returned in a new
AttributeSource instance. |
Methods in org.apache.lucene.util with parameters of type AttributeSource | |
---|---|
void |
AttributeSource.copyTo(AttributeSource target)
Copies the contents of this AttributeSource to the given target AttributeSource . |
Constructors in org.apache.lucene.util with parameters of type AttributeSource | |
---|---|
AttributeSource(AttributeSource input)
An AttributeSource that uses the same attributes as the supplied one. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |