|
||||||||||
PREV CLASS NEXT CLASS | FRAMES NO FRAMES | |||||||||
SUMMARY: NESTED | FIELD | CONSTR | METHOD | DETAIL: FIELD | CONSTR | METHOD |
java.lang.Object org.apache.lucene.util.AttributeSource org.apache.lucene.analysis.TokenStream org.apache.lucene.analysis.Tokenizer org.apache.lucene.analysis.MockTokenizer
public class MockTokenizer
Tokenizer for testing.
This tokenizer is a replacement for WHITESPACE
, SIMPLE
, and KEYWORD
tokenizers. If you are writing a component such as a TokenFilter, its a great idea to test
it wrapping this tokenizer instead for extra checks. This tokenizer has the following behavior:
setEnableChecks(boolean)
.
Nested Class Summary |
---|
Nested classes/interfaces inherited from class org.apache.lucene.util.AttributeSource |
---|
AttributeSource.AttributeFactory |
Field Summary | |
---|---|
static int |
DEFAULT_MAX_TOKEN_LENGTH
|
static int |
KEYWORD
Acts Similar to KeywordTokenizer. |
static int |
SIMPLE
Acts like LetterTokenizer. |
static int |
WHITESPACE
Acts Similar to WhitespaceTokenizer |
Fields inherited from class org.apache.lucene.analysis.Tokenizer |
---|
input |
Constructor Summary | |
---|---|
MockTokenizer(AttributeSource.AttributeFactory factory,
Reader input,
int pattern,
boolean lowerCase,
int maxTokenLength)
|
|
MockTokenizer(Reader input,
int pattern,
boolean lowerCase)
|
|
MockTokenizer(Reader input,
int pattern,
boolean lowerCase,
int maxTokenLength)
|
Method Summary | |
---|---|
void |
close()
By default, closes the input Reader. |
void |
end()
This method is called by the consumer after the last token has been consumed, after TokenStream.incrementToken() returned false
(using the new TokenStream API). |
boolean |
incrementToken()
Consumers (i.e., IndexWriter ) use this method to advance the stream to
the next token. |
protected boolean |
isTokenChar(int c)
|
protected int |
normalize(int c)
|
protected int |
readCodePoint()
|
void |
reset()
Resets this stream to the beginning. |
void |
reset(Reader input)
Expert: Reset the tokenizer to a new reader. |
void |
setEnableChecks(boolean enableChecks)
Toggle consumer workflow checking: if your test consumes tokenstreams normally you should leave this enabled. |
Methods inherited from class org.apache.lucene.analysis.Tokenizer |
---|
correctOffset |
Methods inherited from class org.apache.lucene.util.AttributeSource |
---|
addAttribute, addAttributeImpl, captureState, clearAttributes, cloneAttributes, copyTo, equals, getAttribute, getAttributeClassesIterator, getAttributeFactory, getAttributeImplsIterator, hasAttribute, hasAttributes, hashCode, reflectAsString, reflectWith, restoreState, toString |
Methods inherited from class java.lang.Object |
---|
clone, finalize, getClass, notify, notifyAll, wait, wait, wait |
Field Detail |
---|
public static final int WHITESPACE
public static final int KEYWORD
public static final int SIMPLE
public static final int DEFAULT_MAX_TOKEN_LENGTH
Constructor Detail |
---|
public MockTokenizer(AttributeSource.AttributeFactory factory, Reader input, int pattern, boolean lowerCase, int maxTokenLength)
public MockTokenizer(Reader input, int pattern, boolean lowerCase, int maxTokenLength)
public MockTokenizer(Reader input, int pattern, boolean lowerCase)
Method Detail |
---|
public final boolean incrementToken() throws IOException
TokenStream
IndexWriter
) use this method to advance the stream to
the next token. Implementing classes must implement this method and update
the appropriate AttributeImpl
s with the attributes of the next
token.
The producer must make no assumptions about the attributes after the method
has been returned: the caller may arbitrarily change it. If the producer
needs to preserve the state for subsequent calls, it can use
AttributeSource.captureState()
to create a copy of the current attribute state.
This method is called for every token of a document, so an efficient
implementation is crucial for good performance. To avoid calls to
AttributeSource.addAttribute(Class)
and AttributeSource.getAttribute(Class)
,
references to all AttributeImpl
s that this stream uses should be
retrieved during instantiation.
To ensure that filters and consumers know which attributes are available,
the attributes must be added during instantiation. Filters and consumers
are not required to check for availability of attributes in
TokenStream.incrementToken()
.
incrementToken
in class TokenStream
IOException
protected int readCodePoint() throws IOException
IOException
protected boolean isTokenChar(int c)
protected int normalize(int c)
public void reset() throws IOException
TokenStream
TokenStream.reset()
is not needed for
the standard indexing process. However, if the tokens of a
TokenStream
are intended to be consumed more than once, it is
necessary to implement TokenStream.reset()
. Note that if your TokenStream
caches tokens and feeds them back again after a reset, it is imperative
that you clone the tokens when you store them away (on the first pass) as
well as when you return them (on future passes after TokenStream.reset()
).
reset
in class TokenStream
IOException
public void close() throws IOException
Tokenizer
close
in interface Closeable
close
in class Tokenizer
IOException
public void reset(Reader input) throws IOException
Tokenizer
reset
in class Tokenizer
IOException
public void end() throws IOException
TokenStream
TokenStream.incrementToken()
returned false
(using the new TokenStream
API). Streams implementing the old API
should upgrade to use this feature.
This method can be used to perform any end-of-stream operations, such as
setting the final offset of a stream. The final offset of a stream might
differ from the offset of the last token eg in case one or more whitespaces
followed after the last token, but a WhitespaceTokenizer
was used.
end
in class TokenStream
IOException
public void setEnableChecks(boolean enableChecks)
|
||||||||||
PREV CLASS NEXT CLASS | FRAMES NO FRAMES | |||||||||
SUMMARY: NESTED | FIELD | CONSTR | METHOD | DETAIL: FIELD | CONSTR | METHOD |