@Deprecated public final class SynonymFilter extends TokenFilter
Note that with the current implementation, parsing is greedy, so whenever multiple parses would apply, the rule starting the earliest and parsing the most tokens wins. For example if you have these rules:
a -> x a b -> y b c d -> zThen input
a b c d eparses to
y b c d, ie the 2nd rule "wins" because it started earliest and matched the most input tokens of other rules starting at that point.
A future improvement to this filter could allow non-greedy parsing, such that the 3rd rule would win, and also separately allow multiple parses, such that all 3 rules would match, perhaps even on a rule by rule basis.
NOTE: when a match occurs, the output tokens
associated with the matching rule are "stacked" on top of
the input stream (if the rule had
keepOrig=true) and also on top of another
matched rule's output tokens. This is not a correct
solution, as really the output should be an arbitrary
graph/lattice. For example, with the above match, you
would expect an exact
c" to match the parsed tokens, but it will fail to
do so. This limitation is necessary because Lucene's
TokenStream (and index) cannot yet represent an arbitrary
NOTE: If multiple incoming tokens arrive on the same position, only the first token at that position is used for parsing. Subsequent tokens simply pass through and are not parsed. A future improvement would be to allow these tokens to also be matched.
|Modifier and Type||Field and Description|
|Constructor and Description|
|Modifier and Type||Method and Description|
This method is called by a consumer before it begins consumption using
addAttribute, addAttributeImpl, captureState, clearAttributes, cloneAttributes, copyTo, endAttributes, equals, getAttribute, getAttributeClassesIterator, getAttributeFactory, getAttributeImplsIterator, hasAttribute, hasAttributes, hashCode, reflectAsString, reflectWith, removeAllAttributes, restoreState, toString
public SynonymFilter(TokenStream input, SynonymMap synonyms, boolean ignoreCase)
public boolean incrementToken() throws IOException
IndexWriter) use this method to advance the stream to the next token. Implementing classes must implement this method and update the appropriate
AttributeImpls with the attributes of the next token.
The producer must make no assumptions about the attributes after the method
has been returned: the caller may arbitrarily change it. If the producer
needs to preserve the state for subsequent calls, it can use
AttributeSource.captureState() to create a copy of the current attribute state.
This method is called for every token of a document, so an efficient
implementation is crucial for good performance. To avoid calls to
references to all
AttributeImpls that this stream uses should be
retrieved during instantiation.
To ensure that filters and consumers know which attributes are available,
the attributes must be added during instantiation. Filters and consumers
are not required to check for availability of attributes in
public void reset() throws IOException
Resets this stream to a clean state. Stateful implementations must implement this method so that they can be reused, just as if they had been created fresh.
The default implementation chains the call to the input TokenStream, so
be sure to call
super.reset() when overriding this method.