| Modifier and Type | Method and Description |
|---|---|
AnalyzedTokenReadings[] |
AnalyzedSentence.getPreDisambigTokens() |
AnalyzedTokenReadings[] |
AnalyzedSentence.getPreDisambigTokensWithoutWhitespace() |
AnalyzedTokenReadings[] |
AnalyzedSentence.getTokens()
Returns the
AnalyzedTokenReadings of the analyzed text. |
AnalyzedTokenReadings[] |
AnalyzedSentence.getTokensWithoutWhitespace()
Returns the
AnalyzedTokenReadings of the analyzed text, with
whitespace tokens removed but with the artificial SENT_START
token included. |
| Constructor and Description |
|---|
AnalyzedSentence(AnalyzedTokenReadings[] tokens)
Creates an AnalyzedSentence from the given
AnalyzedTokenReadings. |
AnalyzedSentence(AnalyzedTokenReadings[] tokens,
AnalyzedTokenReadings[] preDisambigTokens) |
AnalyzedSentence(AnalyzedTokenReadings[] tokens,
AnalyzedTokenReadings[] preDisambigTokens) |
AnalyzedTokenReadings(AnalyzedTokenReadings oldAtr,
List<AnalyzedToken> newReadings,
String ruleApplied) |
| Modifier and Type | Method and Description |
|---|---|
AnalyzedTokenReadings |
ChunkTaggedToken.getReadings() |
| Modifier and Type | Method and Description |
|---|---|
void |
Chunker.addChunkTags(List<AnalyzedTokenReadings> sentenceTokenReadings) |
| Constructor and Description |
|---|
ChunkTaggedToken(String token,
List<ChunkTag> chunkTags,
AnalyzedTokenReadings readings) |
| Modifier and Type | Method and Description |
|---|---|
void |
NoopChunker.addChunkTags(List<AnalyzedTokenReadings> tokenReadings) |
| Modifier and Type | Method and Description |
|---|---|
protected abstract AnalyzedTokenReadings |
AbstractStatisticSentenceStyleRule.conditionFulfilled(List<AnalyzedTokenReadings> tokens)
Condition to generate a hint (possibly including all exceptions)
Returns:
< nAnalysedToken, if condition is not fulfilled
>= nAnalysedToken, if condition is not fulfilled; integer is number of token which is the end hint
|
static AnalyzedTokenReadings |
GRPCUtils.fromGRPC(MLServerProto.AnalyzedTokenReadings tokenReadings) |
| Modifier and Type | Method and Description |
|---|---|
protected abstract List<AnalyzedTokenReadings> |
PartialPosTagFilter.tag(String token) |
| Modifier and Type | Method and Description |
|---|---|
RuleMatch |
DateRangeChecker.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractDateCheckFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AddCommasFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
WhitespaceCheckFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractFindSuggestionsFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
ShortenedYearRangeChecker.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
PartialPosTagFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractNewYearDateFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractDateCheckWithSuggestionsFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
ConvertToSentenceCaseFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractAdvancedSynthesizerFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AdaptSuggestionsFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractSuppressMisspelledSuggestionsFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
UnderlineSpacesFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
SuppressIfAnyRuleMatchesFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractNumberInWordFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractTextToNumberFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
IsEnglishWordFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractFutureDateFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> args,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
RuleMatch |
AbstractMakeContractionsFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
protected abstract int |
AbstractStatisticStyleRule.conditionFulfilled(AnalyzedTokenReadings[] tokens,
int nAnalysedToken)
Condition to generate a hint (possibly including all exceptions)
Returns:
< nAnalysedToken, if condition is not fulfilled
>= nAnalysedToken, if condition is not fulfilled; integer is number of token which is the end hint
|
protected String |
AbstractWordCoherencyRule.createReplacement(String marked,
String token,
String otherSpelling,
AnalyzedTokenReadings tmpToken) |
protected RuleMatch |
AbstractSimpleReplaceRule.createRuleMatch(AnalyzedTokenReadings tokenReadings,
List<String> replacements,
AnalyzedSentence sentence,
String originalTokenStr) |
protected List<RuleMatch> |
AbstractSimpleReplaceRule.findMatches(AnalyzedTokenReadings tokenReadings,
AnalyzedSentence sentence) |
protected String |
AbstractStyleTooOftenUsedWordRule.getLemmaForPosTagStartsWith(String startPos,
AnalyzedTokenReadings token) |
protected abstract List<String> |
AbstractFindSuggestionsFilter.getSpellingSuggestions(AnalyzedTokenReadings atr) |
protected List<String> |
WordRepeatBeginningRule.getSuggestions(AnalyzedTokenReadings analyzedToken) |
boolean |
WordRepeatRule.ignore(AnalyzedTokenReadings[] tokens,
int position)
Implement this method to return
true if there's
a potential word repetition at the current position that should be ignored,
i.e. if no error should be created. |
protected boolean |
WordRepeatBeginningRule.isAdverb(AnalyzedTokenReadings token) |
boolean |
ParagraphRepeatBeginningRule.isArticle(AnalyzedTokenReadings token) |
protected abstract boolean |
AbstractStyleTooOftenUsedWordRule.isException(AnalyzedTokenReadings token)
An exception is defined for the token
|
protected boolean |
AbstractFillerWordsRule.isException(AnalyzedTokenReadings[] tokens,
int num) |
protected boolean |
CommaWhitespaceRule.isException(AnalyzedTokenReadings[] tokens,
int tokenIdx) |
protected boolean |
UppercaseSentenceStartRule.isException(AnalyzedTokenReadings[] tokens,
int tokenIdx) |
protected abstract boolean |
AbstractRepeatedWordsRule.isException(AnalyzedTokenReadings[] tokens,
int i,
boolean sentStart,
boolean isCapitalized,
boolean isAllUppercase) |
protected boolean |
AbstractStyleRepeatedWordRule.isExceptionPair(AnalyzedTokenReadings token1,
AnalyzedTokenReadings token2) |
protected boolean |
AbstractStatisticSentenceStyleRule.isMark(AnalyzedTokenReadings token) |
protected boolean |
GenericUnpairedBracketsRule.isNoException(String token,
AnalyzedTokenReadings[] tokens,
int i,
int j,
boolean precSpace,
boolean follSpace,
UnsyncStack<SymbolLocator> symbolStack)
Generic method to specify an exception.
|
protected boolean |
GenericUnpairedQuotesRule.isNotBeginningApostrophe(AnalyzedTokenReadings[] tokens,
int i) |
protected boolean |
GenericUnpairedQuotesRule.isNotEndingApostrophe(AnalyzedTokenReadings[] tokens,
int i) |
protected boolean |
AbstractStatisticSentenceStyleRule.isOpeningQuote(AnalyzedTokenReadings token) |
protected abstract boolean |
AbstractTextToNumberFilter.isPercentage(AnalyzedTokenReadings[] patternTokens,
int i) |
protected boolean |
AbstractFindSuggestionsFilter.isSuggestionException(AnalyzedTokenReadings analyzedSuggestion) |
protected boolean |
AbstractSimpleReplaceRule.isTagged(AnalyzedTokenReadings tokenReadings)
This method allows to override which tags will mark token as tagged
|
protected abstract boolean |
AbstractStyleTooOftenUsedWordRule.isToCountedWord(AnalyzedTokenReadings token)
A token that has to be counted
|
protected boolean |
AbstractSimpleReplaceRule.isTokenException(AnalyzedTokenReadings atr) |
protected boolean |
AbstractSimpleReplaceRule2.isTokenException(AnalyzedTokenReadings atr) |
protected abstract boolean |
AbstractStyleRepeatedWordRule.isTokenPair(AnalyzedTokenReadings[] tokens,
int n,
boolean before) |
protected abstract boolean |
AbstractStyleRepeatedWordRule.isTokenToCheck(AnalyzedTokenReadings token) |
protected abstract boolean |
AbstractStatisticStyleRule.sentenceConditionFulfilled(AnalyzedTokenReadings[] tokens,
int nAnalysedToken)
Condition to generate a hint related to the sentence (possibly including all exceptions)
|
protected URL |
AbstractStyleRepeatedWordRule.setURL(AnalyzedTokenReadings token) |
protected abstract String |
AbstractStyleTooOftenUsedWordRule.toAddedLemma(AnalyzedTokenReadings token)
Gives back the lemma that should be added to the word map
|
static MLServerProto.AnalyzedTokenReadings |
GRPCUtils.toGRPC(AnalyzedTokenReadings readings) |
protected boolean |
WordRepeatRule.wordRepetitionOf(String word,
AnalyzedTokenReadings[] tokens,
int position) |
| Modifier and Type | Method and Description |
|---|---|
protected abstract AnalyzedTokenReadings |
AbstractStatisticSentenceStyleRule.conditionFulfilled(List<AnalyzedTokenReadings> tokens)
Condition to generate a hint (possibly including all exceptions)
Returns:
< nAnalysedToken, if condition is not fulfilled
>= nAnalysedToken, if condition is not fulfilled; integer is number of token which is the end hint
|
| Modifier and Type | Field and Description |
|---|---|
protected AnalyzedTokenReadings[] |
AbstractPatternRulePerformer.unifiedTokens |
| Modifier and Type | Method and Description |
|---|---|
AnalyzedTokenReadings |
MatchState.filterReadings() |
AnalyzedTokenReadings[] |
Unifier.getFinalUnified()
Used for getting a unified sequence in case when simple test method
Unifier.isUnified(AnalyzedToken, Map, boolean)} was used. |
AnalyzedTokenReadings[] |
Unifier.getUnifiedTokens()
Gets a full sequence of filtered tokens.
|
| Modifier and Type | Method and Description |
|---|---|
abstract RuleMatch |
RuleFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions)
Returns the original rule match or a modified one, or
null
if the rule match is filtered out. |
RuleMatch |
ApostropheTypeFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
void |
Unifier.addNeutralElement(AnalyzedTokenReadings analyzedTokenReadings)
Used to add neutral elements (
AnalyzedTokenReadings to the
unified sequence. |
PatternToken |
PatternToken.compile(AnalyzedTokenReadings token,
Synthesizer synth)
Prepare PatternToken for matching by formatting its string token and POS (if the Element is supposed
to refer to some other token).
|
MatchState |
Match.createState(Synthesizer synthesizer,
AnalyzedTokenReadings token)
Creates a state used for actually matching a token.
|
MatchState |
Match.createState(Synthesizer synthesizer,
AnalyzedTokenReadings[] tokens,
int index,
int next)
Creates a state used for actually matching a token.
|
protected void |
AbstractPatternRulePerformer.doMatch(AnalyzedSentence sentence,
AnalyzedTokenReadings[] tokens,
AbstractPatternRulePerformer.MatchConsumer consumer) |
protected int |
RuleFilter.getPosition(String fromStr,
AnalyzedTokenReadings[] patternTokens,
RuleMatch match) |
Map<String,String> |
RuleFilterEvaluator.getResolvedArguments(String filterArgs,
AnalyzedTokenReadings[] patternTokens,
int patternTokenPos,
List<Integer> tokenPositions)
Resolves the backref arguments, e.g. replaces
\1 by the value of the first token in the pattern. |
protected boolean |
RuleFilter.isMatchAtSentenceStart(AnalyzedTokenReadings[] tokens,
RuleMatch match) |
boolean |
PatternTokenMatcher.isMatchedByPreviousException(AnalyzedTokenReadings token) |
boolean |
PatternToken.isMatchedByPreviousException(AnalyzedTokenReadings prevToken)
Checks whether an exception for a previous token matches all readings of a given token (in case
the exception had scope == "previous").
|
boolean |
RuleFilter.matches(Map<String,String> arguments,
AnalyzedTokenReadings[] patternTokens,
int firstMatchToken,
List<Integer> tokenPositions) |
void |
PatternTokenMatcher.resolveReference(int firstMatchToken,
AnalyzedTokenReadings[] tokens,
Language language) |
RuleMatch |
RuleFilterEvaluator.runFilter(String filterArgs,
RuleMatch ruleMatch,
AnalyzedTokenReadings[] patternTokens,
int patternTokenPos,
List<Integer> tokenPositions) |
void |
MatchState.setToken(AnalyzedTokenReadings token) |
void |
MatchState.setToken(AnalyzedTokenReadings[] tokens,
int index,
int next)
Sets the token to be formatted etc. and includes the support for
including the skipped tokens.
|
protected boolean |
AbstractPatternRulePerformer.testAllReadings(AnalyzedTokenReadings[] tokens,
PatternTokenMatcher matcher,
PatternTokenMatcher prevElement,
int tokenNo,
int firstMatchToken,
int prevSkipNext) |
protected boolean |
PatternRuleMatcher.testAllReadings(AnalyzedTokenReadings[] tokens,
PatternTokenMatcher matcher,
PatternTokenMatcher prevElement,
int tokenNo,
int firstMatchToken,
int prevSkipNext) |
| Modifier and Type | Method and Description |
|---|---|
protected boolean |
SpellingCheckRule.ignoreToken(AnalyzedTokenReadings[] tokens,
int idx)
Returns true iff the token at the given position should be ignored by the spell checker.
|
| Modifier and Type | Method and Description |
|---|---|
protected List<RuleMatch> |
MorfologikSpellerRule.getRuleMatches(String word,
int startPos,
AnalyzedSentence sentence,
List<RuleMatch> ruleMatchesSoFar,
int idx,
AnalyzedTokenReadings[] tokens) |
| Modifier and Type | Method and Description |
|---|---|
RuleMatch |
MultitokenSpellerFilter.acceptRuleMatch(RuleMatch match,
Map<String,String> arguments,
int patternTokenPos,
AnalyzedTokenReadings[] patternTokens,
List<Integer> tokenPositions) |
| Modifier and Type | Method and Description |
|---|---|
AnalyzedTokenReadings |
Tagger.createNullToken(String token,
int startPos)
Create the AnalyzedToken used for whitespace and other non-words.
|
AnalyzedTokenReadings |
BaseTagger.createNullToken(String token,
int startPos) |
| Modifier and Type | Method and Description |
|---|---|
List<AnalyzedTokenReadings> |
Tagger.tag(List<String> sentenceTokens)
Returns a list of
AnalyzedTokens that assigns each term in the
sentence some kind of part-of-speech information (not necessarily just one tag). |
List<AnalyzedTokenReadings> |
BaseTagger.tag(List<String> sentenceTokens) |
| Modifier and Type | Method and Description |
|---|---|
protected AnalyzedTokenReadings |
MultiWordChunker2.prepareNewReading(String tokens,
String tok,
AnalyzedTokenReadings token,
String tag) |
| Modifier and Type | Method and Description |
|---|---|
protected boolean |
MultiWordChunker2.matches(String matchText,
AnalyzedTokenReadings inputTokens) |
protected AnalyzedTokenReadings |
MultiWordChunker2.prepareNewReading(String tokens,
String tok,
AnalyzedTokenReadings token,
String tag) |
| Modifier and Type | Method and Description |
|---|---|
AnalyzedTokenReadings |
DemoTagger.createNullToken(String token,
int startPos) |
| Modifier and Type | Method and Description |
|---|---|
List<AnalyzedTokenReadings> |
DemoTagger.tag(List<String> sentenceTokens) |