| Package | Description |
|---|---|
| org.languagetool | |
| org.languagetool.noop | |
| org.languagetool.rules.ngrams | |
| org.languagetool.tokenizers |
| Modifier and Type | Method and Description |
|---|---|
Tokenizer |
Language.createDefaultWordTokenizer()
Creates language specific word tokenizer.
|
Tokenizer |
Language.getWordTokenizer()
Get this language's word tokenizer implementation.
|
| Modifier and Type | Method and Description |
|---|---|
void |
Language.setWordTokenizer(Tokenizer tokenizer)
Set this language's word tokenizer implementation.
|
| Modifier and Type | Method and Description |
|---|---|
Tokenizer |
NoopLanguage.createDefaultWordTokenizer() |
| Modifier and Type | Method and Description |
|---|---|
protected Tokenizer |
NgramProbabilityRule.getGoogleStyleWordTokenizer() |
| Modifier and Type | Interface and Description |
|---|---|
interface |
CompoundWordTokenizer
Interface for components that take compound words and split
them into their parts.
|
interface |
SentenceTokenizer
Tokenizes text into sentences.
|
| Modifier and Type | Class and Description |
|---|---|
class |
SimpleSentenceTokenizer
A very simple sentence tokenizer that splits on {@code [.!?
|
class |
SRXSentenceTokenizer
Class to tokenize sentences using rules from an SRX file.
|
class |
WordTokenizer
Tokenizes a sentence into words.
|