SENTENCE_SEGMENTER_MODEL¶
-
lexnlp.nlp.en.segments.sentences.SENTENCE_SEGMENTER_MODEL= <nltk.tokenize.punkt.PunktSentenceTokenizer object>¶ A sentence tokenizer which uses an unsupervised algorithm to build a model for abbreviation words, collocations, and words that start sentences; and then uses that model to find sentence boundaries. This approach has been shown to work well for many European languages.