author vincent.michel@logilab.fr
Thu, 19 Dec 2013 14:45:26 +0000
changeset 373 77a3a4107f5c
parent 370 named_entities/tokenizer.py@4a79af983c29
child 469 9d9d8c4f2bab
permissions -rw-r--r--
[named entities] Move tokenizer to utils and create a sources module for named entities, related to #187461

# -*- coding: utf-8 -*-
""" Tokenizer for sentences/words segmentation.
import itertools
import collections
import re

Token = collections.namedtuple('Token', ['word', 'start', 'end', 'sentence'])
Sentence = collections.namedtuple('Sentence', ['indice', 'start', 'end'])

class RichStringTokenizer(object):
    """Tokenizer for Yams' RichString content.

    The tokenizer uses a variable-length sliding window, i.e. a sliding
    window yielding tokens of N words.

    def __init__(self, text, token_min_size=1, token_max_size=3):
        :token_min_size: minimum number of words required to be a valid token
        :token_max_size: minimum number of words required to be a valid token
        self.text = text
        self.token_min_size = token_min_size
        self.token_max_size = token_max_size

    def iter_tokens(self, text):
        """ Iterate tokens over a text
        # Compute sentences
        sentences = self.find_sentences(text)
        # Compute words
        words = list([m for m in re.finditer(r'[\w@-]+', text, re.UNICODE)])
        indice = 0
        while indice < len(words):
            # Choose the current sentence of the first word
            current_sentence = [s for s in sentences if s.start<=words[indice].start()][-1]
            # Sliding windows over the different words for each sentence
            remaining = len(words) - indice
            for length in range(min(self.token_max_size, remaining), self.token_min_size-1, -1):
                _words = words[indice:indice+length]
                if _words[-1].start() > current_sentence.end:
                    # The last word in not in the same sentence anymore, split
                normalized_word = ' '.join([w.group() for w in _words]).strip()
                yield Token(normalized_word, _words[0].start(), _words[-1].end(), current_sentence)
            indice += 1

    def find_sentences(self, text):
        """ Find the sentences
        return [Sentence(ind, s.start(), s.end()) for ind, s in
                enumerate(re.finditer(r'[^.!?]+(?:[.!?]|$)', text, re.UNICODE))]

    def load_text(self, text):
        """ Load the text to be tokenized
        self.text = text

    def __iter__(self):
        """ Iterator over the text given in the object instantiation
        for t in self.iter_tokens(self.text):
            yield t