Hot-keys on this page

r m x p   toggle line displays

j k   next/prev highlighted chunk

0   (zero) top of page

1   (one) first highlighted chunk

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

81

82

83

84

85

86

87

88

89

90

91

92

93

94

95

96

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

# Natural Language Toolkit: Tokenizers 

# 

# Copyright (C) 2001-2012 NLTK Project 

# Author: Edward Loper <edloper@gradient.cis.upenn.edu> 

#         Michael Heilman <mheilman@cmu.edu> (re-port from http://www.cis.upenn.edu/~treebank/tokenizer.sed) 

#          

# URL: <http://nltk.sourceforge.net> 

# For license information, see LICENSE.TXT 

 

r""" 

 

Penn Treebank Tokenizer 

 

The Treebank tokenizer uses regular expressions to tokenize text as in Penn Treebank. 

This implementation is a port of the tokenizer sed script written by Robert McIntyre 

and available at http://www.cis.upenn.edu/~treebank/tokenizer.sed. 

""" 

 

import re 

from nltk.tokenize.api import TokenizerI 

 

 

class TreebankWordTokenizer(TokenizerI): 

    """ 

    The Treebank tokenizer uses regular expressions to tokenize text as in Penn Treebank. 

    This is the method that is invoked by ``word_tokenize()``.  It assumes that the 

    text has already been segmented into sentences, e.g. using ``sent_tokenize()``. 

 

    This tokenizer performs the following steps: 

 

    - split standard contractions, e.g. ``don't`` -> ``do n't`` and ``they'll`` -> ``they 'll`` 

    - treat most punctuation characters as separate tokens 

    - split off commas and single quotes, when followed by whitespace 

    - separate periods that appear at the end of line 

 

        >>> from nltk.tokenize import TreebankWordTokenizer 

        >>> s = '''Good muffins cost $3.88\\nin New York.  Please buy me\\ntwo of them.\\n\\nThanks.''' 

        >>> TreebankWordTokenizer().tokenize(s) 

        ['Good', 'muffins', 'cost', '$', '3.88', 'in', 'New', 'York.', 

        'Please', 'buy', 'me', 'two', 'of', 'them', '.', 'Thanks', '.'] 

        >>> s = "They'll save and invest more." 

        >>> TreebankWordTokenizer().tokenize(s) 

        ['They', "'ll", 'save', 'and', 'invest', 'more', '.'] 

 

    NB. this tokenizer assumes that the text is presented as one sentence per line, 

    where each line is delimited with a newline character. 

    The only periods to be treated as separate tokens are those appearing 

    at the end of a line. 

    """ 

 

    # List of contractions adapted from Robert MacIntyre's tokenizer. 

    CONTRACTIONS2 = [re.compile(r"(?i)\b(can)(not)\b"), 

                     re.compile(r"(?i)\b(d)('ye)\b"), 

                     re.compile(r"(?i)\b(gim)(me)\b"), 

                     re.compile(r"(?i)\b(gon)(na)\b"), 

                     re.compile(r"(?i)\b(got)(ta)\b"), 

                     re.compile(r"(?i)\b(lem)(me)\b"), 

                     re.compile(r"(?i)\b(mor)('n)\b"), 

                     re.compile(r"(?i)\b(wan)(na) ")] 

    CONTRACTIONS3 = [re.compile(r"(?i) ('t)(is)\b"), 

                     re.compile(r"(?i) ('t)(was)\b")] 

    CONTRACTIONS4 = [re.compile(r"(?i)\b(whad)(dd)(ya)\b"), 

                     re.compile(r"(?i)\b(wha)(t)(cha)\b")] 

 

    def tokenize(self, text): 

        #starting quotes 

        text = re.sub(r'^\"', r'``', text) 

        text = re.sub(r'(``)', r' \1 ', text) 

        text = re.sub(r'([ (\[{<])"', r'\1 `` ', text) 

 

        #punctuation 

        text = re.sub(r'([:,])([^\d])', r' \1 \2', text) 

        text = re.sub(r'\.\.\.', r' ... ', text) 

        text = re.sub(r'[;@#$%&]', r' \g<0> ', text) 

        text = re.sub(r'([^\.])(\.)([\]\)}>"\']*)\s*$', r'\1 \2\3 ', text) 

        text = re.sub(r'[?!]', r' \g<0> ', text) 

 

        text = re.sub(r"([^'])' ", r"\1 ' ", text) 

 

        #parens, brackets, etc. 

        text = re.sub(r'[\]\[\(\)\{\}\<\>]', r' \g<0> ', text) 

        text = re.sub(r'--', r' -- ', text) 

 

        #add extra space to make things easier 

        text = " " + text + " " 

 

        #ending quotes 

        text = re.sub(r'"', " '' ", text) 

        text = re.sub(r'(\S)(\'\')', r'\1 \2 ', text) 

 

        text = re.sub(r"([^' ])('[sS]|'[mM]|'[dD]|') ", r"\1 \2 ", text) 

        text = re.sub(r"([^' ])('ll|'re|'ve|n't|) ", r"\1 \2 ", text) 

        text = re.sub(r"([^' ])('LL|'RE|'VE|N'T|) ", r"\1 \2 ", text) 

 

        for regexp in self.CONTRACTIONS2: 

            text = regexp.sub(r' \1 \2 ', text) 

        for regexp in self.CONTRACTIONS3: 

            text = regexp.sub(r' \1 \2 ', text) 

 

        # We are not using CONTRACTIONS4 since 

        # they are also commented out in the SED scripts 

        # for regexp in self.CONTRACTIONS4: 

        #     text = regexp.sub(r' \1 \2 \3 ', text) 

 

        text = re.sub(" +", " ", text) 

        text = text.strip() 

 

        #add space at end to match up with MacIntyre's output (for debugging) 

        if text != "": 

            text += " " 

 

        return text.split() 

 

if __name__ == "__main__": 

    import doctest 

    doctest.testmod(optionflags=doctest.NORMALIZE_WHITESPACE)