Hot-keys on this page

r m x p   toggle line displays

j k   next/prev highlighted chunk

0   (zero) top of page

1   (one) first highlighted chunk

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

81

82

83

84

85

86

87

88

89

90

91

92

93

94

95

96

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

165

# Natural Language Toolkit: Parser Utility Functions 

# 

# Author: Ewan Klein <ewan@inf.ed.ac.uk> 

# 

# Copyright (C) 2001-2012 NLTK Project 

# URL: <http://www.nltk.org/> 

# For license information, see LICENSE.TXT 

 

 

""" 

Utility functions for parsers. 

""" 

from __future__ import print_function 

 

from nltk.grammar import ContextFreeGrammar, FeatureGrammar, WeightedGrammar 

from nltk.data import load 

 

from nltk.parse.chart import Chart, ChartParser 

from nltk.parse.pchart import InsideChartParser 

from nltk.parse.featurechart import FeatureChart, FeatureChartParser 

 

def load_parser(grammar_url, trace=0, 

                parser=None, chart_class=None, 

                beam_size=0, **load_args): 

    """ 

    Load a grammar from a file, and build a parser based on that grammar. 

    The parser depends on the grammar format, and might also depend 

    on properties of the grammar itself. 

 

    The following grammar formats are currently supported: 

      - ``'cfg'``  (CFGs: ``ContextFreeGrammar``) 

      - ``'pcfg'`` (probabilistic CFGs: ``WeightedGrammar``) 

      - ``'fcfg'`` (feature-based CFGs: ``ContextFreeGrammar``) 

 

    :type grammar_url: str 

    :param grammar_url: A URL specifying where the grammar is located. 

        The default protocol is ``"nltk:"``, which searches for the file 

        in the the NLTK data package. 

    :type trace: int 

    :param trace: The level of tracing that should be used when 

        parsing a text.  ``0`` will generate no tracing output; 

        and higher numbers will produce more verbose tracing output. 

    :param parser: The class used for parsing; should be ``ChartParser`` 

        or a subclass. 

        If None, the class depends on the grammar format. 

    :param chart_class: The class used for storing the chart; 

        should be ``Chart`` or a subclass. 

        Only used for CFGs and feature CFGs. 

        If None, the chart class depends on the grammar format. 

    :type beam_size: int 

    :param beam_size: The maximum length for the parser's edge queue. 

        Only used for probabilistic CFGs. 

    :param load_args: Keyword parameters used when loading the grammar. 

        See ``data.load`` for more information. 

    """ 

    grammar = load(grammar_url, **load_args) 

    if not isinstance(grammar, ContextFreeGrammar): 

        raise ValueError("The grammar must be a ContextFreeGrammar, " 

                         "or a subclass thereof.") 

    if isinstance(grammar, WeightedGrammar): 

        if parser is None: 

            parser = InsideChartParser 

        return parser(grammar, trace=trace, beam_size=beam_size) 

 

    elif isinstance(grammar, FeatureGrammar): 

        if parser is None: 

            parser = FeatureChartParser 

        if chart_class is None: 

            chart_class = FeatureChart 

        return parser(grammar, trace=trace, chart_class=chart_class) 

 

    else: # Plain ContextFreeGrammar. 

        if parser is None: 

            parser = ChartParser 

        if chart_class is None: 

            chart_class = Chart 

        return parser(grammar, trace=trace, chart_class=chart_class) 

 

 

###################################################################### 

#{ Test Suites 

###################################################################### 

 

class TestGrammar(object): 

    """ 

    Unit tests for  CFG. 

    """ 

    def __init__(self, grammar, suite, accept=None, reject=None): 

        self.test_grammar = grammar 

 

        self.cp = load_parser(grammar, trace=0) 

        self.suite = suite 

        self._accept = accept 

        self._reject = reject 

 

 

    def run(self, show_trees=False): 

        """ 

        Sentences in the test suite are divided into two classes: 

         - grammatical (``accept``) and 

         - ungrammatical (``reject``). 

        If a sentence should parse accordng to the grammar, the value of 

        ``trees`` will be a non-empty list. If a sentence should be rejected 

        according to the grammar, then the value of ``trees`` will be None. 

        """ 

        for test in self.suite: 

            print(test['doc'] + ":", end=' ') 

            for key in ['accept', 'reject']: 

                for sent in test[key]: 

                    tokens = sent.split() 

                    trees = self.cp.parse(tokens) 

                    if show_trees and trees: 

                        print() 

                        print(sent) 

                        for tree in trees: 

                            print(tree) 

                    if key == 'accept': 

                        if trees == []: 

                            raise ValueError("Sentence '%s' failed to parse'" % sent) 

                        else: 

                            accepted = True 

                    else: 

                        if trees: 

                            raise ValueError("Sentence '%s' received a parse'" % sent) 

                        else: 

                            rejected = True 

            if accepted and rejected: 

                print("All tests passed!") 

 

def extract_test_sentences(string, comment_chars="#%;"): 

    """ 

    Parses a string with one test sentence per line. 

    Lines can optionally begin with: 

      - a bool, saying if the sentence is grammatical or not, or 

      - an int, giving the number of parse trees is should have, 

    The result information is followed by a colon, and then the sentence. 

    Empty lines and lines beginning with a comment char are ignored. 

 

    :return: a list of tuple of sentences and expected results, 

        where a sentence is a list of str, 

        and a result is None, or bool, or int 

 

    :param comment_chars: ``str`` of possible comment characters. 

    """ 

    sentences = [] 

    for sentence in string.split('\n'): 

        if sentence == '' or sentence[0] in comment_chars: 

            continue 

        split_info = sentence.split(':', 1) 

        result = None 

        if len(split_info) == 2: 

            if split_info[0] in ['True','true','False','false']: 

                result = split_info[0] in ['True','true'] 

                sentence = split_info[1] 

            else: 

                result = int(split_info[0]) 

                sentence = split_info[1] 

        tokens = sentence.split() 

        if tokens == []: 

            continue 

        sentences += [(tokens, result)] 

    return sentences 

 

# nose thinks it is a test 

extract_test_sentences.__test__ = False