## A New Formalization of Probabilistic GLR Parsing (1997)

Venue: | In Proceedings of the 5th International Workshop on Parsing Technologies |

Citations: | 21 - 6 self |

### BibTeX

@INPROCEEDINGS{Inui97anew,

author = {Kentaro Inui and Virach Sornlertlamvanich and Hozumi Tanaka and Takenobu Tokunaga},

title = {A New Formalization of Probabilistic GLR Parsing},

booktitle = {In Proceedings of the 5th International Workshop on Parsing Technologies},

year = {1997},

pages = {123--134}

}

### OpenURL

### Abstract

This paper presents a new formalization of probabilistic GLR language modeling for statistical parsing. Our model inherits its essential features from Briscoe and Carroll's generalized probabilistic LR model [3], which obtains context-sensitivity by assigning a probability to each LR parsing action according to its left and right context. Briscoe and Carroll's model, however, has a drawback in that it is not formalized in any probabilistically well-founded way, which may degrade its parsing performance. Our formulation overcomes this drawback with a few significant refinements, while maintaining all the advantages of Briscoe and Carroll's modeling. 1 Introduction The increasing availability of text corpora has encouraged researchers to explore statistical approaches for various tasks in natural language processing. Statistical parsing is one of these approaches. In statistical parsing, one of the most straightforward methodologies is to generalize context-free grammars by associating ...

### Citations

438 | A New Statistical Parser Based on Bigram Lexical Dependencies
- Collins
- 1996
(Show Context)
Citation Context ...s and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fact, several attempts to use lexically sensitive models already exist: e.g. =-=[6, 10, 13]-=-. Our future research will also be directed towards this area [7]. Acknowledgements The authors would like to thank the reviewers for their suggestive comments. They would also like to thank Mr. UEKI ... |

192 | Generalized Probabilistic LR Parsing of Natural Language (Corpora) with Uni�cation-Based Grammars
- Briscoe, Carroll
- 1993
(Show Context)
Citation Context ...er presents a new formalization of probabilistic GLR language modeling for statistical parsing. Our model inherits its essential features from Briscoe and Carroll's generalized probabilistic LR model =-=[3]-=-, which obtains context-sensitivity by assigning a probability to each LR parsing action according to its left and right context. Briscoe and Carroll's model, however, has a drawback in that it is not... |

188 |
Ef�cient Parsing for Natural Language
- Tomita
- 1986
(Show Context)
Citation Context ...g is one existing statistical parsing methodology which is more context-sensitive than PCFG-based parsing. Several attempts have been made to incorporate probability into generalized LR (GLR) parsing =-=[17]-=-. For example, Wright and Wrigley proposed an algorithm to distribute probabilities originally associated with CFG rules to LR parsing actions, in such a way that the resulting model is equivalent to ... |

157 | Towards history-based grammars: Using richer models for probabilistic parsing
- Black, Jelinek, et al.
- 1992
(Show Context)
Citation Context ... trees (a) and (b) are the parse trees for input sentence W 1 = fu, x, xg, and (c) and (d) are those for W 2 = fv, x, xg. Table 3 shows the LR table for grammar G3, with the trained S x S S u S x (b) =-=[2]-=- 5 4 3 7 4 re1 (a) [3] 3 sh3 7 5 S u S S x S x 4 5 S x S S v S x (d) [4] 6 4 3 7 4 re2 (c) [1] 3 sh3 7 6 S v S S x S x 4 6 Figure 3. Training set (The square-bracketed number below each tree denotes t... |

120 |
Stochastic Lexicalized Tree-Adjoining Grammars
- Schabes
- 1992
(Show Context)
Citation Context ...s and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fact, several attempts to use lexically sensitive models already exist: e.g. =-=[6, 10, 13]-=-. Our future research will also be directed towards this area [7]. Acknowledgements The authors would like to thank the reviewers for their suggestive comments. They would also like to thank Mr. UEKI ... |

73 | A Corpus-based Probabilistic Grammar with Only Two Non-terminals
- Sekine, Grishman
- 1995
(Show Context)
Citation Context ...erimental comparisons between these two types (for a qualitative comparison, see [8]). Other approaches to statistical parsing using context-sensitive language models have also been proposed, such as =-=[2, 9, 12, 14]-=-. We need to make theoretical and empirical comparisons between these models and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fa... |

22 |
A Probabilistic Chart Parser
- Pearl
- 1991
(Show Context)
Citation Context ...erimental comparisons between these two types (for a qualitative comparison, see [8]). Other approaches to statistical parsing using context-sensitive language models have also been proposed, such as =-=[2, 9, 12, 14]-=-. We need to make theoretical and empirical comparisons between these models and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fa... |

20 |
GLR Parsing with Scoring
- Su, Wang, et al.
- 1991
(Show Context)
Citation Context ...ramework, their language model still failed to capture the context-sensitivity of languages. Su et al. proposed a way of introducing probabilistic distribution into the shift-reduce parsing framework =-=[16]-=-. Unlike Wright and Wrigley's work, the goal of this research was the construction of a mildly contextsensitive model 1 that is effective for statistical parsing. Their model distributes probabilities... |

15 |
LR Parsing: Theory and Practice
- Chapman
- 1987
(Show Context)
Citation Context ... grammar rules, it is to be expected that B&C model tends to strongly prefer parse trees involving fewer grammar rules. transitions may not be accepted (for details of CLR and LALR, see, for example, =-=[1, 5]-=-). However, this fact will never prevent our model from being applicable to LALR. Let T acc ae T be the infinite set of all possible complete and acceptable stack transition sequences. The second fact... |

12 | A probabilistic disambiguation method based on psycholinguistic
- Li
- 1996
(Show Context)
Citation Context ...s and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fact, several attempts to use lexically sensitive models already exist: e.g. =-=[6, 10, 13]-=-. Our future research will also be directed towards this area [7]. Acknowledgements The authors would like to thank the reviewers for their suggestive comments. They would also like to thank Mr. UEKI ... |

12 |
GLR parsing with probability
- Wright, Wrigley
- 1991
(Show Context)
Citation Context ...ght and Wrigley proposed an algorithm to distribute probabilities originally associated with CFG rules to LR parsing actions, in such a way that the resulting model is equivalent to the original PCFG =-=[19]-=-. Perhaps, the most naive way of coupling a PCFG model with the GLR parsing framework would be to assign the probability associated with each CFG rule to the reduce actions for that rule. Wright and W... |

6 |
Probabilistic Normalization and Unpacking of Packed Parse Forests for Unification-based Grammars
- Carrol, Briscoe
- 1992
(Show Context)
Citation Context ...oth shift and reduce actions. Furthermore, since our model assigns a single probabilistic parameter to each action in an LR table similarly to B&C model, the algorithm proposed by Carroll and Briscoe =-=[4]-=- for efficient unpacking of packed parse forests with probability annotations can be equally applicable to our model. Finally, although not explicitly pointed out by Briscoe and Carroll, it should als... |

6 |
A new probabilistic LR language model for statistical parsing
- Inui, Sornlertlamvanich, et al.
- 1997
(Show Context)
Citation Context ... one can rank the parse tree candidates for any given input sentence according to P (T ) and P (W jT ), whether one bases the model on CLR, LALR, or even LR(0) (i.e. SLR). For further discussion, see =-=[8]-=-. To solve this problem, Briscoe and Carroll proposed calculating the geometric mean of the probabilities of the actions involved in each stack transition sequence. However, this solution makes their ... |

5 | Integrated probabilistic language modeling for statistical parsing
- Inui, Shirai, et al.
- 1997
(Show Context)
Citation Context ...odel to estimate the lexical distribution P (W jT ); for example, one could introduce the statistics of word collocations. However, this is beyond the scope of this paper. For further discussion, see =-=[7]-=-. A stack transition sequence T can be described as (4): oe 0 l 1 ;a1 =) oe 1 l 2 ;a2 =) . . . l n01 ;an01 =) oe n01 l n ;an =) oe n (4) where oe i is the i-th stack, whose stack-top state is denoted ... |

4 |
Spoken sentence recognition based on HMM-LR with hybrid language modeling
- Kita
- 1994
(Show Context)
Citation Context ...erimental comparisons between these two types (for a qualitative comparison, see [8]). Other approaches to statistical parsing using context-sensitive language models have also been proposed, such as =-=[2, 9, 12, 14]-=-. We need to make theoretical and empirical comparisons between these models and ours. The significance of introducing lexical sensitivity into language models should also not be underestimated. In fa... |

4 |
Incorporating probabilistic parsing into an LR parser -- LR table engineering (4
- Sornlartlamvanich, Inui, et al.
- 1997
(Show Context)
Citation Context ...ove parsing performance. Those examples may seem to be relatively artificial and forced. However, in our preliminary experiments, we are achieving some promising results, which support our claim (see =-=[15]-=- for a summary of preliminary results). We are now planning to conduct further large-scaled experiments. It should also be noted that our modeling is equally applicable to both CLR tables and LALR tab... |

3 |
A method for integrating the connection constraints into an LR table
- Li, Tanaka
- 1995
(Show Context)
Citation Context ...rminal symbols P (l i jl i01 ). This feature is expected to be significant particularly when one attempts to integrate syntactic parsing with morphological analysis in the GLR parsing framework (e.g. =-=[11]-=-), since the bigram model of terminal symbols has been empirically proven to be effective in morphological analysis. Besides these advantages, which are all shared with B&C model, our model overcomes ... |