• No results found

Efficient Disfluency Detection with Transition based Parsing

N/A
N/A
Protected

Academic year: 2020

Share "Efficient Disfluency Detection with Transition based Parsing"

Copied!
9
0
0

Loading.... (view fulltext now)

Full text

Loading

Figure

Figure 2(b), while the parsing tree is being built,
Figure 3: An example of UT model, where ‘N’means the word is a fluent word and ‘X’ means it isdisfluent
Table 1: Feature templates designed for disfluencydetection and dependency parsing.
Table 2: Disfluency detection and parsing accuracies on English Switchboard data. The accuracy ofM3 N refers to the result reported in (Qian and Liu, 2013)

References

Related documents

Greedy transition-based dependency parsing has been widely deployed because of its speed (Cer et al., 2010); however, state-of-the-art accuracies have been achieved by

(2009) introduced the best transition- based system using synchronous syntactic-semantic parsing (‘Merlo’), and Bohnet (2009) introduced the best graph-based system using a

Section 2 describes our proposed models including the transition-based neural model, the dynamic oracle strategy and the implicit syntax feature extraction approach.. Section 3

T RANX uses a transition system based on the abstract syntax descrip- tion language for the target MR, which gives it two major advantages: (1) it is highly ac- curate,

One issue with the simple seq2seq neural net- work for dependency parsing is that structural lin- guistic information, which plays a key role in clas- sic transition-based

While past work on neural network transition- based parsing has focused exclusively on the arc- standard transition system, it is known that bet- ter results can often be obtained

To mitigate both issues, we pro- pose a novel parsing system based on in-order traversal over syntactic trees, designing a set of transition actions to find a compromise be-

We train six different types of NN classifiers for transition-based dependency parsing: one for each search algorithm with error states, with and with- out pre-trained word