Download presentation
Presentation is loading. Please wait.
Published byLesley Washington Modified over 9 years ago
1
Modelling Human Thematic Fit Judgments IGK Colloquium 3/2/2005 Ulrike Padó
2
Overview (Very) quick introduction to my framework Testing the Semantic Module Different input corpora Smoothing Comparing the Semantic Module to standard selectional preference methods
3
Modelling Semantic Processing General idea: Build a probabilistic large scale broad coverage model of syntactic and semantic sentence processing
4
Semantic Processing Assign thematic roles on the basis of co- occurrence statistics from semantically annotated corpora Corpus-based frequency estimates of: Semantic Subcategorisation (Probability of seeing the role with the verb) Selectional Preferences (Probability of seeing the argument head in a role given the verb frame)
5
Testing the Semantic Module Evaluate just thematic fit of verbs and argument phrases Evaluation: 1.Correlate predictions with human judgments 2.Role labelling (prefer correct role) Try Different input corpora Smoothing
6
Training Data Frequency counts from the PropBank (ca. 3000 verb types) Very specific domain Relatively flat, syntax-based annotation FrameNet (ca. 1500 verb types) Deep semantic annotation: Frames code situations, group verbs that describe similar events and their arguments Extracted from balanced corpus Skewed sample through frame-wise annotation
7
Development/Test Data Development: 60 verb-argument pairs from McRae et al. 98 Two judgments for each data point: Agent/Patient Use to determine optimal parameters of clustering (number of clusters, smoothing) Test: 50 verb-argument pairs, 100 data points
8
Sparse Data Raw frequencies are sparse: 1 (Dev)/2 (Test) pairs seen in PropBank 0 (Dev)/2 (Test) pairs seen in FrameNet Use semantic classes as level of abstraction: Class-based smoothing
9
Smoothing Reconstruct probabilities for unseen data Smoothing by verb and noun classes Count class members instead of word tokens Compare two alternatives : Hand-constructed classes Induced verb classes (clustering)
10
Hand-constructed Verb and Noun classes WordNet: Use top-level ontology and synsets as noun classes VerbNet: Use top-level classes for verbs Presumably correct and reliable Result: No significant correlations with human data for either training corpus
11
Induced Verb Classes Automatically cluster verbs Group by similarities of argument heads, paths from argument to verb, frame, role labels Determine optimal number of clusters and parameters of the clustering algorithm on the development set
12
Induced Classes, PB/FN Data points covered / Significance Raw data 2-/- 2 All Arguments 59ns 12 =0.55/ p<0.05 Just NPs 48ns 16 =0.56/ p<0.05
13
Results Hand-built classes do not work (with this amount of data) Module achieves reliable correlations with FN data: Important result for the overall feasibility of my model
14
Adding Noun Classes (PB/FN) Data points covered / Significance Raw data 2-/- 2 PB, all args, Noun classes 4 =1/ p<0.01 FN, just NPs, Noun classes 18 =0.63/ p<0.01
15
Results Hand-built classes do not work (with this amount of data) Module achieves reliable correlations with FN data Adding noun classes helps yet a little
16
Comparison with Selectional Preference Methods Have established that our system reliably predicts human data How do we do in comparison to standard computational linguistics methods?
17
Selectional Preference Methods Clark & Weir (2002) Add data points by finding the topmost class in WN that still reliably mirrors the target word frequency Resnik (1996) Quantify contribution of WN class n to the overall preference strength of the verb Both rely on WN noun classes, no verb class smoothing
18
Selectional Preference Methods (PB/FN) Data points covered / Significance Labelling (Cov/Acc) Sem. Module 118 =0.63/ p<0.01 38%/47.4% Sem. Module 216 =0.56/ p<0.05 30%/60% Clark & Weir 72ns84%/50% 23ns36%/50% Resnik 75ns74%/48.6% 46ns50%/48%
19
Results Too little input data No results for selectional preference models Small coverage for Semantic Module Semantic module manages to make predictions all the same Relies on verb clusters: Verbs are less sparse than nouns in small corpora Annotate larger corpus with FN roles
20
Annotating the BNC Annotate large, balanced corpus: BNC More data points for verbs covered in FN More verb coverage (though purely syntactic annotation for unknown verbs) Results: Annotation relatively sensible and reliable for non-FN verbs Frame-wise annotation in FN causes problems for FN verbs
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.