Presentation is loading. Please wait.

Presentation is loading. Please wait.

Ali Hakimi Parizi, Paul Cook

Similar presentations


Presentation on theme: "Ali Hakimi Parizi, Paul Cook"— Presentation transcript:

1 Do Character-level Language Models Capture Knowledge of Multiword Expression Compositionality?
Ali Hakimi Parizi, Paul Cook Faculty of Computer Science, University of New Brunswick Multiword Expressions Training Data and Evaluation Dataset Multiword expressions (MWEs) are word combinations that display some form of idiomaticity. For all experiments, we train our models over raw text Wikipedia corpora for either English or German Compositionality refers to the degree to which the meaning of a MWE can be predicted by combining the meanings of its components. 1 Preprocessing steps: Use the WP2TXT toolbox to eliminate XML and HTML tags and hyperlinks. Select a portion of data randomly. Remove all other characters which are not in the language. End User End + User Compositional Its meaning is related to “game” to some degree Game Plan Semi Compositional We evaluate our methods over two datasets: English noun compounds, “ENCs”: contains 90 noun compounds, annotated on a continuous [0,5] scale. The “GNC” dataset consists of 244 German noun compounds, annotated on a continuous [1,7] scale. We cannot infer its meaning from its components Couch Potato Non- Compositional Character Level Language Models Results The probability of the next character is determined based on previous characters. Pearson’s r Model Comp1 Comp2 ENC RNN(1%)  -0.01 p=0.92 -0.11 p=0.28  GRU(1%)   0.25 p=0.016 0.20 p=0.04    0.23 p=0.02 0.18 p=0.08   0.27 p=0.011 0.27 p=0.009  Word-Embedding English 0.717  0.736  GNC German LSTM (5%)   0.03 p=0.618   0.03 p=0.590 Word Embedding German 0.371 0.349 Can estimate the probability of out-of-vocabulary words, unlike word-level language models. Target chars “e” “l” “l” “o” 1 2.2 -3.0 4.1 0.5 0.3 -1 1.2 0.1 0.5 1.9 -1.1 0.2 -1.5 -0.1 2.2 Output layer W-hy W-hh Hidden layer 0.3 -0.1 9 1 0.3 0.1 1 -0.5 -0.3 -0.3 0.7 1 W-hx 1 1 1 1 Input layer Input chars “h” “e” “l” “l” Conclusions and Future Work Character-level language models capture (some) knowledge about MWE compositionality. Compositionality Equations Word embedding models achieve higher correlation, but can't represent out-of-vocabulary or low-frequency words. To calculate the compositionality, we use two different methods: Future work: Explore parameter settings of character-level language models: e.g., embedding size, batch size, learning rate, dropout. MWE, C1 and C2 are vector representations for a multiword expression, its first component and its second componen,t respectively. Each vector is the hidden state of the network after reading the whole word. Consider other neural network architectures such as a bidirectional LSTM and sub-word level language models. α is set to 0.7, and we use cosine similarity as our similarity measure.


Download ppt "Ali Hakimi Parizi, Paul Cook"

Similar presentations


Ads by Google