Abstract
We present a methodology that explores how sentence structure is reflected in neural representations of machine translation systems. We demonstrate our model-agnostic approach with the Transformer English-German translation model. We analyze neuron-level correlation of activations between paraphrases while discussing the methodology challenges and the need for confound analysis to isolate the effects of shallow cues. We find that similarity between activation patterns can be mostly accounted for by similarity in word choice and sentence length. Following that, we manipulate neuron activations to control the syntactic form of the output. We show this intervention to be somewhat successful, indicating that deep models capture sentence-structure distinctions, despite finding no such indication at the neuron level. To conduct our experiments, we develop a semi-automatic method to generate meaning-preserving minimal pair paraphrases (active-passive voice and adverbial clause-noun phrase) and compile a corpus of such pairs.
Original language | English |
---|---|
Title of host publication | CoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 194-212 |
Number of pages | 19 |
ISBN (Electronic) | 9781959429074 |
State | Published - 2022 |
Event | 26th Conference on Computational Natural Language Learning, CoNLL 2022 collocated and co-organized with EMNLP 2022 - Abu Dhabi, United Arab Emirates Duration: 7 Dec 2022 → 8 Dec 2022 |
Publication series
Name | CoNLL 2022 - 26th Conference on Computational Natural Language Learning, Proceedings of the Conference |
---|
Conference
Conference | 26th Conference on Computational Natural Language Learning, CoNLL 2022 collocated and co-organized with EMNLP 2022 |
---|---|
Country/Territory | United Arab Emirates |
City | Abu Dhabi |
Period | 7/12/22 → 8/12/22 |
Bibliographical note
Publisher Copyright:©2022 Association for Computational Linguistics.