Logo: to the web site of Uppsala University

uu.sePublications from Uppsala University
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
What Should/Do/Can LSTMs Learn When Parsing Auxiliary Verb Constructions?
Uppsala University, Disciplinary Domain of Humanities and Social Sciences, Faculty of Languages, Department of Linguistics and Philology.ORCID iD: 0000-0001-8844-2126
Uppsala University, Disciplinary Domain of Humanities and Social Sciences, Faculty of Languages, Department of Linguistics and Philology.ORCID iD: 0000-0003-3726-9399
Uppsala University, Disciplinary Domain of Humanities and Social Sciences, Faculty of Languages, Department of Linguistics and Philology.
2020 (English)In: Computational linguistics - Association for Computational Linguistics (Print), ISSN 0891-2017, E-ISSN 1530-9312, Vol. 46, no 4, p. 763-784Article in journal (Refereed) Published
Abstract [en]

There is a growing interest in investigating what neural NLP models learn about language. A prominent open question is the question of whether or not it is necessary to model hierarchical structure. We present a linguistic investigation of a neural parser adding insights to this question. We look at transitivity and agreement information of auxiliary verb constructions (AVCs) in comparison to finite main verbs (FMVs). This comparison is motivated by theoretical work in dependency grammar and in particular the work of Tesnière (1959), where AVCs and FMVs are both instances of a nucleus, the basic unit of syntax. An AVC is a dissociated nucleus; it consists of at least two words, and an FMV is its non-dissociated counterpart, consisting of exactly one word. We suggest that the representation of AVCs and FMVs should capture similar information. We use diagnostic classifiers to probe agreement and transitivity information in vectors learned by a transition-based neural parser in four typologically different languages. We find that the parser learns different information about AVCs and FMVs if only sequential models (BiLSTMs) are used in the architecture but similar information when a recursive layer is used. We find explanations for why this is the case by looking closely at how information is learned in the network and looking at what happens with different dependency representations of AVCs. We conclude that there may be benefits to using a recursive layer in dependency parsing and that we have not yet found the best way to integrate it in our parsers.

Place, publisher, year, edition, pages
MIT Press, 2020. Vol. 46, no 4, p. 763-784
National Category
Natural Language Processing
Research subject
Computational Linguistics
Identifiers
URN: urn:nbn:se:uu:diva-462769DOI: 10.1162/coli_a_00392OAI: oai:DiVA.org:uu-462769DiVA, id: diva2:1623977
Available from: 2022-01-02 Created: 2022-01-02 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

fulltext(298 kB)256 downloads
File information
File name FULLTEXT01.pdfFile size 298 kBChecksum SHA-512
473c5a91471099313b9c919bc4c78a726065713d22c48db1095764faae67a96967fe21b53522f306019db104c2bbcce0891c1f13ad412877b9c59f8a0de240ee
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

de Lhoneux, MiryamStymne, SaraNivre, Joakim

Search in DiVA

By author/editor
de Lhoneux, MiryamStymne, SaraNivre, Joakim
By organisation
Department of Linguistics and Philology
In the same journal
Computational linguistics - Association for Computational Linguistics (Print)
Natural Language Processing

Search outside of DiVA

GoogleGoogle Scholar
Total: 266 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 138 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf