• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. DIBERT: Dependency Injected Bidirectional Encoder Representations from Transformers
 
  • Details
  • Full
Options
2021
Conference Paper
Title

DIBERT: Dependency Injected Bidirectional Encoder Representations from Transformers

Abstract
Prior research in the area of Natural Language Processing (NLP) has shown that including the syntactic structure of a sentence using a dependency parse tree while training a representation learning model improves the performance on downstream tasks. However, most of these modeling approaches make use of the dependency parse tree of sentences for learning task-specific word representations rather than considering that for learning generic representations. In this paper, we propose a new model named DIBERT which stands for Dependency Injected Bidirectional Encoder Representations from Transformers. DIBERT is a variation of the BERT, that apart from Masked Language Modeling (MLM) and Next Sentence Prediction (NSP) also incorporates an additional third objective called Parent Prediction (PP). PP injects the syntactic structure of a dependency tree while pre-training the DIBERT, which generates syntax-aware generic representations. We use the WikiText-103 benchmark dataset to pre-train both the original BERT (BERT-Base) and the proposed DIBERT models. After fine-tuning, we observe that DIBERT performs better than BERT-Base on various NLP downstream tasks including Semantic Similarity, Natural Language Inference and Sentiment Analysis hinting at the fact that incorporating dependency information when learning textual representations can improve the quality of the learned representations.
Author(s)
Wahab, Abdul
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Sifa, Rafet  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Mainwork
IEEE Symposium Series on Computational Intelligence, SSCI 2021. Proceedings  
Project(s)
ML2R  
Funder
Bundesministerium für Bildung und Forschung -BMBF-
Conference
Symposium Series on Computational Intelligence 2021  
Open Access
DOI
10.1109/SSCI50451.2021.9659898
Language
English
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Keyword(s)
  • Natural Language Processing

  • NLP

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024