Reconstructing Implicit Knowledge with Language Models

13Citations
Citations of this article
62Readers
Mendeley users who have this article in their library.

Abstract

In this work we propose an approach for generating statements that explicate implicit knowledge connecting sentences in text. We make use of pre-trained language models which we refine by fine-tuning them on specifically prepared corpora that we enriched with implicit information, and by constraining them with relevant concepts and connecting commonsense knowledge paths. Manual and automatic evaluation of the generations shows that by refining language models as proposed, we can generate coherent and grammatically sound sentences that explicate implicit knowledge which connects sentence pairs in texts – on both in-domain and out-of-domain test data.

References Powered by Scopus

A large annotated corpus for learning natural language inference

2531Citations
N/AReaders
Get full text

Coherence and coreference

519Citations
N/AReaders
Get full text

Commonsense knowledge aware conversation generation with graph attention

454Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Scientia Potentia Est—On the Role of Knowledge in Computational Argumentation

21Citations
N/AReaders
Get full text

Similarity-weighted Construction of Contextualized Commonsense Knowledge Graphs for Knowledge-intense Argumentation Tasks

7Citations
N/AReaders
Get full text

Open Relation Modeling: Learning to Define Relations between Entities

7Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Becker, M., Liang, S., & Frank, A. (2021). Reconstructing Implicit Knowledge with Language Models. In Deep Learning Inside Out: 2nd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, DeeLIO 2021 - Proceedings, co-located with the Annual Conference of the North American Chapter of the Association for Computational Linguistics, NAACL-HLT 2021 (pp. 11–24). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.deelio-1.2

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 13

59%

Researcher 4

18%

Professor / Associate Prof. 3

14%

Lecturer / Post doc 2

9%

Readers' Discipline

Tooltip

Computer Science 22

79%

Linguistics 4

14%

Neuroscience 1

4%

Social Sciences 1

4%

Save time finding and organizing research with Mendeley

Sign up for free