Can We Survive without Labelled Data in NLP? Transfer Learning for Open Information Extraction

I. Sarhan, M. Spruit, M. Esposito, G. Massala (Editor), A. Minutolo (Editor), M. Pota (Editor)

    Research output: Contribution to journalArticleAcademicpeer-review

    Abstract

    Various tasks in natural language processing (NLP) suffer from lack of labelled training data, which deep neural networks are hungry for. In this paper, we relied upon features learned to generate relation triples from the open information extraction (OIE) task. First, we studied how transferable these features are from one OIE domain to another, such as from a news domain to a bio-medical domain. Second, we analyzed their transferability to a semantically related NLP task, namely, relation extraction (RE). We thereby contribute to answering the question: can OIE help us achieve adequate NLP performance without labelled data? Our results showed comparable performance when using inductive transfer learning in both experiments by relying on a very small amount of the target data, wherein promising results were achieved. When transferring to the OIE bio-medical domain, we achieved an F-measure of 78.0%, only 1% lower when compared to traditional learning. Additionally, transferring to RE using an inductive approach scored an F-measure of 67.2%, which was 3.8% lower than training and testing on the same task. Hereby, our analysis shows that OIE can act as a reliable source task.
    Original languageEnglish
    Article number5758
    Number of pages16
    JournalApplied Sciences
    Volume10
    Issue number17
    DOIs
    Publication statusPublished - 1 Sept 2020

    Keywords

    • Transfer learning
    • Open information extraction
    • Relation extraction
    • Recurrent neutral networks
    • Word embeddings

    Fingerprint

    Dive into the research topics of 'Can We Survive without Labelled Data in NLP? Transfer Learning for Open Information Extraction'. Together they form a unique fingerprint.

    Cite this