A Fully Attention-Based Information Retriever

A.H. Chaim Correia, Jorge Luiz Moreira Silva, Thiago de Castro Martins, Fabio Gagliardi Cozman

    Research output: Contribution to conferencePaperAcademic

    Abstract

    Recurrent neural networks are now the state-of-the-art in natural language processing because they can build rich contextual representations and process texts of arbitrary length. However, recent developments on attention mechanisms have equipped feedforward networks with similar capabilities, hence enabling faster computations due to the increase in the number of operations that can be parallelized. We explore this new type of architecture in the domain of question-answering and propose a novel approach that we call Fully Attention Based Information Retriever (FABIR). We show that FABIR achieves competitive results in the Stanford Question Answering Dataset (SQuAD) while having fewer parameters and being faster at both learning and inference than rival methods.
    Original languageEnglish
    Pages2799
    Number of pages2806
    Publication statusPublished - 8 Jul 2018
    EventInternational Joint Conference on Neural Networks - Rio de Janeiro, Brazil
    Duration: 8 Jul 201813 Jul 2018
    http://www.ecomp.poli.br/~wcci2018/

    Conference

    ConferenceInternational Joint Conference on Neural Networks
    Abbreviated titleIJCNN
    Country/TerritoryBrazil
    CityRio de Janeiro
    Period8/07/1813/07/18
    Internet address

    Keywords

    • Machine Learning
    • Information Retrieval
    • Question answering
    • Natural Language Processing
    • Deep Learning

    Fingerprint

    Dive into the research topics of 'A Fully Attention-Based Information Retriever'. Together they form a unique fingerprint.

    Cite this