Nothing Special   »   [go: up one dir, main page]

loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Paper Unlock

Authors: Cristian David Estupiñán Ojeda ; Cayetano Nicolás Guerra Artal and Francisco Mario Hernández Tejera

Affiliation: University Institute SIANI, University of Las Palmas de Gran Canaria, 35017, Las Palmas de Gran Canaria, Spain

Keyword(s): Deep Learning, Linear Transformer, Informer, Convolution, Self Attention, Organization, Neural Machine Translation.

Abstract: The use of architectures based on transformers presents a state of the art revolution in natural language processing (NLP). The employment of these architectures with high computational costs has increased in the last few months, despite the existing use of parallelization techniques. This is due to the high performance that is obtained by increasing the size of the learnable parameters for these kinds of architectures, while maintaining the models’ predictability. This relates to the fact that it is difficult to do research with limited computational resources. A restrictive element is the memory usage, which seriously affects the replication of experiments. We are presenting a new architecture called Informer, which seeks to exploit the concept of information organization. For the sake of evaluation, we use a neural machine translation (NMT) dataset, the English-Vietnamese IWSLT15 dataset (Luong and Manning, 2015). In this paper, we also compare this proposal with architectures tha t reduce the computational cost to O(n · r), such as Linformer (Wang et al., 2020). In addition, we have managed to improve the SOTA of the BLEU score from 33.27 to 35.11. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 65.254.225.175

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Ojeda, C.; Artal, C. and Tejera, F. (2021). Informer, an Information Organization Transformer Architecture. In Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART; ISBN 978-989-758-484-8; ISSN 2184-433X, SciTePress, pages 381-389. DOI: 10.5220/0010372703810389

@conference{icaart21,
author={Cristian David Estupiñán Ojeda. and Cayetano Nicolás Guerra Artal. and Francisco Mario Hernández Tejera.},
title={Informer, an Information Organization Transformer Architecture},
booktitle={Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART},
year={2021},
pages={381-389},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0010372703810389},
isbn={978-989-758-484-8},
issn={2184-433X},
}

TY - CONF

JO - Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART
TI - Informer, an Information Organization Transformer Architecture
SN - 978-989-758-484-8
IS - 2184-433X
AU - Ojeda, C.
AU - Artal, C.
AU - Tejera, F.
PY - 2021
SP - 381
EP - 389
DO - 10.5220/0010372703810389
PB - SciTePress

<style> #socialicons>a span { top: 0px; left: -100%; -webkit-transition: all 0.3s ease; -moz-transition: all 0.3s ease-in-out; -o-transition: all 0.3s ease-in-out; -ms-transition: all 0.3s ease-in-out; transition: all 0.3s ease-in-out;} #socialicons>ahover div{left: 0px;} </style>