4.6 Article

Metaformer: A Transformer That Tends to Mine Metaphorical-Level Information

Related references

Note: Only part of the references are listed.
Article Computer Science, Artificial Intelligence

Dialogue Relation Extraction with Document-Level Heterogeneous Graph Attention Networks

Hui Chen et al.

Summary: We propose a heterogeneous graph attention network to address the problem of dialogue relation extraction. Compared with several popular sequence-based and graph-based models, our method shows superior performance on the benchmark dataset DialogRE.

COGNITIVE COMPUTATION (2023)

Article Multidisciplinary Sciences

Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences

Alexander Rives et al.

Summary: The deep contextual language model trained through unsupervised learning on protein sequences contains information about biological properties, has a multiscale structural organization, and can be used to improve predictions for protein mutational effects, secondary structure, and long-range contacts.

PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA (2021)

Proceedings Paper Acoustics

DEVELOPING REAL-TIME STREAMING TRANSFORMER TRANSDUCER FOR SPEECH RECOGNITION ON LARGE-SCALE DATASET

Xie Chen et al.

Summary: This research investigates the potential of Transformer Transducer (T-T) models for low latency and fast decoding on large-scale datasets by combining Transformer-XL and chunk-wise streaming processes, showing that it outperforms other models in streaming scenarios. Runtime cost and latency can be optimized with a relatively small look-ahead.

2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021) (2021)

Review Computer Science, Artificial Intelligence

A review on the long short-term memory model

Greg Van Houdt et al.

ARTIFICIAL INTELLIGENCE REVIEW (2020)

Article Chemistry, Multidisciplinary

Molecular Transformer: A Model for Uncertainty-Calibrated Chemical Reaction Prediction

Philippe Schwaller et al.

ACS CENTRAL SCIENCE (2019)

Article Computer Science, Information Systems

T-LSTM: A Long Short-Term Memory Neural Network Enhanced by Temporal Information for Traffic Flow Prediction

Luntian Mou et al.

IEEE ACCESS (2019)

Proceedings Paper Computer Science, Information Systems

Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks

Guokun Lai et al.

ACM/SIGIR PROCEEDINGS 2018 (2018)

Review Computer Science, Hardware & Architecture

A Few Useful Things to Know About Machine Learning

Pedro Domingos

COMMUNICATIONS OF THE ACM (2012)