4.7 Review

A survey on modern trainable activation functions

相关参考文献

注意:仅列出部分参考文献,下载原文获取全部文献信息。
Article Computer Science, Artificial Intelligence

A comparison of deep networks with ReLU activation function and linear spline-type methods

Konstantin Eckle et al.

NEURAL NETWORKS (2019)

Article Computer Science, Artificial Intelligence

Kafnets: Kernel-based non-parametric activation functions for neural networks

Simone Scardapane et al.

NEURAL NETWORKS (2019)

Article Computer Science, Artificial Intelligence

A simple and efficient architecture for trainable activation functions

Andrea Apicella et al.

NEUROCOMPUTING (2019)

Article Automation & Control Systems

Learning with Type-2 Fuzzy activation functions to improve the performance of Deep Neural Networks

Aykut Beke et al.

ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE (2019)

Article Computer Science, Artificial Intelligence

Sigmoid-weighted linear units for neural network function approximation in reinforcement learning

Stefan Elfwing et al.

NEURAL NETWORKS (2018)

Article Computer Science, Artificial Intelligence

A novel type of activation function in artificial neural networks: Trained activation function

Omer Faruk Ertugrul

NEURAL NETWORKS (2018)

Article Computer Science, Artificial Intelligence

Adaptive activation functions in convolutional neural networks

Sheng Qian et al.

NEUROCOMPUTING (2018)

Article Computer Science, Artificial Intelligence

Improving deep neural networks with multi-layer maxout networks and a novel initialization method

Weichen Sun et al.

NEUROCOMPUTING (2018)

Article Computer Science, Artificial Intelligence

Convolution in Convolution for Network in Network

Yanwei Pang et al.

IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS (2018)

Article Mathematics, Applied

Neural network with unbounded activation functions is universal approximator

Sho Sonoda et al.

APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS (2017)

Proceedings Paper Computer Science, Artificial Intelligence

Densely Connected Convolutional Networks

Gao Huang et al.

30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017) (2017)

Article Computer Science, Artificial Intelligence

A novel activation function for multilayer feed-forward neural networks

Aboubakar Nasser Samatin Njikam et al.

APPLIED INTELLIGENCE (2016)

Article Computer Science, Artificial Intelligence

ImageNet Large Scale Visual Recognition Challenge

Olga Russakovsky et al.

INTERNATIONAL JOURNAL OF COMPUTER VISION (2015)

Article Computer Science, Artificial Intelligence

A linear approach for sparse coding by a two-layer neural network

Alessandro Montalto et al.

NEUROCOMPUTING (2015)

Article Physics, Multidisciplinary

Enhanced Higgs Boson to τ+τ- Search with Deep Learning

P. Baldi et al.

PHYSICAL REVIEW LETTERS (2015)

Article Computer Science, Artificial Intelligence

What are Extreme Learning Machines? Filling the Gap Between Frank Rosenblatt's Dream and John von Neumann's Puzzle

Guang-Bin Huang

COGNITIVE COMPUTATION (2015)

Article Multidisciplinary Sciences

Searching for exotic particles in high-energy physics with deep learning

P. Baldi et al.

NATURE COMMUNICATIONS (2014)

Article Computer Science, Artificial Intelligence

Local minima in hierarchical structures of complex-valued neural networks

Tohru Nitta

NEURAL NETWORKS (2013)

Article Computer Science, Artificial Intelligence

Am activation function adapting training algorithm for sigmoidal feedforward networks

P Chandra et al.

NEUROCOMPUTING (2004)

Article Economics

A class+1 sigmoidal activation functions for FFANNs

Y Singh et al.

JOURNAL OF ECONOMIC DYNAMICS & CONTROL (2003)

Article Computer Science, Artificial Intelligence

Recurrent neural networks with trainable amplitude of activation functions

SL Goh et al.

NEURAL NETWORKS (2003)