Multikernel Activation Functions: Formulation and a Case Study

  • Scardapane S
  • Nieddu E
  • Firmani D
  • et al.
N/ACitations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The design of activation functions is a growing research area in the field of neural networks. In particular, instead of using fixed point-wise functions (e.g., the rectified linear unit), several authors have proposed ways of learning these functions directly from the data in a non-parametric fashion. In this paper we focus on the kernel activation function (KAF), a recently proposed framework wherein each function is modeled as a one-dimensional kernel model, whose weights are adapted through standard backpropagation-based optimization. One drawback of KAFs is the need to select a single kernel function and its eventual hyper-parameters. To partially overcome this problem, we motivate an extension of the KAF model, in which multiple kernels are linearly combined at every neuron, inspired by the literature on multiple kernel learning. We provide an application of the resulting multi-KAF on a realistic use case, specifically handwritten Latin OCR, on a large dataset collected in the context of the `In Codice Ratio' project. Results show that multi-KAFs can improve the accuracy of the convolutional networks previously developed for the task, with faster convergence, even with a smaller number of overall parameters.

Cite

CITATION STYLE

APA

Scardapane, S., Nieddu, E., Firmani, D., & Merialdo, P. (2020). Multikernel Activation Functions: Formulation and a Case Study (pp. 320–329). https://doi.org/10.1007/978-3-030-16841-4_33

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free