LEARNING ALGORITHMS IN QUATERNION NEURAL NETWORKS USING GHR CALCULUS

  • Xu D
  • Zhang L
  • Zhang H
N/ACitations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

One difficulty for quaternion neural networks (QNNs) is that quater-nion nonlinear activation functions are usually non-analytic and thus quaternion derivatives cannot be used. In this paper, we derive the quaternion gradient de-scent, approximated quaternion Gauss-Newton and quaternion Levenberg-Mar-quardt algorithms for feedforward QNNs based on the GHR calculus, which is suitable for analytic and non-analytic quaternion functions. Meanwhile, we solve a widely linear quaternion least squares problem in the derivation of quaternion Gauss-Newton algorithm, which is more general than the usual least squares prob-lem. A rigorous analysis of the convergence of the proposed algorithms is provided. Simulations on the prediction of benchmark signals support the approach.

Cite

CITATION STYLE

APA

Xu, D., Zhang, L., & Zhang, H. (2017). LEARNING ALGORITHMS IN QUATERNION NEURAL NETWORKS USING GHR CALCULUS. Neural Network World, 27(3), 271–282. https://doi.org/10.14311/nnw.2017.27.014

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free