Abstract
Regarding the sequence learning of neural networks, there exists a problem of how to capture long-term dependencies and alleviate the gradient vanishing phenomenon. To manage this problem, we proposed a neural network with random connections via a scheme of a neural architecture search. First, a dense network was designed and trained to construct a search space, and then another network was generated by random sampling in the space, whose skip connections could transmit information directly over multiple periods and capture long-term dependencies more efficiently. Moreover, we devised a novel cell structure that required less memory and computational power than the structures of long short-term memories (LSTMs), and finally, we performed a special initialization scheme on the cell parameters, which could permit unhindered gradient propagation on the time axis at the beginning of training. In the experiments, we evaluated four sequential tasks: adding, copying, frequency discrimination, and image classification; we also adopted several state-of-the-art methods for comparison. The experimental results demonstrated that our proposed model achieved the best performance.
Author supplied keywords
Cite
CITATION STYLE
Shan, D., Zhang, X., Shi, W., & Li, L. (2020). Neural architecture search for a highly efficient network with random skip connections. Applied Sciences (Switzerland), 10(11). https://doi.org/10.3390/app10113712
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.