Influence of feature scaling on convergence of gradient iterative algorithm

65Citations
Citations of this article
118Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Feature scaling is a method to unify self-variables or feature ranges in data. In data processing, it is usually used in data pre-processing. Because in the original data, the range of variables is very different. Feature scaling is a necessary step in the calculation of stochastic gradient descent. This paper takes the computer hardware data set maintained by UCI as an example, and compares the influence of normalization method and interval scaling method on the convergence of stochastic gradient descent by algorithm simulation. The result of study has a certain value on feature scaling.

Cite

CITATION STYLE

APA

Wan, X. (2019). Influence of feature scaling on convergence of gradient iterative algorithm. In Journal of Physics: Conference Series (Vol. 1213). Institute of Physics Publishing. https://doi.org/10.1088/1742-6596/1213/3/032021

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free