Abstract
Given its demonstrated ability in analyzing and revealing patterns underlying data, Deep Learning (DL) has been increasingly investigated to complement physics-based models in various aspects of smart manufacturing, such as machine condition monitoring and fault diagnosis, complex manufacturing process modeling, and quality inspection. However, successful implementation of DL techniques relies significantly on the amount, variety, and veracity of data for robust network training. Also, the distributions of data used for network training and application should be identical to avoid the internal covariance shift problem that reduces the network performance applicability. As a promising solution to address these challenges, Transfer Learning (TL) enables DL networks trained on a source domain and task to be applied to a separate target domain and task. This paper presents a domain adversarial TL approach, based upon the concepts of generative adversarial networks. In this method, the optimizer seeks to minimize the loss (i.e., regression or classification accuracy) across the labeled training examples from the source domain while maximizing the loss of the domain classifier across the source and target data sets (i.e., maximizing the similarity of source and target features). The developed domain adversarial TL method has been implemented on a 1D CNN backbone network and evaluated for prediction of tool wear propagation, using NASA's milling dataset. The experimental results indicate that domain adversarial TL can successfully allow DL models trained on certain scenarios to be applied to other scenarios.
Cite
CITATION STYLE
Russell, M., & Wang, P. (2020). Domain adversarial transfer learning for generalized tool wear prediction. In Proceedings of the Annual Conference of the Prognostics and Health Management Society, PHM (Vol. 12). Prognostics and Health Management Society. https://doi.org/10.36001/phmconf.2020.v12i1.1137
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.