期刊文献+

Hyperparameter on-line learning of stochastic resonance based threshold networks

原文传递
导出
摘要 Aiming at training the feed-forward threshold neural network consisting of nondifferentiable activation functions, the approach of noise injection forms a stochastic resonance based threshold network that can be optimized by various gradientbased optimizers. The introduction of injected noise extends the noise level into the parameter space of the designed threshold network, but leads to a highly non-convex optimization landscape of the loss function. Thus, the hyperparameter on-line learning procedure with respective to network weights and noise levels becomes of challenge. It is shown that the Adam optimizer, as an adaptive variant of stochastic gradient descent, manifests its superior learning ability in training the stochastic resonance based threshold network effectively. Experimental results demonstrate the significant improvement of performance of the designed threshold network trained by the Adam optimizer for function approximation and image classification.
作者 Weijin Li Yuhao Ren Fabing Duan 李伟进;任昱昊;段法兵(College of Automation,Qingdao University,Qingdao 266071,China)
机构地区 College of Automation
出处 《Chinese Physics B》 SCIE EI CAS CSCD 2022年第8期289-295,共7页 中国物理B(英文版)
基金 Project supported by the Natural Science Foundation of Shandong Province,China(Grant No.ZR2021MF051)。
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部