跳至主導覽 跳至搜尋 跳過主要內容

SinP[N]: A fast convergence activation function for convolutional neural networks

研究成果: Conference contribution同行評審

6 引文 斯高帕斯(Scopus)

摘要

Convolutional Neural Networks (CNNs) are currently the most advanced machine learning architecture for visual data classification. The choice of activation functions has a significant impact on the performance of a training task. In order to overcome the vanishing gradient problem, we propose a new activation function for the classification system. The activation function makes use of the properties of periodic functions, where the derivative of a periodic function is also periodic. Furthermore, a linear combination is introduced to prevent the derivative from becoming zero. We verify this novel activation function by training an empirical analysis and comparing with the currently discovered activation functions. Experimental results show that our activation function SinP[N](x) = sin(x)+Nx, leads to very fast convergence even without the normalization layer. As a result, this new activation function enhances the training accuracy significantly, and can be easily deployed in the current systems built upon the standard CNN architecture.

原文English
主出版物標題Proceedings - 11th IEEE/ACM International Conference on Utility and Cloud Computing Companion, UCC Companion 2018
編輯Alan Sill, Josef Spillner
發行者Institute of Electrical and Electronics Engineers Inc.
頁面359-364
頁數6
ISBN(電子)9781728103594
DOIs
出版狀態Published - 2 7月 2018
事件11th IEEE/ACM International Conference on Utility and Cloud Computing Companion, UCC Companion 2018 - Zurich, Switzerland
持續時間: 17 12月 201820 12月 2018

出版系列

名字Proceedings - 11th IEEE/ACM International Conference on Utility and Cloud Computing Companion, UCC Companion 2018

Conference

Conference11th IEEE/ACM International Conference on Utility and Cloud Computing Companion, UCC Companion 2018
國家/地區Switzerland
城市Zurich
期間17/12/1820/12/18

指紋

深入研究「SinP[N]: A fast convergence activation function for convolutional neural networks」主題。共同形成了獨特的指紋。

引用此