ログイン 新規登録
言語:

WEKO3

  • トップ
  • ランキング
To
lat lon distance
To

Field does not validate



インデックスリンク

インデックスツリー

メールアドレスを入力してください。

WEKO

One fine body…

WEKO

One fine body…

アイテム

  1. 論文誌(トランザクション)
  2. 数理モデル化と応用(TOM)
  3. Vol.15
  4. No.3

Modeling Pooling Layers For CNN Initialization

https://ipsj.ixsq.nii.ac.jp/records/219072
https://ipsj.ixsq.nii.ac.jp/records/219072
afa73583-ac92-4240-a1ae-bc29da0d1d95
名前 / ファイル ライセンス アクション
IPSJ-TOM1503005.pdf IPSJ-TOM1503005.pdf (1.3 MB)
Copyright (c) 2022 by the Information Processing Society of Japan
オープンアクセス
Item type Trans(1)
公開日 2022-07-26
タイトル
タイトル Modeling Pooling Layers For CNN Initialization
タイトル
言語 en
タイトル Modeling Pooling Layers For CNN Initialization
言語
言語 eng
キーワード
主題Scheme Other
主題 [オリジナル論文] deep neural network, convolutional neural network, initialization, signal variance
資源タイプ
資源タイプ識別子 http://purl.org/coar/resource_type/c_6501
資源タイプ journal article
著者所属
Graduate School of Science and Technology, Gunma University
著者所属
Graduate School of Science and Technology, Gunma University
著者所属(英)
en
Graduate School of Science and Technology, Gunma University
著者所属(英)
en
Graduate School of Science and Technology, Gunma University
著者名 Takahiko, Henmi

× Takahiko, Henmi

Takahiko, Henmi

Search repository
Tsuyoshi, Kato

× Tsuyoshi, Kato

Tsuyoshi, Kato

Search repository
著者名(英) Takahiko, Henmi

× Takahiko, Henmi

en Takahiko, Henmi

Search repository
Tsuyoshi, Kato

× Tsuyoshi, Kato

en Tsuyoshi, Kato

Search repository
論文抄録
内容記述タイプ Other
内容記述 Deep convolutional neural networks (CNNs) have achieved consistent excellent performance on image processing tasks. The CNN architecture consists of different types of layers, including the convolution layer and the max pooling layer. It is widely understood among CNN practitioners that the stability of learning depends on the initialization of the model parameters in each layer. Currently, the de facto standard scheme for initialization is the Kaiming initialization, which was developed by He et al. The Kaiming scheme was derived from a much simpler model than the currently used CNN structure that evolved since the emergence of the Kaiming scheme. It consists only of the convolution and fully connected layers, and does not include the max pooling or global average pooling layers. In this study, we derive an initialization scheme, not from the simplified Kaiming model, but from modern CNN architectures consisting not only of the convolution and the fully connected layers but also of the max pooling and the global average pooling. Furthermore, the new model expresses the padding which is not considered in the existing models. We empirically investigate the performance of the new initialization methods compared to the de facto standard methods that are widely used today. Source code: https://github.com/hecwitane/ASV-pub/
論文抄録(英)
内容記述タイプ Other
内容記述 Deep convolutional neural networks (CNNs) have achieved consistent excellent performance on image processing tasks. The CNN architecture consists of different types of layers, including the convolution layer and the max pooling layer. It is widely understood among CNN practitioners that the stability of learning depends on the initialization of the model parameters in each layer. Currently, the de facto standard scheme for initialization is the Kaiming initialization, which was developed by He et al. The Kaiming scheme was derived from a much simpler model than the currently used CNN structure that evolved since the emergence of the Kaiming scheme. It consists only of the convolution and fully connected layers, and does not include the max pooling or global average pooling layers. In this study, we derive an initialization scheme, not from the simplified Kaiming model, but from modern CNN architectures consisting not only of the convolution and the fully connected layers but also of the max pooling and the global average pooling. Furthermore, the new model expresses the padding which is not considered in the existing models. We empirically investigate the performance of the new initialization methods compared to the de facto standard methods that are widely used today. Source code: https://github.com/hecwitane/ASV-pub/
書誌レコードID
収録物識別子タイプ NCID
収録物識別子 AA11464803
書誌情報 情報処理学会論文誌数理モデル化と応用(TOM)

巻 15, 号 3, p. 29-37, 発行日 2022-07-26
ISSN
収録物識別子タイプ ISSN
収録物識別子 1882-7780
出版者
言語 ja
出版者 情報処理学会
戻る
0
views
See details
Views

Versions

Ver.1 2025-01-19 14:56:03.921951
Show All versions

Share

Mendeley Twitter Facebook Print Addthis

Cite as

エクスポート

OAI-PMH
  • OAI-PMH JPCOAR
  • OAI-PMH DublinCore
  • OAI-PMH DDI
Other Formats
  • JSON
  • BIBTEX

Confirm


Powered by WEKO3


Powered by WEKO3