Item type |
Trans(1) |
公開日 |
2015-07-27 |
タイトル |
|
|
タイトル |
Audio-Visual Speech Recognition Using Convolutive Bottleneck Networks for a Person with Severe Hearing Loss |
タイトル |
|
|
言語 |
en |
|
タイトル |
Audio-Visual Speech Recognition Using Convolutive Bottleneck Networks for a Person with Severe Hearing Loss |
言語 |
|
|
言語 |
eng |
キーワード |
|
|
主題Scheme |
Other |
|
主題 |
[Research Paper - Express Paper] multimodal, lip reading, deep-learning, assistive technology |
資源タイプ |
|
|
資源タイプ識別子 |
http://purl.org/coar/resource_type/c_6501 |
|
資源タイプ |
journal article |
著者所属 |
|
|
|
Graduate School of System Informatics, Kobe University |
著者所属 |
|
|
|
Graduate School of System Informatics, Kobe University |
著者所属 |
|
|
|
Graduate School of System Informatics, Kobe University |
著者所属 |
|
|
|
Graduate School of System Informatics, Kobe University |
著者所属 |
|
|
|
Graduate School of System Informatics, Kobe University |
著者所属 |
|
|
|
Hyogo Institute of Assistive Technology |
著者所属 |
|
|
|
Hyogo Institute of Assistive Technology |
著者所属 |
|
|
|
Hyogo Institute of Assistive Technology |
著者所属(英) |
|
|
|
en |
|
|
Graduate School of System Informatics, Kobe University |
著者所属(英) |
|
|
|
en |
|
|
Graduate School of System Informatics, Kobe University |
著者所属(英) |
|
|
|
en |
|
|
Graduate School of System Informatics, Kobe University |
著者所属(英) |
|
|
|
en |
|
|
Graduate School of System Informatics, Kobe University |
著者所属(英) |
|
|
|
en |
|
|
Graduate School of System Informatics, Kobe University |
著者所属(英) |
|
|
|
en |
|
|
Hyogo Institute of Assistive Technology |
著者所属(英) |
|
|
|
en |
|
|
Hyogo Institute of Assistive Technology |
著者所属(英) |
|
|
|
en |
|
|
Hyogo Institute of Assistive Technology |
著者名 |
Yuki, Takashima
Yasuhiro, Kakihara
Ryo, Aihara
Tetsuya, Takiguchi
Yasuo, Ariki
Nobuyuki, Mitani
Kiyohiro, Omori
Kaoru, Nakazono
|
著者名(英) |
Yuki, Takashima
Yasuhiro, Kakihara
Ryo, Aihara
Tetsuya, Takiguchi
Yasuo, Ariki
Nobuyuki, Mitani
Kiyohiro, Omori
Kaoru, Nakazono
|
論文抄録 |
|
|
内容記述タイプ |
Other |
|
内容記述 |
In this paper, we propose an audio-visual speech recognition system for a person with an articulation disorder resulting from severe hearing loss. In the case of a person with this type of articulation disorder, the speech style is quite different from with the result that of people without hearing loss that a speaker-independent model for unimpaired persons is hardly useful for recognizing it. We investigate in this paper an audio-visual speech recognition system for a person with severe hearing loss in noisy environments, where a robust feature extraction method using a convolutive bottleneck network (CBN) is applied to audio-visual data. We confirmed the effectiveness of this approach through word-recognition experiments in noisy environments, where the CBN-based feature extraction method outperformed the conventional methods. |
論文抄録(英) |
|
|
内容記述タイプ |
Other |
|
内容記述 |
In this paper, we propose an audio-visual speech recognition system for a person with an articulation disorder resulting from severe hearing loss. In the case of a person with this type of articulation disorder, the speech style is quite different from with the result that of people without hearing loss that a speaker-independent model for unimpaired persons is hardly useful for recognizing it. We investigate in this paper an audio-visual speech recognition system for a person with severe hearing loss in noisy environments, where a robust feature extraction method using a convolutive bottleneck network (CBN) is applied to audio-visual data. We confirmed the effectiveness of this approach through word-recognition experiments in noisy environments, where the CBN-based feature extraction method outperformed the conventional methods. |
書誌レコードID |
|
|
収録物識別子タイプ |
NCID |
|
収録物識別子 |
AA12628065 |
書誌情報 |
IPSJ Transactions on Computer Vision and Applications (CVA)
巻 7,
p. 64-68,
発行日 2015-07-27
|
ISSN |
|
|
収録物識別子タイプ |
ISSN |
|
収録物識別子 |
1882-6695 |
出版者 |
|
|
言語 |
ja |
|
出版者 |
情報処理学会 |