<?xml version='1.0' encoding='UTF-8'?>
<OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd">
  <responseDate>2026-03-16T15:52:11Z</responseDate>
  <request metadataPrefix="oai_dc" verb="GetRecord" identifier="oai:ipsj.ixsq.nii.ac.jp:00231272">https://ipsj.ixsq.nii.ac.jp/oai</request>
  <GetRecord>
    <record>
      <header>
        <identifier>oai:ipsj.ixsq.nii.ac.jp:00231272</identifier>
        <datestamp>2025-01-19T10:49:43Z</datestamp>
        <setSpec>1164:4179:11237:11430</setSpec>
      </header>
      <metadata>
        <oai_dc:dc xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:oai_dc="http://www.openarchives.org/OAI/2.0/oai_dc/" xmlns="http://www.w3.org/2001/XMLSchema" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd">
          <dc:title>Enhancing Dysarthric Speech Recognition with Auxiliary Feature Fusion Module: Exploring Articulatory-related Features from Foundation Models</dc:title>
          <dc:title>Enhancing Dysarthric Speech Recognition with Auxiliary Feature Fusion Module: Exploring Articulatory-related Features from Foundation Models</dc:title>
          <dc:creator>Yuqin, Lin</dc:creator>
          <dc:creator>Longbiao, Wang</dc:creator>
          <dc:creator>Jianwu, Dang</dc:creator>
          <dc:creator>Nobuaki, Minematsu</dc:creator>
          <dc:creator>Yuqin, Lin</dc:creator>
          <dc:creator>Longbiao, Wang</dc:creator>
          <dc:creator>Jianwu, Dang</dc:creator>
          <dc:creator>Nobuaki, Minematsu</dc:creator>
          <dc:subject>ポスター</dc:subject>
          <dc:description>Addressing dysarthric speech variability in Automatic Speech Recognition (ASR) is crucial for improving human-computer interactions for everyone. This paper proposes the Auxiliary Features Fusion (AFFusion) module, which leverages phonetic and articulatory-related features from models like wav2vec to compensate for distorted acoustics in dysarthric ASR. Experimental results using AFFusion with various feature models demonstrate its effectiveness on dysarthric databases. Interestingly, the analysis suggests that AFFusion shares similarities with human speech perception processes, offering potential insights into addressing fuzzy recognition in dysarthric ASR based on the motor theory of speech perception.</dc:description>
          <dc:description>Addressing dysarthric speech variability in Automatic Speech Recognition (ASR) is crucial for improving human-computer interactions for everyone. This paper proposes the Auxiliary Features Fusion (AFFusion) module, which leverages phonetic and articulatory-related features from models like wav2vec to compensate for distorted acoustics in dysarthric ASR. Experimental results using AFFusion with various feature models demonstrate its effectiveness on dysarthric databases. Interestingly, the analysis suggests that AFFusion shares similarities with human speech perception processes, offering potential insights into addressing fuzzy recognition in dysarthric ASR based on the motor theory of speech perception.</dc:description>
          <dc:description>technical report</dc:description>
          <dc:publisher>情報処理学会</dc:publisher>
          <dc:date>2023-11-25</dc:date>
          <dc:format>application/pdf</dc:format>
          <dc:identifier>研究報告自然言語処理（NL）</dc:identifier>
          <dc:identifier>14</dc:identifier>
          <dc:identifier>2023-NL-258</dc:identifier>
          <dc:identifier>1</dc:identifier>
          <dc:identifier>6</dc:identifier>
          <dc:identifier>2188-8779</dc:identifier>
          <dc:identifier>AN10115061</dc:identifier>
          <dc:identifier>https://ipsj.ixsq.nii.ac.jp/record/231272/files/IPSJ-NL23258014.pdf</dc:identifier>
          <dc:language>eng</dc:language>
        </oai_dc:dc>
      </metadata>
    </record>
  </GetRecord>
</OAI-PMH>
