<?xml version='1.0' encoding='UTF-8'?>
<OAI-PMH xmlns="http://www.openarchives.org/OAI/2.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.openarchives.org/OAI/2.0/ http://www.openarchives.org/OAI/2.0/OAI-PMH.xsd">
  <responseDate>2026-03-16T11:34:25Z</responseDate>
  <request metadataPrefix="jpcoar_1.0" verb="GetRecord" identifier="oai:ipsj.ixsq.nii.ac.jp:00231272">https://ipsj.ixsq.nii.ac.jp/oai</request>
  <GetRecord>
    <record>
      <header>
        <identifier>oai:ipsj.ixsq.nii.ac.jp:00231272</identifier>
        <datestamp>2025-01-19T10:49:43Z</datestamp>
        <setSpec>1164:4179:11237:11430</setSpec>
      </header>
      <metadata>
        <jpcoar:jpcoar xmlns:datacite="https://schema.datacite.org/meta/kernel-4/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:dcndl="http://ndl.go.jp/dcndl/terms/" xmlns:dcterms="http://purl.org/dc/terms/" xmlns:jpcoar="https://github.com/JPCOAR/schema/blob/master/1.0/" xmlns:oaire="http://namespace.openaire.eu/schema/oaire/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:rioxxterms="http://www.rioxx.net/schema/v2.0/rioxxterms/" xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns="https://github.com/JPCOAR/schema/blob/master/1.0/" xsi:schemaLocation="https://github.com/JPCOAR/schema/blob/master/1.0/jpcoar_scm.xsd">
          <dc:title>Enhancing Dysarthric Speech Recognition with Auxiliary Feature Fusion Module: Exploring Articulatory-related Features from Foundation Models</dc:title>
          <dc:title xml:lang="en">Enhancing Dysarthric Speech Recognition with Auxiliary Feature Fusion Module: Exploring Articulatory-related Features from Foundation Models</dc:title>
          <jpcoar:creator>
            <jpcoar:creatorName>Yuqin, Lin</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>Longbiao, Wang</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>Jianwu, Dang</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName>Nobuaki, Minematsu</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Yuqin, Lin</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Longbiao, Wang</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Jianwu, Dang</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:creator>
            <jpcoar:creatorName xml:lang="en">Nobuaki, Minematsu</jpcoar:creatorName>
          </jpcoar:creator>
          <jpcoar:subject subjectScheme="Other">ポスター</jpcoar:subject>
          <datacite:description descriptionType="Other">Addressing dysarthric speech variability in Automatic Speech Recognition (ASR) is crucial for improving human-computer interactions for everyone. This paper proposes the Auxiliary Features Fusion (AFFusion) module, which leverages phonetic and articulatory-related features from models like wav2vec to compensate for distorted acoustics in dysarthric ASR. Experimental results using AFFusion with various feature models demonstrate its effectiveness on dysarthric databases. Interestingly, the analysis suggests that AFFusion shares similarities with human speech perception processes, offering potential insights into addressing fuzzy recognition in dysarthric ASR based on the motor theory of speech perception.</datacite:description>
          <datacite:description descriptionType="Other">Addressing dysarthric speech variability in Automatic Speech Recognition (ASR) is crucial for improving human-computer interactions for everyone. This paper proposes the Auxiliary Features Fusion (AFFusion) module, which leverages phonetic and articulatory-related features from models like wav2vec to compensate for distorted acoustics in dysarthric ASR. Experimental results using AFFusion with various feature models demonstrate its effectiveness on dysarthric databases. Interestingly, the analysis suggests that AFFusion shares similarities with human speech perception processes, offering potential insights into addressing fuzzy recognition in dysarthric ASR based on the motor theory of speech perception.</datacite:description>
          <dc:publisher xml:lang="ja">情報処理学会</dc:publisher>
          <datacite:date dateType="Issued">2023-11-25</datacite:date>
          <dc:language>eng</dc:language>
          <dc:type rdf:resource="http://purl.org/coar/resource_type/c_18gh">technical report</dc:type>
          <jpcoar:identifier identifierType="URI">https://ipsj.ixsq.nii.ac.jp/records/231272</jpcoar:identifier>
          <jpcoar:sourceIdentifier identifierType="ISSN">2188-8779</jpcoar:sourceIdentifier>
          <jpcoar:sourceIdentifier identifierType="NCID">AN10115061</jpcoar:sourceIdentifier>
          <jpcoar:sourceTitle>研究報告自然言語処理（NL）</jpcoar:sourceTitle>
          <jpcoar:volume>2023-NL-258</jpcoar:volume>
          <jpcoar:issue>14</jpcoar:issue>
          <jpcoar:pageStart>1</jpcoar:pageStart>
          <jpcoar:pageEnd>6</jpcoar:pageEnd>
          <jpcoar:file>
            <jpcoar:URI label="IPSJ-NL23258014.pdf">https://ipsj.ixsq.nii.ac.jp/record/231272/files/IPSJ-NL23258014.pdf</jpcoar:URI>
            <jpcoar:mimeType>application/pdf</jpcoar:mimeType>
            <jpcoar:extent>1.7 MB</jpcoar:extent>
          </jpcoar:file>
        </jpcoar:jpcoar>
      </metadata>
    </record>
  </GetRecord>
</OAI-PMH>
