2024-03-29T07:19:19Zhttps://ipsj.ixsq.nii.ac.jp/ej/?action=repository_oaipmhoai:ipsj.ixsq.nii.ac.jp:000595822017-03-31T05:36:57Z05471:05473
Extraction of Bilingual Terminology from a Multilingual Web-based EncyclopediaExtraction of Bilingual Terminology from a Multilingual Web-based EncyclopediaengRegular Paperhttp://id.nii.ac.jp/1001/00059582/Articlehttps://ipsj.ixsq.nii.ac.jp/ej/?action=repository_action_common_download&item_id=59582&item_no=1&attribute_id=1&file_no=1Copyright (c) 2008 by the Information Processing Society of JapanData MiningDepartment of Multimedia Engineering Graduate School of Information Science and Technology Osaka UniversityDepartment of Multimedia Engineering Graduate School of Information Science and Technology Osaka UniversityDepartment of Multimedia Engineering Graduate School of Information Science and Technology Osaka UniversityDepartment of Multimedia Engineering Graduate School of Information Science and Technology Osaka UniversityMaike, ErdmannKotaro, NakayamaTakahiro, HaraShojiro, NishioWith the demand for bilingual dictionaries covering domain-specific terminology research in the field of automatic dictionary extraction has become popular. However the accuracy and coverage of dictionaries created based on bilingual text corpora are often not sufficient for domain-specific terms. Therefore we present an approach for extracting bilingual dictionaries from the link structure of Wikipedia a huge scale encyclopedia that contains a vast number of links between articles in different languages. Our methods analyze not only these interlanguage links but extract even more translations from redirect page and link text information. In an experiment which we have interpreted in detail we proved that the combination of redirect page and link text information achieves much better results than the traditional approach of extracting bilingual terminology from parallel corpora.With the demand for bilingual dictionaries covering domain-specific terminology, research in the field of automatic dictionary extraction has become popular. However, the accuracy and coverage of dictionaries created based on bilingual text corpora are often not sufficient for domain-specific terms. Therefore, we present an approach for extracting bilingual dictionaries from the link structure of Wikipedia, a huge scale encyclopedia that contains a vast number of links between articles in different languages. Our methods analyze not only these interlanguage links but extract even more translations from redirect page and link text information. In an experiment which we have interpreted in detail, we proved that the combination of redirect page and link text information achieves much better results than the traditional approach of extracting bilingual terminology from parallel corpora.AA00700121Journal of Information Processing 1668792008-07-091882-66522009-06-30