@techreport{oai:ipsj.ixsq.nii.ac.jp:00208954, author = {Yiqi, Shao and Mizuho, Iwaihara and Yiqi, Shao and Mizuho, Iwaihara}, issue = {7}, month = {Dec}, note = {Wikipedia has been one of the most famous online encyclopedias. Articles' contents in Wikipedia are constantly changing, where new content should be added into an article when a related new event happens. However, news articles about the event are usually too long to be inserted into the article directly, so summarization is necessary. When generating a summary, the title and subtitles that represent the topic of the article, and the context up to the inserting location should be considered. In this paper, we focus on topic- and context-focused extractive summarization which extracts valuable sentences from a news article according to a given topic and context, to form a summary to be inserted. As one of the most popular pretrained language models, BERT has been widely used in various natural language processing tasks, and BERT has been proved to greatly improve the performance of single-document extractive summarization. In this paper, we propose a two-step BERT-based model that can encode the topic and context into their representations for guiding generation of a summary. We conduct evaluations of our model on the benchmark dataset WikiCite and achieved the current state-of-the-art performance., Wikipedia has been one of the most famous online encyclopedias. Articles' contents in Wikipedia are constantly changing, where new content should be added into an article when a related new event happens. However, news articles about the event are usually too long to be inserted into the article directly, so summarization is necessary. When generating a summary, the title and subtitles that represent the topic of the article, and the context up to the inserting location should be considered. In this paper, we focus on topic- and context-focused extractive summarization which extracts valuable sentences from a news article according to a given topic and context, to form a summary to be inserted. As one of the most popular pretrained language models, BERT has been widely used in various natural language processing tasks, and BERT has been proved to greatly improve the performance of single-document extractive summarization. In this paper, we propose a two-step BERT-based model that can encode the topic and context into their representations for guiding generation of a summary. We conduct evaluations of our model on the benchmark dataset WikiCite and achieved the current state-of-the-art performance.}, title = {Topic- and Context-Focused Extractive Summarization of Wikipedia sentences by Fine-tuning BERT and Similarity Calculation}, year = {2020} }