ParaSCI: A Large Scientific Paraphrase Dataset for Longer Paraphrase Generation
Abstract
We propose ParaSCI, the first large-scale paraphrase dataset in the scientific field, including 33,981 paraphrase pairs from ACL (ParaSCI-ACL) and 316,063 pairs from arXiv (ParaSCI-arXiv). Digging into characteristics and common patterns of scientific papers, we construct this dataset though intra-paper and inter-paper methods, such as collecting citations to the same paper or aggregating definitions by scientific terms. To take advantage of sentences paraphrased partially, we put up PDBERT as a general paraphrase discovering method. The major advantages of paraphrases in ParaSCI lie in the prominent length and textual diversity, which is complementary to existing paraphrase datasets. ParaSCI obtains satisfactory results on human evaluation and downstream tasks, especially long paraphrase generation.
1 Introduction
A paraphrase is a restatement of meaning with different expressions Bhagat and Hovy (2013). Being very common in our daily language expressions, it can also be applied to multiple downstream tasks of natural language processing (NLP), such as generating diverse text or adding richness to a chatbot.
At present, paraphrase recognition or paraphrase generation are largely limited to the deficiency of paraphrase corpus. Especially, due to the permanent vacancy of paraphrase corpus in the scientific field, scientific paraphrase generation advances slowly. Scientific paraphrases can not only be helpful for data augmentation of challenging scientific machine translation, but is also effective for polishing scientific papers. However, existing paraphrase datasets are mainly from news, novels, or social media platforms. Most of them remain short sentences and interrogative or oral style. As a result, none of such training data can train out a scientific paraphrase generator. Taking the sentence “we used pos tags predicted by the stanford pos tagger” as an example, the generated sentences from Transformer Vaswani et al. (2017) models trained on existing paraphrase datasets111Here we use Quora Question Pairs and MSCOCO, they are introduced in Section 2 are “level basic topics : what is the basic purpose of stanford traditional hmo” and “a picture of a street sign with a sign on it”, far from ground-truth paraphrases.
We have noticed that the structure of scientific papers is nearly fixed. Paraphrase sentence pairs appear not only within a paper (intra-paper) but also across different papers (inter-paper), which makes it possible to construct a paraphrase dataset in the scientific field. For example, repetitions of the same crucial contribution in a paper or explanations of the same term in different papers are potential paraphrases. Based on such characteristics, we design different methods to extract paraphrase pairs (shown in Section 4).
In terms of the construction methods, existing methods merely focus on the paraphrase relationship between entire sentences, while hardly handle sentences with partial paraphrase parts, leaving much original corpus idle. We find that if part of a sentence paraphrases another short sentence, such sequences will be filtered out because the overall semantic similarity is not high enough. For paraphrase discovering in this case, we fine-tune BERT to extract semantically equivalent parts of two sentences and name it PDBERT. In order to train PDBERT, we construct pseudo training data by stitching existing paraphrase sentences, and train a paraphrase extraction model using the pseudo training data. In the end, this model performs well on real scientific texts.
After filtering, we obtain 350,044 paraphrase pairs and name this dataset ParaSCI. It consists of two parts: ParaSCI-ACL (33,981 pairs) and ParaSCI-arXiv (316,063 pairs). Compared with other paraphrase datasets, sentences in ParaSCI are longer and more sententially divergent. ParaSCI can be used for training paraphrase generation models. Furthermore, we hope that it can be applied to enlarge training data for other NLP tasks in the scientific domain.
Our main contributions include:
-
1.
We propose the first large-scale paraphrase dataset in the scientific field (ParaSCI), including 33,981 pairs in ParaSCI-ACL and 316,063 pairs in ParaSCI-arXiv. Our dataset has been released to the public222https://github.com/dqxiu/ParaSCI.
-
2.
We propose a general method for paraphrase discovering. By fine-tuning BERT innovatively, our PDBERT can extract paraphrase pairs from partially paraphrased sentences.
-
3.
The model trained on ParaSCI can generate longer paraphrases, and sentences are enriched with scientific knowledge, such as terms and abbreviations.
2 Related Work
Paraphrases capture the essence of language diversity Pavlick et al. (2015) and play significant roles in many challenging NLP tasks, such as question answering Dong et al. (2017), semantic parsing Su and Yan (2017) and machine translation Cho et al. (2014). Development in paraphrases relies heavily on the construction of paraphrase datasets.
Paraphrase Identification Datasets
Dolan and Brockett (2005) proposed MSR Paraphrase Corpus [MSRP], a paraphrase dataset of 5,801 sentence pairs, by clustering news articles with an SVM classifier and human annotations. As is discovered, platforms such as Twitter also contain many paraphrase pairs. Twitter Paraphrase Corpus [PIT-2015] Xu et al. (2015) contains 14,035 paraphrase pairs on more than 400 distinct topics. Two years later, Twitter Url Corpus [TUC] Lan et al. (2017) was proposed as a development of PIT-2015. TUC contains 51,524 sentence pairs, collected from Twitter by linking tweets through shared URLs and do not leverage any classifier or human intervention. Datasets such as MSRP or PIT-2015 encourage a series of work in paraphrase identification Das and Smith (2009); Mallinson et al. (2017) but the size limitation hinders complex generation models.
Paraphrase Generation Datasets
MSCOCO Lin et al. (2014) was originally described as a large-scale object detection dataset. It contains human-annotated captions of over 120K images, and each image is associated with five captions from five different annotators. In most cases, annotators describe the most prominent object/action in an image, which makes this dataset suitable for paraphrase-related tasks. Consequently, MSCOCO makes great contribution to paraphrase generation. Quora released a new dataset333website:https://data.quora.com/First-Quora-Dataset-Release-Question-Pairs in January 2017, which consists of over 400K lines of potential question duplicate pairs. Wieting and Gimpel (2018) constructed ParaNMT-50M, a dataset of more than 50 million paraphrase pairs. The pairs were generated automatically by translating the non-English side of a large parallel corpus. Nowadays, MSCOCO and Quora are mainly used for paraphrase generation Fu et al. (2019); Gupta et al. (2018). Nevertheless, their sentence lengths or related domains are restricted.
3 Dataset
3.1 Source Materials
Our ParaSCI dataset is constructed based on the following source materials:
ACL Anthology Sentence Corpus (AASC)
AASC Aizawa et al. (2018) is a corpus of natural language text extracted from scientific papers. It contains 2,339,195 sentences from 44,481 PDF-format papers from the ACL Anthology, a comprehensive scientific paper repository on computational linguistics and NLP.
ArXiv Bulk Data
ArXiv444website:https://arxiv.org/help/bulk_data is an open-access repository of electronic preprints. It consists of scientific papers in the fields of mathematics, physics, astronomy, etc.. As the complete set is too large to process, we randomly select 202,125 PDF files as our original data and convert them to TXT files, arranged by sentence.
Semantic Scholar Open Research Corpus (S2ORC)
S2ORC Lo et al. (2020) is a large contextual citation graph of scientific papers from multiple scientific domains, consisting of 81.1M papers, 380.5M citation edges. We select all the citation edges of ACL and arXiv from S2ORC for subsequent processing.
3.2 Basic Information
According to the source materials, ParaSCI includes two subsets, ParaSCI-ACL and ParaSCI-arXiv. Paraphrase pairs in ParaSCI-ACL focus on the NLP field, while paraphrase pairs in ParaSCI-arXiv are more general. Some cases are shown in Table 1. ParaSCI show three main highlights: 1) Sentences included are long, nearly 19 words per sentence; 2) Sentences are more sententially divergent; 3) It provides rich scientific knowledge.
Name | Sentence A | Sentence B |
---|---|---|
ParaSCI-ACL | Word sense disambiguation (wsd) is the task of identifying the correct meaning of a word in context. | The process of identifying the correct meaning, or sense of a word in context, is known as word sense disambiguation (wsd). |
ParaSCI-ACL | In this paper, we study the use of standard continuous representations for words to generate translation rules for infrequent phrases. | In this work, we show how simple continuous representations of phrases can be successfully used to induce translation rules for infrequent phrases. |
ParaSCI-arXiv | Simon and Ronder propose a constellation model to localize parts of objects, which utilizes cnn to find the constellations of neural activation patterns. | Simon et al. propose a neural activation constellations part model to localize parts with constellation model. |
ParaSCI-arXiv | Here we will concentrate only on those aspects that are directly relevant to the odderon. | We will put some emphasis on those aspects that are immediately relevant to the odderon. |
3.3 Statistic Characteristic
To assess the characteristics of ParaSCI, we compare its statistic characteristics with five main sentential paraphrase datasets in Table 2. The source genre of ParaSCI is scientific papers. Therefore, sentences are more formal and scholastic, and they differ from oral TUC or newsy MSRP. The average sentence length of ParaSCI is almost twice as long as that of ParaNMT-50M, MSCOCO and Quora, and also much longer than that of TUC. Its average length is only a little shorter than that of MSRP. As MSRP only contains 3,900 gold-standard paraphrases, our ParaSCI is complementary to the vacancy of large-scale long paraphrase pairs.
The degree of alteration is another important aspect of paraphrases. To compare our ParaSCI with other existing paraphrase datasets in this aspect, we propose to calculate the BLEU4 score Papineni et al. (2002) between the source and target sentences of each paraphrase pair and name it Self-BLEU. In Table 2, ParaSCI, especially ParaSCI-ACL, shows a relatively low Self-BLEU, which means sentences are significantly changed.
Name | Genre | Size (pairs) | Gold Size555gold-standard paraphrase (pairs) | Len | Char Len | Self-BLEU |
---|---|---|---|---|---|---|
MSRP | news | 5,801 | 3,900 | 22.48 | 119.62 | 47.98 |
TUC | 56,787 | 21,287 | 15.55 | 85.10 | 12.53 | |
ParaNMT-50M | Novels, laws | 51,409,585 | 51,409,585 | 12.94 | 59.18 | 28.60 |
MSCOCO | Description | 493,186 | 493,186 | 10.48 | 51.56 | 31.97 |
Quora | Question | 404,289 | 149,263 | 11.14 | 52.89 | 29.46 |
ParaSCI-ACL | Scientific Papers | 59,402 | 33,981 | 19.10 | 113.76 | 26.52 |
ParaSCI-arXiv | Scientific Papers | 479,526 | 316,063 | 18.84 | 114.46 | 29.90 |
4 Method
4.1 Extracting Paraphrase Candidates
Based on the unique characteristics of scientific papers, we extract the paraphrase sentences from a same paper (intra-paper) and across different papers (inter-paper). In most cases, we develop a simple but practical model to discover paraphrases in different sections effectively. For a more challenging case, when sentences are paraphrased partially rather than entirely, we propose PDBERT to collect more paraphrase candidates.
4.1.1 Intra-paper Extraction of Paraphrase Candidates
Authors usually write down the same information with transformed expressions repeatedly in different parts of the paper to emphasize critical information or echo back and forth. This kind of feature is the premise of our intra-paper extraction methods.
Sentence BERT for Paraphrase Extraction across Different Sections
Noting that sentences with shared semantics appear in different parts of one paper. For instance, the following sentences are from different parts of a same paper, and they are paraphrases:
: we propose a simple yet robust stochastic answer network (SAN) that simulates multi-step reasoning in machine reading comprehension. (abstract, Liu et al. (2018))
: we introduce Stochastic Answer Networks (SAN), a simple yet robust model for machine reading comprehension. (introduction, Liu et al. (2018))
However, sentences in Method, Data and Result sections are semantically different even when they only have minor changes. For example, the strings other than numbers may be very similar when presenting the two experimental results, but the semantics are completely different. Therefore, we mainly focus on six sections (Abstract, Introduction, Background, Discussion, Preamble and Conclusion). We directly obtain embeddings of sentences through BERT Devlin et al. (2018). Then, we calculate the cosine similarity pair by pair and retain sentence pairs with a similarity score higher than 0.931 as favorable paraphrase candidates. 16,563 paraphrase candidates from ACL and 158,227 paraphrase candidates from arXiv are obtained efficiently in this way.
It is noteworthy that as information is hardly repeated in the same section, we exclude the comparison between sentences in the same section.
PDBERT for Paraphrase Discovering
In fact, two sentences, even if they are not semantically equivalent, may share some parts with common semantics. For example:
: rationales are never given during training.
: in other words, target rationales are never provided during training; the intermediate step of rationale generation is guided only by the two desiderata discussed above.
and the bold part of constitute a paraphrase. However, since the similarity between the entire and is only 0.88, this pair of sentences will be filtered out. This phenomenon frequently occurs when a sentence is much longer than the other, so that part of the long sentence paraphrases the short sentence. To resolve this problem, we propose a new paraphrase discovering model, PDBERT.

Our model architecture is shown in Figure 1. We use the paraphrase sentence pairs recognized based on sentence similarity (our first method mentioned above) to construct pseudo-labeled data for model training. For each genuine paraphrase pairs, we take one as Sentence A and the other as Sentence B. Then we randomly pick out two sentences from the whole sentence set, as Sentence C and Sentence D. We use Sentence A as input 1, and concatenate Sentence C, Sentence B and Sentence D, with the 80%, 50%, and 80% appearance probabilities respectively, as input 2. Input 1 and input 2 are further concatenated by adding a token [CLS] at the beginning of input 1, and inserting a [SEP] token between input 1 and input 2. The start and end positions of Sentence B in the concatenated string are recorded as the ground-truth. It is worth mentioning that while concatenating, the ending punctuations of Sentence C and Sentence B are removed. Besides, the random selection of sentences in input 2 guarantees our model to cover various situations, for example, the first part of input 2 is the ground-truth paraphrase of input 1, or input 2 is just Sentence B.
In this way, we generate a great number of training pairs. Input 1 represents a short sentence and input 2 represents a long sentence, which includes the paraphrase of the short one. We fine-tune the model to predict the start and end positions of Sentence B using a softmax function.
Although our training data are pseudo, the fine-tuned model performs well on the real data. For a given real sentence pair, we take the shorter sentence as input 1 and the longer one as input 2, and extract from the long sentence according to the predicted positions. Actually, there is a tiny probability that the entire long sentence is the paraphrase of the short sentence. We exclude such a result because we have already obtained it via sentence BERT. Here we obtain 9,915 paraphrase candidates from ACL and 45,397 pairs from arXiv.
We compare our PDBERT model with a best-clause baseline. The latter is simply splitting the long sentence into several clauses by punctuation and selecting the combination of clauses that have a maximum similarity with the short sentence.
Method | Speed | Size | BERTScore |
---|---|---|---|
Best-clause | 7.16 | 341 | 74.06 |
PDBERT | 32.34 | 387 | 88.23 |
We implement PDBERT and the baseline method on the same corpus and compare the processing speed, size of valuable extractions and BERTScoreZhang et al. (2019). Table 3 demonstrates the comparison. PDBERT has a higher extraction speed because, in the baseline method, we have to embed each possible combination of clauses. For valuable extractions, PDBERT extracts a larger size of candidate paraphrase pairs. Moreover, the BERTScore of PDBERT’s results is higher, indicating its semantic advantage in paraphrase discovering.
4.1.2 Inter-paper Extraction of Paraphrase Candidates
Paraphrases also exist across different papers in the same field, including explanations of the same concept in different papers and citations to the same paper.
Explanations of the Same Concept
In scientific papers, in order to introduce the definition of a task or a scientific terminology, the authors often explain it in one sentence. Therefore, various definitions of the same term in different papers become paraphrase candidates naturally, just as the following case:
: Sentence compression is the task of producing a shorter form of a single given sentence, so that the new form is grammatical and retains the most important information of the original one.
: Sentence compression is a task of creating a short grammatical sentence by removing extraneous words or phrases from an original sentence while preserving its meaning.
In order to extract the definition sentences from different papers, we design a series of possible patterns (regular expressions) of definition sentences and tag the terms in them. In the same subject or area (provided by meta-data of source materials), the extracted sentences are aggregated according to terms, so we obtain multiple explanations of the same concept. In order to ensure the same semantics, we combine them into pairs and adopt the method in Section 4.1.1 to filter out the sentence pairs that are semantically different. Here we get 5,912 paraphrase candidates from ACL and 63,258 paraphrase candidates from arXiv.
Citations to the Same Paper
Scientific papers often need to cite previous works. Besides, authors tend to give a brief introduction (i.e., citation text) to the cited paper. If different papers cite the same one, the introductory sentences to this cited paper in different papers also naturally constitute a sentence set with possible paraphrase relationships. Figure LABEL:f2 provides an example.
We locate the citations in the paper through S2ORC data and extract the citation sentence and the cited article. All the extracted results are aggregated according to the same cited paper. Then we match sentences in the same group. Similarly, in order to ensure the same semantics, we use the method in Section 4.1.1 to filter out semantically inconsistent sentence pairs. In this way, we obtain 27,016 pairs of candidates for ACL and 212,644 pairs for arXiv.
4.2 Selecting High-quality Paraphrases
Due to insufficient computing resources, we have used a rough but fast filtering method to construct paraphrase candidate set as mentioned above. It includes 59,406 pairs from ACL and 479,526 pairs from arXiv. To obtain the high-quality paraphrase corpus, we implement domain-related BERTScore and paraphrase length rates to determine if a candidate pair is really paraphrase. Our two-stage construction process works in a coarse-to-fine way.
BERTScore leverages the pre-trained contextual embeddings from BERT and matches words in candidate and reference sentences by cosine similarity. It has been shown to correlate with human judgment on sentence-level and system-level evaluation. We calculate domain-related BERTScore for each paraphrase candidate, with the concrete setting of scibert-scivocab-uncased_L8_no-idf_version=0.3.3.
We design paraphrase length rate (PLR) as another filter because the numbers of words in two paraphrase sentences usually do not vary too much. PLR is simply calculated as:
and stand for lengths of the corresponding sentences.
For paraphrase candidates extracted from explanations of the same concept, they consist of more abstract knowledge so we set a loose restriction. We select those with a BERTScore higher than 0.6 and PLR lower than 2. Therefore, we get 4,566 definition paraphrase pairs from ACL candidates and 49,052 pairs from arXiv candidates. For paraphrase candidates extracted from other methods, we change the threshold of BERTScore to be 0.7 and PLR to be 1.0. In this way, we get another 29,415 pairs from ACL candidates and 267,106 pairs from arXiv candidates.
5 Manual Evaluation
We conduct a manual analysis of our dataset in order to quantify its semantic consistency and literal variation lexically, phrasally and sententially. We employ 12 volunteers who are proficient in English to rate the instances. Three human judgements are obtained for every sample and the final scores are averaged across different judges.
Consistency Evaluation Criterion
For semantic consistency of paraphrase pairs, we design 5 degrees to distinguish. For a sentence pair to have a rating of 5, the sentences must have exactly the same meaning with all the same details. To have a rating of 4, the sentences are mostly equivalent, but some unimportant details can differ. To have a rating of 3, the sentences are roughly equivalent, with some important information missing or that differs slightly. For a rating of 2, the sentences are not equivalent, even if they share minor details. For a rating of 1, the sentences are totally different. (Examples are shown in the appendix)
Variation Evaluation Criterion
For literal variation of paraphrase pairs lexically, phrasally and sententially, we use the following criterion respectively: 5 means there are more than five variations of this level, 4 means four or five, 3 means two or three, 2 means it has only one change and 1 means no change.
Name | Lexical | Phrasal | Sentential |
---|---|---|---|
ParaSCI-ACL | 3.82 | 2.73 | 2.01 |
ParaSCI-arXiv | 3.67 | 2.68 | 1.48 |
Quality Control
We evaluate the annotation quality of each worker using Cohen’s kappa agreement Artstein and Poesio (2008) against the majority vote of other workers. We asked the best worker to label more data by republishing the questions done by workers with low reliability (Cohen’s kappa <0.4). Finally, the average Cohen’s kappa of semantic consistency evaluation is 0.71 and that of literal variation is 0.62 (0.66 lexically, 0.59 phrasally and 0.61 sententially).
Evaluation Results
The average semantic consistency of ParaSCI-ACL is 4.17 and that of ParaSCI-arXiv is 3.94, both around 4, which means most paraphrase pairs are nearly equivalent, only some unimportant details may differ. Besides, the average semantic consistency of ParaSCI-ACL is higher than that of ParaSCI-arXiv. In terms of literal variation, Table 4 summarizes the annotations. ParaSCI-ACL and ParaSCI-arXiv show similar distributions. Paraphrase sentences usually change a lot lexically, because lexical variation is easier to realize. Although sentential variation scores are lower than lexical or phrasal scores, nearly one or two sentential variations for each pair are already rare and valuable for a paraphrase dataset. The long average length makes such sentential transformation possible, which is complementary to other datasets of short paraphrases.
6 Paraphrase Phenomenon Occurrence
In order to show the differences across paraphrase datasets, we sample 100 sentential paraphrases from each dataset and count occurrences of each phenomenon. Boonthum (2004) grouped common paraphrase phenomenon into 6 categories : Synonym (substitute a word with its synonym), Voice (change the voice of sentence from active to passive or vice versa), Word-Form (change a word into a different form), Break (break a long sentence down into small sentences), Definition (substitute a word with its definition or meaning), Structure (use different sentence structures to express the same thing). We report the average number of occurrences of each paraphrase type per sentence pair for each corpus in Table 5 and visualize that in Figure LABEL:f3.
Name | Syn | Voice | Form | Break | Def | Struct |
---|---|---|---|---|---|---|
MSRP | 0.80 | 0.19 | 0.15 | 0.15 | 0.31 | 0.28 |
TUC | 0.50 | 0.10 | 0.10 | 0.09 | 0.53 | 0.29 |
ParaNMT-50M | 0.87 | 0.15 | 0.20 | 0.13 | 0.40 | 0.25 |
MSCOCO | 0.72 | 0.05 | 0.12 | 0.10 | 0.36 | 0.26 |
Quora | 1.02 | 0.16 | 0.22 | 0.23 | 0.74 | 0.46 |
ParaSCI-ACL | 0.97 | 0.14 | 0.12 | 0.28 | 0.57 | 0.45 |
ParaSCI-arXiv | 1.04 | 0.11 | 0.15 | 0.32 | 0.68 | 0.41 |
As shown in Table 5 and Figure LABEL:f3, ParaSCI-ACL, ParaSCI-arXiv and Quora share a similar paraphrase phenomenon distribution. Besides, with respect to four categories of paraphrase phenomena (Synonym, Break, Definition, Structure), they all rank top 3. It reveals that our ParaSCI and Quora datasets contain more valuable paraphrase phenomena or paraphrase patterns. However, the Quora dataset is limited to questions pairs and our ParaSCI provides declarative sentence pairs.
7 Paraphrase Generation
To demonstrate further application of ParaSCI, we train paraphrase generation models on Quora, MSCOCO and ParaSCI. We then test their generation ability on the same scientific corpus, including sentences from ACL and arXiv respectively. For paraphrase generation from ParaSCI-ACL, we use 20,388 pairs for training, 6,796 pairs for validation and 6,797 pairs for test. For paraphrase generation from ParaSCI-arXiv, we use 189,639 pairs for training, 63,212 pairs for validation and 63,121 pairs for test. The BLEU scores and average lengths of generated sentences are shown in Table 6.
Training | Test | BLEU | Len |
---|---|---|---|
Quora | ParaSCI-ACL | 6.31 | 14.23 |
Quora | ParaSCI-arXiv | 8.06 | 13.92 |
ParaSCI-ACL | ParaSCI-ACL | 15.70 | 18.45 |
ParaSCI-arXiv | ParaSCI-arXiv | 27.18 | 18.82 |
Although there are many technology-related or scientific questions on Quora, the paraphrase generation model trained on Quora still fails to perform well in the scientific field, with low BLEU scores and short average length. On the contrary, the paraphrase generation model trained on ParaSCI keeps generating longer sentences. The BLEU scores also demonstrate that the quality of the generated sentences is higher. This reflects the significant value of ParaSCI on paraphrase generation.
We show the generated paraphrases on different datasets in Table 7. To be fair, we still use the same Transformer architecture in the experiment. The generated paraphrases vary a lot. In most situations, the generated sentences from MSCOCO is incomplete, more like a phrase. The model trained on Quora only generates short questions. Whether trained on MSCOCO or Quora, the generated sentences usually share similar structures and have a large portion of entities in them. On the contrary, models trained on ParaSCI handle the paraphrase generation of longer sequence, including more modifiers and conjunctions.
Apart from that, generation models trained on ParaSCI manifest another characteristic. As ParaSCI consists of quantities of scientific terms and expressions, generation models trained on ParaSCI bring valuable scientific knowledge to the output sentence.
One thing to mention is that some abbreviations can be understood and utilized in the generation process. For instance, we generated “we ran mt experiments using the moses phrase-based translation system.” for the sentence “we used moses as the phrase-based machine translation system.”. As we use domain-related terms and corresponding abbreviations in scientific texts frequently, this advantage can add conciseness, technicality and naturality to the generated sentences.
Another thing to mention is that some common sense or scientific knowledge is also taken into paraphrase generation. For example, we input “the penn discourse treebank is the largest corpus richly annotated with explicit and implicit discourse relations and their senses.” as the original sentence. It introduces the Penn Discourse Treebank Miltsakaki et al. (2004) without any information related to its size and source, but the information is added to the generated sentences: “the penn discourse treebank is the largest available annotated corpora of discourse relations over 2,312 wall street journal articles.”
These cases reveal different aspects of ParaSCI’s advantages in paraphrase generation. In further work, we hope that ParaSCI will contribute more to scientific paraphrase generation and subsequently, be applied to more downstream tasks in the scientific field.
Name | Original | Paraphrase |
---|---|---|
MSCOCO | a group of people watch a dog ride a motorcycle. | an old photo of people riding on a motorcycle and waving. |
Quora | how can i get saved wifi password? | how can i see a saved wifi password? |
ParaSCI-ACL | relation extraction ( re ) is the task of determining semantic relations between entities mentioned in text. | relation extraction ( re ) is the task of recognizing the assertion of a particular relationship between two or more entities in text. |
ParaSCI-arXiv | cosmic strings are linear topological defects that can form in the early universe as a result of symmetry-breaking phase transitions. | cosmic strings are one-dimensional massive objects, which may appear as topological defects at the spontaneous symmetry breaking in the early universe. |
8 Conclusion and Future Work
In this paper, we describe the characteristics and construction process of ParaSCI, a large-scale paraphrase dataset in the scientific field. It shows favorable results in the either automatic or manual evaluation.
For future work, although we filter out more than 200 thousand paraphrase candidates to promise the quality of ParaSCI, most candidates include valuable paraphrase patterns lexically or phrasally. Therefore, more paraphrase patterns are remaining to be discovered. Similarly, compared to the bulk data on arXiv or other scientific websites, we only use the tip of an iceberg to construct this dataset, and we are expecting to implement the methods in other scientific domains. For instance, we can obtain a biomedical paraphrase dataset from PubMed.
We hope that ParaSCI can be used to augment training data for various NLP tasks, such as machine translation in scientific field, and make more contributions to the development of NLP.
Acknowledgments
This work was supported by National Natural Science Foundation of China (61772036), Beijing Academy of Artificial Intelligence (BAAI) and Key Laboratory of Science, Technology and Standard in Press Industry (Key Laboratory of Intelligent Press Media Technology). We appreciate the anonymous reviewers for their helpful comments. Xiaojun Wan is the corresponding author.
References
- Aizawa et al. (2018) Akiko Aizawa, Takeshi Sagara, Kenichi Iwatsuki, and Goran Topic. 2018. Construction of a new acl anthology corpus for deeper analysis of scientific papers. In Third International Workshop on SCIentific DOCument Analysis (SCIDOCA-2018), Nov. 2018.
- Artstein and Poesio (2008) Ron Artstein and Massimo Poesio. 2008. Inter-coder agreement for computational linguistics. Computational Linguistics, 34(4):555–596.
- Bhagat and Hovy (2013) Rahul Bhagat and Eduard Hovy. 2013. What is a paraphrase? Computational Linguistics, 39(3):463–472.
- Boonthum (2004) Chutima Boonthum. 2004. iSTART: Paraphrase recognition. In Proceedings of the ACL Student Research Workshop, pages 31–36, Barcelona, Spain. Association for Computational Linguistics.
- Cho et al. (2014) Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078.
- Das and Smith (2009) Dipanjan Das and Noah A Smith. 2009. Paraphrase identification as probabilistic quasi-synchronous recognition. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 1-Volume 1, pages 468–476. Association for Computational Linguistics.
- Devlin et al. (2018) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
- Dolan and Brockett (2005) William B Dolan and Chris Brockett. 2005. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the Third International Workshop on Paraphrasing (IWP2005).
- Dong et al. (2017) Li Dong, Jonathan Mallinson, Siva Reddy, and Mirella Lapata. 2017. Learning to paraphrase for question answering. arXiv preprint arXiv:1708.06022.
- Fu et al. (2019) Yao Fu, Yansong Feng, and John P Cunningham. 2019. Paraphrase generation with latent bag of words. In Advances in Neural Information Processing Systems, pages 13623–13634.
- Gupta et al. (2018) Ankush Gupta, Arvind Agarwal, Prawaan Singh, and Piyush Rai. 2018. A deep generative framework for paraphrase generation. In Thirty-Second AAAI Conference on Artificial Intelligence.
- Lan et al. (2017) Wuwei Lan, Siyu Qiu, Hua He, and Wei Xu. 2017. A continuously growing dataset of sentential paraphrases. In Proceedings of The 2017 Conference on Empirical Methods on Natural Language Processing (EMNLP), pages 1235–1245. Association for Computational Linguistics.
- Lin et al. (2014) Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. 2014. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740–755. Springer.
- Liu et al. (2018) Xiaodong Liu, Yelong Shen, Kevin Duh, and Jianfeng Gao. 2018. Stochastic answer networks for machine reading comprehension. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1694–1704, Melbourne, Australia. Association for Computational Linguistics.
- Lo et al. (2020) Kyle Lo, Lucy Lu Wang, Mark Neumann, Rodney Kinney, and Daniel S. Weld. 2020. S2ORC: The Semantic Scholar Open Research Corpus. In Proceedings of ACL.
- Mallinson et al. (2017) Jonathan Mallinson, Rico Sennrich, and Mirella Lapata. 2017. Paraphrasing revisited with neural machine translation. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 881–893.
- Miltsakaki et al. (2004) Eleni Miltsakaki, Rashmi Prasad, Aravind K Joshi, and Bonnie L Webber. 2004. The penn discourse treebank. In LREC.
- Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics, pages 311–318. Association for Computational Linguistics.
- Pavlick et al. (2015) Ellie Pavlick, Pushpendre Rastogi, Juri Ganitkevitch, Benjamin Van Durme, and Chris Callison-Burch. 2015. Ppdb 2.0: Better paraphrase ranking, fine-grained entailment relations, word embeddings, and style classification. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 425–430.
- Su and Yan (2017) Yu Su and Xifeng Yan. 2017. Cross-domain semantic parsing via paraphrasing. arXiv preprint arXiv:1704.05974.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems, pages 5998–6008.
- Wieting and Gimpel (2018) John Wieting and Kevin Gimpel. 2018. ParaNMT-50M: Pushing the limits of paraphrastic sentence embeddings with millions of machine translations. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 451–462, Melbourne, Australia. Association for Computational Linguistics.
- Xu et al. (2015) Wei Xu, Chris Callison-Burch, and Bill Dolan. 2015. Semeval-2015 task 1: Paraphrase and semantic similarity in twitter (pit). In Proceedings of the 9th international workshop on semantic evaluation (SemEval 2015), pages 1–11.
- Zhang et al. (2019) Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. 2019. Bertscore: Evaluating text generation with bert. arXiv preprint arXiv:1904.09675.
Appendix A Appendices
Examples of semantic consistency evaluation are shown in the following table.
Sentence A | Sentence B | Score |
---|---|---|
Task-oriented dialog systems help users to achieve specific goals with natural language. | We use a set of 318 English function words from the scikit-learn package. | 1 |
End-to-end task-oriented dialog systems usually suffer from the challenge of incorporating knowledge bases. | Task-oriented dialog systems help users to achieve specific goals with natural language. | 2 |
Opinion mining has recently received considerable attentions. | Analysis has received much attention in recent years. | 3 |
We evaluated all agents on 57 Atari 2600 games from the arcade learning environment. | We evaluated EMDQN on the benchmark suite of 57 Atari 2600 games from the arcade learning environment. | 4 |
Here we will concentrate only on those aspects that are directly relevant to the odderon. | We will put some emphasis on those aspects that are immediately relevant to the odderon. | 5 |