Skip to content
/ Pira Public

Pirá: A Bilingual Portuguese-English Dataset for Question-Answering about the Ocean

Notifications You must be signed in to change notification settings

C4AI/Pira

Repository files navigation

Pirá: A Bilingual Portuguese-English Dataset for Question-Answering about the Ocean, the Brazilian coast, and climate change

Pirá is a crowdsourced reading comprehension dataset on the ocean, the Brazilian coast, and climate change. QA sets are presented in both Portuguese and English; the dataset also contains human paraphrases and assessments.

The original paper was published at CIKM'21 and can be found here. As a subsequent project, we have produced a curated version of the dataset, which we refer to as Pirá 2.0. In this step, we have also defined a number of benchmarks and reported the corresponding baselines. Pirá 2.0's preprint is available here.

Pirá is, to the best of our knowledge, the first QA dataset with supporting texts in Portuguese, and, perhaps more importantly, the first bilingual QA dataset that includes Portuguese as one of its languages. Pirá is also the first QA dataset in Portuguese with unanswerable questions so as to allow the study of answer triggering. Finally, it is the first QA dataset that tackles scientific knowledge about the ocean, climate change, and marine biodiversity.

This Githubg Page gathers the different versions of Pirá as well as the Codes for the experiments described in the papers. You can also download Pirá directly from Hugging Face.

Methodology

The dataset generation process is depicted below. Two different corpora of texts were collected: corpus 1 is composed of 3891 abstracts of scientific papers about the Brazilian coast from the Scopus database, and corpus 2 is composed of 189 small excerpts of two reports on the ocean organized by the United Nations (World Ocean Assessment I and World Ocean Assessment II). Undergraduate and graduate volunteers then created questions based on these texts, both in English and Portuguese. Participants were instructed to produce questions that could be answered with the use of the texts and no other source of information. In a second step, the same volunteers assessed these question sets in a number of ways. They were asked to: i) answer the question in both languages without having access to the original answer; ii) assess the whole original question-answer pair (QA sets) according to a number of aspects; and iii) paraphrase the original question.



Here is an example of a question-answer pair set from Pirá:

{
    "id_qa": "B2142",
    "corpus": 2,
    "question_en_origin": "What are the proportion of men and women employed in the fishery sector worlwide?",
    "question_pt_origin": "Qual é a proporção de homens e mulheres empregados no setor pesqueiro em todo o mundo?",
    "question_en_paraphase": "Which share of the fishery sector workers of the world are women?",
    "question_pt_paraphase": "Qual parcela dos trabalhadores do setor da pesca no mundo são mulheres?",
    "answer_en_origin": "85 per cent men and 15 per cent women.",
    "answer_pt_origin": "85 por cento homens e 15 por cento mulheres.",
    "answer_en_validate": "It is estimated that more than fifteen per cent of the fishing sector workers are women.",
    "answer_pt_validate": "Estima-se que mais de quinze por cento dos trabalhadores do setor da pesca são mulheres.",
    "eid_article_scopus": "",
    "text_excerpts_un_reports": "Distribution of ocean benefits and disbenefits Developments in employment and income from fisheries and aquaculture The global harvest of marine capture fisheries has expanded rapidly since the early 1950s and is currently estimated to be about 80 million tons a year. That harvest is estimated to have a first (gross) value on the order of 113 billion dollars. Although it is difficult to produce accurate employment statistics, estimates using a fairly narrow definition of employment have put the figure of those employed in fisheries and aquaculture at 58.3 million people (4.4 per cent of the estimated total of economically active people), of which 84 per cent are in Asia and 10 per cent in Africa. Women are estimated to account for more than 15 per cent of people employed in the fishery sector. Other estimates, probably taking into account a wider definition of employment, suggest that capture fisheries provide direct and indirect employment for at least 120 million persons worldwide. Small-scale fisheries employ more than 90 per cent of the world’s capture fishermen and fish workers, about half of whom are women. When all dependants of those taking full- or part-time employment in the full value chain and support industries (boatbuilding, gear construction, etc.) of fisheries and aquaculture are included, one estimate concludes that between 660 and 820 million persons have some economic or livelihood dependence on fish capture and culture and the subsequent direct value chain. No sound information appears to be available on the levels of death and injury of those engaged in capture fishing or aquaculture, but capture fishing is commonly characterized as a dangerous occupation. Over time, a striking shift has occurred in the operation and location of capture fisheries. In the 1950s, capture fisheries were largely undertaken by developed fishing States. Since then, developing countries have increased their share. As a broad illustration, in the 1950s, the southern hemisphere accounted for no more than 8 per cent of landed values. By the last decade, the southern hemisphere’s share had risen to 20 per cent. In 2012, international trade represented 37 per cent of the total fish production in value, with a total export value of 129 billion dollars, of which 70 billion dollars (58 per cent) was exports by developing countries. Aquaculture is responsible for the bulk of the production of seaweeds. Worldwide, reports show that 24.9 million tons was produced in 2012, valued at about 6 billion dollars. In addition, about 1 million tons of wild seaweed were harvested. Few data were found on international trade in seaweeds, but their culture is concentrated in countries where consumption of seaweeds is high.",
    "question_generic": false,
    "answer_in_text": true,
    "answer_difficulty": 1,
    "question_meaningful": 5,
    "answer_equivalent": 5,
    "question_type": "None of the above"
  }

Dataset description

Pirá contains 2258 QA sets, as well as supporting texts associated with them. Each QA set contains at least four elements: a question in Portuguese and in English, and an answer in Portuguese and in English. Around 90% of the QA sets also contain human evaluations.

For Pirá 2.0, the original dataset has been completely revised for grammar issues, misplaced entries, repeated questions, plus other minor flaws. In addition to that, we extended the dataset in several directions. First, the new dataset includes automatic translations of the supporting texts into Portuguese, allowing us to run the benchmarks in this language as well. Second, we provide classification labels indicating whether a question can be answered or not, which are essential to the answer triggering benchmark. Third, Pirá 2.0 offers a multiple-choice QA extension, where each question has five candidate answers, of which only one is correct. Finally, Pirá 2.0 brings automatically-generated paraphrases for questions and answers in both Portuguese and English, a valuable data augmentation addition for training larger models.

We organize the dataset in two different files. In both cases, the dataset is split into train, validation, and test sets.

  • Contains the supporting texts, QA sets, manual paraphrases, human assessments, and automatic paraphrases.

  • Benchmarks supported: machine reading comprehension, information retrieval, open question answering, and answer triggering.

Multiple-choice QA

  • Contains the supporting texts, candidate answers, and the label for the correct answer.

  • Benchmark supported: multiple-choice question answering.

Pirá 1.0

The original dataset is available here.

Baselines

Five benchmarks have been created for Pirá 2.0: machine reading comprehension, information retrieval, open question answering, answer triggering, and multiple-choice question answering. Codes for the experiments reported in Pirá 2.0 are available at this page.

A brief description of each bechmark is provided above, as well as the results for the best baselines in Portuguese and English:

In Machine Reading Comprehension (MRC), the goal is to predict a span in the supporting text that answers a given question.

Results in English. In bold, best results for extractive and generative models based on F1-score.

Model type Model Fine-tuned MSL F1 EM
Extractive BERT Base SQuAD 1 512 41.54 11.01
Extractive BERT Large SQuAD 2 512 46.96 12.34
Extractive RoBERTa Base SQuAD 2 512 47.65 13.66
Extractive RoBERTa Large SQuAD 2 512 48.22 12.78
Extractive ELECTRA Large SQuAD 2 512 46.20 11.89
Generative T5 Base SQuAD 2/Pirá 2 512 49.12 9.78
Generative T5 Base SQuAD 2/Pirá 2 1024 50.50 11.56
Generative T5 Base SQuAD 2/Pirá 2 1536 51.27 13.33
Generative T5 Large SQuAD 2/Pirá 2 512 41.22 10.67

Results in Portuguese. In bold, best results for extractive and generative models based on F1-score.

Model type Model Fine-tuned MSL F1 EM
Extractive BERTimbau Squad 1.1 512 37.53 4.44
Generative PTT5 Base Pirá 2 1536 27.90 4.44
Generative mT5 Base SQuAD 2/Pirá 2 512 14.23 0.00

Information Retrieval is the task of traversing a corpus C and delivering the k most relevant documents for a query q.

Performance comparison between BM25 (sparse, in blue) and DPR (dense, in red) for increasing number of retrieved documents k. In yellow, the simulated performance of a random retriever.

Performance comparison between the two different implementations of the algorithm for Portuguese. In the first, the system searches for passages that contain the correct answer in a Portuguese-translated version of the corpus (blue line). In the second, the system translates the question into English and searches in the original English corpus. In yellow, we plot the performance of a simulated random retriever.

Open Question Answering combines the two previous tasks in one: given a question, one has to find the corresponding texts and generate an answer based on them.

Results in English. In bold, best results for extractive and generative models based on F1-score.

Model type Model Fine-tuned k F1 EM
Extractive RoBERTa Large SQuAD 2 5 41.65 11.01
Extractive RoBERTa Large SQuAD 2 10 40.87 11.01
Extractive RoBERTa Large SQuAD 2 15 40.48 10.57
Generative T5 Base SQuAD 2/Pirá 2 5 45.99 12.33
Generative T5 Base SQuAD 2/Pirá 2 10 47.12 11.45
Generative T5 Base SQuAD 2/Pirá 2 15 48.11 11.45

Results in Portuguese. In bold, best results for extractive and generative models based on F1-score.

Model type Model Fine-tuned k F1 EM
Extractive BERTimbau Squad 1.1 15 31.23 3.08
Generative PTT5 Base Pirá 2 15 24.47 1.76

Answer Triggering is the problem of finding which questions should be answered and which ones should not; the “answerability” label are created from the manual assessments for question meaningfulness found on Pirá.

For this benchmark, we use three types of dataset. Standard is the regular Pirá 2.0 dataset after the aswer triggering labeling process. In Human Paraphrase, the training set is augmented with the validation answers and question paraphrases produced in the assessment phase. In Automatic Paraphrase, the training set is augmented with automatic paraphrases, using PEGASUS for English and PTT5-Paraphraser for Portuguese.

F1-score for class assignment and Naive Bayes. In bold, best results for Class assignment and Naive Bayes.

Dataset Class Assignment Naive Bayes
Standard 45.32 45.32 (1)
Hum. Par. 45.32 43.90 (3)
Autom. Par. 45.32 44.34 (2)

F1-score for transformer models in English (DistilBERT, RoBERTa, and ALBERT) and Portuguese (BERTimbau). In bold, best results for Emglish and Portuguese.

Dataset DistilBERT RoBERTa ALBERT BERTimbau
Standard 48.33 (2) 51.61 (1) 52.31 (2) 46.09 (3)
Hum. Par. 47.84 (3) 49.72 (3) 54.53 (1) 48.55 (2)
Autom. Par. 48.59 (1) 50.76 (2) 51.32 (3) 49.18 (1)

In Multiple Choice Question Answer, each question has five alternative answers, and the goals is to find the correct one.

Accuracy for Multiple Choice Question Answer. Best result in bold.

Model Accuracy
Random 20
IR-Text 25.80
IR-Base 22.58
UnifiedQA-NST 51.98
UnifiedQA-ST 73.12

Web application

As part of the process of creating Pirá, we developed a web application for conducting the two phases of the experiment: QA creation and evaluation.

The code for the two parts of the application can be found in data-set-builder and data-set-validator.

Citation

If you use or discuss this dataset in your work, please cite it as follows.

For the original paper:

@inproceedings{10.1145/3459637.3482012,
author = {Paschoal, Andr\'{e} F. A. and Pirozelli, Paulo and Freire, Valdinei and Delgado, Karina V. and Peres, Sarajane M. and Jos\'{e}, Marcos M. and Nakasato, Fl\'{a}vio and Oliveira, Andr\'{e} S. and Brand\~{a}o, Anarosa A. F. and Costa, Anna H. R. and Cozman, Fabio G.},
title = {Pir\'{a}: A Bilingual Portuguese-English Dataset for Question-Answering about the Ocean},
year = {2021},
isbn = {9781450384469},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3459637.3482012},
doi = {10.1145/3459637.3482012},
abstract = {Current research in natural language processing is highly dependent on carefully produced
corpora. Most existing resources focus on English; some resources focus on languages
such as Chinese and French; few resources deal with more than one language. This paper
presents the Pir\'{a} dataset, a large set of questions and answers about the ocean and
the Brazilian coast both in Portuguese and English. Pir\'{a} is, to the best of our knowledge,
the first QA dataset with supporting texts in Portuguese, and, perhaps more importantly,
the first bilingual QA dataset that includes this language. The Pir\'{a} dataset consists
of 2261 properly curated question/answer (QA) sets in both languages. The QA sets
were manually created based on two corpora: abstracts related to the Brazilian coast
and excerpts of United Nation reports about the ocean. The QA sets were validated
in a peer-review process with the dataset contributors. We discuss some of the advantages
as well as limitations of Pir\'{a}, as this new resource can support a set of tasks in
NLP such as question-answering, information retrieval, and machine translation.},
booktitle = {Proceedings of the 30th ACM International Conference on Information & Knowledge Management},
pages = {4544–4553},
numpages = {10},
keywords = {Portuguese-English dataset, question-answering dataset, bilingual dataset, ocean dataset},
location = {Virtual Event, Queensland, Australia},
series = {CIKM '21}
}

Applications

Rodrigues, L. A., & Vieira, M. H. D. S. (2021). Autosumm: Architecture for a Multi-Document Abstractive Summarization System.

Pellicer, L. F. A. O., Pirozelli, P., Costa, A. H. R., & Inoue, A. (2022). PTT5-Paraphraser: Diversity and Meaning Fidelity in Automatic Portuguese Paraphrasing. In International Conference on Computational Processing of the Portuguese Language. Springer, Cham, p. 299-309.

Pirozelli, P., Brandão, A. A. F., Peres, S. M., Cozman, F. G. (2022). To Answer or not to Answer? Filtering Questions for QA Systems. 11th Brazilian Conference on Intelligent Systems (BRACIS).

In case you cite our work, please contact us: we will be very happy of referencing it here. E-mail: paulo.pirozelli.silva@usp.br.

License

Pirá dataset is licensed under CC BY 4.0.

Scopus adopts a license–based approach which automatically enables researchers at subscribing institutions to text mine for non-commercial research purposes and to gain access to full text content in XML for this purpose. More information can be found here.

The United Nations' Reports on the ocean are freely accessible through the links above.

This work was carried out at the Center for Artificial Intelligence at the University of São Paulo (C4AI-USP), with support by the São Paulo Research Foundation (FAPESP grant #2019/07665-4) and by the IBM Corporation. It is part of the Knowledge-Enhanced Machine Learning for Reasoning about Ocean Data (KEML), a project that aims to developed a BLue Amazon Brain.

About

Pirá: A Bilingual Portuguese-English Dataset for Question-Answering about the Ocean

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •