| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """OrangeSum dataset""" |
| |
|
| |
|
| | import datasets |
| |
|
| |
|
| | _CITATION = """\ |
| | @article{eddine2020barthez, |
| | title={BARThez: a Skilled Pretrained French Sequence-to-Sequence Model}, |
| | author={Eddine, Moussa Kamal and Tixier, Antoine J-P and Vazirgiannis, Michalis}, |
| | journal={arXiv preprint arXiv:2010.12321}, |
| | year={2020} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | The OrangeSum dataset was inspired by the XSum dataset. It was created by scraping the "Orange Actu" website: https://actu.orange.fr/. Orange S.A. is a large French multinational telecommunications corporation, with 266M customers worldwide. Scraped pages cover almost a decade from Feb 2011 to Sep 2020. They belong to five main categories: France, world, politics, automotive, and society. The society category is itself divided into 8 subcategories: health, environment, people, culture, media, high-tech, unsual ("insolite" in French), and miscellaneous. |
| | |
| | Each article featured a single-sentence title as well as a very brief abstract, both professionally written by the author of the article. These two fields were extracted from each page, thus creating two summarization tasks: OrangeSum Title and OrangeSum Abstract. |
| | """ |
| |
|
| | _URL_DATA = { |
| | "abstract": "https://raw.githubusercontent.com/Tixierae/OrangeSum/main/data/docs/splits/abstract.tgz", |
| | "title": "https://raw.githubusercontent.com/Tixierae/OrangeSum/main/data/docs/splits/title.tgz", |
| | } |
| |
|
| | _DOCUMENT = "text" |
| | _SUMMARY = "summary" |
| |
|
| |
|
| | class OrangeSum(datasets.GeneratorBasedBuilder): |
| | """OrangeSum: a french abstractive summarization dataset""" |
| |
|
| | VERSION = datasets.Version("1.1.0") |
| |
|
| | BUILDER_CONFIGS = [ |
| | datasets.BuilderConfig(name="abstract", description="Abstracts used as summaries", version=VERSION), |
| | datasets.BuilderConfig(name="title", description="Titles used as summaries", version=VERSION), |
| | ] |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | _DOCUMENT: datasets.Value("string"), |
| | _SUMMARY: datasets.Value("string"), |
| | } |
| | ), |
| | supervised_keys=(_DOCUMENT, _SUMMARY), |
| | homepage="https://github.com/Tixierae/OrangeSum/", |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | """Returns SplitGenerators.""" |
| | archive = dl_manager.download(_URL_DATA[self.config.name]) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | |
| | gen_kwargs={ |
| | "source_files": dl_manager.iter_archive(archive), |
| | "target_files": dl_manager.iter_archive(archive), |
| | "split": "train", |
| | }, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | |
| | gen_kwargs={ |
| | "source_files": dl_manager.iter_archive(archive), |
| | "target_files": dl_manager.iter_archive(archive), |
| | "split": "test", |
| | }, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | |
| | gen_kwargs={ |
| | "source_files": dl_manager.iter_archive(archive), |
| | "target_files": dl_manager.iter_archive(archive), |
| | "split": "valid", |
| | }, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, source_files, target_files, split): |
| | """Yields examples.""" |
| | expected_source_path = f"{self.config.name}/{split}.source" |
| | expected_target_path = f"{self.config.name}/{split}.target" |
| | for source_path, f_source in source_files: |
| | if source_path == expected_source_path: |
| | for target_path, f_target in target_files: |
| | if target_path == expected_target_path: |
| | for idx, (document, summary) in enumerate(zip(f_source, f_target)): |
| | yield idx, {_DOCUMENT: document.decode("utf-8"), _SUMMARY: summary.decode("utf-8")} |
| | break |
| | break |
| |
|