| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| import os |
| import re |
| import warnings |
|
|
| import datasets |
| import requests |
|
|
| _DESCRIPTION = """\ |
| CCMatrix: Mining Billions of High-Quality Parallel Sentences on the WEB |
| |
| We show that margin-based bitext mining in LASER's multilingual sentence space can be applied to |
| monolingual corpora of billions of sentences to produce high quality aligned translation data. |
| We use thirty-two snapshots of a curated common crawl corpus [1] totaling 69 billion unique sentences. |
| Using one unified approach for 80 languages, we were able to mine 10.8 billion parallel sentences, |
| out of which only 2.9 billion are aligned with English. |
| |
| IMPORTANT: Please cite reference [2][3] if you use this data. |
| |
| [1] Guillaume Wenzek, Marie-Anne Lachaux, Alexis Conneau, Vishrav Chaudhary, Francisco Guzmán, Armand Jouli |
| and Edouard Grave, CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data |
| |
| [2] Holger Schwenk, Guillaume Wenzek, Sergey Edunov, Edouard Grave and Armand Joulin, |
| CCMatrix: Mining Billions of High-Quality Parallel Sentences on the WEB |
| |
| [3] Angela Fan, Shruti Bhosale, Holger Schwenk, Zhiyi Ma, Ahmed El-Kishky, Siddharth Goyal, Mandeep Baines, |
| Onur Celebi, Guillaume Wenzek, Vishrav Chaudhary, Naman Goyal, Tom Birch, Vitaliy Liptchinsky, |
| Sergey Edunov, Edouard Grave, Michael Auli, and Armand Joulin. |
| Beyond English-Centric Multilingual Machine Translation |
| |
| 90 languages, 1,197 bitexts |
| total number of files: 90 |
| total number of tokens: 112.14G |
| total number of sentence fragments: 7.37G |
| """ |
| _HOMEPAGE_URL = "https://opus.nlpl.eu/CCMatrix.php" |
| _CITATION = """\ |
| Guillaume Wenzek, Marie-Anne Lachaux, Alexis Conneau, Vishrav Chaudhary, Francisco Guzmán, Armand Jouli and Edouard Grave, CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data |
| """ |
|
|
| _VERSION = "1.0.0" |
| _FILE = "CCMatrix.{}.{}" |
| _URL = "https://opus.nlpl.eu/CCMatrix.php" |
| _DOWNLOAD_URL = "https://object.pouta.csc.fi/OPUS-CCMatrix/v1/moses/{}.txt.zip" |
|
|
|
|
| def get_language_pairs(): |
| try: |
| response = requests.get(_URL) |
| except requests.exceptions.RequestException: |
| warnings.warn( |
| "Unable to download language pairs from '{}'. Using cached version".format( |
| _URL |
| ) |
| ) |
| from language_pairs_cache import language_pairs |
|
|
| return language_pairs |
|
|
| html = response.text |
|
|
| ccmatrix_hrefs = [ |
| href |
| for href in re.findall(r'href=[\'"]?([^\'" >]+)', html) |
| if href.startswith("CCMatrix/") |
| ] |
|
|
| language_pairs = [] |
| for href in ccmatrix_hrefs: |
| match = re.search(r"CCMatrix/v1/(\w+-\w+)_sample.html", href) |
| if match: |
| language1, language2 = match.group(1).split("-") |
| language_pairs.append((language1, language2)) |
| language_pairs.append((language2, language1)) |
| language_pairs.sort() |
| return language_pairs |
|
|
|
|
| _CONFIGS = get_language_pairs() |
|
|
|
|
| class CCMatrixConfig(datasets.BuilderConfig): |
| def __init__(self, **kwargs): |
| super().__init__(**kwargs) |
| lang1, lang2 = kwargs["name"].split("-") |
| self.lang1 = lang1 |
| self.lang2 = lang2 |
| x, y = (lang1, lang2) if lang1 < lang2 else (lang2, lang1) |
| self.download_pair = f"{x}-{y}" |
|
|
|
|
| class CCMatrix(datasets.GeneratorBasedBuilder): |
| BUILDER_CONFIGS = [ |
| CCMatrixConfig( |
| name=f"{lang1}-{lang2}", |
| description=f"Translating {lang1} to {lang2} or vice versa", |
| version=datasets.Version(_VERSION), |
| ) |
| for lang1, lang2 in _CONFIGS |
| ] |
| BUILDER_CONFIG_CLASS = CCMatrixConfig |
|
|
| def __init__(self, *args, **kwargs): |
| if "max_train_samples" in kwargs and kwargs.get("cache_dir", None) is None: |
| kwargs["cache_dir"] = os.path.join( |
| str(datasets.config.HF_DATASETS_CACHE), |
| "trainsamples_{}".format(kwargs["max_train_samples"]), |
| ) |
| self.max_samples = { |
| "train": kwargs.get("max_train_samples", 2**64), |
| } |
| kwargs = { |
| k: v |
| for k, v in kwargs.items() |
| if k not in ["max_train_samples", "id_filter"] |
| } |
| super().__init__(*args, **kwargs) |
|
|
| def _info(self): |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| features=datasets.Features( |
| { |
| "id": datasets.Value("int32"), |
| "score": datasets.Value("float"), |
| "translation": datasets.Translation( |
| languages=(self.config.lang1, self.config.lang2) |
| ), |
| }, |
| ), |
| supervised_keys=None, |
| homepage=_HOMEPAGE_URL, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| download_url = _DOWNLOAD_URL.format(self.config.download_pair) |
| path = dl_manager.download_and_extract(download_url) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| gen_kwargs={"datapath": path, "max_samples": self.max_samples["train"]}, |
| ) |
| ] |
|
|
| def _generate_examples(self, datapath, max_samples): |
| l1_path = os.path.join( |
| datapath, _FILE.format(self.config.download_pair, self.config.lang1) |
| ) |
| l2_path = os.path.join( |
| datapath, _FILE.format(self.config.download_pair, self.config.lang2) |
| ) |
| scores_path = os.path.join( |
| datapath, _FILE.format(self.config.download_pair, "scores") |
| ) |
| with open(l1_path, encoding="utf-8") as f1, open( |
| l2_path, encoding="utf-8" |
| ) as f2, open(scores_path, encoding="utf-8") as f3: |
| for sentence_counter, (x, y, score) in enumerate(zip(f1, f2, f3)): |
| if sentence_counter == max_samples: |
| return |
| result = ( |
| sentence_counter, |
| { |
| "id": sentence_counter, |
| "score": score, |
| "translation": { |
| self.config.lang1: x.strip(), |
| self.config.lang2: y.strip(), |
| }, |
| }, |
| ) |
| yield result |
|
|