| import shutil |
|
|
| import pandas as pd |
| import os |
|
|
| import huggingface_hub |
| import pytest |
| from datasets import load_dataset |
|
|
|
|
| @pytest.mark.parametrize( |
| "dataset_name, link_to_source", |
| [ |
| ( |
| "h2ogpt-oig-instruct-cleaned", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/b8f15efcc305a953c52a0ee25b8b4897ceb68c0a/scrape_dai_docs.py) |
| """ |
| ), |
| ( |
| "h2ogpt-oig-instruct-cleaned-v2", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/40c217f610766715acec297a5535eb440ac2f2e2/create_data.py) |
| """ |
| ), |
| ( |
| "h2ogpt-oig-instruct-cleaned-v3", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/bfc3778c8db938761ce2093351bf2bf82159291e/create_data.py) |
| """ |
| ), |
| ( |
| "openassistant_oasst1", |
| """ |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/45e6183171fb16691ad7d3ab006fad973f971e98/create_data.py#L1253) |
| """ |
| ), |
| ( |
| "h2ogpt-oig-oasst1-instruct-cleaned-v1", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
| |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/5fc91911bc2bfaaf3b6c2de577c4b0ae45a07a4a/create_data.py#L1253) |
| """ |
| ), |
| ( |
| "h2ogpt-oig-oasst1-instruct-cleaned-v2", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
| |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/0e70c2fbb16410bd8e6992d879b4c55cd981211f/create_data.py#L1375-L1415) |
| """ |
| ), |
| ( |
| "h2ogpt-oig-oasst1-instruct-cleaned-v3", |
| """ |
| - [Original LAION OIG Dataset](https://github.com/LAION-AI/Open-Instruction-Generalist) |
| - [LAION OIG data detoxed and filtered down by scripts in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/main/docs/FINETUNE.md#high-quality-oig-based-instruct-data) |
| |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/6728938a262d3eb5e8db1f252bbcd7de838da452/create_data.py#L1415) |
| """ |
| ), |
| ( |
| "openassistant_oasst1_h2ogpt", |
| """ |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/83857fcf7d3b712aad5db32207e6db0ab0f780f9/create_data.py#L1252) |
| """ |
| ), |
| ( |
| "openassistant_oasst1_h2ogpt_graded", |
| """ |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/d1f8ce975a46056d41135d126dd33de8499aa26e/create_data.py#L1259) |
| """ |
| ), |
| ( |
| "h2ogpt-fortune2000-personalized", |
| """ |
| - [Fortune 2000 companies from Wikipedia](https://github.com/h2oai/h2ogpt/blob/b1ea74c0088884ebff97f1ccddbfb3f393e29e44/create_data.py#L1743) |
| """ |
| ), |
| ( |
| "openassistant_oasst1_h2ogpt_llama2_chat", |
| """ |
| - [Original Open Assistant data in tree structure](https://huggingface.co/datasets/OpenAssistant/oasst1) |
| - [This flattened dataset created by script in h2oGPT repository](https://github.com/h2oai/h2ogpt/blob/0bee5f50a74f489ca3fc81486f9322078360f2cb/src/create_data.py#L1296) |
| """ |
| ), |
| ], |
| ) |
| def test_create_data_cards(dataset_name, link_to_source): |
| if dataset_name != "openassistant_oasst1_h2ogpt_llama2_chat": |
| return |
| |
| assert os.path.exists("README-template.md"), "must be running this test from the data dir." |
| shutil.rmtree(dataset_name, ignore_errors=True) |
| try: |
| repo = huggingface_hub.Repository( |
| local_dir=dataset_name, |
| clone_from="h2oai/%s" % dataset_name, |
| repo_type="dataset", |
| skip_lfs_files=True, |
| token=True, |
| ) |
| repo.git_pull() |
| except Exception as e: |
| print(str(e)) |
| print("call 'huggingface_cli login' first and provide access token with write permission") |
| dataset = load_dataset("h2oai/%s" % dataset_name)["train"] |
|
|
| pd.set_option('display.max_columns', None) |
| with open("README-template.md", "r") as f: |
| content = f.read() |
| assert "<<DATASET_NAME>>" in content |
| content = content.replace("<<DATASET_NAME>>", dataset_name) |
|
|
| assert "<<NROWS>>" in content |
| content = content.replace("<<NROWS>>", str(dataset.num_rows)) |
|
|
| assert "<<NCOLS>>" in content |
| content = content.replace("<<NCOLS>>", str(dataset.num_columns)) |
|
|
| assert "<<COLNAMES>>" in content |
| content = content.replace("<<COLNAMES>>", str(dataset.column_names)) |
|
|
| |
| |
|
|
| assert "<<SOURCE_LINK>>" in content |
| content = content.replace("<<SOURCE_LINK>>", link_to_source) |
|
|
| assert "<<" not in content |
| assert ">>" not in content |
|
|
| with open(os.path.join(dataset_name, "README.md"), "w") as f: |
| f.write(content) |
| try: |
| repo.commit("Update README.md") |
| repo.push_to_hub() |
| except Exception as e: |
| print(str(e)) |
|
|