| --- |
| annotations_creators: |
| - no-annotation |
| language: |
| - en |
| language_creators: |
| - found |
| license: |
| - mit |
| multilinguality: |
| - monolingual |
| size_categories: |
| - n<1K |
| source_datasets: |
| - original |
| task_categories: |
| - text-classification |
| task_ids: |
| - text-classification |
| paperswithcode_id: null |
| pretty_name: Text360 Sample Dataset |
| tags: |
| - text-classification |
| - arxiv |
| - wikipedia |
| dataset_info: |
| data_files: |
| train: |
| - dir1/subdir1/s1.jsonl |
| - dir2/subdir2/s2.jsonl |
| config_name: default |
| --- |
| |
| # Dataset Card for Text360 Sample Dataset |
|
|
| ## Dataset Description |
|
|
| - **Repository:** [Add your repository URL here] |
| - **Paper:** [Add paper URL if applicable] |
| - **Point of Contact:** [Add contact information] |
|
|
| ### Dataset Summary |
|
|
| This dataset contains text samples from two sources (arXiv and Wikipedia) organized in a hierarchical directory structure. Each sample includes a text field and a subset identifier. |
|
|
| ### Data Files Structure |
|
|
| The dataset maintains its original directory structure: |
| ``` |
| . |
| ├── dir1/ |
| │ └── subdir1/ |
| │ └── sample1.jsonl # Contains arXiv samples |
| └── dir2/ |
| └── subdir2/ |
| └── sample2.jsonl # Contains Wikipedia samples |
| ``` |
|
|
| ### Data Fields |
|
|
| Each JSONL file contains records with the following fields: |
| - `text`: string - The main text content |
| - `subset`: string - Source identifier ("arxiv" or "wikipedia") |
|
|
| ### Data Splits |
|
|
| All data is included in the train split, distributed across the JSONL files in their respective directories. |
|
|
| ### Example Instance |
|
|
| ```json |
| { |
| "text": "This is a long text sample from arxiv about quantum computing...", |
| "subset": "arxiv" |
| } |
| ``` |
|
|
| ## Additional Information |
|
|
| ### Dataset Creation |
|
|
| The dataset is organized in its original directory structure, with JSONL files containing text samples from arXiv and Wikipedia sources. Each file maintains its original location and format. |
|
|
| ### Curation Rationale |
|
|
| The dataset was created to provide a sample of text data from different sources for text classification tasks. |
|
|
| ### Source Data |
|
|
| #### Initial Data Collection and Normalization |
|
|
| The data was collected from two sources: |
| 1. arXiv papers |
| 2. Wikipedia articles |
|
|
| #### Who are the source language producers? |
|
|
| - arXiv: Academic researchers and scientists |
| - Wikipedia: Community contributors |
|
|
| ### Annotations |
|
|
| #### Annotation process |
|
|
| No additional annotations were added to the source data. |
|
|
| #### Who are the annotators? |
|
|
| N/A |
|
|
| ### Personal and Sensitive Information |
|
|
| The dataset does not contain any personal or sensitive information. |
|
|
| ## Considerations for Using the Data |
|
|
| ### Social Impact of Dataset |
|
|
| This dataset can be used for educational and research purposes in text classification tasks. |
|
|
| ### Discussion of Biases |
|
|
| The dataset may contain biases inherent to the source materials (arXiv papers and Wikipedia articles). |
|
|
| ### Other Known Limitations |
|
|
| The dataset is a small sample and may not be representative of all content from the source materials. |
|
|
| ### Dataset Curators |
|
|
| [Add curator information] |
|
|
| ### Licensing Information |
|
|
| This dataset is released under the MIT License. |
|
|
| ### Citation Information |
|
|
| [Add citation information] |
|
|
| ### Contributions |
|
|
| [Add contribution information] |
|
|
| ### Contact |
|
|
| [Add contact information] |