--- annotations_creators: - expert-generated language: - en license: - apache-2.0 pretty_name: Watsonx Docs Document Type Classification size_categories: - n<1K source_datasets: - ibm-research/watsonxDocsQA task_categories: - text-classification task_ids: - multi-class-classification --- # Watsonx Docs Document Type Classification This dataset is a balanced binary document-level classification subset derived from `ibm-research/watsonxDocsQA`. ## Task Classify IBM Watsonx documentation pages by their dominant user-facing purpose: - `conceptual`: documents primarily used to understand or look up information. - `how-to`: documents primarily used to complete a procedure or fix a problem. ## Splits | Split | conceptual | how-to | Total | |---|---:|---:|---:| | train | 140 | 140 | 280 | | validation | 30 | 30 | 60 | | test | 30 | 30 | 60 | ## Fields - `doc_id`: original document ID from the source dataset. - `url`: source documentation URL. - `title`: documentation page title. - `text`: model input text, constructed as `title + "\n" + first 800 words of document`. The title is preserved in full; the document body is truncated to keep inputs manageable for embedding-based classifiers. - `label`: string label, either `conceptual` or `how-to`. - `label_id`: numeric label ID, where `conceptual = 0` and `how-to = 1`. - `split`: dataset split. ## Usage ```python from datasets import load_dataset data_files = { "train": "train.csv", "validation": "validation.csv", "test": "test.csv", } dataset = load_dataset("csv", data_files=data_files) ``` ## Curation Notes IBM technical documentation has traditionally been structured around DITA (Darwin Information Typing Architecture), which classifies documents into four types: `task`, `concept`, `reference`, and `troubleshooting`. This dataset adapts that taxonomy into two classes: `conceptual` merges `concept` and `reference` (both primarily information-seeking); `how-to` merges `task` and `troubleshooting` (both action- or fix-oriented). The binary schema was chosen because `troubleshooting` was too rare to form a reliable standalone class, and `reference` and `concept` were difficult to separate consistently. Annotation followed a semi-automatic process. Labelling rules were first defined based on IBM Writing Style guidelines, then applied by a heuristic script to generate candidate labels. Each candidate was assigned a confidence tier: `title_high` (strong title signal), `body_medium` (body-text signal only, no strong title match), or `default_low` (no strong signal in either title or body). All tiers except `body_medium` how-to rows were manually reviewed. The `body_medium` how-to subset (333 rows) was left unreviewed because the remaining manually checked data was sufficient to construct a balanced 400-example dataset; retaining unreviewed borderline rows would have introduced noise without benefit. Rows marked `X` during manual review were removed because the source document was incomplete or too ambiguous to label reliably. Rows marked `?` were interpreted as belonging to the opposite binary class. The final subset contains 400 examples, sampled with random seed `42` after manual correction and filtering. This dataset is derived from [ibm-research/watsonxDocsQA](https://huggingface.co/datasets/ibm-research/watsonxDocsQA), which is licensed under Apache 2.0. This dataset inherits the same license.