Datasets:
File size: 14,929 Bytes
ab7f078 f435f7a ab7f078 f435f7a ab7f078 f435f7a 1e860a9 ab7f078 f435f7a ab7f078 a94ee2b ab7f078 f435f7a a94ee2b ab7f078 a94ee2b ab7f078 a94ee2b f435f7a a94ee2b ab7f078 f435f7a ab7f078 f435f7a ab7f078 1e860a9 f435f7a ab7f078 1e860a9 ab7f078 f435f7a ab7f078 f435f7a ab7f078 f435f7a ab7f078 1e860a9 ab7f078 f435f7a ab7f078 f435f7a a94ee2b f435f7a a94ee2b f435f7a a94ee2b f435f7a a94ee2b f435f7a ab7f078 a94ee2b ab7f078 a94ee2b f435f7a a94ee2b f435f7a 1e860a9 f435f7a a94ee2b ab7f078 a94ee2b ab7f078 a94ee2b ab7f078 a94ee2b ab7f078 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 | ---
language:
- en
license: cc-by-4.0
size_categories:
- 1K<n<10K
task_categories:
- text-classification
- feature-extraction
tags:
- patent
- prior-art
- freedom-to-operate
- fto
- office-action
- uspto
- epo
- multi-jurisdiction
- intellectual-property
- legal-nlp
pretty_name: "Layer A — Office Action Triples for FTO Eval"
configs:
- config_name: default
data_files:
- split: train
path: "**/jurisdiction=*/filing_year=*/*.parquet"
- config_name: us
data_files:
- split: train
path: "US/jurisdiction=US/filing_year=*/*.parquet"
- config_name: ep
data_files:
- split: train
path: "EP/jurisdiction=EP/filing_year=*/*.parquet"
# v1.2 — sibling partition aggregating per-(ref_id, citing_jurisdiction)
# citation counts. See "Cross-citation index" below.
- config_name: prior_art_index
data_files:
- split: train
path: "prior_art_index/*/index.parquet"
---
# Layer A — Office Action Triples for FTO Evaluation
A public dataset of `(invention → cited prior art → outcome)` triples
extracted from the **USPTO Office Action Research Dataset (OARD)**
+ **USPTO Open Data Portal (ODP)** API (US slice) and the **EPO
Open Patent Services (OPS)** Register service (EP slice). Built as
the agent-evaluation substrate for
[Parallax](https://parallax.3mergen.com), an AI-native Freedom-to-
Operate (FTO) and defensive-publication platform for individual
inventors and small teams.
> **Curated by [Vox](https://vox.delivery)** (org: `v13s`).
> Parallax is a Vox product; the curation layer (annotations,
> severity tagging, schema, manifest) is © Vox 2026 under
> CC-BY-4.0. The underlying USPTO patent data is public domain.
## TL;DR
- **5 011 rows** (v1.1.20260513): 5 000 US Office Actions (filing years
2011–2017) + 11 EP search reports (filing years 2014–2018)
- **14 Parquet shards**, partitioned by `jurisdiction × filing_year`
- Schema: `(case_id, invention, examination, prior_art[], outcome,
provenance)` — see [Schema](#schema) below
- Two jurisdictions live (US, EP); JP gated on INPIT bulk
credential issuance
- License: **CC-BY-4.0** on the curation; underlying patent
documents remain in the public domain
- SHA-256 manifest at `MANIFEST.json` for byte-level
reproducibility
## Quick start
```python
from datasets import load_dataset
# Default config returns both jurisdictions (5 011 rows).
ds = load_dataset("v13s/golden-fto-layer-a", split="train")
print(len(ds)) # 5011
# Per-jurisdiction configs are also available.
ds_us = load_dataset("v13s/golden-fto-layer-a", "us", split="train")
ds_ep = load_dataset("v13s/golden-fto-layer-a", "ep", split="train")
print(len(ds_us), len(ds_ep)) # 5000, 11
row = ds[0]
print(row["case_id"]) # e.g. "US-13004847-0"
print(row["invention"]["title"]) # "SYSTEM AND METHOD FOR ..."
print(row["examination"]["oa_type"]) # "rejection" | "search_report"
print(row["examination"]["rejection_reasons"]) # ["obviousness_103"]
for ref in row["prior_art"]:
print(ref["ref_id"], ref["severity"])
# US: "US9123456B2", "obviousness"
# EP: "US6825941", "novelty_destroying"
```
## Schema
Each row is a single Office Action event linked to its prior-art
citations. The full schema lives at
[`data-pipeline/src/layer_a/schema.py`](https://github.com/masterleopold/parallax/blob/main/data-pipeline/src/layer_a/schema.py)
in the source repo.
| Field | Type | Description |
|---|---|---|
| `case_id` | string | Stable id: `<jurisdiction>-<application_number>-<oa_seq>` |
| `schema_version` | string | Per-row schema version (`1.0` legacy / `1.2` post-2026-05-07) |
| `jurisdiction` | string | `US` or `EP`; `JP` in future versions |
| `source_dataset` | string | `uspto_oard` (US rows) or `epo_ops` (EP rows) |
| `extracted_at` | timestamp[s, UTC] | When this row was emitted |
| `invention` | struct | Application metadata — title, abstract, IPC/CPC codes, claims, applicant |
| `examination` | struct | OA event — `oa_date`, `oa_type` ∈ {rejection, allowance, search_report}, `rejection_reasons[]`, `examiner_id` |
| `prior_art` | list<struct> | Cited references — `ref_id`, `ref_type`, `source` (examiner/applicant), `rejection_basis`, `claims_blocked[]`, `severity`, `categories[]` (v1.2+), `metadata` |
| `outcome` | struct | Final disposition — `final_disposition`, `disposition_date`, `granted_claims[]`, `amendments_made`, `decision_source` (v1.1+) |
| `provenance` | struct | Audit trail — `parser_version`, `source_file`, `manifest_sha256`, `validation_status`, `validation_notes[]` |
### Severity enum (`prior_art[].severity`)
A 3-value severity enum that downstream consumers can join across
jurisdictions. Each jurisdiction has its own source signal:
| Severity | US (OARD signal) | EP (WIPO ST.14 search-report category) |
|---|---|---|
| `novelty_destroying` | `rejection_102=1` AND `citation_in_oa=1` | `X` or `E` (incl. multi-char `XY`, `XYI`) |
| `obviousness` | `rejection_103=1` AND `citation_in_oa=1` | `Y` |
| `background` | otherwise (PTO-892, PTO-1449 IDS) | `A`, `P`, `D`, `T`, `L`, `O`, `I` |
The EP search-report category sometimes concatenates multiple
codes (e.g. `"XY"` means the citation is BOTH novelty-relevant AND
obviousness-relevant). The lowering preserves the raw string and
the extractor maps the most-severe component to `severity`.
### `prior_art[].categories` (v1.2+)
The single-string `severity` collapses multi-character ST.14 codes
to one band (e.g. `XY` → `novelty_destroying`, dropping the
inventive-step signal). To preserve the full set, v1.2 adds a
`categories: list<string>` field with each code as its own
alphabetically-sorted entry:
| Source category | `severity` | `categories` |
|---|---|---|
| `X` | `novelty_destroying` | `["X"]` |
| `Y` | `obviousness` | `["Y"]` |
| `XY` | `novelty_destroying` | `["X", "Y"]` |
| `XYI` | `novelty_destroying` | `["I", "X", "Y"]` |
| `A` | `background` | `["A"]` |
Legacy v1.0 / v1.1 rows have `categories = []` (empty).
Jurisdictions whose source data doesn't expose ST.14 codes
(US OARD uses 35 USC § sections, not ST.14) also leave the field
empty. Filter for `len(categories) > 0` to query only ST.14-
exposed rows.
### Cross-citation index (v1.2+)
A sibling partition `prior_art_index/<jurisdiction>/index.parquet`
aggregates the cases partition by `(ref_id, citing_jurisdiction)`
so consumers can ask "how often has document X been cited" without
walking the cases data row-by-row.
```python
from datasets import load_dataset
idx = load_dataset(
"v13s/golden-fto-layer-a", "prior_art_index", split="train",
)
# Top-cited refs in EP search reports
top = sorted(
[r for r in idx if r["citing_jurisdiction"] == "EP"],
key=lambda r: r["citation_count"], reverse=True,
)[:10]
for r in top:
print(r["ref_id"], r["citation_count"], r["citing_case_ids"])
```
Index schema:
| Field | Type | Description |
|---|---|---|
| `ref_id` | string | Cited document id (e.g. `US10721059`) |
| `citing_jurisdiction` | string | Where the citing examiner sits (`EP`, `US`) |
| `citation_count` | int32 | Total times this ref appears in `prior_art[]` across cases |
| `citing_case_ids` | list<string> | Sorted set of `case_id` values that cite this ref |
| `severity_distribution` | struct | Count by severity band (`novelty_destroying`, `obviousness`, `background`) |
| `first_cited_date` | date | Earliest `examination.oa_date` across citing cases |
| `last_cited_date` | date | Latest `examination.oa_date` across citing cases |
Per-jurisdiction subdirs (`prior_art_index/EP/`, `prior_art_index/US/`)
keep the index sharded by which extractor produced it. To get a
cross-jurisdiction view, union the partition or use the default
config above which includes both.
### Rejection reason codes
Canonical 3-letter codes consistent across jurisdictions:
| Code | USC § | Description |
|---|---|---|
| `anticipation_102` | 35 USC §102 | Lack of novelty (single-reference) |
| `obviousness_103` | 35 USC §103 | Obviousness (multi-reference combination) |
| `subject_matter_101` | 35 USC §101 | Patent-eligible subject matter (Alice/Mayo/Bilski) |
| `indefiniteness_112` | 35 USC §112 | Written description / definiteness |
| `double_patenting` | non-statutory | Same invention claimed twice |
Future EP/JP releases add their statute-equivalent codes
(`novelty_epc_54`, `inventive_step_epc_56`, `novelty_jp_29_1`,
etc.) without breaking the schema.
## How was this built?
### US slice (5 000 rows)
1. **OARD bulk download** (the 4M-row USPTO Office Action
Research Dataset, frozen at the 2017 release): manually
browser-downloaded from
[research.uspto.gov](https://www.uspto.gov/ip-policy/economic-research/research-datasets/office-action-research-dataset-patents),
mirrored to [v13s/oard-2017-mirror](https://huggingface.co/datasets/v13s/oard-2017-mirror)
for repeatable fetches
2. **office_actions.csv scan** for the first 5 000 unique
application IDs in chronological order
3. **citations.csv filter pass** to keep only those 5 000 apps'
citation rows (~50 MB filtered from a 4 M-row, 5 GB unfiltered
source)
4. **USPTO ODP API** enrichment per app (60 RPM rate limit; ~85
minutes wall-clock for the full pass)
5. **Triple construction** — the OARD's pre-classified
`rejection_*` boolean columns + the citation rows + the ODP
metadata combine into a `LayerATriple` per OA event
### EP slice (11 rows, new in v1.0.2)
1. **EP publications list** curated from Espacenet IPC searches
(`G06F`, `H04L`, `A61K`), filing years 2014–2018
2. **OPS published-data full-cycle** for biblio + claims
(epodoc/docdb format, kind-suffix fallback for older
publications)
3. **OPS Register service** (`/rest-services/register/publication/
epodoc/{pub}/biblio`) for search-report citations — these
carry the WIPO ST.14 category codes, mapped to `severity` via
the table above
4. **Two-endpoint merge per publication**: full-cycle gives the
bibliographic context; the Register service gives the
`prior_art[]` list. Filtered to `@cited-phase == "search"` to
keep the high-signal X/Y/A subset
5. **Triple construction** — same `LayerATriple` shape as the US
slice; `oa_type = "search_report"`, `outcome.final_disposition
= "pending"` (EP grants land in a separate legal-status
endpoint, planned for v1.1)
### Common steps (both slices)
6. **Validation**: every row passes a linking validator that
checks temporal sanity (cited prior art filed before the
invention), severity coherence (novelty-destroying citations
on a granted+unamended application would be an inconsistency),
and schema round-trip
7. **Parquet emit** partitioned by jurisdiction × filing_year,
with a SHA-256 manifest for byte-level reproducibility
8. **HuggingFace push** under [v13s/golden-fto-layer-a](https://huggingface.co/datasets/v13s/golden-fto-layer-a)
The full pipeline source lives in the public repo at
[parallax/data-pipeline](https://github.com/masterleopold/parallax/tree/main/data-pipeline).
The release runner is
[`bin/local-extract-v1.sh`](https://github.com/masterleopold/parallax/blob/main/bin/local-extract-v1.sh).
## Known limitations
- **Sample size**: 5 011 rows are a first cut. The full OARD has
4 M+ Office Actions; ramp-up to 50 K+ US rows in v1.1+ is
planned. The EP slice is intentionally small (11 publications)
to validate the OPS Register integration end-to-end before
scaling.
- **Sparse claim text**: The ODP search endpoint returns
bibliographic metadata (title, applicant, IPC) but not full
claim text. Some rows have `invention.claims = []` or
placeholder markers; full claim extraction needs a separate
ODP call (planned for v1.1).
- **JP not yet shipped**: JP slice gated on INPIT bulk
credential approval; see
[docs/07-partnerships/inpit-bulk-data-application.md](https://github.com/masterleopold/parallax/blob/main/docs/07-partnerships/inpit-bulk-data-application.md).
- **EP claim ranges**: The Register service embeds claim ranges
in the citation's bibliographic text annotation (`[Y] 5,12`).
v1.0.3+ extracts these into `prior_art[].claims_blocked`; legacy
v1.0.2 rows leave the list empty.
- **Mixed schema_version partition**: rows from v1.0 / v1.1 cron
cycles carry `schema_version="1.0"` and an empty `categories[]`,
while v1.2+ rows carry `schema_version="1.2"` and populated
`categories[]` (when the source supports ST.14). Filter on
`schema_version` if you need a single-version partition.
- **EP outcome field is conservative**: Without joining the OPS
legal-status endpoint, `outcome.final_disposition` defaults to
`pending` for EP rows. v1.1 will resolve to
`granted` / `rejected` / `withdrawn`.
- **US outcome field is conservative**: Without joining USPTO
PEDS (Patent Examination Data System), the same defaulting
applies on the US slice.
## Versioning
Semantic versioning per
[golden-dataset-plan.md](https://github.com/masterleopold/parallax/blob/main/docs/06-evaluation/golden-dataset-plan.md):
- **MAJOR** — schema-incompatible (field removed, type changed)
- **MINOR** — new fields, new jurisdictions, ≥10 % data growth
- **PATCH** — parser bugfix, individual case re-validation
The HuggingFace dataset repo's git history is the canonical
release ledger. To pin a specific version in your code:
```python
ds = load_dataset("v13s/golden-fto-layer-a", revision="v1.1.20260513")
```
## Citation
If you use this dataset in academic work, please cite:
```bibtex
@dataset{vox_layer_a_2026,
author = {Hara, Yoichiro and {Vox}},
title = {Layer A — Office Action Triples for
Freedom-to-Operate Evaluation},
year = 2026,
publisher = {Hugging Face},
version = {{1.1.20260513}},
url = {https://huggingface.co/datasets/v13s/golden-fto-layer-a},
note = {Curated under CC-BY-4.0; underlying patent
data in the public domain}
}
```
## License
- **Curation layer (this dataset)**: [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/)
— the schema, severity tagging, and triple construction are
© Vox 2026 and may be used / redistributed with attribution.
- **Underlying patent documents**: public domain (USPTO).
- **OARD source data**: public domain (USPTO Office of the Chief
Economist).
## Contact
- Curator: **Yoichiro Hara** (`taisei@vox.delivery`)
- Org: [Vox](https://huggingface.co/v13s) (HF: `v13s`)
- Source repo: <https://github.com/masterleopold/parallax>
- Issues: <https://github.com/masterleopold/parallax/issues>
- Product surface: <https://parallax.3mergen.com>
For takedown requests on specific patent applications, file an
issue or email the curator. Public-domain patent data is
included in good faith; the curation layer can be redacted on
request.
|