Dataset Viewer
The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code:   ConfigNamesError
Exception:    FileNotFoundError
Message:      Couldn't find any data file at /src/services/worker/Sashank-810/crisisnet-dataset. Couldn't find 'Sashank-810/crisisnet-dataset' on the Hugging Face Hub either: LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response
                  config_names = get_dataset_config_names(
                                 ^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 161, in get_dataset_config_names
                  dataset_module = dataset_module_factory(
                                   ^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/load.py", line 1203, in dataset_module_factory
                  raise FileNotFoundError(
              FileNotFoundError: Couldn't find any data file at /src/services/worker/Sashank-810/crisisnet-dataset. Couldn't find 'Sashank-810/crisisnet-dataset' on the Hugging Face Hub either: LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

CrisisNet — Corporate Default Risk Dataset

"Every cancer screening programme works on one insight: the disease speaks before the patient feels it. CrisisNet applies the same logic to corporate finance."

CrisisNet is a multi-modal, network-aware dataset for building early-warning systems for corporate financial distress. It covers 40 U.S. Energy sector companies (S&P 500) over 10 years (2015–2025) across three parallel signal types: time series financials, NLP text from filings and earnings calls, and a supply-chain network graph.


Dataset Structure

The dataset is organised into four top-level folders mirroring the four analytical modules in the CrisisNet architecture:

Module_1/   ← Time Series & Credit Risk (Module A)
Module_2/   ← NLP Text: 10-K Filings + Earnings Calls (Module B)
Module_3/   ← Supply Chain Network Graph (Module C)
Labels/     ← Default & Distress Events (ground truth for all modules)
splits/     ← Pre-computed train / validation / test splits
data/       ← Master company list

Module_1 — Time Series & Credit Risk Engine

Purpose: Feeds the X_ts(c,t) feature vector — the financial heartbeat monitor.

Module_1/market_data/

File Description
all_prices.parquet Daily OHLCV stock prices for all 40 tickers, 2015–present (2,821 rows × 205 cols)
all_prices.csv Same as above in CSV format
financials/{TICKER}_income.csv Quarterly income statement per company
financials/{TICKER}_balance_sheet.csv Quarterly balance sheet per company
financials/{TICKER}_cashflow.csv Quarterly cash flow statement per company
financials/{TICKER}_info.csv Company metadata (sector, market cap, description)

Note: CHK, HES, MRO, PXD, SWN have only _info.csv — these companies were acquired or delisted (Pioneer→Exxon, Hess→Chevron, Marathon Oil→ConocoPhillips, Southwestern→Chesapeake). They are intentionally kept as distress/exit label cases.

Module_1/credit_spreads/

22 FRED series (2005–present) covering credit spreads, treasury yields, macro indicators, and energy prices:

Series Description
BAMLH0A0HYM2 ICE BofA US High Yield OAS — primary distress signal
BAA10Y, AAA10Y Moody's corporate bond spreads
VIXCLS CBOE VIX — market fear gauge
DCOILWTICO, DCOILBRENTEU WTI & Brent crude oil prices
DHHNGSP Henry Hub natural gas spot price
T10Y2Y 10Y-2Y Treasury spread (recession predictor)
DGS10, DGS2, DGS3MO Treasury yields
UNRATE, CPIAUCSL, FEDFUNDS, INDPRO Macro indicators
fred_all_series.parquet All 22 series combined (5,681 rows × 22 cols)

Module_1/sec_xbrl/

Structured XBRL financial data from SEC EDGAR for 35/40 companies:

  • company_facts/{TICKER}_facts.json — All XBRL-reported line items (Assets, Liabilities, Revenue, EPS, etc.) with full quarterly history
  • submissions/{TICKER}_submissions.json — Filing history (dates, form types, accession numbers)
  • ticker_cik_mapping.csv — Ticker ↔ SEC CIK number mapping

Module_2 — NLP: 10-K Filings & Earnings Calls

Purpose: Feeds the X_nlp(c,t) feature vector via LDA topic modelling and FinBERT sentiment.

Module_2/10k_extracted/10-K/

353 structured JSON files, one per company per year (2015–2024). Each file contains:

{
  "item_1":  "Business description — supply chain, customers, operations...",
  "item_1a": "Risk factors — debt levels, commodity exposure, going concern...",
  "item_7":  "Management Discussion & Analysis — earnings narrative...",
  "item_7a": "Market risk disclosures — interest rate, commodity hedging...",
  "item_8":  "Financial statements narrative..."
}

Naming: {CIK}_{FormType}_{Year}_{AccessionNumber}.json

Usage for NLP:

  • Run LDA (Gensim) or BERTopic on item_7 (MD&A) to extract latent distress topics
  • Apply FinBERT sentence-by-sentence to item_1a (Risk Factors) for sentiment time series
  • Track KL-divergence of topic distributions quarter-over-quarter as a leading signal

Module_2/transcripts/

Earnings call Q&A transcripts from HuggingFace (lamini/earnings-calls-qa):

  • 860,164 Q&A records from public company earnings calls
  • Fields: question, answer, ticker, date
  • Re-download: datasets.load_dataset("lamini/earnings-calls-qa")

Module_3 — Supply Chain Network Graph

Purpose: Feeds the X_graph(c,t) feature vector via community detection and contagion simulation.

Module_3/edges_template.csv

30 pre-populated directed edges representing known Energy sector supplier-customer relationships:

source, target, relationship_type, description
SLB, XOM, service_provider, oilfield services
HAL, CVX, service_provider, oilfield services
EPD, VLO, pipeline_supplier, NGL supply
...

Usage: Load into NetworkX → run Louvain community detection → compute DebtRank contagion scores.

Module_3/customer_disclosures_raw.csv

660 customer/supplier disclosure mentions extracted from 10-K Item 1 and Item 7 sections. Use these to augment the graph edges with NLP-extracted relationships.


Labels — Default & Distress Events

Purpose: Ground truth labels for all three modules.

Labels/energy_defaults_curated.csv

24 curated bankruptcy/default events (2001–2021):

company, ticker, event_date, event_type, details
Chesapeake Energy, CHK, 2020-06-28, Chapter 11, COVID + legacy debt...
Whiting Petroleum, WLL, 2020-04-01, Chapter 11, COVID oil crash
...

Labels/distress_from_drawdowns.csv

76 mechanically detected distress episodes from stock price drawdowns (>50% peak-to-trough within 6 months) — useful as soft labels for the ML model.

Labels/lopucki_brd_reference.json

Reference pointer to the Florida-UCLA LoPucki Bankruptcy Research Database (1,000+ cases, 1979–2022) for cross-referencing additional default events.


Train / Validation / Test Splits

Split strategy: temporal walk-forward (no lookahead leakage)

Split Period Rationale
train 2015–2021 Includes 2015–16 oil crash + 2020 COVID wave defaults
validation 2022 Post-COVID recovery, hyperparameter tuning
test 2023–2025 Held-out, never seen during training

Pre-split parquet files are in splits/:

splits/
  stock_prices/     train.parquet, validation.parquet, test.parquet
  fred_macro/       train.parquet, validation.parquet, test.parquet
  labels/
    energy_defaults/    train.parquet, validation.parquet, test.parquet
    distress_drawdowns/ train.parquet, validation.parquet, test.parquet
  10k_filings/      train_manifest.json, validation_manifest.json, test_manifest.json

Recommended Usage

Module A — Time Series Credit Risk Engine

import pandas as pd

# Load training data
prices_train = pd.read_parquet("splits/stock_prices/train.parquet")
fred_train   = pd.read_parquet("splits/fred_macro/train.parquet")
labels_train = pd.read_parquet("splits/labels/distress_drawdowns/train.parquet")

# Feature engineering: rolling volatility, Merton Distance-to-Default
# 30-day rolling log-return volatility per ticker
log_ret = prices_train.xs("Close", axis=1, level=0).pct_change().apply(lambda x: (1+x).apply(pd.np.log))
vol_30d = log_ret.rolling(30).std() * (252**0.5)

# Altman Z-Score (benchmark) — requires balance sheet data:
# Z = 1.2*X1 + 1.4*X2 + 3.3*X3 + 0.6*X4 + 1.0*X5
# where X1=Working Capital/TA, X2=Retained Earnings/TA, X3=EBIT/TA,
#       X4=Market Cap/Book Liabilities, X5=Revenue/TA

# Walk-forward cross-validation (never use future data in training window)
# Use expanding window: train on t-36m to t, predict t+1m to t+6m

Module B — NLP Topic Modelling

import json, os
from gensim import corpora, models  # LDA baseline
# from bertopic import BERTopic     # upgraded model

# Load 10-K MD&A sections for training period
train_manifest = json.load(open("splits/10k_filings/train_manifest.json"))
corpus = []
for fpath in train_manifest["files"]:
    filing = json.load(open(fpath))
    text = filing.get("item_7", "") + " " + filing.get("item_1a", "")
    corpus.append(text)

# Preprocessing: tokenise, remove Safe Harbor boilerplate, financial stopwords
# financial_stopwords = ["forward-looking", "may", "could", "believe", "expect", ...]

# Train LDA (K=15 topics typical for earnings text)
# Compare coherence scores across K=10,15,20,25 to find optimal

# FinBERT sentiment on risk factor sentences:
# from transformers import pipeline
# finbert = pipeline("text-classification", model="ProsusAI/finbert")

Module C — Supply Chain Network

import pandas as pd, networkx as nx
import community as community_louvain  # pip install python-louvain

# Build directed graph from known edges
edges = pd.read_csv("Module_3/edges_template.csv")
G = nx.from_pandas_edgelist(edges, "source", "target",
                             edge_attr="relationship_type",
                             create_using=nx.DiGraph())

# Add nodes from company list
companies = pd.read_csv("data/company_list.csv")
for _, row in companies.iterrows():
    G.nodes[row["ticker"]]["subsector"] = row["subsector"]

# Louvain community detection
partition = community_louvain.best_partition(G.to_undirected())

# Centrality metrics
betweenness = nx.betweenness_centrality(G)
pagerank    = nx.pagerank(G)

# DebtRank contagion: mark one node as defaulted, propagate stress
# proportional to edge weights through the graph

Module D — Fusion & Health Score

import lightgbm as lgb
from sklearn.calibration import CalibratedClassifierCV

# Concatenate feature vectors from all three modules:
# X = pd.concat([X_ts, X_nlp, X_graph], axis=1)  # (company × quarter) index

# Walk-forward split (expanding window)
# Train: 2015Q1 → 2020Q4 | Val: 2021Q1 → 2021Q4 | Test: 2022Q1 →

# LightGBM with early stopping on val AUC
model = lgb.LGBMClassifier(n_estimators=500, learning_rate=0.05,
                            num_leaves=31, min_child_samples=10)
model.fit(X_train, y_train,
          eval_set=[(X_val, y_val)],
          eval_metric="auc",
          callbacks=[lgb.early_stopping(50), lgb.log_evaluation(50)])

# Calibrate to produce true probabilities (Platt scaling)
calibrated = CalibratedClassifierCV(model, cv="prefit", method="sigmoid")
calibrated.fit(X_val, y_val)

# SHAP for interpretability
import shap
explainer   = shap.TreeExplainer(model)
shap_values = explainer.shap_values(X_test)
shap.summary_plot(shap_values, X_test)

Company Universe

40 S&P 500 Energy sector companies across 6 subsectors:

Subsector Tickers
Integrated Oil XOM, CVX, OXY
Exploration & Production COP, EOG, PXD*, DVN, FANG, MRO*, APA, OVV, HES*, CTRA, MTDR, PR, CHRD
Oilfield Services SLB, HAL, BKR, FTI, NOV
Refining VLO, MPC, PSX, DK, PBF
Midstream/Pipelines KMI, WMB, OKE, ET, EPD, TRGP, DTM, AM
Natural Gas / LNG EQT, AR, RRC, SWN*, CHK*, LNG

* Delisted/acquired/bankrupt — useful as distress/exit label cases


Research Questions (from project proposal)

  • RQ1 — Prediction: Can we predict corporate default events 3–6 months in advance using time series + NLP + network features with higher AUC-ROC than Altman Z-Score?
  • RQ2 — Contagion: Which companies act as 'super-spreaders' of financial distress — and can community detection identify them before a crisis?
  • RQ3 — Narrative Signal: Does sentiment and topic shift in earnings call language provide statistically significant leading signal for credit deterioration?

Citation

@dataset{crisisnet2025,
  title     = {CrisisNet: A Multi-Modal Corporate Default Risk Dataset},
  author    = {Sashank and team},
  year      = {2025},
  publisher = {HuggingFace},
  url       = {https://huggingface.co/datasets/Sashank-810/crisisnet-dataset}
}

License

CC BY 4.0 — Free to use for research and commercial purposes with attribution.

Data sourced from: Yahoo Finance (yfinance), FRED API (St. Louis Fed), SEC EDGAR (public domain), HuggingFace lamini/earnings-calls-qa.

Downloads last month
710