FinBERT-FOMC Aspect Sentiment v5

This is a single aspect-aware sentiment model fine-tuned from ZiweiChen/FinBERT-FOMC on FOMC_sentences_expanded_zeroshot_labeled.xlsx.

Model Details

  • Base model: ZiweiChen/FinBERT-FOMC
  • Training mode: all
  • Train rows: 28377
  • Eval rows: 3123

Labels

  • Negative (0)
  • Neutral (1)
  • Positive (2)

Inference Contract

Use one model for all targets:

  • text = mode
  • text_pair = FOMC sentence

Valid mode values:

  • growth
  • employment
  • inflation

Python Usage

from transformers import BertForSequenceClassification, BertTokenizer, pipeline

model_id = "Wonseong/FinBERT-FOMC-aspects"
tokenizer = BertTokenizer.from_pretrained(model_id)
model = BertForSequenceClassification.from_pretrained(model_id)
clf = pipeline("text-classification", model=model, tokenizer=tokenizer)

sentence = "Spending on cars and light trucks increased somewhat in July."
mode = "growth"  # growth | employment | inflation

result = clf({"text": mode, "text_pair": sentence})
print(result)

Local Helper Usage

from predict_aspect import FinBERTFOMCModeClassifier

clf = FinBERTFOMCModeClassifier("Wonseong/FinBERT-FOMC-aspects")
print(clf.predict("Spending on cars and light trucks increased somewhat in July.", mode="growth"))

Metrics

Holdout evaluation:

  • Accuracy: 0.8789625360230547
  • Loss: 0.5731372237205505
  • Macro F1: 0.7980614874405157
  • Macro Precision: 0.7708534213383098
  • Macro Recall: 0.8321941498302832
  • MAE: 0.1399295549151457
  • MSE: 0.1777137367915466
  • MAPE: 8.09051126054008

Full training data fit:

  • Accuracy: 0.9348888888888889
  • Macro F1: 0.8977282349978499
  • Macro Precision: 0.8621019267617939
  • Macro Recall: 0.9424642336852936
  • MAE: 0.06847619047619048
  • MSE: 0.07520634920634921
  • MAPE: 3.6275132275132274

Training Data Contract

Expected spreadsheet columns in sheet expanded_10500:

  • sentence
  • growth_sentiment_final
  • employment_sentiment_final
  • inflation_sentiment_final

Label values are normalized from negative, neutral, and positive.

Dataset Provenance

The training file FOMC_sentences_expanded_zeroshot_labeled.xlsx contains sentence-level labels for the expanded_10500 sheet. These aspect labels were generated through zero-shot annotation with oss-20b using the prompt template stored in zeroshot_fomc_sentiment_prompt.txt.

The labeled dataset can be provided upon reasonable request.

The zero-shot prompt asks for one label per aspect:

  • growth_sentiment_final
  • employment_sentiment_final
  • inflation_sentiment_final

Each aspect is restricted to:

  • positive
  • neutral
  • negative

The prompt design is derived from the human-labeling framework described in:

Kim, W., Sporer, J., Lee, C. L., & Handschuh, S. (2024, November). Is small really beautiful for central Bank communication? Evaluating language models for finance: Llama-3-70B, GPT-4, FinBERT-FOMC, FinBERT, and VADER. In Proceedings of the 5th ACM International Conference on AI in Finance (pp. 626-633).

Intended Use

This model is intended for aspect-conditioned sentiment classification on FOMC-related sentences. The caller supplies the target aspect through mode and the sentence through text_pair.

Limitations

  • This is not a general finance sentiment model for arbitrary entities or tasks.
  • Reported metrics are from a holdout split of the labeled spreadsheet, not from external benchmark evaluation.
  • The model predicts one of three ordinal sentiment classes only: Negative, Neutral, Positive.

Training Summary

  • Runtime device: Tesla T4
  • CUDA available: true
  • BF16: true
  • FP16: false

Additional artifacts in this repo:

  • training_metadata.json
  • metric_history.csv
  • mae_mse_curve.png

Citation

@inproceedings{kim2024small,
  author = {Kim, W. and Sporer, J. and Lee, C. L. and Handschuh, S.},
  title = {Is small really beautiful for central Bank communication? Evaluating language models for finance: Llama-3-70B, GPT-4, FinBERT-FOMC, FinBERT, and VADER},
  booktitle = {Proceedings of the 5th ACM International Conference on AI in Finance},
  pages = {626--633},
  year = {2024}
}
Downloads last month
68
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Wonseong/FinBERT-FOMC-aspects

Finetuned
(2)
this model

Evaluation results

  • Holdout Accuracy on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.879
  • Holdout Macro F1 on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.798
  • Holdout Macro Precision on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.771
  • Holdout Macro Recall on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.832
  • Holdout MAE on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.140
  • Holdout MSE on FOMC_sentences_expanded_zeroshot_labeled.xlsx / expanded_10500
    self-reported
    0.178