Dummy BERT Model

This is a test model created for experimental upload testing to Hugging Face using dmf-ng.

Model Details

Model Description

A minimal BERT model for testing artifact upload workflows with dmf-ng to Hugging Face Hub.

  • Developed by: dmf-ng Test Suite
  • Model type: Transformer-based language model
  • Library: Transformers
  • License: MIT

Model Architecture

  • Architecture: BERT (Bidirectional Encoder Representations from Transformers)
  • Hidden Size: 768
  • Number of Hidden Layers: 12
  • Number of Attention Heads: 12
  • Intermediate Size: 3,072
  • Maximum Position Embeddings: 512
  • Vocabulary Size: 30,522

Model Configuration

{
  "model_type": "bert",
  "hidden_size": 768,
  "num_hidden_layers": 12,
  "num_attention_heads": 12,
  "intermediate_size": 3072,
  "hidden_act": "gelu",
  "hidden_dropout_prob": 0.1,
  "attention_probs_dropout_prob": 0.1,
  "max_position_embeddings": 512,
  "type_vocab_size": 2,
  "initializer_range": 0.02,
  "layer_norm_eps": 1e-12,
  "pad_token_id": 0
}

Files

  • model.pt - PyTorch model weights (placeholder)
  • config.json - Model configuration in HuggingFace format
  • tokenizer.json - Tokenizer configuration
  • vocab.txt - Vocabulary file with token mappings
  • README.md - This model card

Intended Use

This model is for testing purposes only and should not be used for actual inference or production workloads.

Primary Intended Use

  • Testing artifact upload workflows with dmf-ng
  • Validating model card metadata
  • Experimenting with Hugging Face Hub integration
  • Testing lineage tracking with MLflow

Out-of-Scope Use Cases

  • Production inference
  • Real-world text classification tasks
  • Fine-tuning on real datasets
  • Deploying to inference endpoints

Technical Details

Model Inputs

  • input_ids: Token IDs (shape: [batch_size, sequence_length])
  • attention_mask: Binary mask for padding (shape: [batch_size, sequence_length])
  • token_type_ids: Segment IDs for sentence pairs (shape: [batch_size, sequence_length])

Model Outputs

  • Hidden states from the last transformer layer (shape: [batch_size, sequence_length, 768])
  • [CLS] token representation for sequence classification tasks

Limitations and Biases

This is a dummy model created for testing purposes and does not represent a real, trained model. It has not been trained on any data and produces random outputs.

Training Data

None - this model was generated as test data.

Evaluation Results

Not applicable - this is a test model.

Environmental Impact

Minimal environmental impact - this is a test model used only for software development and testing.

How to Get Started

from transformers import AutoTokenizer, AutoModelForMaskedLM

model_id = "your-username/test-model"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForMaskedLM.from_pretrained(model_id)

# This model is not trained, so outputs are random
inputs = tokenizer("Hello, world!", return_tensors="pt")
outputs = model(**inputs)

Model Card Contact

For issues related to this test model, please open an issue on the dmf-ng repository.


Note: This is a test artifact. For production models, ensure comprehensive model cards with real training data, evaluation metrics, and bias analysis.

Downloads last month
14
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support