| --- |
| tags: |
| - text-generation |
| - gpt2 |
| license: gpl-3.0 |
| datasets: |
| - bookcorpus |
| metrics: |
| - perplexity |
| model-index: |
| - name: basic-text-generator |
| results: |
| - task: |
| type: text-generation |
| dataset: |
| name: bookcorpus |
| type: bookcorpus |
| metrics: |
| - name: Perplexity |
| type: perplexity |
| value: 25.3 |
| --- |
| |
| # Basic Text Generator |
|
|
| ## Overview |
|
|
| This is a fine-tuned GPT-2 model for general text generation. It can continue prompts, generate stories, or create coherent paragraphs based on input text. Trained on a diverse corpus for broad applicability. |
|
|
| ## Model Architecture |
|
|
| - Base Model: GPT-2 |
| - Layers: 12 |
| - Hidden Size: 768 |
| - Attention Heads: 12 |
| - Context Window: 1024 tokens |
|
|
| ## Intended Use |
|
|
| Suitable for creative writing, content generation, or prototyping language-based applications. |
|
|
| ## Limitations |
|
|
| - May generate biased or inappropriate content based on training data. |
| - Outputs can be repetitive or nonsensical for long generations. |
| - Not optimized for specific domains like code or math. |
|
|
| ## Example Code |
|
|
| ```python |
| from transformers import pipeline |
| |
| generator = pipeline("text-generation", model="user/basic-text-generator") |
| result = generator("Once upon a time,", max_length=50) |
| print(result[0]['generated_text']) |
| # "Once upon a time, in a land far away..." |