Model Card for telecomadm1145/Qwen3-8B-Novel-AdaLoRA

This is an AdaLoRA adapter for the Qwen/Qwen3-8B model, fine-tuned for creative writing tasks (such as novels, stories, and poetry) in both Chinese and English.

Model Details

Model Description

This model is a Parameter-Efficient Fine-Tuning (PEFT) version of Qwen/Qwen3-8B, utilizing the AdaLoRA (Adaptive LoRA) method. It was trained on the telecomadm1145/creative_writing dataset, which contains a rich collection of creative texts. As a result, the model excels at generating imaginative, coherent, and stylistically rich narratives. It is designed to assist users with tasks like novel writing, story continuation, and poetry generation in both English and Chinese.

  • Developed by: telecomadm1145
  • Funded by: None.
  • Model type: Causal Language Model with AdaLoRA adapter
  • Language(s) (NLP): English, Chinese
  • License: The base model Qwen/Qwen3-8B uses the Tongyi Qianwen LICENSE AGREEMENT. This adapter is likely subject to the same license terms. Please refer to the base model's license for details.
  • Finetuned from model: Qwen/Qwen3-8B

Uses

This model can be used directly for text generation, especially in creative contexts. Potential uses include:

  • Generating story openings, plot developments, or endings.
  • Composing poems, lyrics, or short stories.
  • Serving as a source of inspiration to overcome writer's block.
  • Role-playing and dialogue generation.

Downstream Use

This model is suitable as a foundation for downstream applications, such as:

  • Creative Writing Assistant Apps: Integrate into writing software to provide real-time suggestions and content generation.
  • Game Development: Generate in-game character dialogue, quest descriptions, and lore.
  • Interactive Narratives: Build text-based adventure games or interactive fiction.
  • Content Marketing: Automatically generate creative ad copy or social media posts.

Out-of-Scope Use

This model should not be used for:

  • Generating factually accurate content (e.g., news articles, scientific papers, medical or legal advice).
  • Any malicious purposes, including generating hate speech, discriminatory content, or disinformation.
  • Critical decision-making systems without rigorous evaluation and safety measures.
  • Tasks requiring high-fidelity logical reasoning, such as complex programming or mathematical problem-solving.

Bias, Risks, and Limitations

  • Inherited Bias: The model was trained on the telecomadm1145/creative_writing dataset and may learn and amplify biases present in the data, such as stereotypes in character portrayal or cultural depictions.
  • Factual Unreliability: The generated content is fictional and should never be treated as factual information.
  • Risk of Harmful Content: Although the base model has undergone safety alignment, it may still be possible to prompt the model to generate inappropriate or offensive content.
  • Limited Stylistic Range: The model's writing style is heavily influenced by its training data and may not cover all genres or styles of creative writing.

Recommendations

Users (both direct and downstream) should be made aware of the risks, biases, and limitations of the model. It is recommended to implement content filtering and safety mechanisms when deploying this model in user-facing applications and to clearly disclose that the content is AI-generated.

How to Get Started with the Model

Use the code below to get started with the model.

import torch
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer

# Check for available GPU
device = "cuda" if torch.cuda.is_available() else "cpu"

# Define model IDs
base_model_id = "Qwen/Qwen3-8B"
adapter_id = "telecomadm1145/Qwen3-8B-Novel-AdaLoRA"

# Load tokenizer
tokenizer = AutoTokenizer.from_pretrained(base_model_id)

# Load base model
model = AutoModelForCausalLM.from_pretrained(
    base_model_id,
    torch_dtype=torch.bfloat16, # Use torch.float16 if your GPU doesn't support bfloat16
    device_map="auto",
)

# Load and merge the AdaLoRA adapter
print(f"Loading adapter from {adapter_id}...")
model = PeftModel.from_pretrained(model, adapter_id)
# For maximum inference speed, you can merge the adapter into the base model
# model = model.merge_and_unload() 

# Set up a streamer for continuous output
streamer = TextStreamer(tokenizer, skip_prompt=True)

# --- Example 1: English Creative Writing ---
prompt_en = "In a city powered by starlight, a lone detective uncovers a conspiracy that reaches the highest echelons of the celestial government. The first clue was a single, dark feather..."
messages_en = [
    {"role": "system", "content": "You are a creative and imaginative storyteller."},
    {"role": "user", "content": prompt_en}
]
text_en = tokenizer.apply_chat_template(
    messages_en,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs_en = tokenizer([text_en], return_tensors="pt").to(device)

print("\n--- English Story ---")
_ = model.generate(
    **model_inputs_en,
    max_new_tokens=256,
    streamer=streamer,
)

# --- Example 2: Chinese Creative Writing ---
prompt_zh = "写一段故事的开头,背景设定在一个古代赛博朋克世界,主角是一位使用机械义肢的侠客。"
messages_zh = [
    {"role": "system", "content": "你是一位富有想象力的创意作家。"},
    {"role": "user", "content": prompt_zh}
]
text_zh = tokenizer.apply_chat_template(
    messages_zh,
    tokenize=False,
    add_generation_prompt=True
)
model_inputs_zh = tokenizer([text_zh], return_tensors="pt").to(device)

print("\n\n--- Chinese Story ---")
_ = model.generate(
    **model_inputs_zh,
    max_new_tokens=256,
    streamer=streamer,
)

Training Details

Training Data

This model was fine-tuned on the telecomadm1145/creative_writing dataset. The dataset consists of diverse creative texts, such as novels and stories, designed to enhance the model's narrative and imaginative capabilities. The raw data was processed using a GPT-based model to convert it into high-quality instruction-response pairs for supervised fine-tuning.

Training Procedure

The model was trained using the AdaLoRA method from the Hugging Face peft library. AdaLoRA is an adaptive version of LoRA that dynamically allocates ranks to weight matrices based on their importance, enabling efficient fine-tuning with fewer parameters.

Preprocessing

The training data was formatted into instruction-response pairs. During training, the loss was calculated only on the response (completion) part of the text, ignoring the instruction. This technique encourages the model to focus on learning to generate the desired output.

Training Hyperparameters

  • Training regime: bf16 mixed precision
  • PEFT Method: AdaLoRA
  • Epochs: 3
  • Learning rate: 1e-4
  • LR Scheduler: linear

Speeds, Sizes, Times

  • The model was trained for approximately 6 hours on a single NVIDIA A100 40GB GPU.

Evaluation

No formal evaluation or testing has been conducted on this model.

Testing Data, Factors & Metrics

Testing Data

No formal testing data was used. The model has not been evaluated on any standard benchmarks.

Results

No quantitative results are available as the model has not been formally evaluated. Its performance is best assessed qualitatively by generating text for creative writing tasks.

Summary

No formal evaluation summary is available.

Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

  • Hardware Type: 1 x NVIDIA A100 40GB
  • Hours used: ~6 hours
  • Cloud Provider: Google Colab
  • Compute Region: N/A
  • Carbon Emitted: 0.84 kg CO2 eq.

Technical Specifications

Hardware

  • NVIDIA A100 40GB

Software

  • PyTorch
  • Transformers
  • PEFT

Framework versions

  • PEFT 0.17.1
Downloads last month
3
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for telecomadm1145/Qwen3-8B-Novel-AdaLoRA

Finetuned
Qwen/Qwen3-8B
Adapter
(1072)
this model

Dataset used to train telecomadm1145/Qwen3-8B-Novel-AdaLoRA

Paper for telecomadm1145/Qwen3-8B-Novel-AdaLoRA