qlora-mistral / README.md
Savianto's picture
Create README.md
fdfdfde verified
---
{}
---
This model card provides details on a fine-tuned version of Savianto/qlora-mistral, a language model trained using the QLoRA technique on conversational data for enhanced text generation, particularly in question-answering and conversational tasks.
## Model Details
### Model Description
Model Description
This is a fine-tuned version of the Savianto/qlora-mistral model using the QLoRA technique. The fine-tuning was done to improve the model’s ability to generate coherent and context-aware responses in conversational and question-answering tasks. QLoRA allows for efficient fine-tuning of large models while optimizing for memory usage.
Developed by: Yash Sawant
Model type: Causal Language Model (AutoModelForCausalLM)
Language(s) (NLP): English
License: [Specify License Type Here]
Finetuned from model: Savianto/qlora-mistral
- **Developed by:** Yash Sawant
- **Model type:** Causal Language Model (AutoModelForCausalLM)
- **Language(s) (NLP):** English
- **Finetuned from model [optional]:** teknium/OpenHermes-2-Mistral-7B
## Uses
This model can be directly used for:
Question answering
Conversational agents (chatbots)
Text generation tasks (summarization, text completion)
### Direct Use
This model can be fine-tuned further for specific tasks such as:
Domain-specific question answering
Custom chatbot agents
Document summarization
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the fine-tuned model and tokenizer
model = AutoModelForCausalLM.from_pretrained("Savianto/qlora-mistral-finetuned")
tokenizer = AutoTokenizer.from_pretrained("Savianto/qlora-mistral-finetuned")
# Example prompt
prompt = "What is the capital of France?"
# Tokenize and generate output
input_ids = tokenizer(prompt, return_tensors="pt").input_ids
output = model.generate(input_ids, max_length=50)
# Decode the response
response = tokenizer.decode(output[0], skip_special_tokens=True)
print(response)
[More Information Needed]
## Training Details
Training Details
Training Data
The model was fine-tuned using a conversational dataset, focusing on question-answer pairs and dialogue examples. This enhances the model's ability to generate contextually relevant and coherent responses.
Training Procedure
Hardware: GPU (NVIDIA A100, 40GB)
Training Time: 5 epochs with early stopping
Optimizer: AdamW
Learning Rate: 2e-5
Batch Size: 16
Training regime: Mixed Precision (fp16)
Preprocessing
Tokenized the input text with padding and truncation for consistent input lengths.
Speeds, Sizes, Times
Training Time: ~3 hours
Model Size: ~7B parameters (Base Model: Savianto/qlora-mistral)
Evaluation
Testing Data
The model was evaluated on a validation split of the fine-tuning dataset, with question-answer pairs and conversational exchanges.
Metrics
Perplexity: Evaluated using standard perplexity for text generation models.
Coherence: Human-evaluated coherence in generated responses.
Results
The model exhibited low perplexity scores on the validation set and performed well in conversational coherence during testing.
Summary
The model is well-suited for question-answering tasks, conversational agents, and general text generation tasks but may require additional tuning for domain-specific applications.
Model Examination
No further interpretability analysis was conducted on this model.
Environmental Impact
Carbon emissions for this model can be estimated using the Machine Learning Impact calculator based on the following parameters:
Hardware Type: NVIDIA A100
Training Hours: ~3 hours
Cloud Provider: Google Cloud
Compute Region: US-Central
Carbon Emitted: 0.98 kg CO2eq (estimated)
Technical Specifications
Model Architecture and Objective
This model is based on Mistral architecture, with the objective of generating coherent and contextually aware responses in conversation and question-answering tasks.
Compute Infrastructure
Hardware
NVIDIA A100 40GB GPU
Software
Python 3.8
Transformers (Hugging Face) v4.x
PyTorch 1.10+
Accelerate