Savianto commited on
Commit
fdfdfde
·
verified ·
1 Parent(s): aad9631

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +125 -0
README.md ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+
3
+ {}
4
+ ---
5
+
6
+ This model card provides details on a fine-tuned version of Savianto/qlora-mistral, a language model trained using the QLoRA technique on conversational data for enhanced text generation, particularly in question-answering and conversational tasks.
7
+ ## Model Details
8
+
9
+ ### Model Description
10
+
11
+ Model Description
12
+ This is a fine-tuned version of the Savianto/qlora-mistral model using the QLoRA technique. The fine-tuning was done to improve the model’s ability to generate coherent and context-aware responses in conversational and question-answering tasks. QLoRA allows for efficient fine-tuning of large models while optimizing for memory usage.
13
+
14
+ Developed by: Yash Sawant
15
+ Model type: Causal Language Model (AutoModelForCausalLM)
16
+ Language(s) (NLP): English
17
+ License: [Specify License Type Here]
18
+ Finetuned from model: Savianto/qlora-mistral
19
+
20
+
21
+
22
+ - **Developed by:** Yash Sawant
23
+ - **Model type:** Causal Language Model (AutoModelForCausalLM)
24
+ - **Language(s) (NLP):** English
25
+ - **Finetuned from model [optional]:** teknium/OpenHermes-2-Mistral-7B
26
+
27
+
28
+ ## Uses
29
+
30
+ This model can be directly used for:
31
+
32
+ Question answering
33
+ Conversational agents (chatbots)
34
+ Text generation tasks (summarization, text completion)
35
+
36
+ ### Direct Use
37
+
38
+ This model can be fine-tuned further for specific tasks such as:
39
+
40
+ Domain-specific question answering
41
+ Custom chatbot agents
42
+ Document summarization
43
+
44
+
45
+
46
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
47
+
48
+ ## How to Get Started with the Model
49
+
50
+ from transformers import AutoModelForCausalLM, AutoTokenizer
51
+
52
+ # Load the fine-tuned model and tokenizer
53
+ model = AutoModelForCausalLM.from_pretrained("Savianto/qlora-mistral-finetuned")
54
+ tokenizer = AutoTokenizer.from_pretrained("Savianto/qlora-mistral-finetuned")
55
+
56
+ # Example prompt
57
+ prompt = "What is the capital of France?"
58
+
59
+ # Tokenize and generate output
60
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids
61
+ output = model.generate(input_ids, max_length=50)
62
+
63
+ # Decode the response
64
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
65
+ print(response)
66
+
67
+ [More Information Needed]
68
+
69
+ ## Training Details
70
+ Training Details
71
+ Training Data
72
+ The model was fine-tuned using a conversational dataset, focusing on question-answer pairs and dialogue examples. This enhances the model's ability to generate contextually relevant and coherent responses.
73
+
74
+ Training Procedure
75
+ Hardware: GPU (NVIDIA A100, 40GB)
76
+ Training Time: 5 epochs with early stopping
77
+ Optimizer: AdamW
78
+ Learning Rate: 2e-5
79
+ Batch Size: 16
80
+ Training regime: Mixed Precision (fp16)
81
+ Preprocessing
82
+ Tokenized the input text with padding and truncation for consistent input lengths.
83
+ Speeds, Sizes, Times
84
+ Training Time: ~3 hours
85
+ Model Size: ~7B parameters (Base Model: Savianto/qlora-mistral)
86
+ Evaluation
87
+ Testing Data
88
+ The model was evaluated on a validation split of the fine-tuning dataset, with question-answer pairs and conversational exchanges.
89
+
90
+ Metrics
91
+ Perplexity: Evaluated using standard perplexity for text generation models.
92
+ Coherence: Human-evaluated coherence in generated responses.
93
+ Results
94
+ The model exhibited low perplexity scores on the validation set and performed well in conversational coherence during testing.
95
+
96
+ Summary
97
+ The model is well-suited for question-answering tasks, conversational agents, and general text generation tasks but may require additional tuning for domain-specific applications.
98
+
99
+ Model Examination
100
+ No further interpretability analysis was conducted on this model.
101
+
102
+ Environmental Impact
103
+ Carbon emissions for this model can be estimated using the Machine Learning Impact calculator based on the following parameters:
104
+
105
+ Hardware Type: NVIDIA A100
106
+ Training Hours: ~3 hours
107
+ Cloud Provider: Google Cloud
108
+ Compute Region: US-Central
109
+ Carbon Emitted: 0.98 kg CO2eq (estimated)
110
+ Technical Specifications
111
+ Model Architecture and Objective
112
+ This model is based on Mistral architecture, with the objective of generating coherent and contextually aware responses in conversation and question-answering tasks.
113
+
114
+ Compute Infrastructure
115
+ Hardware
116
+ NVIDIA A100 40GB GPU
117
+ Software
118
+ Python 3.8
119
+ Transformers (Hugging Face) v4.x
120
+ PyTorch 1.10+
121
+ Accelerate
122
+
123
+
124
+
125
+