hz3519 commited on
Commit ·
09b8939
1
Parent(s): d0b3b6e
Add models
Browse files
.ipynb_checkpoints/README-checkpoint.md
ADDED
|
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
language:
|
| 3 |
+
- "en"
|
| 4 |
+
thumbnail: "https://example.com/path/to/your/thumbnail.jpg" # URL to a thumbnail used in social sharing
|
| 5 |
+
tags:
|
| 6 |
+
- "tag1" # For example, "sentiment-analysis"
|
| 7 |
+
- "tag2" # For example, "machine-translation"
|
| 8 |
+
license: "mit"
|
| 9 |
+
datasets:
|
| 10 |
+
- "dataset1" # For example, "imdb"
|
| 11 |
+
- "dataset2" # For example, "wmt16"
|
| 12 |
+
metrics:
|
| 13 |
+
- "metric1" # For example, "accuracy"
|
| 14 |
+
- "metric2" # For example, "f1"
|
| 15 |
+
---
|
| 16 |
+
|
| 17 |
+
# Your Model Name
|
| 18 |
+
|
| 19 |
+
## Introduction
|
| 20 |
+
|
| 21 |
+
This is a brief introduction about your transformer-based model. Here, you can mention the type of the model, the task it was trained for, its performance, and other key features or highlights.
|
| 22 |
+
|
| 23 |
+
## Training
|
| 24 |
+
|
| 25 |
+
Here, give detailed information about how the model was trained:
|
| 26 |
+
|
| 27 |
+
- Dataset(s) used for training
|
| 28 |
+
- Preprocessing techniques used
|
| 29 |
+
- Training configuration such as the batch size, learning rate, optimizer, number of epochs, etc.
|
| 30 |
+
- Any specific challenges or notable aspects of the training process
|
| 31 |
+
|
| 32 |
+
## Usage
|
| 33 |
+
|
| 34 |
+
Provide examples of how to use the model for inference. You can provide both a simple usage case and a more complex one if necessary. Make sure to explain what the inputs and outputs are.
|
| 35 |
+
|
| 36 |
+
Here's a basic example:
|
| 37 |
+
|
| 38 |
+
from transformers import AutoTokenizer, AutoModel
|
| 39 |
+
|
| 40 |
+
tokenizer = AutoTokenizer.from_pretrained("your-model-name")
|
| 41 |
+
model = AutoModel.from_pretrained("your-model-name")
|
| 42 |
+
|
| 43 |
+
inputs = tokenizer("Your example sentence", return_tensors="pt")
|
| 44 |
+
outputs = model(**inputs)
|
| 45 |
+
|
| 46 |
+
# Explain what the outputs are
|
| 47 |
+
|
| 48 |
+
## Evaluation
|
| 49 |
+
|
| 50 |
+
Discuss how the model was evaluated, which metrics were used, and what results it achieved.
|
| 51 |
+
|
| 52 |
+
## Limitations and Bias
|
| 53 |
+
|
| 54 |
+
Every model has its limitations and may have certain biases due to the data it was trained on. Explain those here.
|
| 55 |
+
|
| 56 |
+
## About Us
|
| 57 |
+
|
| 58 |
+
A small introduction about you or your team.
|
| 59 |
+
|
| 60 |
+
## Acknowledgments
|
| 61 |
+
|
| 62 |
+
Thank people, organizations or mention the resources that helped you in this work.
|
| 63 |
+
|
| 64 |
+
## License
|
| 65 |
+
|
| 66 |
+
This model is distributed under the MIT license.
|
| 67 |
+
|
| 68 |
+
## Contact
|
| 69 |
+
|
| 70 |
+
Provide a contact method (e.g., email or GitHub issues) for people to reach out with questions, comments, or concerns.
|
| 71 |
+
|
| 72 |
+
## References
|
| 73 |
+
|
| 74 |
+
List any relevant references for your model here.
|
.ipynb_checkpoints/model_evaluate-checkpoint.ipynb
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model_M/config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
|
model_M/model_weights.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:197a52e0a47711602b2ae60f2f9ee34b15fdc20b195aad66ab8df0bda9aa3256
|
| 3 |
+
size 176577875
|
model_M_retrain/config.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
|
model_M_retrain/model_weights.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:596171bfa552fd963cf7e3f1bd8b6c6f35f7ea14d129d9f7b451d0ecb4014ddc
|
| 3 |
+
size 176577875
|