hz3519
commited on
Commit
•
09b8939
1
Parent(s):
d0b3b6e
Add models
Browse files
.ipynb_checkpoints/README-checkpoint.md
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- "en"
|
4 |
+
thumbnail: "https://example.com/path/to/your/thumbnail.jpg" # URL to a thumbnail used in social sharing
|
5 |
+
tags:
|
6 |
+
- "tag1" # For example, "sentiment-analysis"
|
7 |
+
- "tag2" # For example, "machine-translation"
|
8 |
+
license: "mit"
|
9 |
+
datasets:
|
10 |
+
- "dataset1" # For example, "imdb"
|
11 |
+
- "dataset2" # For example, "wmt16"
|
12 |
+
metrics:
|
13 |
+
- "metric1" # For example, "accuracy"
|
14 |
+
- "metric2" # For example, "f1"
|
15 |
+
---
|
16 |
+
|
17 |
+
# Your Model Name
|
18 |
+
|
19 |
+
## Introduction
|
20 |
+
|
21 |
+
This is a brief introduction about your transformer-based model. Here, you can mention the type of the model, the task it was trained for, its performance, and other key features or highlights.
|
22 |
+
|
23 |
+
## Training
|
24 |
+
|
25 |
+
Here, give detailed information about how the model was trained:
|
26 |
+
|
27 |
+
- Dataset(s) used for training
|
28 |
+
- Preprocessing techniques used
|
29 |
+
- Training configuration such as the batch size, learning rate, optimizer, number of epochs, etc.
|
30 |
+
- Any specific challenges or notable aspects of the training process
|
31 |
+
|
32 |
+
## Usage
|
33 |
+
|
34 |
+
Provide examples of how to use the model for inference. You can provide both a simple usage case and a more complex one if necessary. Make sure to explain what the inputs and outputs are.
|
35 |
+
|
36 |
+
Here's a basic example:
|
37 |
+
|
38 |
+
from transformers import AutoTokenizer, AutoModel
|
39 |
+
|
40 |
+
tokenizer = AutoTokenizer.from_pretrained("your-model-name")
|
41 |
+
model = AutoModel.from_pretrained("your-model-name")
|
42 |
+
|
43 |
+
inputs = tokenizer("Your example sentence", return_tensors="pt")
|
44 |
+
outputs = model(**inputs)
|
45 |
+
|
46 |
+
# Explain what the outputs are
|
47 |
+
|
48 |
+
## Evaluation
|
49 |
+
|
50 |
+
Discuss how the model was evaluated, which metrics were used, and what results it achieved.
|
51 |
+
|
52 |
+
## Limitations and Bias
|
53 |
+
|
54 |
+
Every model has its limitations and may have certain biases due to the data it was trained on. Explain those here.
|
55 |
+
|
56 |
+
## About Us
|
57 |
+
|
58 |
+
A small introduction about you or your team.
|
59 |
+
|
60 |
+
## Acknowledgments
|
61 |
+
|
62 |
+
Thank people, organizations or mention the resources that helped you in this work.
|
63 |
+
|
64 |
+
## License
|
65 |
+
|
66 |
+
This model is distributed under the MIT license.
|
67 |
+
|
68 |
+
## Contact
|
69 |
+
|
70 |
+
Provide a contact method (e.g., email or GitHub issues) for people to reach out with questions, comments, or concerns.
|
71 |
+
|
72 |
+
## References
|
73 |
+
|
74 |
+
List any relevant references for your model here.
|
.ipynb_checkpoints/model_evaluate-checkpoint.ipynb
ADDED
The diff for this file is too large to render.
See raw diff
|
|
model_M/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
|
model_M/model_weights.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:197a52e0a47711602b2ae60f2f9ee34b15fdc20b195aad66ab8df0bda9aa3256
|
3 |
+
size 176577875
|
model_M_retrain/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 24, "query_size": 512, "key_size": 512, "value_size": 512, "num_hiddens": 512, "num_layers": 6, "dropout": 0.2, "lr": 0.0004, "training_steps": 300000, "batch_size": 4096, "label_smoothing": 0.1, "ffn_num_input": 512, "ffn_num_hiddens": 2048, "num_heads": 8, "norm_shape": [512], "device": "cpu"}
|
model_M_retrain/model_weights.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:596171bfa552fd963cf7e3f1bd8b6c6f35f7ea14d129d9f7b451d0ecb4014ddc
|
3 |
+
size 176577875
|