Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,9 @@ language:
|
|
12 |
tags:
|
13 |
- qwen
|
14 |
inference: false
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
# `rinna/nekomata-7b`
|
@@ -44,7 +47,7 @@ The name `nekomata` comes from the Japanese word [`猫又/ねこまた/Nekomata`
|
|
44 |
- [Wikipedia](https://dumps.wikimedia.org/other/cirrussearch)
|
45 |
- rinna curated Japanese dataset
|
46 |
|
47 |
-
* **
|
48 |
|
49 |
- [Tianyu Zhao](https://huggingface.co/tianyuz)
|
50 |
- [Akio Kaga](https://huggingface.co/rakaga)
|
@@ -114,15 +117,24 @@ We compared the `Qwen` tokenizer (as used in `nekomata`) and the `llama-2` token
|
|
114 |
|
115 |
# How to cite
|
116 |
~~~
|
117 |
-
@misc{
|
118 |
-
|
119 |
-
title={rinna/nekomata-7b},
|
120 |
author={Zhao, Tianyu and Kaga, Akio and Sawada, Kei}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
121 |
}
|
122 |
~~~
|
123 |
---
|
124 |
|
125 |
-
#
|
126 |
~~~
|
127 |
@software{gpt-neox-library,
|
128 |
title = {{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
|
|
|
12 |
tags:
|
13 |
- qwen
|
14 |
inference: false
|
15 |
+
license: other
|
16 |
+
license_name: tongyi-qianwen-license-agreement
|
17 |
+
license_link: https://github.com/QwenLM/Qwen/blob/main/Tongyi%20Qianwen%20LICENSE%20AGREEMENT
|
18 |
---
|
19 |
|
20 |
# `rinna/nekomata-7b`
|
|
|
47 |
- [Wikipedia](https://dumps.wikimedia.org/other/cirrussearch)
|
48 |
- rinna curated Japanese dataset
|
49 |
|
50 |
+
* **Contributors**
|
51 |
|
52 |
- [Tianyu Zhao](https://huggingface.co/tianyuz)
|
53 |
- [Akio Kaga](https://huggingface.co/rakaga)
|
|
|
117 |
|
118 |
# How to cite
|
119 |
~~~
|
120 |
+
@misc{rinna-nekomata-7b,
|
121 |
+
title = {rinna/nekomata-7b},
|
|
|
122 |
author={Zhao, Tianyu and Kaga, Akio and Sawada, Kei}
|
123 |
+
url = {https://huggingface.co/rinna/nekomata-7b},
|
124 |
+
}
|
125 |
+
|
126 |
+
@inproceedings{sawada2024release,
|
127 |
+
title = {Release of Pre-Trained Models for the {J}apanese Language},
|
128 |
+
author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
|
129 |
+
booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
|
130 |
+
month = {5},
|
131 |
+
year = {2024},
|
132 |
+
url = {https://arxiv.org/abs/2404.01657},
|
133 |
}
|
134 |
~~~
|
135 |
---
|
136 |
|
137 |
+
# References
|
138 |
~~~
|
139 |
@software{gpt-neox-library,
|
140 |
title = {{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
|