update
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ tags:
|
|
10 |
license: gpl-3.0
|
11 |
---
|
12 |
|
13 |
-
# CKIP
|
14 |
|
15 |
This model provides word segmentation for the ancient Chinese language. Our training dataset covers four eras of the Chinese language.
|
16 |
|
@@ -36,14 +36,14 @@ The copyright of the datasets belongs to the Institute of Linguistics, Academia
|
|
36 |
AutoModel,
|
37 |
)
|
38 |
|
39 |
-
tokenizer = AutoTokenizer.from_pretrained("ckiplab/
|
40 |
-
model = AutoModel.from_pretrained("ckiplab/
|
41 |
```
|
42 |
|
43 |
* Using our model for inference
|
44 |
```python
|
45 |
>>> from transformers import pipeline
|
46 |
-
>>> classifier = pipeline("token-classification", model="ckiplab/
|
47 |
>>> classifier("帝堯曰放勳")
|
48 |
|
49 |
# output
|
|
|
10 |
license: gpl-3.0
|
11 |
---
|
12 |
|
13 |
+
# CKIP BERT Base Han Chinese WS
|
14 |
|
15 |
This model provides word segmentation for the ancient Chinese language. Our training dataset covers four eras of the Chinese language.
|
16 |
|
|
|
36 |
AutoModel,
|
37 |
)
|
38 |
|
39 |
+
tokenizer = AutoTokenizer.from_pretrained("ckiplab/bert-base-han-chinese-ws")
|
40 |
+
model = AutoModel.from_pretrained("ckiplab/bert-base-han-chinese-ws")
|
41 |
```
|
42 |
|
43 |
* Using our model for inference
|
44 |
```python
|
45 |
>>> from transformers import pipeline
|
46 |
+
>>> classifier = pipeline("token-classification", model="ckiplab/bert-base-han-chinese-ws")
|
47 |
>>> classifier("帝堯曰放勳")
|
48 |
|
49 |
# output
|