Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ language:
|
|
15 |
|
16 |
# Poro 34B Model Card
|
17 |
|
18 |
-
Poro is a 34B parameter decoder-only transformer pretrained on Finnish, English and code. It
|
19 |
|
20 |
Poro was created in a collaboration between [SiloGen](https://www.silo.ai/silogen) from [Silo AI](https://www.silo.ai/), the [TurkuNLP group](https://turkunlp.org/) of the University of Turku, and [High Performance Language Technologies](https://hplt-project.org/) (HPLT). Training was conducted on the [LUMI supercomputer](https://www.lumi-supercomputer.eu/), using compute resources generously provided by [CSC](https://csc.fi/) - IT Center for Science, Finland.
|
21 |
|
|
|
15 |
|
16 |
# Poro 34B Model Card
|
17 |
|
18 |
+
Poro is a 34B parameter decoder-only transformer pretrained on Finnish, English and code. It was trained on 1 trillion tokens. Poro is a fully open source model and is made available under the Apache 2.0 License.
|
19 |
|
20 |
Poro was created in a collaboration between [SiloGen](https://www.silo.ai/silogen) from [Silo AI](https://www.silo.ai/), the [TurkuNLP group](https://turkunlp.org/) of the University of Turku, and [High Performance Language Technologies](https://hplt-project.org/) (HPLT). Training was conducted on the [LUMI supercomputer](https://www.lumi-supercomputer.eu/), using compute resources generously provided by [CSC](https://csc.fi/) - IT Center for Science, Finland.
|
21 |
|