Update README.md
Browse files
README.md
CHANGED
@@ -2,16 +2,23 @@
|
|
2 |
license: cc-by-sa-4.0
|
3 |
datasets:
|
4 |
- bigcode/the-stack-dedup
|
|
|
|
|
5 |
---
|
6 |
|
7 |
|
8 |
# replit-code-v1-3b
|
9 |
|
10 |
-
`replit-code-v1-3b` is a 2.7B
|
11 |
|
|
|
|
|
|
|
12 |
|
|
|
13 |
|
14 |
-
|
|
|
15 |
|
16 |
|
17 |
```python
|
@@ -101,4 +108,4 @@ Coming soon.
|
|
101 |
Coming soon.
|
102 |
|
103 |
## Model Hash
|
104 |
-
5bc28ce32c6f9aec935ead7b60ea1c46
|
|
|
2 |
license: cc-by-sa-4.0
|
3 |
datasets:
|
4 |
- bigcode/the-stack-dedup
|
5 |
+
tags:
|
6 |
+
- code
|
7 |
---
|
8 |
|
9 |
|
10 |
# replit-code-v1-3b
|
11 |
|
12 |
+
`replit-code-v1-3b` is a 2.7B Causal Language Model focused on Code Completion. The model has been trained on a subset of the Stack Dedup v1.2 dataset.
|
13 |
|
14 |
+
The training mixture includes 20 different languages, listed here in descending order of number of tokens:
|
15 |
+
<br/>
|
16 |
+
`Markdown`, `Java`, `JavaScript`, `Python`, `TypeScript`, `PHP`, `SQL`, `JSX`, `reStructuredText`, `Rust`, `C`, `CSS`, `Go`, `C++`, `HTML`, `Vue`, `Ruby`, `Jupyter Notebook`, `R`, `Shell`
|
17 |
|
18 |
+
In total, the training dataset contains 175B tokens, which were repeated over 3 epochs -- in total, `replit-code-v1-3b` has been trained on 525B tokens (~195 tokens per parameter).
|
19 |
|
20 |
+
|
21 |
+
## How to use the model
|
22 |
|
23 |
|
24 |
```python
|
|
|
108 |
Coming soon.
|
109 |
|
110 |
## Model Hash
|
111 |
+
5bc28ce32c6f9aec935ead7b60ea1c46
|