Update README.md
Browse files
README.md
CHANGED
@@ -22,11 +22,10 @@ This model was fine-tuned on over 24 hours of transcribed upper sorbian speech t
|
|
22 |
## Training Details
|
23 |
- **Hyperparameters**:
|
24 |
- Batch size: 64
|
25 |
-
- Learning rate: 3e-6
|
26 |
-
- Linear Decay
|
27 |
- **Optimizer**: AdamW
|
28 |
-
- **Warmup**: 1000
|
29 |
-
- **Additional Techniques**: BF16
|
30 |
|
31 |
|
32 |
## Performance
|
|
|
22 |
## Training Details
|
23 |
- **Hyperparameters**:
|
24 |
- Batch size: 64
|
25 |
+
- Learning rate: 3e-6, linear decay
|
|
|
26 |
- **Optimizer**: AdamW
|
27 |
+
- **Warmup**: 1000 steps
|
28 |
+
- **Additional Techniques**: BF16 training, initial 15 layers frozen
|
29 |
|
30 |
|
31 |
## Performance
|