File size: 635 Bytes
96bd8fc 6f1d0ba 96bd8fc 7a62ac3 96bd8fc 081c115 96bd8fc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
---
datasets:
- Aeala/ShareGPT_Vicuna_unfiltered
---
## LoRA Info:
Please note that this is a highly experimental LoRA model. It may do some good stuff, it might do some undesirable stuff. Training is paused for now. Feel free to try it!~
**Important Note**: While this is trained on a cleaned ShareGPT dataset like Vicuna used, this was trained in the *Alpaca* format, so prompting should be something like:
```
### Instruction:
<prompt> (without the <>)
### Response:
```
Current upload: checkpoint of step 1200 in training.
## Benchmarks
**wikitext2:** Coming soon...
**ptb-new:** Coming soon...
**c4-new:** Coming soon... |