|
--- |
|
language: |
|
- en |
|
tags: |
|
- text generation |
|
- pytorch |
|
- causal-lm |
|
- gpt_neox |
|
license: mit |
|
datasets: |
|
- hoskinson-center/proof-pile |
|
--- |
|
|
|
# ProofGPT-v0.1 |
|
|
|
# Model Description |
|
ProofGPT-v0.1 is a 6.7B parameter language model based on the GPT-NeoX architecture and trained on the [proof-pile](https://huggingface.co./datasets/hoskinson-center/proof-pile) (v1.1). |
|
We initiailized training with pythia-6.7b weights, a precursor to the [pythia-6.9b](https://huggingface.co./EleutherAI/pythia-6.9b) model that has roughly equivalent performance. |
|
|
|
Detailed evaluations coming soon :) |