--- language: - en tags: - text generation - pytorch - causal-lm - gpt_neox license: mit datasets: - hoskinson-center/proof-pile --- # ProofGPT-v0.1 # Model Description ProofGPT-v0.1 is a 6.7B parameter language model based on the GPT-NeoX architecture and trained on the [proof-pile](https://huggingface.co./datasets/hoskinson-center/proof-pile) (v1.1). We initiailized training with pythia-6.7b weights, a precursor to the [pythia-6.9b](https://huggingface.co./EleutherAI/pythia-6.9b) model that has roughly equivalent performance. Detailed evaluations coming soon :)