proofGPT-v0.1-6.7B / README.md
zhangirazerbayev's picture
Create README.md
02f405f
|
raw
history blame
573 Bytes
metadata
language:
  - en
tags:
  - text generation
  - pytorch
  - causal-lm
  - gpt_neox
license: mit
datasets:
  - hoskinson-center/proof-pile

ProofGPT-v0.1

Model Description

ProofGPT-v0.1 is a 6.7B parameter language model based on the GPT-NeoX architecture and trained on the proof-pile (v1.1). We initiailized training with pythia-6.7b weights, a precursor to the pythia-6.9b model that has roughly equivalent performance.

Detailed evaluations coming soon :)