--- license: apache-2.0 --- The *TokenFormer* is a **fully attention-based architecture** that unifies the computations of token-token and token-parameter interactions by entirely employing the attention mechanism, **maximizes the flexibility of neural network**.[(see paper)](https://github.com/Haiyang-W/TokenFormer). It contains four models of sizes 150M, 450M, 900M, 1.5B. For each size, it's trained based on [gpt-neox](https://github.com/EleutherAI/gpt-neox) code base and uses [Pile](https://huggingface.co./datasets/EleutherAI/pile) with 300B tokens. All 4 model sizes are trained on the exact same data, in the exact same order. # TokenFormer-450M ## Model Details - Developed by: [Haiyang Wang](https://haiyang-w.github.io/) - Model type: ToeknFormer-based Language Model - Language: English - Learn more: [TokenFormer's GitHub repository](https://github.com/Haiyang-W/TokenFormer) for training procedure, config files, and details on how to use. [See paper](https://github.com/Haiyang-W/TokenFormer) for more evals and implementation details. - Library: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox) - License: Apache 2.0 - Contact: to ask questions about this model, please email Haiyang Wang.