# ai-msgbot GPT2-L + daily dialogues _NOTE: this model card is a WIP_ GPT2-L (774M parameters) trained on the Wizard of Wikipedia dataset for 40k steps with 34/36 layers frozen using `aitextgen`. This model was then subsequently trained on the [Daily Dialogues](http://yanran.li/dailydialog) dataset for an additional 40k steps, this time with **35** of 36 layers frozen. Designed for use with [ai-msgbot](https://github.com/pszemraj/ai-msgbot).