mamba-vs-gpt2
Collection
This repository contains the code for comparing the performance of Mamba vs GPT2 on text classification task
•
4 items
•
Updated
This model is a fine-tuned version of state-spaces/mamba-130m-hf on an unknown dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
---|---|---|---|---|---|
0.188 | 0.9994 | 781 | 0.1786 | 0.9307 | 0.9313 |
Base model
state-spaces/mamba-130m-hf