Edit model card

t5e-mini-nl24-flan

25k steps on FLAN as an initial test/validation that code works. Not practically useful.

from transformers import pipeline

pipe = pipeline(
    "text2text-generation",
    model="pszemraj/t5e-mini-nl24-flan",
)
res = pipe(
    "true or false: water is wet.",
    top_k=4,
    penalty_alpha=0.6,
    max_new_tokens=128,
)
print(res[0]["generated_text"])

Quick eval

Quick eval for: pszemraj/t5e-mini-nl24-flan

hf (pretrained=pszemraj/t5e-mini-nl24-flan,trust_remote_code=True,dtype=bfloat16,trust_remote_code=True), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 8

Tasks Version Filter n-shot Metric Value Stderr
boolq 2 none 0 acc ↑ 0.4541 ± 0.0087
openbookqa 1 none 0 acc ↑ 0.1300 ± 0.0151
none 0 acc_norm ↑ 0.2700 ± 0.0199
piqa 1 none 0 acc ↑ 0.6159 ± 0.0113
none 0 acc_norm ↑ 0.6077 ± 0.0114
social_iqa 0 none 0 acc ↑ 0.3705 ± 0.0109
tinyArc 0 none 25 acc_norm ↑ 0.2913 ± N/A
tinyGSM8k 0 flexible-extract 5 exact_match ↑ 0.0269 ± N/A
strict-match 5 exact_match ↑ 0.0055 ± N/A
tinyHellaswag 0 none 10 acc_norm ↑ 0.3538 ± N/A
tinyMMLU 0 none 0 acc_norm ↑ 0.2551 ± N/A
winogrande 1 none 0 acc ↑ 0.5217 ± 0.0140
base model evals: click to expand

Quick eval for: google/t5-efficient-mini-nl24

hf (pretrained=google/t5-efficient-mini-nl24,trust_remote_code=True,dtype=bfloat16,trust_remote_code=True), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 8

Tasks Version Filter n-shot Metric Value Stderr
boolq 2 none 0 acc ↑ 0.3783 ± 0.0085
openbookqa 1 none 0 acc ↑ 0.1280 ± 0.0150
none 0 acc_norm ↑ 0.2660 ± 0.0198
piqa 1 none 0 acc ↑ 0.5473 ± 0.0116
none 0 acc_norm ↑ 0.5267 ± 0.0116
social_iqa 0 none 0 acc ↑ 0.3536 ± 0.0108
tinyArc 0 none 25 acc_norm ↑ 0.3101 ± N/A
tinyGSM8k 0 flexible-extract 5 exact_match ↑ 0.0145 ± N/A
strict-match 5 exact_match ↑ 0.0055 ± N/A
tinyHellaswag 0 none 10 acc_norm ↑ 0.2616 ± N/A
tinyMMLU 0 none 0 acc_norm ↑ 0.2839 ± N/A
winogrande 1 none 0 acc ↑ 0.4996 ± 0.0141
Downloads last month
8
Safetensors
Model size
126M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for pszemraj/t5e-mini-nl24-flan

Finetuned
(1)
this model

Dataset used to train pszemraj/t5e-mini-nl24-flan