Text Generation
Transformers
PyTorch
code
gpt2
custom_code
Eval Results
text-generation-inference
Inference Endpoints

fix sequence length in santacoder and introduce new model type

#23

Adds a new model_type to the config. Currently this is gpt2 which creates problems with huggingface/optimum.
Fix sequence length bug which is not seen in transformers but in ONNX because transformers's generate method passes position_ids itself whereas when running with ONNX, the model needs to infer it itself.

mayank-mishra changed pull request status to open
Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment