daT5-large
A smaller version of Google's mt5-large model, where the original model is reduced to only include Danish embeddings.
How to use
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("emillykkejensen/daT5-large")
model = AutoModel.from_pretrained("emillykkejensen/daT5-large")
Further reading
Gist showing (in Danish) how the embeddings are extracted (for mt5-base)
Article explaining how to do it by David Dale
Also check out
- Downloads last month
- 13
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.