translation

T5 for belarusian language

model image

This model is based on T5-small with sequence length equal 128 tokens. Model trained from scratch on RTX 3090 24GB.

Supported tasks:

Metrics:

How to Get Started with the Model

<details> <summary> Click to expand </summary>

from transformers import T5TokenizerFast, T5ForConditionalGeneration

tokenizer = T5TokenizerFast.from_pretrained("WelfCrozzo/T5-L128-belarusian")
model = T5ForConditionalGeneration.from_pretrained("WelfCrozzo/T5-L128-belarusian")

x = tokenizer.encode('<extra_id_1>да зорак праз цяжкасці', return_tensors='pt')

result = model.generate(x, return_dict_in_generate=True, output_scores=True,max_length=128)
print(tokenizer.decode(result["sequences"][0]))

</details>

References