SloBERTa model, fine-tuned for natural language inference on 50,000 samples from ESNLI dataset, machine translated to Slovene.
Usage
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("timkmecl/sloberta-esnli")
model = AutoModelForMaskedLM.from_pretrained("timkmecl/sloberta-esnli")
Expected inputs are of the form
Premisa: {premise}
Hipoteza: {hypothesis}
with strings {premise}
and {hypothesis}
being replaced with premise and hypothesis in Slovene.
Class 0 is entailment, class 1 neutral and 2 contradiction.