SloBERTa model, fine-tuned for natural language inference on 50,000 samples from ESNLI dataset, machine translated to Slovene.

Usage

from transformers import AutoTokenizer, AutoModelForMaskedLM

tokenizer = AutoTokenizer.from_pretrained("timkmecl/sloberta-esnli")
model = AutoModelForMaskedLM.from_pretrained("timkmecl/sloberta-esnli")

Expected inputs are of the form

Premisa: {premise}
Hipoteza: {hypothesis}

with strings {premise} and {hypothesis} being replaced with premise and hypothesis in Slovene.

Class 0 is entailment, class 1 neutral and 2 contradiction.