bert-base for KLUE Relation Extraction task.

Fine-tuned klue/bert-base using KLUE RE dataset.

<br>

Usage

<pre><code> from transformers import AutoTokenizer, AutoModelForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("ainize/klue-bert-base-re") model = AutoModelForSequenceClassification.from_pretrained("ainize/klue-bert-base-re")

Add "&ltsubj&gt", "&lt/subj&gt" to both ends of the subject object and "&ltobj&gt", "&lt/obj&gt" to both ends of the object object.

sentence = "&ltsubj&gt손흥민&lt/subj&gt은 &ltobj&gt대한민국&lt/obj&gt에서 태어났다."

encodings = tokenizer(sentence, max_length=128, truncation=True, padding="max_length", return_tensors="pt")

outputs = model(**encodings)

logits = outputs['logits']

preds = torch.argmax(logits, dim=1) </code></pre>

<br>

About us