This model provides a MobileBERT (Sun et al., 2020) fine-tuned on the SST data with three sentiments (0 -- negative, 1 -- neutral, and 2 -- positive).

Example Usage

Below, we provide illustrations on how to use this model to make sentiment predictions.

import torch
from transformers import AutoTokenizer, AutoConfig, MobileBertForSequenceClassification
# load model
model_name = r'cambridgeltl/sst_mobilebert-uncased'
tokenizer = AutoTokenizer.from_pretrained(model_name)
config = AutoConfig.from_pretrained(model_name)
model = MobileBertForSequenceClassification.from_pretrained(model_name, config=config)
model.eval()
'''
    labels: 
        0 -- negative
        1 -- neutral
        2 -- positive
'''

# prepare exemplar sentences
batch_sentences = [
    "in his first stab at the form , jacquot takes a slightly anarchic approach that works only sporadically .",
    "a valueless kiddie paean to pro basketball underwritten by the nba .",
    "a very well-made , funny and entertaining picture .",
]

# prepare input
inputs = tokenizer(batch_sentences, max_length=256, truncation=True, padding=True, return_tensors='pt')
input_ids, attention_mask = inputs.input_ids, inputs.attention_mask

# make predictions
outputs = model(input_ids=input_ids, attention_mask=attention_mask)
predictions = torch.argmax(outputs.logits, dim = -1)
print (predictions)
# tensor([1, 0, 2])

Citation:

If you find this model useful, please kindly cite our model as

@misc{susstmobilebert,
  author = {Su, Yixuan},
  title = {A MobileBERT Fine-tuned on SST},
  howpublished = {\url{https://huggingface.co/cambridgeltl/sst_mobilebert-uncased}},
  year = 2022
}