luxembourgish lëtzebuergesch text generation

LuxGPT-2

GPT-2 model for Text Generation in luxembourgish language, trained on 667 MB of text data, consisting of RTL.lu news articles, comments, parlament speeches, the luxembourgish Wikipedia, Newscrawl, Webcrawl and subtitles. The training took place on a 32 GB Nvidia Tesla V100

Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("laurabernardy/LuxGPT2")

model = AutoModelForCausalLM.from_pretrained("laurabernardy/LuxGPT2")

Limitations and Biases

See the GPT2 model card for considerations on limitations and bias. See the GPT2 documentation for details on GPT2.