electra-ka is first of its kind, Transformer based, open source Georgian language model.

The model is trained on 33GB of Georgian text collected from 4854621 pages in commoncrowl archive.