electra-ka is first of its kind, Transformer based, open source Georgian language model.
The model is trained on 33GB of Georgian text collected from 4854621 pages in commoncrowl archive.
The model is trained on 33GB of Georgian text collected from 4854621 pages in commoncrowl archive.