base model: bert-base-multilingual-cased

trained in MLM task with 0.15 probability

trained with data from DGLAB ANTT data totalizing more than 351k text examples of 512chars each

1 epoch corresponding to 6-8h hours of training


license: other language: