base model: bert-base-multilingual-cased
trained in MLM task with 0.15 probability
trained with data from DGLAB ANTT data totalizing more than 351k text examples of 512chars each
1 epoch corresponding to 6-8h hours of training
license: other language:
- pt