generated_from_trainer

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

20230830015435

This model is a fine-tuned version of bert-large-cased on the super_glue dataset. It achieves the following results on the evaluation set:

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Accuracy
No log 1.0 35 0.5691 0.6058
No log 2.0 70 0.5906 0.5962
No log 3.0 105 0.5703 0.625
No log 4.0 140 0.4604 0.6154
No log 5.0 175 0.4834 0.5962
No log 6.0 210 0.4973 0.4135
No log 7.0 245 0.9046 0.3654
No log 8.0 280 0.3669 0.6346
No log 9.0 315 0.3828 0.4231
No log 10.0 350 0.4207 0.5769
No log 11.0 385 0.7596 0.3654
No log 12.0 420 0.9833 0.6346
No log 13.0 455 0.3754 0.6346
No log 14.0 490 0.4325 0.6346
0.6565 15.0 525 0.4163 0.6346
0.6565 16.0 560 0.7707 0.3654
0.6565 17.0 595 0.4262 0.6346
0.6565 18.0 630 0.3547 0.5
0.6565 19.0 665 0.3355 0.6346
0.6565 20.0 700 0.3787 0.4423
0.6565 21.0 735 0.3718 0.6346
0.6565 22.0 770 0.4742 0.3846
0.6565 23.0 805 0.3361 0.6827
0.6565 24.0 840 0.4078 0.6346
0.6565 25.0 875 0.3701 0.6346
0.6565 26.0 910 0.3726 0.6346
0.6565 27.0 945 0.7422 0.6346
0.6565 28.0 980 0.6071 0.6346
0.5512 29.0 1015 0.4255 0.4038
0.5512 30.0 1050 0.3393 0.6346
0.5512 31.0 1085 0.3556 0.6346
0.5512 32.0 1120 0.4493 0.3846
0.5512 33.0 1155 0.4296 0.4231
0.5512 34.0 1190 0.3491 0.625
0.5512 35.0 1225 0.3334 0.625
0.5512 36.0 1260 0.3552 0.6346
0.5512 37.0 1295 0.3297 0.6538
0.5512 38.0 1330 0.3939 0.4231
0.5512 39.0 1365 0.4693 0.3846
0.5512 40.0 1400 0.4533 0.3942
0.5512 41.0 1435 0.3350 0.6346
0.5512 42.0 1470 0.7957 0.3654
0.4482 43.0 1505 0.3327 0.6346
0.4482 44.0 1540 0.3572 0.6346
0.4482 45.0 1575 0.3303 0.6346
0.4482 46.0 1610 0.3398 0.6058
0.4482 47.0 1645 0.3778 0.4135
0.4482 48.0 1680 0.3528 0.5962
0.4482 49.0 1715 0.3447 0.6346
0.4482 50.0 1750 0.3419 0.5673
0.4482 51.0 1785 0.4567 0.3846
0.4482 52.0 1820 0.3662 0.4712
0.4482 53.0 1855 0.3298 0.6827
0.4482 54.0 1890 0.3821 0.4423
0.4482 55.0 1925 0.3533 0.6346
0.4482 56.0 1960 0.3340 0.6346
0.4482 57.0 1995 0.3546 0.5288
0.3923 58.0 2030 0.3321 0.6442
0.3923 59.0 2065 0.3440 0.5481
0.3923 60.0 2100 0.3326 0.6442
0.3923 61.0 2135 0.3378 0.6635
0.3923 62.0 2170 0.3346 0.625
0.3923 63.0 2205 0.3460 0.5
0.3923 64.0 2240 0.3517 0.6346
0.3923 65.0 2275 0.3301 0.6346
0.3923 66.0 2310 0.3306 0.6635
0.3923 67.0 2345 0.3362 0.6346
0.3923 68.0 2380 0.3382 0.5865
0.3923 69.0 2415 0.3314 0.6731
0.3923 70.0 2450 0.3283 0.6538
0.3923 71.0 2485 0.3304 0.6635
0.3611 72.0 2520 0.3445 0.5385
0.3611 73.0 2555 0.3291 0.6731
0.3611 74.0 2590 0.3414 0.5577
0.3611 75.0 2625 0.3334 0.6058
0.3611 76.0 2660 0.3304 0.6731
0.3611 77.0 2695 0.3415 0.5481
0.3611 78.0 2730 0.3314 0.6346
0.3611 79.0 2765 0.3315 0.6346
0.3611 80.0 2800 0.3304 0.6538

Framework versions