generated_from_trainer

<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

20230829213025

This model is a fine-tuned version of bert-large-cased on the super_glue dataset. It achieves the following results on the evaluation set:

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

Training results

Training Loss Epoch Step Validation Loss Accuracy
No log 1.0 35 0.7603 0.5962
No log 2.0 70 1.3915 0.4038
No log 3.0 105 0.7099 0.625
No log 4.0 140 0.8092 0.4615
No log 5.0 175 0.6872 0.5192
No log 6.0 210 0.6200 0.5673
No log 7.0 245 0.9617 0.3558
No log 8.0 280 0.6118 0.625
No log 9.0 315 0.9262 0.3846
No log 10.0 350 0.8814 0.4231
No log 11.0 385 0.7562 0.3942
No log 12.0 420 0.6150 0.6346
No log 13.0 455 0.8615 0.4135
No log 14.0 490 0.7234 0.4038
0.8691 15.0 525 0.6146 0.5577
0.8691 16.0 560 0.6651 0.625
0.8691 17.0 595 0.5924 0.5962
0.8691 18.0 630 0.6120 0.5673
0.8691 19.0 665 1.0046 0.3558
0.8691 20.0 700 0.6139 0.6058
0.8691 21.0 735 0.6114 0.5577
0.8691 22.0 770 0.7540 0.3846
0.8691 23.0 805 0.8208 0.3654
0.8691 24.0 840 0.5966 0.6538
0.8691 25.0 875 0.6293 0.6154
0.8691 26.0 910 0.5975 0.5962
0.8691 27.0 945 1.0532 0.3558
0.8691 28.0 980 0.6205 0.5385
0.7391 29.0 1015 0.6774 0.4038
0.7391 30.0 1050 0.6520 0.3846
0.7391 31.0 1085 0.8306 0.3846
0.7391 32.0 1120 0.7331 0.3846
0.7391 33.0 1155 0.7668 0.6346
0.7391 34.0 1190 0.6067 0.6346
0.7391 35.0 1225 0.7269 0.4231
0.7391 36.0 1260 0.6287 0.5
0.7391 37.0 1295 0.5856 0.6442
0.7391 38.0 1330 0.6163 0.6346
0.7391 39.0 1365 0.7107 0.6346
0.7391 40.0 1400 0.6136 0.6346
0.7391 41.0 1435 1.5113 0.3654
0.7391 42.0 1470 0.6669 0.375
0.7392 43.0 1505 0.7104 0.3558
0.7392 44.0 1540 0.6053 0.6346
0.7392 45.0 1575 0.8003 0.375
0.7392 46.0 1610 0.5931 0.6635
0.7392 47.0 1645 0.5956 0.6442
0.7392 48.0 1680 0.6996 0.3942
0.7392 49.0 1715 0.5921 0.6346
0.7392 50.0 1750 0.6349 0.4423
0.7392 51.0 1785 0.5973 0.6731
0.7392 52.0 1820 0.6991 0.4135
0.7392 53.0 1855 0.8069 0.375
0.7392 54.0 1890 0.5853 0.6346
0.7392 55.0 1925 0.6038 0.6538
0.7392 56.0 1960 0.5989 0.6346
0.7392 57.0 1995 0.6352 0.4423
0.6889 58.0 2030 0.6508 0.4327
0.6889 59.0 2065 0.6231 0.5096
0.6889 60.0 2100 0.6778 0.6346
0.6889 61.0 2135 0.6168 0.6346
0.6889 62.0 2170 0.5914 0.6346
0.6889 63.0 2205 0.6067 0.625
0.6889 64.0 2240 0.6539 0.6346
0.6889 65.0 2275 0.6046 0.6346
0.6889 66.0 2310 0.5865 0.6346
0.6889 67.0 2345 0.6015 0.6346
0.6889 68.0 2380 0.5962 0.6635
0.6889 69.0 2415 0.5856 0.6346
0.6889 70.0 2450 0.5992 0.6346
0.6889 71.0 2485 0.6123 0.5096
0.6617 72.0 2520 0.6096 0.5288
0.6617 73.0 2555 0.5906 0.6346
0.6617 74.0 2590 0.6423 0.4519
0.6617 75.0 2625 0.6075 0.5962
0.6617 76.0 2660 0.6052 0.5962
0.6617 77.0 2695 0.6619 0.4519
0.6617 78.0 2730 0.6185 0.5192
0.6617 79.0 2765 0.6038 0.6635
0.6617 80.0 2800 0.5982 0.6538

Framework versions