<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->
LKD_Experience_all_labels
This model is a fine-tuned version of distilbert-base-uncased on an a dataset of Reddit comments and posts related to Living Kidney Donation (LKD). The first word/token of each document is the name of the subreddit where the post was written. If an original post, the document is the title of the post, as well as the post description, where avaialble. At present, the labels are unevenly distributed so the model does not perform well (to be remedied in the near future).
It achieves the following results on the evaluation set:
- Loss: 1.0036
- Accuracy: 0.7429
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 8
- eval_batch_size: 4
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
Training results
Training Loss | Epoch | Step | Validation Loss | Accuracy |
---|---|---|---|---|
No log | 1.0 | 66 | 1.6537 | 0.3286 |
No log | 2.0 | 132 | 1.3005 | 0.7 |
No log | 3.0 | 198 | 1.1173 | 0.7429 |
No log | 4.0 | 264 | 1.0296 | 0.7429 |
No log | 5.0 | 330 | 1.0036 | 0.7429 |
Framework versions
- Transformers 4.27.1
- Pytorch 1.13.1+cu116
- Datasets 2.10.1
- Tokenizers 0.13.2