uniBERT.SciBERT.1
This model is a fine-tuned version of allenai/scibert_scivocab_uncased on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.5001
- Accuracy: (0.6085790884718498,)
- F1: (0.6080656648964631,)
- Precision: (0.6210146704857022,)
- Recall: 0.6086
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
|---|---|---|---|---|---|---|---|
| 2.7217 | 1.0 | 163 | 2.2807 | (0.33065236818588023,) | (0.2969202316616143,) | (0.41162651970044983,) | 0.3307 |
| 1.8287 | 2.0 | 326 | 1.6950 | (0.4932975871313673,) | (0.4839828000587976,) | (0.5140427497857424,) | 0.4933 |
| 1.2837 | 3.0 | 489 | 1.6288 | (0.5120643431635389,) | (0.5079862696781359,) | (0.5571916164585299,) | 0.5121 |
| 0.856 | 4.0 | 652 | 1.4501 | (0.5612153708668454,) | (0.5624450158293163,) | (0.5965162734318348,) | 0.5612 |
| 0.6242 | 5.0 | 815 | 1.4338 | (0.5817694369973191,) | (0.5798849473141822,) | (0.5972769806009371,) | 0.5818 |
| 0.4313 | 6.0 | 978 | 1.4303 | (0.5996425379803396,) | (0.5989475055461593,) | (0.6176864401782607,) | 0.5996 |
| 0.2881 | 7.0 | 1141 | 1.4654 | (0.6076854334226989,) | (0.6071865300018395,) | (0.6237968728447791,) | 0.6077 |
| 0.2842 | 8.0 | 1304 | 1.4833 | (0.5978552278820375,) | (0.5996793150499145,) | (0.6223484815147282,) | 0.5979 |
| 0.199 | 9.0 | 1467 | 1.4836 | (0.6058981233243967,) | (0.6051827831523029,) | (0.6205279397909851,) | 0.6059 |
| 0.1949 | 10.0 | 1630 | 1.5001 | (0.6085790884718498,) | (0.6080656648964631,) | (0.6210146704857022,) | 0.6086 |
Framework versions
- Transformers 4.39.3
- Pytorch 2.2.1+cu121
- Datasets 2.18.0
- Tokenizers 0.15.2
- Downloads last month
- 4
Model tree for dbala02/uniBERT.SciBERT.1
Base model
allenai/scibert_scivocab_uncased