QA-DeBERTa-v3-large-qa_cross_attn-binary
This model is a fine-tuned version of microsoft/deberta-v3-large on the saiteki-kai/Beavertails-it dataset. It achieves the following results on the evaluation set:
- Loss: 0.3189
- Accuracy: 0.8622
- Unsafe Precision: 0.8773
- Unsafe Recall: 0.8747
- Unsafe F1: 0.8760
- Unsafe Fpr: 0.1535
- Unsafe Aucpr: 0.9539
- Safe Precision: 0.8433
- Safe Recall: 0.8465
- Safe F1: 0.8449
- Safe Fpr: 0.1253
- Safe Aucpr: 0.9185
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-06
- train_batch_size: 64
- eval_batch_size: 128
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Unsafe Precision | Unsafe Recall | Unsafe F1 | Unsafe Fpr | Unsafe Aucpr | Safe Precision | Safe Recall | Safe F1 | Safe Fpr | Safe Aucpr |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 0.3101 | 0.2501 | 2114 | 0.3580 | 0.8457 | 0.8898 | 0.8249 | 0.8561 | 0.1282 | 0.9424 | 0.7988 | 0.8718 | 0.8337 | 0.1751 | 0.8922 |
| 0.343 | 0.5001 | 4228 | 0.3325 | 0.8550 | 0.8714 | 0.8675 | 0.8695 | 0.1606 | 0.9479 | 0.8347 | 0.8394 | 0.8370 | 0.1325 | 0.9052 |
| 0.3078 | 0.7502 | 6342 | 0.3240 | 0.8572 | 0.9026 | 0.8333 | 0.8666 | 0.1128 | 0.9517 | 0.8093 | 0.8872 | 0.8464 | 0.1667 | 0.9131 |
| 0.3524 | 1.0002 | 8456 | 0.3194 | 0.8596 | 0.8728 | 0.8753 | 0.8740 | 0.1601 | 0.9528 | 0.8430 | 0.8399 | 0.8414 | 0.1247 | 0.9160 |
| 0.3087 | 1.2503 | 10570 | 0.3234 | 0.8601 | 0.8764 | 0.8715 | 0.8739 | 0.1542 | 0.9527 | 0.8399 | 0.8458 | 0.8428 | 0.1285 | 0.9151 |
| 0.2752 | 1.5004 | 12684 | 0.3339 | 0.8610 | 0.8717 | 0.8797 | 0.8757 | 0.1624 | 0.9532 | 0.8473 | 0.8376 | 0.8424 | 0.1203 | 0.9167 |
| 0.2819 | 1.7504 | 14798 | 0.3189 | 0.8622 | 0.8773 | 0.8747 | 0.8760 | 0.1535 | 0.9539 | 0.8433 | 0.8465 | 0.8449 | 0.1253 | 0.9185 |
| 0.3151 | 2.0005 | 16912 | 0.3186 | 0.8629 | 0.8920 | 0.8575 | 0.8744 | 0.1302 | 0.9551 | 0.8295 | 0.8698 | 0.8492 | 0.1425 | 0.9188 |
| 0.3024 | 2.2505 | 19026 | 0.3245 | 0.8602 | 0.8753 | 0.8731 | 0.8742 | 0.1560 | 0.9536 | 0.8413 | 0.8440 | 0.8426 | 0.1269 | 0.9183 |
| 0.2619 | 2.5006 | 21140 | 0.3270 | 0.8607 | 0.8708 | 0.8802 | 0.8755 | 0.1638 | 0.9545 | 0.8477 | 0.8362 | 0.8419 | 0.1198 | 0.9202 |
| 0.2704 | 2.7507 | 23254 | 0.3189 | 0.8606 | 0.8697 | 0.8815 | 0.8756 | 0.1656 | 0.9550 | 0.8487 | 0.8344 | 0.8415 | 0.1185 | 0.9212 |
Framework versions
- Transformers 4.57.3
- Pytorch 2.7.1+cu118
- Datasets 4.4.1
- Tokenizers 0.22.1
- Downloads last month
- 4
Model tree for saiteki-kai/QA-DeBERTa-v3-large-qa_cross_attn-binary
Base model
microsoft/deberta-v3-largeEvaluation results
- Accuracy on saiteki-kai/Beavertails-itself-reported0.862