Edit model card

bart_left_fulltext_torch

This model is a fine-tuned version of facebook/bart-large on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.9329

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20

Training results

Training Loss Epoch Step Validation Loss
No log 1.0 8 9.2222
No log 2.0 16 8.1910
No log 3.0 24 7.8851
No log 4.0 32 7.5382
No log 5.0 40 7.2200
No log 6.0 48 6.7294
No log 7.0 56 6.1038
No log 8.0 64 5.5721
No log 9.0 72 5.1404
No log 10.0 80 4.4875
No log 11.0 88 3.8785
No log 12.0 96 3.3749
No log 13.0 104 2.9965
No log 14.0 112 2.6608
No log 15.0 120 2.3879
No log 16.0 128 2.2349
No log 17.0 136 2.0993
No log 18.0 144 2.0023
No log 19.0 152 1.9480
No log 20.0 160 1.9329

Framework versions

  • Transformers 4.31.0
  • Pytorch 2.0.1+cu118
  • Datasets 2.14.1
  • Tokenizers 0.13.3
Downloads last month
17
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for tobijen/bart_left_fulltext_torch

Finetuned
(139)
this model