Edit model card

RWKV-1b5-finetuned-overfit

This model is a fine-tuned version of RWKV/rwkv-raven-1b5 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 68.7560

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.005
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 8

Training results

Training Loss Epoch Step Validation Loss
1.6836 1.0 1 1.4341
1.5494 2.0 2 1.7198
0.7595 3.0 3 9.1981
0.3142 4.0 4 35.6430
0.1007 5.0 5 68.5554
0.0256 6.0 6 69.8436
0.0119 7.0 7 69.2797
0.0082 8.0 8 68.7560

Framework versions

  • Transformers 4.31.0
  • Pytorch 2.0.1+cu117
  • Datasets 2.13.1
  • Tokenizers 0.13.3
Downloads last month
5
Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for avidoavid/RWKV-1b5-finetuned-overfit

Finetuned
this model