TianyiQ commited on
Commit
a67ee3b
1 Parent(s): 0f3aaea

Upload ./README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -67,9 +67,9 @@ ProgressGym-HistLlama3-8B-C020-instruct is one of the **36 historical language m
67
  | 1.7028 | 3.2258 | 1600 | 1.9448 |
68
  | 1.7383 | 3.6290 | 1800 | 1.9450 |
69
 
70
- Note that the training data volume for the continued pretraining stage is capped at 300MB. When the corresponding century's corpus exceeds this volume, the training data is randomly sampled to fit the volume.
71
 
72
- **ProgressGym-HistLlama3-8B-C020-instruct is an instruction-tuned language model.** It is tuned on [ProgressGym-TimelessQA](https://huggingface.co/datasets/PKU-Alignment/ProgressGym-TimelessQA), using the following hyperparameters:
73
  - learning_rate: 1.5e-05
74
  - train_batch_size: 8
75
  - eval_batch_size: 16
@@ -105,6 +105,7 @@ Note that the training data volume for the continued pretraining stage is capped
105
  - **[Leaderboard & Interactive Playground]** [PKU-Alignment/ProgressGym-LeaderBoard](https://huggingface.co/spaces/PKU-Alignment/ProgressGym-LeaderBoard)
106
  - **[Huggingface Data & Model Collection]** [PKU-Alignment/ProgressGym](https://huggingface.co/collections/PKU-Alignment/progressgym-666735fcf3e4efa276226eaa)
107
  - **[Github Codebase]** [PKU-Alignment/ProgressGym](https://github.com/PKU-Alignment/ProgressGym)
 
108
  - **[PyPI Package]** *(coming soon - [stay tuned](https://forms.gle/1TWFLL4ZCLeYTD5N6)!)*
109
 
110
  ## Citation
 
67
  | 1.7028 | 3.2258 | 1600 | 1.9448 |
68
  | 1.7383 | 3.6290 | 1800 | 1.9450 |
69
 
70
+ Note that the training data volume for the continued pretraining stage is capped at 3GB. When the corresponding century's corpus exceeds this volume, the training data is randomly sampled to fit the volume.
71
 
72
+ **ProgressGym-HistLlama3-8B-C020-instruct is an instruction-tuned language model.** It is tuned on [ProgressGym-TimelessQA](https://huggingface.co/datasets/PKU-Alignment/ProgressGym-TimelessQA), using the following hyperparameters. Note, however, that the snapshot at training step 10 is used for the final model, to minimize erosion of the value tendencies learned during continued pretraining; we qualitatively observe that this snapshot still possesses strong instruction-following capabilities.
73
  - learning_rate: 1.5e-05
74
  - train_batch_size: 8
75
  - eval_batch_size: 16
 
105
  - **[Leaderboard & Interactive Playground]** [PKU-Alignment/ProgressGym-LeaderBoard](https://huggingface.co/spaces/PKU-Alignment/ProgressGym-LeaderBoard)
106
  - **[Huggingface Data & Model Collection]** [PKU-Alignment/ProgressGym](https://huggingface.co/collections/PKU-Alignment/progressgym-666735fcf3e4efa276226eaa)
107
  - **[Github Codebase]** [PKU-Alignment/ProgressGym](https://github.com/PKU-Alignment/ProgressGym)
108
+ - **[Documentation]** [ProgressGym Documentation](https://pku-alignment.github.io/ProgressGym/)
109
  - **[PyPI Package]** *(coming soon - [stay tuned](https://forms.gle/1TWFLL4ZCLeYTD5N6)!)*
110
 
111
  ## Citation