metadata
license: mit
- Check out the demo: https://huggingface.co/spaces/winglian/llama-adapter
- Read the paper: https://arxiv.org/abs/2303.16199
- PEFT PR: https://github.com/huggingface/peft/pull/268
training hyperparamters:
--batch_size 64 --micro_batch_size 8 --num_epochs 5 --learning_rate 9e-3 --cutoff_len 2048 --val_set_size 0.05 --train_on_inputs 0
training dataset: https://github.com/tloen/alpaca-lora/blob/main/alpaca_data_gpt4.json