Description
This repo contains GGUF files of Dawn-70B, a merge I have done with the new layer shuffle method from mergekit.
UtopiaXL was a huge success for me, I really liked it, so I took the same path to do this 70B: A good base, some psychologic data, some medical data, a little bit of this, of that, and LimaRP at the end as always.
Models and loras used
- Sao10K/Euryale-1.3-L2-70B
- Xwin-LM/Xwin-LM-70B-V0.1
- ehartford/Samantha-1.11-70b
- NousResearch/Nous-Hermes-Llama2-70b
- augtoma/qCammel-70-x
- jondurbin/airoboros-l2-c70b-3.1.2
- fangloveskari/ORCA_LLaMA_70B_QLoRA
- Doctor-Shotgun/limarpv3-llama2-70b-qlora
The sauce
!mergekit-layershuffle ./Dawn-v2-70B \
--model Sao10K/Euryale-1.3-L2-70B --weight 0.3 \
--model Xwin-LM/Xwin-LM-70B-V0.1 --weight 0.2 \
--model ehartford/Samantha-1.11-70b --weight 0.1 \
--model NousResearch/Nous-Hermes-Llama2-70b --weight 0.05 \
--model augtoma/qCammel-70-x --weight 0.05 \
--model jondurbin/airoboros-l2-c70b-3.1.2 --weight 0.2 \
--model fangloveskari/ORCA_LLaMA_70B_QLoRA --weight 0.1 \
--write-yaml Dawn-v2-70B.yaml
=========================
merge_method: passthrough
slices:
- sources:
- layer_range:
- 0
- 1
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 1
- 2
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 2
- 3
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 3
- 4
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 4
- 5
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 5
- 6
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 6
- 8
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 8
- 9
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 9
- 10
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 10
- 11
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 11
- 12
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 12
- 13
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 13
- 14
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 14
- 15
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 15
- 16
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 16
- 17
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 17
- 18
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 18
- 19
model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
- layer_range:
- 19
- 20
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 20
- 21
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 21
- 22
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 22
- 23
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 23
- 24
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 24
- 25
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 25
- 27
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 27
- 28
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 28
- 29
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 29
- 30
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 30
- 32
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 32
- 33
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 33
- 34
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 34
- 35
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 35
- 37
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 37
- 38
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 38
- 39
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 39
- 40
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 40
- 41
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 41
- 42
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 42
- 43
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 43
- 44
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 44
- 45
model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
- layer_range:
- 45
- 46
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 46
- 48
model: ehartford/Samantha-1.11-70b
- sources:
- layer_range:
- 48
- 49
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 49
- 50
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 50
- 51
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 51
- 54
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 54
- 55
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 55
- 56
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 56
- 58
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 58
- 59
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 59
- 60
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 60
- 62
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 62
- 63
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 63
- 64
model: fangloveskari/ORCA_LLaMA_70B_QLoRA
- sources:
- layer_range:
- 64
- 65
model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
- layer_range:
- 65
- 66
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 66
- 67
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 67
- 68
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 68
- 70
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 70
- 71
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 71
- 72
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 72
- 73
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 73
- 75
model: jondurbin/airoboros-l2-c70b-3.1.2
- sources:
- layer_range:
- 75
- 76
model: Sao10K/Euryale-1.3-L2-70B
- sources:
- layer_range:
- 76
- 77
model: augtoma/qCammel-70-x
- sources:
- layer_range:
- 77
- 78
model: Xwin-LM/Xwin-LM-70B-V0.1
- sources:
- layer_range:
- 78
- 79
model: NousResearch/Nous-Hermes-Llama2-70b
- sources:
- layer_range:
- 79
- 80
model: Xwin-LM/Xwin-LM-70B-V0.1
=========================
=> Applying Doctor-Shotgun/limarpv3-llama2-70b-qlora x 0.35
Prompt template: Alpaca
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:
A big thanks to Charles for adding the layer shuffle method to his tool mergekit and Henky/KoboldAI for the machine he let me use.
If you want to support me, you can here.
- Downloads last month
- 0