Files changed (1) hide show
  1. README.md +126 -10
README.md CHANGED
@@ -1,10 +1,7 @@
1
  ---
2
- base_model:
3
- - mistralai/Mistral-7B-v0.1
4
- - SanjiWatsuki/Kunoichi-DPO-v2-7B
5
- - maywell/PiVoT-0.1-Evil-a
6
- - mlabonne/ArchBeagle-7B
7
- - LakoMoor/Silicon-Alice-7B
8
  library_name: transformers
9
  tags:
10
  - mergekit
@@ -17,9 +14,115 @@ tags:
17
  - roleplay
18
  - rp
19
  - not-for-all-audiences
20
- license: apache-2.0
21
- language:
22
- - en
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  ---
24
  # Konstanta-Alpha-V2-7B
25
 
@@ -100,4 +203,17 @@ parameters:
100
  int8_mask: true
101
  normalize: true
102
  dtype: float16
103
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
 
 
 
5
  library_name: transformers
6
  tags:
7
  - mergekit
 
14
  - roleplay
15
  - rp
16
  - not-for-all-audiences
17
+ base_model:
18
+ - mistralai/Mistral-7B-v0.1
19
+ - SanjiWatsuki/Kunoichi-DPO-v2-7B
20
+ - maywell/PiVoT-0.1-Evil-a
21
+ - mlabonne/ArchBeagle-7B
22
+ - LakoMoor/Silicon-Alice-7B
23
+ model-index:
24
+ - name: Konstanta-Alpha-V2-7B
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: AI2 Reasoning Challenge (25-Shot)
31
+ type: ai2_arc
32
+ config: ARC-Challenge
33
+ split: test
34
+ args:
35
+ num_few_shot: 25
36
+ metrics:
37
+ - type: acc_norm
38
+ value: 69.62
39
+ name: normalized accuracy
40
+ source:
41
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
42
+ name: Open LLM Leaderboard
43
+ - task:
44
+ type: text-generation
45
+ name: Text Generation
46
+ dataset:
47
+ name: HellaSwag (10-Shot)
48
+ type: hellaswag
49
+ split: validation
50
+ args:
51
+ num_few_shot: 10
52
+ metrics:
53
+ - type: acc_norm
54
+ value: 87.14
55
+ name: normalized accuracy
56
+ source:
57
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: MMLU (5-Shot)
64
+ type: cais/mmlu
65
+ config: all
66
+ split: test
67
+ args:
68
+ num_few_shot: 5
69
+ metrics:
70
+ - type: acc
71
+ value: 65.11
72
+ name: accuracy
73
+ source:
74
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
75
+ name: Open LLM Leaderboard
76
+ - task:
77
+ type: text-generation
78
+ name: Text Generation
79
+ dataset:
80
+ name: TruthfulQA (0-shot)
81
+ type: truthful_qa
82
+ config: multiple_choice
83
+ split: validation
84
+ args:
85
+ num_few_shot: 0
86
+ metrics:
87
+ - type: mc2
88
+ value: 61.08
89
+ source:
90
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: Winogrande (5-shot)
97
+ type: winogrande
98
+ config: winogrande_xl
99
+ split: validation
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 81.22
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: GSM8k (5-shot)
114
+ type: gsm8k
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 69.9
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Inv/Konstanta-Alpha-V2-7B
125
+ name: Open LLM Leaderboard
126
  ---
127
  # Konstanta-Alpha-V2-7B
128
 
 
203
  int8_mask: true
204
  normalize: true
205
  dtype: float16
206
+ ```
207
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
208
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Inv__Konstanta-Alpha-V2-7B)
209
+
210
+ | Metric |Value|
211
+ |---------------------------------|----:|
212
+ |Avg. |72.35|
213
+ |AI2 Reasoning Challenge (25-Shot)|69.62|
214
+ |HellaSwag (10-Shot) |87.14|
215
+ |MMLU (5-Shot) |65.11|
216
+ |TruthfulQA (0-shot) |61.08|
217
+ |Winogrande (5-shot) |81.22|
218
+ |GSM8k (5-shot) |69.90|
219
+