m8than commited on
Commit
ad7133e
1 Parent(s): e4ff26d

readme update

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -26,7 +26,7 @@ tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_c
26
 
27
  The following demonstrates the improvements from Eagle 7B to Finch 14B
28
 
29
- | | [Eagle 7B](https://huggingface.co/RWKV/v5-Eagle-7B-HF) | [Finch 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 14B](https://huggingface.co/RWKV/v6-Finch-14B-HF) |
30
  | --- | --- | --- | --- |
31
  | [ARC](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/arc) | 39.59% | 41.47% | 46.33% |
32
  | [HellaSwag](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/hellaswag) | 53.09% | 55.96% | 57.69% |
@@ -59,8 +59,8 @@ User: {instruction}
59
  Assistant:"""
60
 
61
 
62
- model = AutoModelForCausalLM.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True).to(torch.float32)
63
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True)
64
 
65
  text = "请介绍北京的旅游景点"
66
  prompt = generate_prompt(text)
@@ -115,8 +115,8 @@ User: {instruction}
115
  Assistant:"""
116
 
117
 
118
- model = AutoModelForCausalLM.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True, torch_dtype=torch.float16).to(0)
119
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True)
120
 
121
  text = "介绍一下大熊猫"
122
  prompt = generate_prompt(text)
@@ -162,8 +162,8 @@ User: {instruction}
162
 
163
  Assistant:"""
164
 
165
- model = AutoModelForCausalLM.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True).to(torch.float32)
166
- tokenizer = AutoTokenizer.from_pretrained("RWKV/v5-Eagle-7B-HF", trust_remote_code=True)
167
 
168
  texts = ["请介绍北京的旅游景点", "介绍一下大熊猫", "乌兰察布"]
169
  prompts = [generate_prompt(text) for text in texts]
 
26
 
27
  The following demonstrates the improvements from Eagle 7B to Finch 14B
28
 
29
+ | | [Eagle 7B](https://huggingface.co/RWKV/v6-Finch-14B-HF) | [Finch 7B](https://huggingface.co/RWKV/v6-Finch-7B-HF) | [Finch 14B](https://huggingface.co/RWKV/v6-Finch-14B-HF) |
30
  | --- | --- | --- | --- |
31
  | [ARC](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/arc) | 39.59% | 41.47% | 46.33% |
32
  | [HellaSwag](https://github.com/EleutherAI/lm-evaluation-harness/tree/main/lm_eval/tasks/hellaswag) | 53.09% | 55.96% | 57.69% |
 
59
  Assistant:"""
60
 
61
 
62
+ model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True).to(torch.float32)
63
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True)
64
 
65
  text = "请介绍北京的旅游景点"
66
  prompt = generate_prompt(text)
 
115
  Assistant:"""
116
 
117
 
118
+ model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True, torch_dtype=torch.float16).to(0)
119
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True)
120
 
121
  text = "介绍一下大熊猫"
122
  prompt = generate_prompt(text)
 
162
 
163
  Assistant:"""
164
 
165
+ model = AutoModelForCausalLM.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True).to(torch.float32)
166
+ tokenizer = AutoTokenizer.from_pretrained("RWKV/v6-Finch-14B-HF", trust_remote_code=True)
167
 
168
  texts = ["请介绍北京的旅游景点", "介绍一下大熊猫", "乌兰察布"]
169
  prompts = [generate_prompt(text) for text in texts]