xiaol's picture
Update README.md
c569e07
---
license: apache-2.0
datasets:
- Norquinal/claude_multiround_chat_30k
- ehartford/dolphin
- BAAI/COIG-PC
- Open-Orca/OpenOrca
- vikp/textbook_quality_programming
---
# RWKV v4 world 7B 65k context
This is the model to replace the old rwkv 65k claude model, with special token and lower learning rate to maintain model former abilities.
and trained a lots of English high quality textbooks and chinese novels with 65k context length.
using it with rwkv runner only need 16G vram.(https://github.com/josStorer/RWKV-Runner)
## contributor
[@KevinMr](https://huggingface.co/KevinMr)
[@Remixa](https://huggingface.co/Remixa)
## trainning details
https://wandb.ai/one-/one-rwkv-64k/runs/jn05hyc4
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/mpq2VrOaMZ_nXvV_yL-6o.png)
## Testcase
https://rwkv-next-web.ai-creator.net/ (temporary)
https://rwkv.ai-creator.net/risu
## how to use
use vocabs files in runner config
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/9V3J6uxaJESCC7WhIOD7p.png)