RakutenAI-7B: Extending Large Language Models for Japanese
Abstract
We introduce RakutenAI-7B, a suite of Japanese-oriented large language models that achieve the best performance on the Japanese LM Harness benchmarks among the open 7B models. Along with the foundation model, we release instruction- and chat-tuned models, RakutenAI-7B-instruct and RakutenAI-7B-chat respectively, under the Apache 2.0 license.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Breeze-7B Technical Report (2024)
- Nemotron-4 15B Technical Report (2024)
- H2O-Danube-1.8B Technical Report (2024)
- LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation (2024)
- SaulLM-7B: A pioneering Large Language Model for Law (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Seem there's not pretty much details in retrofitting process... not sure is this mean they did the standard practice on doing this or just they did not disclosed it. One clear thing about is that they did use huge scale of filtered japanese corpus (~175 B tokens).
Models citing this paper 7
Browse 7 models citing this paperDatasets citing this paper 0
No dataset linking this paper