Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,8 @@ datasets:
|
|
12 |
|
13 |
# Humanish-RP-Llama-3.1-8B
|
14 |
|
|
|
|
|
15 |
A DPO-tuned Llama-3.1 to behave more "humanish", i.e., avoiding all the AI assistant slop. It also works for role-play (RP). To achieve this, the model was fine-tuned over a series of datasets:
|
16 |
* General conversations from Claude Opus
|
17 |
* `Undi95/Weyaxi-humanish-dpo-project-noemoji`, to make the model react as a human, rejecting assistant-like or too neutral responses.
|
|
|
12 |
|
13 |
# Humanish-RP-Llama-3.1-8B
|
14 |
|
15 |
+
![image/webp](https://cdn-uploads.huggingface.co/production/uploads/5fad8602b8423e1d80b8a965/VPwtjS3BtjEEEq7ck4kAQ.webp)
|
16 |
+
|
17 |
A DPO-tuned Llama-3.1 to behave more "humanish", i.e., avoiding all the AI assistant slop. It also works for role-play (RP). To achieve this, the model was fine-tuned over a series of datasets:
|
18 |
* General conversations from Claude Opus
|
19 |
* `Undi95/Weyaxi-humanish-dpo-project-noemoji`, to make the model react as a human, rejecting assistant-like or too neutral responses.
|