Feedback

#3
by Clevyby - opened

Hello, in a moment, I'll give a few thoughts I had when I was using this model. In general, it was a good model albeit with a few faults.

Firstly though, here's my setup: I use latest staging sillytavern on my android with roleplay alpaca advanced formatting and no system prompt. I used Lewdiculous' Q5_K_M Imatrix gguf quant and use koboldcpp on free colab as backend. I also use a peculiar character card system wherein I put example dialogue in desc:

Screenshot_2024-03-26_105536.jpg

And I put all info from card into author note at depth 1:

Screenshot_2024-03-26_105542.jpg

Here's what I found during minimal testing:

  1. Prose is great: wording from the model's not too simplistic nor complicated in wording, I like that.

  2. Inconsistent Formatting: Formatting is consistent across all responses, though sometimes model mixes up formatting.

  3. Great understanding of semi-complicated author note: Model can grasp author note's info albeit there is a slight difficulty in grasping her character (For instance, for the character Victoria, model assumes her character to refer other people as 'humans' despite the fact she is supposed to be a former human forcibly converted into a furry so char's not supposed to use such terminology.)

  4. Decent personality grasping: While personality grasping from the model is decent, I feel the model needs improvement as such. For instance while testing said character card, the model did not accentuate the psychological struggle more clearly in 'inner thoughts' against char's embedded brainwashing despite being emphasized in author note and first message. In later messages, char's inner thoughts are read like a scientist's observation rather than an 18 year old tsundere's mind in conflict.

  5. Traces of gptism and weird wording: The classic 'shivers down your spine' and variations thereof, and wording such as 'tingles' looks strange.

  6. A bit of struggle in pov: Usually model's responses are in 3rd pov, but sometimes model mixes it up with 1st pov.

  7. Context Limit: As said by others, the model's context is limited to 8k, though it can be remedied by using 'alpha value' roping via oobabooga.

  8. Slight incoherence: probably sampling (Min_P: 0.05, Smoothing: 0.08) but sometimes, responses by model can become incoherent context wise, sometimes confusing user's responses with model's own.

  9. Good instruct following: It's a given.

Overall, very decent model. Just needs some improvement in generating responses and understanding responses.

Responses (Min_P: 0.05, Smoothing: 0.08):

Screenshot_2024-03-26_112435.jpg

Screenshot_2024-03-26_112945.jpg

Gptisms:

Screenshot_2024-03-26_110618.jpg

Screenshot_2024-03-26_111608.jpg

Screenshot_2024-03-26_113511.jpg

Context limit (This is 9k filled out of 10k):

Screenshot_2024-03-26_111845.jpg

Formatting errors:

Screenshot_2024-03-26_113250.jpg

Screenshot_2024-03-26_113750.jpg

Incoherence:

Screenshot_2024-03-26_113720.jpg

Response that reads like a scientific observation:

Screenshot_2024-03-26_114458.jpg

Screenshot_2024-03-26_114214.jpg

Sign up or log in to comment