Perfect Storyteller
Hi Steelskull! Decided to leave feedback, since the model has very intriguing replies. I used a quantized GGUF 5_K_M version. Mostly tested with Koboldcpp v1.64 but also v1.64.1 and lmstudio v0.2.22. It should also be said that I used different sys prompts, from my own to something more standard.
- Side notes:
- Long-form writing indeed. During RP somewhere around 250-400+ tokens per reply, but the length is justified by the level of details. For those who love to read, the narrative it provides is really enjoyable, or use TTS.
- Narrative focused: The level of narrative quality is impressive. It is able to immerse the user in the scene itself, in the moment, not only the character's actions in some bubble space. Probably one of the best among all smaller models I've tried so far. This model doesn't shy away from vividly describing and interacting with surroundings, which slow down the pacing and creates a sense of real time interaction. Even with that amount of text, it doesn't hallucinate logic, although some small details may seem strange, maybe due lower quants or untuned ST settings, but the fact that the model provides this small details is impressive.
- It works better when it has enough context from the user input to work with and not 15 t of input to generate 300-400 t of output. It can be shortened slightly by adding to sys prompt something like "Prioritize exclusively responses that are VERY SHORT in length." or something similar, to make it around 180-260 tokens (with the characters that require shorter messages).
- The model feels really smart enough and behaves realistically, without exaggerated emotions... which some may like and some may not.
- It may require markdown editing multiple times when mixing both asterisks with quotes. But separately (only * or "), it mostly works fine. Works best with quotes.
- The model has a wide range of behaviors and able to be violent depending on the character. So it doesn't stay only on positive traits, like some other models do, although it also isn't drawn to unwarranted cruelty.
- Sensitive to symbols inside system prompt; other users may not encounter this. It doesn't like when sys prompt uses [ ] and may finish reply with something like '....] and then fake link'. Or use of ``` with some OOC thoughts; with literally no reason. For now this issues encountered only during the first message tho.
- Seems confident with simple math or logic puzzles.
- Since I tested everything else I decided to test erp too. I would say it was very descriptive and the model preferred more literature language than vulgar.
- Technical issue: Almost with each message and any kind of character, it generates EOS
</s>
at the end of reply. Yeah, 'almost', because sometimes it doesn't. After that, it stops generation, and as we know ST hides EOS, so not a big deal, but still.
- I have attached a screenshot with examples that showcases its behavior with different characters.
Agoness (meant to be playful/sadistic) - superb
Seraphina (tender/supportive) - superb
Shiroha (Blind/insecure) - Hmm, quite good, I believe I envisioned her as more stubborn, with some violent outbursts stemming from insecurities.
Nexa (pragmatic and mocking motivator) - superb, good reasoning and not only humiliation or motivational quotes.
I also tried Cindy Aurum from FFXV that intended to have like 70-150 tokens per reply, but the model doesn't manage to keep short messages with any prompt. As a simple 40t assistant in lmstudio_v0.2.22 when I asked a math questions it structured replies in a weird way making a lot of new lines, and didn't worked well, maybe it was trying to stretch its replies, dunno; and again backticks during the first message.
- Strengths: reasoning, descriptive, knowledgeable, focused, creative, good storyteller. The vivid description of the surroundings and not only the character's actions, is what sets this model apart from others I tried so far. Slow paced approach. Knowing that this model works best with long replies, I would consider this as a strong side.
Weaknesses: Well, I will not critique any aspect of its behavior since everyone has their own taste and I like its adequacy and believable behavior. EOS at the end of the message is the main technical issue. Tested with a standard L3 prompt.
In conclusion, this model has exceptional story writing capabilities and is interesting to talk with. I've added it to my personal list of the top models. Thank you for making this model, wish you luck with your next models and MoE merges!
Thank you very much for the extremely detailed feedback! Im very glad to see you enjoying the model!
I have noticed the < s / > error, it is kind of on and off at times. I think its an issue that can be fixed with some tokenizer modifications but I am unsure at the moment. May end up being a "Next version" Problem.
Thanks Again! π
maybe this have something to do with the formatting? IDK ... https://www.reddit.com/r/LocalLLaMA/comments/1cltac3/part3_cause_to_issue_found_possible_bug_llama3/
maybe this have something to do with the formatting? IDK ... https://www.reddit.com/r/LocalLLaMA/comments/1cltac3/part3_cause_to_issue_found_possible_bug_llama3/
sadly I don't think so. When the model was created, the tokenizer didnt take on the stop token < s / > , normally I would modify the tokenizer to fix this but the GGUFs have already been made by others. Im looking into options as i may have a way to fix it
np, it's actually enjoyable to leave positive feedback when you come across something unique that impresses you, and when there is something to talk about. Glad you know where the issue might be, hidden eos is not a big deal compared to advantages. Thank you for taking time to make your models!
Came to Arcania looking for a smaller model to run in place of Astoria (per
@Steelskull
's recommendation) and I am not let down. I can run this on my system, it's reasonably fast, and the results are still one of the best models I've messed with so far. Very good at following instructions, both explicit and more subtle steeing, keeps track of details, obeys lorebook/world info well.
Only issue is the random symbol vomit at the end of it's responses sometimes, but that's easily overlooked/edited out.
Steel is the best.
Came to Arcania looking for a smaller model to run in place of Astoria (per @Steelskull 's recommendation) and I am not let down. I can run this on my system, it's reasonably fast, and the results are still one of the best models I've messed with so far. Very good at following instructions, both explicit and more subtle steeing, keeps track of details, obeys lorebook/world info well.
Only issue is the random symbol vomit at the end of it's responses sometimes, but that's easily overlooked/edited out.
Steel is the best.
Thank you very much! π
As for the </s>
issue - could it just be one of the source models was using the Mistral prompt formatting? Looks like <s>[INST] Instruction [/INST] Model answer</s>[INST] Follow-up instruction [/INST]