view the rest of the comments
Perchance - Create a Random Text Generator
⚄︎ Perchance
This is a Lemmy Community for perchance.org, a platform for sharing and creating random text generators.
Feel free to ask for help, share your generators, and start friendly discussions at your leisure :)
This community is mainly for discussions between those who are building generators. For discussions about using generators, especially the popular AI ones, the community-led Casual Perchance forum is likely a more appropriate venue.
See this post for the Complete Guide to Posting Here on the Community!
Rules
1. Please follow the Lemmy.World instance rules.
- The full rules are posted here: (https://legal.lemmy.world/)
- User Rules: (https://legal.lemmy.world/fair-use/)
2. Be kind and friendly.
- Please be kind to others on this community (and also in general), and remember that for many people Perchance is their first experience with coding. We have members for whom English is not their first language, so please be take that into account too :)
3. Be thankful to those who try to help you.
- If you ask a question and someone has made a effort to help you out, please remember to be thankful! Even if they don't manage to help you solve your problem - remember that they're spending time out of their day to try to help a stranger :)
4. Only post about stuff related to perchance.
- Please only post about perchance related stuff like generators on it, bugs, and the site.
5. Refrain from requesting Prompts for the AI Tools.
- We would like to ask to refrain from posting here needing help specifically with prompting/achieving certain results with the AI plugins (
text-to-image-pluginandai-text-plugin) e.g. "What is the good prompt for X?", "How to achieve X with Y generator?" - See Perchance AI FAQ for FAQ about the AI tools.
- You can ask for help with prompting at the 'sister' community Casual Perchance, which is for more casual discussions.
- We will still be helping/answering questions about the plugins as long as it is related to building generators with them.
6. Search through the Community Before Posting.
- Please Search through the Community Posts here (and on Reddit) before posting to see if what you will post has similar post/already been posted.
I thought I was imagining things, but since others seem to be doing better, I guess that the update really improved the model then! That's awesome
From my side, at least two things have improved: the English no longer decays into caveman speak, and the head-start is infinitely easier with minimal directions to the model. Also, some contradicting descriptions tend to work better. This all is actually a great improvement, but I'd be lying if I'd say that on my side I tested them thoroughly.
Something I tried as a quick test was to check how the model reacts with long logs and... yep, it still get stuck and running in circles due to weave patterns that repeat ad nauseam. It may be me having bad samples, but problems are still lingering past the 200kB, heavy past the 500kB mark, and unbearable on the 1Mb mark. By this I just mean having to deal with unsticking the LLM by editing heavily, not that it is impossible to continue. If someone has a long log that is fluid, please share what conditions allow for it.
But yeah, Basti0n is right! There was indeed a notorious improvement even if we are not there yet. Maybe there is future for DeepSeek after all!
Hello my friend! Thank you for your comment and your opinion! I had also checked out the post you shared earlier, and I mentioned there as well that it was extremely helpful. So thank you again for that too.
As for the model’s current state of course it’s not perfect or flawless. I want to clarify that so it’s not misunderstood. It still needs improvement, and it should be improved. But if we think back to when the model first came out with all those hallucinations, weird nonsense characters appearing out of nowhere, not even getting the character’s name right, and until just a few weeks ago barely being able to form proper sentences its current state is, for me, real progress. And I just wanted to highlight that. So this post is more like saying to the dev, “Whatever you’re doing, you’re on the right track (for now).”
The repetition issue still continues, especially as the topic or story goes on. Based on my own observations (I’m talking about ACC here), its creativity drops after about every 10 messages of three paragraphs each, and it keeps declining. That initial creativity and originality start to disappear. Those parts still definitely need improvement. And I still strongly believe the context window needs to be increased significantly.
But other than that, considering the version it started with and the awful state it was in even just a few weeks ago, there has been progress.
Edit: Today I tried again and I think the quality has dropped again. While I was writing my post, it was giving high-quality answers
You are awesome, thank you for your work, I love perchance. With the AI Chat, I can not say that it's been an improvement. The model we had a few months ago was really good, and th new one seems to insist on making people, such as lovers, strangely cold or snappy/angry. I have also not noticed any improvement in the ai chat model but maybe its just me. Just to give feedback, cheers.
I guess that the drop is the luck of the draw, my friend! Wrangling an LLM is very tricky so as the dev said, we are in for a bumpy ride for the next couple of months!🤣
But you are on point with the diagnostic. I use more AI Chat, so I can't speak much of the particularities of ACC, at least in AI Chat the decay seems to be at the 20-30 input, and then spaced three paragraphs as you said. It could be due to the raw input in ACC is significantly longer than the one in AI Chat, but then you compare it to AI RPG where the raw input is even shorter, and the decay happens even in the fifth input and sticks forever. It's hard to tell, and most of the times it's actually due to what is being "played" as the moment, as like with the old LLM, some topics and write-styles were easier than others.
Just for personal experience, personally the current model "peaked" two times: right after release when the "ultra violencia" mode was patched two months ago, and then yesterday, but it could have been the luck of the draw too, so it could be that the waters are still being tested to know how to lead the model in a proper way without falling into its pitfalls. But hey! At least we know that the project is not being abandoned, and that some stuff that we thought (at least me personally) was impossible, may be actually possible!
Also, something that most people don't realize, is how hard is to debug this, because while I keep referencing numbers of log sizes and all, I don't know the rest of the people that use this service, but due to time and since I treat this just as a game and not in any sort of "professional" usage, the most I can produce a day is just 30kB, 70kB if I'm lucky and locked in playing a run, so imagine how rough it would be to the dev to try going past 1Mb in different scenarios while maintaining the site and trying to wrangle the LLM. Personally, I wouldn't even try! 🤣
I know that many of the people complaining on the new model latch on it being unable to run "comfort scenarios" which... in some runs I had absolutely no problem! (Except of course the issue of repetition and running in circles, which is still universal) So what I think would be an excellent exercise, as well as a proper debug tool to know when and how things break with the current LLM is to try different runs in different topics and check what conditions in particular make things break and when (with when I mean after what input, or log size), since I have the feeling that as now, the LLM breaks faster in certain contexts and decides to stay focused and creative with one particular style, that could point to bias in the training (BTW, is not the violent ones, I tried and those break like paper very quick).
But overall, posts and threads like this do aid a lot. Input, positive or negative, is always good so long it is supported and not just "all is perfect, lol" or "all is crap, lmao". Otherwise, how to know what is working or not? 😅