view the rest of the comments
Games

Welcome to the largest gaming community on Lemmy! Discussion for all kinds of games. Video games, tabletop games, card games etc.
Rules
1. Submissions have to be related to games
Video games, tabletop, or otherwise. Posts not related to games will be deleted.
This community is focused on games, of all kinds. Any news item or discussion should be related to gaming in some way.
2. No bigotry or harassment, be civil
No bigotry, hardline stance. Try not to get too heated when entering into a discussion or debate.
We are here to talk and discuss about one of our passions, not fight or be exposed to hate. Posts or responses that are hateful will be deleted to keep the atmosphere good. If repeatedly violated, not only will the comment be deleted but a ban will be handed out as well. We judge each case individually.
3. No excessive self-promotion
Try to keep it to 10% self-promotion / 90% other stuff in your post history.
This is to prevent people from posting for the sole purpose of promoting their own website or social media account.
4. Stay on-topic; no memes, funny videos, giveaways, reposts, or low-effort posts
This community is mostly for discussion and news. Remember to search for the thing you're submitting before posting to see if it's already been posted.
We want to keep the quality of posts high. Therefore, memes, funny videos, low-effort posts and reposts are not allowed. We prohibit giveaways because we cannot be sure that the person holding the giveaway will actually do what they promise.
5. Mark Spoilers and NSFW
Make sure to mark your stuff or it may be removed.
No one wants to be spoiled. Therefore, always mark spoilers. Similarly mark NSFW, in case anyone is browsing in a public space or at work.
6. No linking to piracy
Don't share it here, there are other places to find it. Discussion of piracy is fine.
We don't want us moderators or the admins of lemmy.world to get in trouble for linking to piracy. Therefore, any link to piracy will be removed. Discussion of it is of course allowed.
Authorized Regular Threads
Related communities
PM a mod to add your own
Video games
Generic
- !gaming@Lemmy.world: Our sister community, focused on PC and console gaming. Meme are allowed.
- !photomode@feddit.uk: For all your screenshots needs, to share your love for games graphics.
- !vgmusic@lemmy.world: A community to share your love for video games music
Help and suggestions
By platform
By type
- !AutomationGames@lemmy.zip
- !Incremental_Games@incremental.social
- !LifeSimulation@lemmy.world
- !CityBuilders@sh.itjust.works
- !CozyGames@Lemmy.world
- !CRPG@lemmy.world
- !horror_games@piefed.world
- !OtomeGames@ani.social
- !Shmups@lemmus.org
- !space_games@piefed.world
- !strategy_games@piefed.world
- !turnbasedstrategy@piefed.world
- !tycoon@lemmy.world
- !VisualNovels@ani.social
By games
- !Baldurs_Gate_3@lemmy.world
- !Cities_Skylines@lemmy.world
- !CassetteBeasts@Lemmy.world
- !Fallout@lemmy.world
- !FinalFantasyXIV@lemmy.world
- !Minecraft@Lemmy.world
- !NoMansSky@lemmy.world
- !Palia@Lemmy.world
- !Pokemon@lemm.ee
- !Silksong@indie-ver.se
- !Skyrim@lemmy.world
- !StardewValley@lemm.ee
- !Subnautica2@Lemmy.world
- !WorkersAndResources@lemmy.world
Language specific
- !JeuxVideo@jlai.lu: French
I mean, I get if you wanna use AI for that, it's your project, it's free, you're a volunteer, etc. I'm just not sure I like the idea that they're obscuring what AI was involved with. I imagine it was done to reduce constant arguments about it, but I'd still prefer transparency.
I tried fitting AI into my workloads just as an experiment and failed. It'll frequently reference APIs that don't even exist or over engineer the shit out of something could be written in just a few lines of code. Often it would be a combo of the two.
Yeah I mean. It's not like AI can think. It's just a glorified text predictor, the same you have on your phone keyboard
It's like having an idiot employee that works for free. Depending on how you manage them, that employee can either do work to benefit you or just get in your way.
Only it's not free. If you run it in the cloud, it's heavily subsidized and proactively destroying the planet, and if you run it at home, you're still using a lot of increasingly unaffordable power, and if you want something smarter than the average American politician, the upfront investment is still very significant.
Yeah I'm not buying the "proactively destroying the planet" angle. I'd imagine there's a lot of misinformation around AI, given that the products surrounding it are mostly Western, like vaccines...
Vaccines are misinformation? What.
Not even free, just cheaper than an actual employee for now, but greed is inevitable and AI is computationally expensive, it's only a matter of time before these AI companies start cranking up the prices.
You might genuinely be using it wrong.
At work we have a big push to use Claude, but as a tool and not a developer replacement. And it's working pretty damn well when properly setup.
Mostly using Claude Sonnet 4.6 with Claude Code. It's important to run /init and check the output, that will produce a CLAUDE.md file that describes your project (which always gets added to your context).
Important: Review everything the AI writes, this is not a hands-off process. For bigger changes use the planning mode and split tasks up, the smaller the task the better the output.
Claude Code automatically uses subagents to fetch information, e.g. API documentation. Nowadays it's extremely rare that it hallucinates something that doesn't exist. It might use outdated info and need a nudge, like after the recent upgrade to .NET 10 (But just adding that info to the project context file is enough).
Agreed, I don't understand people not even giving it a chance. They try it for five minutes, it doesn't do exactly what they want, they give up on it, and shout how shit it is.
Meanwhile, I put the work in, see it do amazing shit after figuring out the basics of how the tech works, write rules and skills for it, have it figure out complex problems, etc.
It's like handing your 90-year-old grandpa the Internet, and they don't know what the fuck to do with it. It's so infuriating.
Probably because, like your 90-year-old grandpa with the Internet, you have to know how to use the search engine. You have to know how to communicate ideas to an LLM, in detail, with fucking context, not just "me needs problem solvey, go do fix thing!"
Just yesterday I had one of those moments of grace that are becoming commonplace.
Basically I have to migrate a service from a n8n workflow to an actual nodejs server for performance reasons. I spent 15 minutes carefully scoping the migration, telling it exactly what tools to use and code style to adopt. Gave it the original brief and access to the n8n workflows.
The whole thing was done in 4 minutes and 30 seconds. It even noticed a bug which has been in production unnoticed for the past year. Gave me some good documentation on how to setup the Google service account, the kind of memory usage to expect so I can dimension the instant accordingly. Another five minutes and I had a whole test suite with decent coverage. I had negotiated with the client that it would take around a week, well that was the under promise of the year...
People who go around telling it doesn't work are incompetent, out of their minds or straight up lying.
Most people on Lemmy probably haven’t given it a single minute let alone 5 minutes.
It's not really that simple. Yes, it's a great tool when it works, but in the end it boils down to being a text prediction machine.
So a nice helper to throw shit at, but I trust the output as much as a random Stackoverflow reply with no votes :)
I feel like there needs to be a dedicated post (and I don't want to write it, but maybe I eventually will) that outlines what a model really is. It is not just a statistical text prediction machine unless you are being so loose with the definition of "statistical" that it doesn't even mean anything anymore.
A decent example of a statistical text prediction machine is the middle word suggested by your phone when you're using the keyboard. An LLM is not that.
In the most general terms, this kind of language model tokenizes a corpus of text based on a vocabulary (which is probably more than just the words in the dictionary), uses an embedding model to translate these tokens into a vector of semantic "meaning" which minimized loss in a bidirectional encoding (probably), that is then trained against a rubric for one or more topic area questions, retrained for instruction and explainability, retrained with reinforcement learning and human feedback to provide guardrails, and retrained again to make use of supplemental materials not part of the original training corpus (resource augmented generation), then distilled, then probably scaled and fine tuned against topic areas of choice (like coding or Korean or whatever) and maybe THEN made available to people to use. There are generally more parts to curriculum learning even than that but it's a representative-ish start.
My point being that, yes, it would be nuts to pose ANY question to a predictor that says "with 84% probability, the word that is most likely follows 'I really like' is 'gooning' on reddit", but even Grok is wildly more sophisticated than that and Grok is terrible.
Edit: And also I really like your take at the start of this thread: user error is a pretty huge problem in this space.
The training is sophisticated, but inference is unfortunately really a text prediction machine. Technically token prediction, but you get the idea.
For every single token/word. You input your system prompt, context, user input, then the output starts.
The
Feed the entire context back in and add the reply "The" at the end.
The capital
Feed everything in again with "The capital"
The capital of
Feed everything in again...
The capital of Austria
...
It literally works like that, which sounds crazy :)
The only control you as a user can have is the sampling, like temperature, top-k and so on. But that's just to soften and randomize how deterministic the model is.
Edit: I should add that tool and subagent use makes this approach a bit more powerful nowadays. But it all boils down to text prediction again. Even the tools are described per text for what they are for.
And we're barely smarter than a bunch of monkeys throw piles of shit at each other. Being reductive about its origins doesn't really explain anything.
Yeah, but that's why there's unit tests. Let it run its own tests and solve its own bugs. How many mistakes have you or I made because we hate making unit tests? At least the LLM has no problems writing the tests, after you know it works.
I've had better luck with using it in a TDD style. "Write a test for this issue, watch it fail, then make it pass."
At a minimum, the agent should be compiling the code and running tests before handing things back to you. "It references non-existent APIs" isn't a modern problem.
I don't know what they are using cause all agents routinely do that. I suspect they are fibbing or tested things out in 2024 and never updated their opinion.
The symptoms you describe are caused by bad prompting. If an AI is providing over-complicated solutions, 9 times out of 10 it's because you didn't constrain your problem enough. If it's referencing tools that don't exist, then you either haven't specified which tools are acceptable or you haven't provided the context required for it to find the tools. You may also be wanting too much out of AI. You can't expect it to do everything for you. You still have to do almost all the thinking and engineering if you want a quality project - the AI is just there to write the code. Sure, you can use an AI to help you learn how to be a better engineer, but AIs typically don't make good high-level decisions. Treat AI like an intern, not like a principal engineer.
"It can't be that stupid, you must be prompting it wrong."
It's not about stupid or smart. It's a tool, not a person. If you don't get the same results that other people get with the same tool, then what could possibly be the problem other than how the person is using the tool?
"it's your fault that it just made up tools that don't exist" is a bold statement, bro.
The junior analogy comes to mind. If you hire a fresh face and they ship code that doesn't work, it's definitely on you, bro.
No, it's not. It doesn't have intention. It's literally just a tool. If you don't get the results you expect with a tool when other people do get those results, then the problem isn't the tool.
If the tool can't be consistent in it's output, it's not a reliable or worthwhile tool to use.
There is such a thing as a bad tool.
Good thing that's not the case then.
I create custom embedded devices with displays and I've found it very useful for laying things out. Like asking it to take secondly wind speed and direction updates and build a Wind Rose out of it, with colored sections in each petal denoting the speed... it makes mistakes but then you just go back and reiterate on those mistakes. I'm able to do so much more, so much faster.
I had the same experience. Asked a local LLM about using sole Qt Wayland stuff for keyboard input, a the only documentation was the official one (which wasn't a lot for a noob), no.examples of it being used online, and with all my attempts at making it work failing. it hallucinated some functions that didn't exist, even when I let it do web search (NOT via my browser). This was a few years ago.
That's 50 years in LLM terms. You might as well have been banging two rocks together.
Yeah, now we're in the iron age!
...Where we get to bang two ingots together
I expect because it wasn't a user - just a random passer by throwing stones on their own personal crusade. The project only has two major contributors who are now being harassed in the issues for the choices they make about how to run their project.
Someone might fork it and continue with pure artisanal human crafted code but such forks tend to die off in the long run.
Considering the amount of damage AI has done to well-funded projects like Windows and Amazon's services, I agree with this entirely. It might be crucial to help fix bigger issues down the line.
I'm the opposite. Its weird to me for someone to add an AI as a co author. Submit it as normal.
It's mostly not a thing developers do. It's a thing the tools themselves do when asked to make a commit.