228
you are viewing a single comment's thread
view the rest of the comments
[-] symbolstumble@lemmy.zip 1 points 7 hours ago

I see what you're saying, my issue with this is the product is (as I understand) no more than an amalgam of its inputs. I do understand the similarity to human artists, where one's art is building from reference (either directly/indirectly/cumulatively). The difference here for me is that current models, don't/can't comprehend the meaning behind the components of their construction. They also don't or aren't able to add any additional meaning to what they produce. I'm not sure that makes much sense. What I'm trying to communicate is more of a feeling behind the art, which is either really difficult to describe, or I lack the words. Maybe you can help with your own thoughts/corrections?

That second paragraph makes perfect sense, especially tying in to the first sentence of your first paragraph. I wonder if it might be possible to escape the necessity for human produced data for training? That would certainly alleviate a lot of my concerns with the tech, especially when talking local.

[-] mindbleach@sh.itjust.works 1 points 4 hours ago* (last edited 4 hours ago)

Consider this image. It's full of blatant tells, like the bouncer becoming a real boy from the knees down, or the improbable stacking inside that trenchcoat. Yet it obviously conveys meaning in a clever way. You wouldn't commend whoever made it for their drawing skills, but the image transmits an idea from their brain to yours.

The model did not have to comprehend anything. That's the user's job. A person used the tool's ability to depict these visual elements, in order to communicate their own message.

If some guy spends days tweaking out the exact right combination of fifteen unforgivable fetishes, that amalgamation is his fault. You would not blame the computer for your immediate revulsion. It tried its best to draw a generic pretty lady in center frame. But that guy kept doodling a ball-gag onto Shrek until image-to-image got the leather strap right, and once he copy-pasted Frieren behind him, it just made her lighting match.

Neural networks are universal approximators, so you're always going to need human art to approximate human art. However, there are efforts to produce models using only public-domain, explicitly licensed, and/or bespoke examples. (Part of the 'do words matter' attitude is that several outspoken critics scoff at that anyway. 'Like that changes anything!' They'll complain about the source of the data, but when that's addressed, they don't actually care about the source of the data.)

Personally, though... I don't have a problem with using whatever's public. For properly published works, especially: so what if the chatbot read every book in the library? That's what libraries are for. And for images, the more they use, the less each one matters. If you show a billion drawings to an eight-gig model then every image contributes eight bytes. The word "contributes" is eleven.

this post was submitted on 08 Feb 2026
228 points (97.5% liked)

Programming

25415 readers
314 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 2 years ago
MODERATORS