399
submitted 1 month ago by [email protected] to c/[email protected]
you are viewing a single comment's thread
view the rest of the comments
[-] [email protected] 13 points 1 month ago* (last edited 1 month ago)

LLMs, in fact, have slop profiles (aka overused tokens/phrases) common to the family/company, often from “inbreeding” by training on their own output.

Sometimes you can tell if new model “stole” output from another company this way. For instance, Deepseek R1 is suspiciously similar to Google Gemini, heh.

This longform writing benchmark tries to test/measure this (click the I on each model for infographics):

https://eqbench.com/creative_writing_longform.html

As well as some some disparate attempts on GitHub (actually all from the eqbench dev): https://github.com/sam-paech/slop-forensics

https://github.com/sam-paech/antislop-vllm

this post was submitted on 11 May 2025
399 points (98.8% liked)

LinkedinLunatics

4873 readers
12 users here now

A place to post ridiculous posts from linkedIn.com

(Full transparency.. a mod for this sub happens to work there.. but that doesn't influence his moderation or laughter at a lot of posts.)

founded 2 years ago
MODERATORS