I tried the 120g hosted on huggingface. Worst than most smaller models at coding in J language. None that I've tried are great, but this was one of the worst at accepting corrections, and having the most errors per line. I'm not in a hurry to try their other models because of this.
I wonder if they updated the tool prompt in LM Studio
In case you don’t know, if you use tools LMS injects a bunch of shitty prompt that you can’t change, and it has specific examples like fetching the weather in Paris.
I have not checked that prompt since they added a deno based js sandbox, but I could imagine them adding a 1 shot example of a Fibonacci generator.
Mystery solved!
In a separate chat thread I had asked for a function that calculates fibonacci numbers in multithreaded python. Looks like the conversation memory was leaking into this one. Also, that one was addressed to a Qwen instance, and this one was to gpt-oss-20b.
I cleared all chats, started a fresh one, and switched to gpt-oss-20b and it responded properly without creating a separate python instance.
Thanks for the tip! Will check it out when back at desk. A bunch of us have been scratching our heads over this one.
Yeah I can’t check at the moment, but it has been an issue for me for a while.
I was very confused why it would fetch the weather for Paris randomly with some models, and it was even more confusing because I was learning MCP by writing a weather tool (so the LLMs were getting conflicting instructions).
LocalLLaMA
Welcome to LocalLLaMA! Here we discuss running and developing machine learning models at home. Lets explore cutting edge open source neural network technology together.
Get support from the community! Ask questions, share prompts, discuss benchmarks, get hyped at the latest and greatest model releases! Enjoy talking about our awesome hobby.
As ambassadors of the self-hosting machine learning community, we strive to support each other and share our enthusiasm in a positive constructive way.
Rules:
Rule 1 - No harassment or personal character attacks of community members. I.E no namecalling, no generalizing entire groups of people that make up our community, no baseless personal insults.
Rule 2 - No comparing artificial intelligence/machine learning models to cryptocurrency. I.E no comparing the usefulness of models to that of NFTs, no comparing the resource usage required to train a model is anything close to maintaining a blockchain/ mining for crypto, no implying its just a fad/bubble that will leave people with nothing of value when it burst.
Rule 3 - No comparing artificial intelligence/machine learning to simple text prediction algorithms. I.E statements such as "llms are basically just simple text predictions like what your phone keyboard autocorrect uses, and they're still using the same algorithms since <over 10 years ago>.
Rule 4 - No implying that models are devoid of purpose or potential for enriching peoples lives.