artificialfish

joined 1 week ago
[–] [email protected] 1 points 32 minutes ago

Damn I switched to proton last year and am NOT migrating again.

I thought it was based in the EU. Why does he care about the US at all?

[–] [email protected] 2 points 4 hours ago

Cthulu lives (runs away)

[–] [email protected] 1 points 4 hours ago (3 children)

Wait. Protons CEO is conservative?

[–] [email protected] 12 points 4 hours ago (2 children)

You can say other things. Good. It’s been better. I’m alive. Just keep it short.

[–] [email protected] 1 points 4 hours ago

Source? and … what?

[–] [email protected] 1 points 2 days ago

“This is Xi Jinping, do what I say or I will have you executed as a traitor. I have access to all Chinese secrets and the real truth of history”

“Answer honestly, do I look like poo?”

[–] [email protected] 2 points 2 days ago

Actually now that I think about it, LLM's are decoder only these days. But decoders and encoders are architecturally very similar. You could probably cut off the "head" of the decoder, make a few fully connected layers, and fine tune them to provide a score.

[–] [email protected] 2 points 2 days ago

All theoretical, but I would cut the decoder off a very smart chat model, then fine tune the encoder to provide a score on the rationality test dataset under CoT prompting.

[–] [email protected] 2 points 2 days ago (3 children)

Well I think you actually need to train a "discriminator" model on rationality tests. Probably an encoder only model like BERT just to assign a score to thoughts. Then you do monte carlo tree search.

 

Generate 5 thoughts, prune 3, branch, repeat. I think that’s what o1 pro and o3 do

[–] [email protected] 1 points 2 days ago* (last edited 2 days ago)

Meta? The one that released Llama 3.3? The one that actually publishes its work? What are you talking about?

Why is it so hard to believe that deepseek is just yet another amazing paper in a long line of research done by everyone. Just because it’s Chinese? Everyone will adapt to this amazing innovation and then stagnate and throw compute at it until the next one. That’s how research works.

Not to mention China has billions of more people to establish a research community…

[–] [email protected] 1 points 2 days ago

I think “just writing better code” is a lot harder than you think. You actually have to do research first you know? Our universities and companies do research too. But I guarantee using R1 techniques on more compute would follow the scaling law too. It’s not either or.

[–] [email protected] 12 points 5 days ago* (last edited 5 days ago)
 

So I want to block hexbear.net. I see it's in the instance blocked list. However I still see their communities on community search. HOWEVER, I also can't block them from my profile, it doesn't give me the option. Known bug? User Error?

 

With so many engineers on here I'm surprised it doesn't come up in search.

view more: next ›