[-] yogthos@lemmy.ml 3 points 37 minutes ago

vassals will vassal

[-] yogthos@lemmy.ml 3 points 1 hour ago

I expect so as well, and China also has a lot of incentive to invest in alternative substrates since they're behind on silicon. If one of these moonshot projects they're pursuing delivers that would make current silicon chips look like vacuum tubes by comparison.

20
submitted 1 hour ago by yogthos@lemmy.ml to c/technology@lemmy.ml
12
submitted 2 hours ago by yogthos@lemmy.ml to c/worldnews@lemmy.ml
31
submitted 2 hours ago by yogthos@lemmy.ml to c/worldnews@lemmy.ml
[-] yogthos@lemmy.ml 2 points 2 hours ago

There are a few different tracks here. One is software optimizations where models require less energy to use. That's been moving really fast over the past few years, and there are still a lot of papers that haven't been integrated into production systems that are really promising.

Another track is hardware architecture where the substrate stays the same, but chip design improves. A general example of this is SoC architecture like M series from Apple of Kirin 9000 from Huawei. The architecture eliminates the memory bus which is one of the main bottlenecks, and RISC instruction set facilitates parallelism much better than SISC. A more specific example would be ASIC chips like what Taalas is making which print the model directly on the chip.

And the last track is the one you mention with using a more efficient substrate. Notably this will directly benefit from the other two tracks as well. Whatever software and hardware architecture improvements people come up with, will directly apply to chips made out of graphene or other materials.

[-] yogthos@lemmy.ml 5 points 2 hours ago

That's very cool!

[-] yogthos@lemmy.ml 2 points 2 hours ago* (last edited 2 hours ago)

Some people care about the continued survival of humanity and look beyond their own nose. Since avoiding extinction requires avoiding barbarism, I don't see any conflict here. 🤷

[-] yogthos@lemmy.ml 4 points 2 hours ago

Right, GDR wasn't fascist.

10
submitted 2 hours ago by yogthos@lemmy.ml to c/science@lemmy.ml
7
submitted 3 hours ago by yogthos@lemmy.ml to c/technology@lemmy.ml
[-] yogthos@lemmy.ml 7 points 3 hours ago* (last edited 3 hours ago)

I expect that software will continue to get optimized, and we'll see new algorithms that are more efficient than what people are doing currently. However, it's possible we'll start seeing hardware specifically built for models as well. For example, there's already a startup that uses ASIC chips to print the model directly to the chip. Since each transistor acts as a state, it doesn't need DRAM and the whole chip requires a small amount of SRAM which isn't in short supply right now https://www.anuragk.com/blog/posts/Taalas.html

The limitation with this approach is that the chip is made for a specific model, but that's not really that different from the way regular chips work either. You buy a chip and if it does what you need, it keeps working. When new models come out, new chips get printed, and if you need the new capabilities then you upgrade.

You can see how absurdly fast their hardware version of llama 3 is here https://chatjimmy.ai/

9
submitted 3 hours ago by yogthos@lemmy.ml to c/science@lemmy.ml
13
submitted 3 hours ago by yogthos@lemmy.ml to c/technology@lemmy.ml
12
submitted 4 hours ago by yogthos@lemmy.ml to c/science@lemmy.ml
48
submitted 4 hours ago by yogthos@lemmy.ml to c/privacy@lemmy.ml
34
submitted 4 hours ago by yogthos@lemmy.ml to c/privacy@lemmy.ml
7
submitted 4 hours ago by yogthos@lemmy.ml to c/technology@lemmy.ml
[-] yogthos@lemmy.ml 18 points 4 hours ago* (last edited 4 hours ago)

I think by the time AI becomes efficient enough to be profitable, it's going to be efficient enough to run locally and the whole AI as a service business model is going to collapse. We're basically in the mainframe era of AI right now, and we've seen this happen with many technologies before. There's no reason to think this case will be different.

Just to give you an idea of how fast this stuff is moving. Qwen 3.6 was just released and can be run on a high end laptop, it outperforms Qwen 3.5 from February which required a commercial grade server to run. https://qwen.ai/blog?id=qwen3.6-27b

[-] yogthos@lemmy.ml 3 points 4 hours ago* (last edited 3 hours ago)

Every thread about China's energy production will inevitably have a comment like this. I really wonder, are you people just trolling or you're just genuinely that ignorant?

[-] yogthos@lemmy.ml 5 points 17 hours ago

You'd think it would be obvious that public services exist because they're necessary rather than to bring in profit, yet here we are.

39
submitted 19 hours ago by yogthos@lemmy.ml to c/usa@lemmy.ml
[-] yogthos@lemmy.ml 22 points 1 day ago

Nobody is judging people by their native language here. What's being judged is the fact that Germany has conducted a genocide during WW2 and is now actively facilitating and supporting a genocide in Gaza. Go do your artless trolling elsewhere little fash.

view more: next ›

yogthos

0 post score
0 comment score
joined 6 years ago
MODERATOR OF