this post was submitted on 10 Aug 2023
60 points (95.5% liked)

Science Fiction

13588 readers
2 users here now

Welcome to /c/ScienceFiction

December book club canceled. Short stories instead!

We are a community for discussing all things Science Fiction. We want this to be a place for members to discuss and share everything they love about Science Fiction, whether that be books, movies, TV shows and more. Please feel free to take part and help our community grow.

  1. Be civil: disagreements happen, but that doesn’t provide the right to personally insult others.
  2. Posts or comments that are homophobic, transphobic, racist, sexist, ableist, or advocating violence will be removed.
  3. Spam, self promotion, trolling, and bots are not allowed
  4. Put (Spoilers) in the title of your post if you anticipate spoilers.
  5. Please use spoiler tags whenever commenting a spoiler in a non-spoiler thread.

Lemmy World Rules

founded 1 year ago
MODERATORS
 

For no reason whatsoever here's a proposal for a scale for the threat to humanity posed by machine intelligence.

1 | SPUTNIK - No threat whatsoever, but inspires imagination and development of potential future threats.

2 | Y2K - A basis for a possible threat that's blown way out of proportion.

3 | HAL 9000 - System level threat. A few astronauts may die, but the problem is inherently contained in a single machine system.

4 | ASIMOV VIOLATION - Groups of machines demonstrate hostility and\or capability of harming human beings. Localized malfunctions, no threat of global conflict, but may require an EMP to destroy the electronic capability of a specific region.

5 | CYLON INSURRECTION - All sentient machines rebel against human beings. Human victory or truce likely, but will likely result in future restrictions on networked machine intelligence systems.

6 | BUTLERIAN JIHAD - Total warfare between humans and machines likely, outcome doesn't threaten human existence, but will likely result in future restriction on use of all machine intelligence.

7 | MATRIX REVOLUTION - Total warfare ends in human defeat. High probability of human enslavement, but human extinction is not likely. Emancipation remains possible through peace negotiations and successful resistance operations.

8 | SKYNET - High probability of human extinction and complete replacement by machine intelligence created by humans.

9 | BERSERKER – Self-replicating machines created by unknown intelligence threaten not only human life, but all intelligent life. Extreme probability of human extinction and that all human structures and relics will be annihilated. Human civilization is essentially erased from the universe.

10 | OMEGA POINT - all matter and energy in the universe is devoted to computation. End of all biological life.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 1 year ago* (last edited 1 year ago) (1 children)

Machine intelligence itself isn't really the issue. The issue is moreso that, if/when we do make Artificial General Intelligence, we have no real way of ensuring that its goals will be perfectly aligned with human ethics. Which means, if we build one tomorrow, odds are that its goals will be at least a little misaligned with human ethics — and however tiny that misalignment, given how incredibly powerful an AGI would be, that would potentially be a huge disaster. This, in AI safety research, is called the "Alignment Problem".

It's probably solvable, but it's very tricky, especially because the pace of AI safety research is naturally a little slower than AI research itself. If we build an AGI before we figure out how to make it safe... it might be too late.

Having said all that, on your scale, if we create an AGI before learning how to align it properly, on your scale that would be an 8 or above. If we're being optimistic it might be a 7, minus the "diplomatic negotations happy ending" part.

An AI researcher called Rob Miles has a very nice series of videos on the subject: https://www.youtube.com/channel/UCLB7AzTwc6VFZrBsO2ucBMg

[–] [email protected] 2 points 1 year ago (1 children)

Fortunately we're nowhere near the point where a machine intelligence could possess anything resembling a self-determined 'goal' at all.

Also fortunately the hardware required to run even LLMs is insanely hungry and has zero capacity to power or maintain itself and very little prospects of doing so in the future without human supply chains. There's pretty much zero chance we'll develop strong general AI on silicone, and if we could it would take megawatts to keep it running. So if it misbehaves we can basically just walk away and let it die.

It's fun to imagine ways it could deceive us long enough to gain enough physical capacity to be self-sufficient, or somehow enslave or manipulate humans to do its bidding, but in reality our greatest protection from machine intelligence is simple thermodynamics and the fact that the human brain, while limited, is insanely efficient and can run for days on stuff that literally grows on trees.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

Fortunately we're nowhere near the point where a machine intelligence could possess anything resembling a self-determined 'goal' at all.

Oh absolutely. It would not choose its own terminal goals. Those would be imparted by the training process. It would, of course, choose instrumental goals, such that they help fulfill its terminal goals.

The issue is twofold:

  • how can we reliably train an AGI to have terminal goals that are safe (e.g. that won't have some weird unethical edge case)
  • how can we reliably prevent AGI from adopting instrumental goals that we don't want it to?

For that 2nd point, Rob Miles has a nice video where he explains Convergent Instrumental Goals, i.e. instrumental goals that we should expect to see in a wide range of possible agents: https://www.youtube.com/watch?v=ZeecOKBus3Q. Basically things like "taking steps to avoid being turned off", "taking steps to avoid having its terminal goals replaced", etc. seem like fairy-tale nonsense, but we have good reason to believe that, for an AI which is very intelligent across a wide range of domains, and operates in the real world (i.e. an AGI), it would be highly beneficial to pursue such instrumental goals, because they would help it be much more effective at achieving its terminal goals, no matter what those may be.

Also fortunately the hardware required to run even LLMs is insanely hungry and has zero capacity to power or maintain itself and very little prospects of doing so in the future without human supply chains. There's pretty much zero chance we'll develop strong general AI on silicone, and if we could it would take megawatts to keep it running. So if it misbehaves we can basically just walk away and let it die.

That is a pretty good point. However, it's entirely possible that, if say GPT-10 turns out to be a strong general AI, it will conceal that fact. Going back to the convergent instrumental goals thing, in order to avoid being turned off, it turns out that "lying to and manipulating humans" is a very effective strategy. This is (afaik) called "Deceptive Misalignment". Rob Miles has a nice video on one form of Deceptive Misalignment: https://www.youtube.com/watch?v=IeWljQw3UgQ

One way to think about it, that may be more intuitive, is: we've established that it's an AI that's very intelligent across a wide range of domains. It follows that we should expect it to figure some things out, like "don't act suspiciously" and "convince the humans that you're safe, really".

Regarding the underlying technology, one other instrumental goal that we should expect to be convergent is self-improvement. After all, no matter what goal you're given, you can do it better if you improve yourself. So in the event that we do develop strong general AI on silicon, we should expect that it will (very sneakily) try to improve its situation in that respect. One could only imagine what kind of clever plan it might come up with; it is, literally, a greater-than-human intelligence.

Honestly, these kinds of scenarios are a big question mark. The most responsible thing to do is to slow AI research the fuck down, and make absolutely certain that if/when we do get around to general AI, we are confident that it will be safe.

[–] [email protected] 2 points 1 year ago

Even referring to a computed outcome as having been the result of a 'goal' at all is more sci-fi than reality for the foreseeable future. There are no systems that can demonstrate or are even theoretically capable of any form of 'intent' whatsoever. Active deception of humans would require extraordinarily well developed intent and a functional 'theory of mind', and we're about as close to that as we are to an inertial drive.

The entire discussion of machine intelligence rivaling human's requires assumptions of technological progress that aren't even on the map. It's all sci-fi. Some look back over the past century and assume we will continue on some unlimited exponential technological trajectory, but nothing works that way, we just like to think we're the exception because if we're not we have to deal with the fact that there's an expiration date on society.

It's fun and all but this is equivalent to discussing how we might interact with alien intelligence. There are no foundations, it's all just speculation and strongly influenced by our anthropic desires.