zlacker

[return to "Stargate Project: SoftBank, OpenAI, Oracle, MGX to build data centers"]
1. belter+X71[view] [source] 2025-01-22 07:56:09
>>tedsan+(OP)
This is a Military project. Have no doubts about it.
◧◩
2. Gud+ie1[view] [source] 2025-01-22 08:55:57
>>belter+X71
This is a money making scheme.
◧◩◪
3. arisAl+we1[view] [source] 2025-01-22 08:57:52
>>Gud+ie1
This has cosmological significance if it leads to superintelligence
◧◩◪◨
4. Cthulh+al1[view] [source] 2025-01-22 09:56:46
>>arisAl+we1
It won't unless there's another (r)evolution in the underlying technology / science / algorithms, at this point scaling up just means they use bigger datasets or more iterations, but it's more finetuning and improving the existing output then coming up with a next generation / superintelligence.
◧◩◪◨⬒
5. Fillig+3o1[view] [source] 2025-01-22 10:24:10
>>Cthulh+al1
Okay, but let’s be pessimistic for a moment. What can we do if that revolution does happen, and they’re close to AGI?

I don’t believe the control problem is solved, but I’m not sure it would matter if it is.

◧◩◪◨⬒⬓
6. ForHac+zo1[view] [source] 2025-01-22 10:29:33
>>Fillig+3o1
Being pessimistic, how come no human supergeniuses ever took over the world? Why didn't Leibniz make everyone else into his slaves?

I don't even understand what the proposed mechanism for "rouge AI enslaves humanity" is. It's scifi (and not hard scifi) as far as I can see.

◧◩◪◨⬒⬓⬔
7. Philpa+Zp1[view] [source] 2025-01-22 10:43:31
>>ForHac+zo1
Once you have one AGI, you can scale it to many AGI as long as you have the necessary compute. An AGI never needs to take breaks, can work non-stop on a problem, has access to all of the world's information simultaneously, and can interact with any system it's connected to.

To put it simply, it could outcompete humanity on every metric that matters, especially given recent advancements in robotics.

◧◩◪◨⬒⬓⬔⧯
8. ForHac+Us1[view] [source] 2025-01-22 11:18:25
>>Philpa+Zp1
...so it can think really hard all the time and come up with lots of great, devious evil ideas?

Again, I wonder why no group of smart people with brilliant ideas has unilaterally imposed those ideas on the rest of humanity through sheer force of genius.

◧◩◪◨⬒⬓⬔⧯▣
9. jprete+CP1[view] [source] 2025-01-22 14:04:08
>>ForHac+Us1
Quite a few have succeeded in conquering large fractions of the Earth's population: Napoleon, Hitler, Genghis Khan, the Roman emperors, Alexander the Great, Mao Zedong. America and Britain as systems did so for long periods of time.

All of these entities would have been enormously more powerful with access to an AGI's immortality, sleeplessness, and ability to clone itself.

◧◩◪◨⬒⬓⬔⧯▣▦
10. Sketch+aV1[view] [source] 2025-01-22 14:37:01
>>jprete+CP1
I can see what you're trying to say, but I cannot for the life of me figure out how an AGI would have helped Alexander the Great.
◧◩◪◨⬒⬓⬔⧯▣▦▧
11. jprete+JZ1[view] [source] 2025-01-22 15:05:13
>>Sketch+aV1
Alexander the Great made his conquests by building a really good reputation for war, then leveraging it to get tribute agreements while leaving the local governments intact. This is a good way to do it when communication lines are slow and unreliable, because the emperor just needs to check tribute once a year to enforce the agreements, but it's weak control.

If Alexander could have left perfectly aligned copies of himself in every city he passed, he could have gotten much more control and authority, and still avoided a fight by agreeing to maintain the local power structure with himself as the new head of state.

◧◩◪◨⬒⬓⬔⧯▣▦▧▨
12. Sketch+W02[view] [source] 2025-01-22 15:11:00
>>jprete+JZ1
Oh, you're assuming an entire networking infrastructure as well. That makes way more sense, but the miracle there isn't AGI - without networking they'd lose alignment over time. Honestly, I feel like it would devolve in a patchwork of different kingdoms run by an Alexander figurehead... where have I seen this before?

The problem you're proposing could be solved via a high quality cellular network.

[go to top]