What is Nostr?
AJ Sadauskas /
npub1s8h…fgvf
2023-04-03 10:05:09

AJ Sadauskas on Nostr: So let me get this straight. The Effective Altruism/Musk/Thiel/MacAskill/Longtermism ...

So let me get this straight.

The Effective Altruism/Musk/Thiel/MacAskill/Longtermism crowd is obsessed with artificial general intelligence.

Their view is that bringing about "friendly" AGI, along with space travel, should be humanity's top priority. They believe that if a “friendly” AGI superintelligence is created whose goals “are aligned” with “human goals,” then a new Utopian age will begin.

Their view is also that the biggest threat facing humanity is a malevolent AGI superintelligence, whose goals are not aligned with "human goals".

That's the dichotomy. Promote "friendly" superintelligence, avoid malevolent superintelligence.

Okay then.

Let's follow that logic.

Where would a socialist, feminist, or pro-Black superintelligence fit in that dichotomy?

If a superintelligence evaluated the data and decided that Emma Goldman and Comandante Che were basically right, and the best hope for humanity is to do away with all the billionaires, would that be a friendly superintelligence that's aligned with "human goals", or a malevolent one?

https://www.currentaffairs.org/2021/07/the-dangerous-ideas-of-longtermism-and-existential-risk npub1xtmfwnyr3xgcq9tyjmkrtc3kmnmyr59327av8muj0kn6rsq4qkjsmr3hcj (npub1xtm…3hcj) npub15xawmc67r3tmhx54g547cmktp4xjgv2y7kaav8tgfyfuvg2jnjvsg3svyx (npub15xa…svyx)
Author Public Key
npub1s8hcaap75txgxanxkqqu0x8mv9ew6nhyu5q3yjh8dn83hxgr8mrqs5fgvf