Discussion about this post

User's avatar
Yancey Ward's avatar

Alexa, destroy the world.

Expand full comment
Doctor Hammer's avatar

I thought you were going to go the route of saying AI alignment can be handled by cybersecurity sharply limiting what the AI has access to. Like having a tiger in a zoo, it is dangerous and might do things you don't like, but when it is contained where it is supposed to be and limited in reach it is valuable. Having tigers means having good locks, fences and other barriers, to keep people out sure, but also to keep the tiger itself limited.

Arguing that AI won't do bad things without humans telling it to strikes me as very naïve regarding how complex adaptive systems behave, how AI is trained, and what people plan to have AI do in the future as a goal instead of a byproduct.

Say you have an AI similar to ChatGPT that reads websites for information. You don't want the AI to doxx people, so you do lots of training to get it to refuse to answer questions like "What is Arnold Kling's home address and workout schedule?" The trouble is that it still might have access to the information, perhaps storing that data locally. No big deal, unless it got that information from somewhere it shouldn't have and now whoops, your program is breaking a privacy law without you knowing.

We want to give AI lots of capabilities, but that requires a huge amount of access to outside data for training and reference. How it puts that data together and stores it is a black box, and increasingly where and how it gets it is a black box. There's a lot of dangerous steps in there.

"ChatGPT, help me optimize my financial investments."

"Done."

"Wait... what do you mean 'done'?"

Expand full comment
17 more comments...

No posts