See 2023-04-30 Patreon post - Ethics of AI-based invention - a personal inquiry
I’m tremendously excited about the prospects of powerful AI systems for humanity. If we get them right, we’ll have wildly expanded our capacity. We’ll become unrecognizable. But these systems also have the potential to do unbelievable amounts of harm. That doesn’t necessarily mean that we should never create such systems. I want to see AGI in my lifetime. But I also want to not live through debilitating catastrophe. What has me quite concerned at the moment is that these models’ capacities are growing at a pace which far outstrips our abilities to confidently contend with their significant potential harms. I’m really thinking about Large language models in particular here.
I’ve idly read plenty on AI risk over the years, but it’s always been a mild background interest. I’ve never felt like the problem is a good match for my own capacities, and I’ve never had any ideas that felt like they could turn into interesting contributions. So I’ve always thought about it as a problem, but “someone else’s problem”, in much the same sense as nuclear non-proliferation.
But now these models are advanced enough that I’m having lots of ideas about interesting systems one could design involving them. And I realized: oh. If I design systems which integrate these models, and talk about them publicly, am I making the problem worse? Worse enough that I have a moral obligation to desist? Does even talking publicly about these models’ potential make the problem worse enough to create a moral obligation?
Suddenly the issue becomes immediate, and personal. The stakes are obvious, and tremendous. If I conclude that I can’t have any part in designing systems which incorporate these models, I may be more or less giving up my career. They’re going to be everywhere, faster than I know it.
2023-04-21: hm, Cavendish fellowship… Cavendish Labs Fellowship