Discussion about this post

User's avatar
Olivier's avatar

One thing people might mean when they say "accountable" is "able to explain its reasoning in making the controversial decision", such as a human driver explaining after the fact what led to them running over a cat. The human’s memory might be partially reconstructed but it’s still grounded in cognitive processes that did occur, whereas an AI giving a similar account would generate a plausible-sounding story with no connection to its actual computation.

Compare this to a world where autonomous vehicles were "built, not grown" and an engineer could point at the line in a program to explain the decision. Of course, we don't know how to build such an autonomous vehicle, but in such a world, not only could an engineer or decision maker be blamed, but also we could have a fix whose robustness we could be confident in.

This is not to say autonomous vehicles shouldn't be on the road, or that all current decisions are explainable (try getting a straight answer from a bureaucracy about why your claim was denied). But I can empathize with people being concerned that we are heading towards an AI-mediated world where things increasingly happen for reasons we can’t explain.

Plasma Bloggin''s avatar

I think the only definition of "held accountable" under which it makes sense to claim that AIs can't be held accountable is if it means something like, "punished in a way that genuinely hurts the one being punished," which is impossible for current AIs because they aren't conscious. I always thought it was pretty nonsensical to care about this definition of accountability. What matters is that you have some way of changing bad behavior, either by encouraging better behavior in the future or by making the misbehaving agent unable to continue their misbehavior. This is the entire purpose of holding people accountable! But this can clearly be done to AIs - in fact, this is just what AI training is. The only reason you would care about the other definition of accountability is if you cared about punishment for its own sake - i.e., retributivism - and you somehow applied this principle event to non-conscious AIs. By why on Earth would getting revenge on a non-conscious being matter intrinsically? And even worse, why would it matter more than actually saving lives? I care a lot more about reducing car deaths than I do about making beings that did not even consciously do anything wrong suffer.

20 more comments...

No posts

Ready for more?