(Replying to PARENT post)
The current most-widely-accepted answer to this is "indirect normativity". This is basically saying that human values are complex enough that we should instruct an AI to study humans to figure out what they are, rather than try to program values in directly and risk getting them wrong.
I'd be a lot more comfortable if metaethics and related philosophy advanced to the point where we didn't have to rely on indirect normativity. But I don't think this is something you can just throw money and people at, because this is a field which is sufficiently difficult and hard to judge, that work which isn't first-rate tends to detract from progress by adding noise rather than help.
๐คjimrandomh๐10y๐ผ0๐จ๏ธ0
(Replying to PARENT post)
Use ethic monoids to combine seemingly disparate ethical systems. https://medium.com/@yanazendo/ethic-monoids-913c3046079c
๐คyanazendo๐10y๐ผ0๐จ๏ธ0
(Replying to PARENT post)
If the research is aimed at ensuring that AI is aligned with human interests, another priority should be researching what we deem those interests to be.
The Law and Ethics research section of the research priorities document is centered on mechanisms, with a couple of nods to policy, but with the seeming assumption that that question of what is in the human interest has already been answered.
But the answer certainly isn't obvious. Different sets of humans have wildly diverging and fundamental disagreements over what is right or wrong for humans and our future.
I doubt we're going to find a consensus in the time we have. It will probably make more sense to just have smart people define a set of good principles. This is hard enough that it should be considered a research task, and I would think it would be at the top of the priority list.
Otherwise, how do we know that the research we are doing on other problems even makes sense? If you just stub out the "human wishes" component of all this as one simple "human approves" button, then you haven't solved anything. We need a robust definition of what "aligned with human interests" means.