Alignment's hardest question is whose values AI follows
βAnd the question is, to what or to whom should the AIs be aligned? In what situation should the AI defer to the model company versus the end user versus the law versus to its own sense of morality. This is maybe the most important question about what happens in the future with powerful AI systems, and we barely talk about it. And it's understandable why, because if you're a model company, you don't really wanna be advertising the fact that you have complete control over the preferences and the character of the entire future labor force.β

