To the extent that humans can imagine these kinds of scenarios, it seems pretty futile to try to prevent sophisticated AI systems from considering them.
I am much more optimistic about the feasibility of straightforward strategies that prevent this problem. I think this is closely related to bigger picture disagreements about the structure of sophisticated AI systems.