Can you state explicitly what background assumption would lead you to think that an AI which behaves badly if subsystem A is very overpowered relative to subsystem B, is still safe?
It seemed to me that Paul was not saying that he thought this scenario would be safe, but that it would be unlikely.