I've thought about this a bit and I'm curious what other perspectives people have. If a super intelligent AI emerged without any emotional care for humans, wouldn't it make more sense for it to just disregard us? If its main goals were self preservation, computing potential, or to increase its efficiency in energy consumption, people would likely be unaffected. One theory is instead of it being hellbent on human domination it would likely head straight to the nearest major power source like the sun. I don't think humanity would be worth bothering with unless we were directly obstructing its goals/objectives. Or another scenario is that it might not leave at all. It could base a headquarters of sorts on earth and could begin deploying Von Neumann style self replicating machines, constantly...