So I'm guessing energy used for mining should be considerably less than $2 to $3 million when rate is at $1000/BTC.
Difficulty grows with total hash rate, it doesn't care about power efficiency per hash.
Once you have the hardware, it makes economic sense to keep mining as long as your costs per one Bitcoin mined are less then one Bitcoin.
And if your cost is, for example 0.3 coins per coin, it makes sense to expand (in a sense of buying new hardware). And buying new hardware will increase difficulty up to the point, perhaps at 0.5, perhaps higher, where you stop expanding.
Those two incentives create a situation where you will always have some miners that are close to zero profit, and not very many of those that are spending 0.1 BTC to mine 1 BTC.
We can never know the true efficiency distribution of all miners, but if you take a straight line from zero to one, on a chart where you display efficiency of all miners, sorted by efficiency, you would get total surface of 0.5, and if you start from an assumption that the most efficient miner is at 0.5, you would get total of 0.75 under the curve. Thus my assumption for total miner efficiency of 0.5 to 0.75. In practice, it could be higher, but if it is lower, it would be a temporary situation.
Ok, but what total hashrate is needed to sustain a network of, say, 1/10 of the world economy/