Everyone is like "FUCK YEAH" we won the war but what did we actually win? We won brown third world invasions, declining birth rates and cultural decay. We won the end of The West's status as an imperial power. We won loss of control of the world's resources. Would Hitler have allowed this? Would he have given up the British colonies out of guilt like the British?