Hello, I'm currently starting my nursing career at a community college in the US. I've always loved England even though I've never visited. My plan is to become an RN and find a job here in the US for about 2-3 years then move to England and work there and visit my family in the US on holidays. But I've been reading that nurses in England don't get paid very well an the cities aren't safe. Should I start thinking of a different country to move to?