What does the "western" mean here? Does it refer to North America and Europe? Or simply Occident that contradicts Orient?

Native speaker’s answer
Rebecca
Good guess! In general, you can understand "Western" as meaning that which originates or relates to the West, which usually refers to the majority of Europe (Western, Northern, and Southern Europe), North America, and Australasia (Australia and New Zealand). Ex: There are often great cultural and ideological differences between the West and the rest of the world. Ex: Immigrants are often caught between Western values and the culture of their home country.