When did the concept of ‘the West’ emerge? How is it related to what (until recently) was referred to as the ‘Orient’? And what has ‘the West’ come to connote in the twentieth century?