Initially used as a term to empower awareness of systemic inequalities in society, wokeism is now a deeply divisive term. The media's perpetuation of woke culture has made this term a buzzword. For some, being woke is part of the antidote of acknowledging the instruments of oppression. For others, it is a dangerously absolutist ideology, a sort of reverse McCarthyism, corroding liberal society and encouraging self-imposed victimhood. Is the 'war on woke' a legitimate phenomenon, or a reactionary distraction from the real problems being 'woke' addresses?