Saturday, August 18, 2018

The Left Has Won The Culture War

2 comments:

Anonymous said...

You mean the left has destroyed the culture. It's not a war, it's a siege.

Western culture is everything of value that has been built and accumulated by our forebears. The left works with single minded determination to bring it all to ruin. They are worse than animals since animals never purposefully ruin their own habitat. They are Satanic. No other explanation will suffice.

swimologist said...

Yes, Anon, you nailed it. And who are the Left? From bankers to the media to academia to the activist/legal/political class-- all dominated by a agitating, no borders, one world order cabal-- International Jewry.