Deleted member 2119
Master
- Joined
- Jun 20, 2019
- Posts
- 1,516
- Reputation
- 1,749
I get that I’m not saying anything new to most of you, but I just want to explain how feminism has single handily destroyed the west. Feminism leads to the death of societies. Feminism is against the natural order. Both studies and observable reality prove this. Women married with kids are much happier off. Women are not meant to be in politics, or have important jobs that require decision making. Women are driven by emotion. This is why leftist men are thought of as beta, they think like women. Women in politics lead to leftist ideas like progressivism. Progressivism is based on emotion and not logic. Ever since women have got the right to vote and entered the workforce, the west has fallen apart. It’s only getting worse.