The belief that women are superior to men. In order to maintain credibility among the general population, this belief is usually portrayed as making women equal to men. This is false, as women are not equal to men. This is called biology. Women are the intellectual equals of men and have always been treated as such, feminism seeks to make women "equal" in other ways, usually by demonisng men at every turn.
A woman chops a man's penis off because he wants a divorce, woman is celebrated.
A man asks a woman for coffee, man is a rapist somehow.
Isn't feminism wonderful..?