adj. Of or pertaining to the practices and institutions that legitimize and privilege heterosexuality, heterosexual relationships, and traditional gender roles as fundamental and "natural" within society.
Word Usage
"We know that the word 'heteronormative' is bad because people who are good said so."