Right? Islam is the single most destructive mainstream culture with how it treats women, relegating them to the same level as slaves and normalizing abuse, and yet people are just expected to go along with it because it’s just their values. Really, really worrying that something that evil is becoming accepted in America to this degree.