It does seem bleak. I think you’re right about everything in the OP.
Fact is, you have to change culture with kids before they reach adulthood, and the institutions that prevented the things you describe are gone. The leftists won the culture war by destroying those institutions.
Just as an example, parents and schools don’t expect celibacy before marriage, and society doesn’t enforce it. Schools don’t have wholesome dances, but twerk fests with all black music. There’s literally no more music ever played or promoted by adults in our society derived from white culture that’s fit for actual pair dancing. And once in college, there are no more dorm supervisors, no sex segregation, no hard rules on heavy drinking or anything that leads to all the moral malaise you describe.
All of this began to fall apart in the 1960s and it was essentially complete by the mid-1990s.