The medical industry is by and large a Godless entity. Especially in modern times. The vast majority of doctors out there are like horse with blinders on. They know nothing about the healing properties of food and nutrition. Barely even consider it in their analysis. The only thing they know is prescription drugs. I can't think of another industry that has damaged the United States more than the medical and pharmaceuticals.
And that is taking into consideration the war mongering politicians. Even they don't stand a chance against the mass murder, addiction and mental health destruction that "doctors" cause every year.
And the left went from "health care is a right!" To "health Care is a privilege of the vAcCinAtEd"
The end won't be for everyone makes more and more sense daily.
The left have always known and told themselves and each other that if they control the food and the medical care they control the person.