Definitions of health care:

  • noun:   social insurance for the ill and injured
  • noun:   the preservation of mental and physical health by preventing or treating illness through services offered by the health profession