Definitions of dermatology:

  • noun:   the branch of medicine dealing with the skin and its diseases