Definitions of Wesleyanism:

  • noun:   evangelical principles taught by John Wesley