\Na"tur*ism\, n. (Med.) The belief or doctrine that attributes everything to nature as a sanative agent.