Associations to the word «Naturism»

Wiktionary

NATURISM, noun. The belief in or practice of going nude or unclad in social and usually mixed-gender groups, specifically either in cultures where this is not the norm or for health reasons.
NATURISM, noun. The belief or doctrine that attributes everything to nature as a sanative agent.

Dictionary definition

NATURISM, noun. Going without clothes as a social practice.

Wise words

Words, words, words! They shut one off from the universe. Three quarters of the time one's never in contact with things, only with the beastly words that stand for them.
Aldous Huxley