From the sun-drenched shores of California to the picturesque coastlines of Florida, the United States offers a variety of nudist beaches for those seeking a liberating and au naturel experience.