While not legally mandated at a federal level, health insurance is generally considered to be an essential employee benefit in the U.S. As a result, employer-sponsored health insurance remains the ...