Food and Drug Administration (FDA)


Definition - What does Food and Drug Administration (FDA) mean?

The Food and Drug Administration (FDA) is a U.S. agency that was founded in 1906 to promote and protect public health by implementing laws and regulations. The main sectors that fall under the FDA's jurisdiction are food safety, dairy products, sanitary conditions and tobacco products.

Safeopedia explains Food and Drug Administration (FDA)

The FDA is a U.S. agency that has the responsibility of looking into the safety of food and other associated products. The FDA also oversees the safety of dietary supplements, tobacco products, vaccines, blood transfusions, medical equipment and veterinary products.

Share this:

Connect with us

Email Newsletter

Join thousands receiving the latest content and insights on health and safety industry.