Snowflake Cortex AI Launches Cortex Guard for LLM Safeguards

Once you activate Cortex Guard, language model responses associated with harmful content — such as violent crimes, hate, sexual content, self-harm and others — will be automatically filtered out, and the model will return a  “Response filtered by Cortex Guard” message. 

Under the hood, Cortex Guard, currently powered by Llama Guard 2 from Meta, works by evaluating the responses of a language model before that output is returned to the application. Snowflake’s partnership with Meta brings best-in-class safety across all models available for use in the Cortex AI COMPLETE function. Llama Guard 2 protects against Violent Crimes, Non-Violent Crimes, Sex-Related Crimes, Child Sexual Exploitation, Specialized Advice, Privacy, Intellectual Property, Indiscriminate Weapons, Hate, Suicide & Self-Harm, and Sexual Content.