NVIDIA Launches NIM AI Fence Service to Prevent Models from Being "Jailbroken" by Users

Nvidiaannounced an AI guardrail service called "NIM," which is now available as NVIDIA's NeMo Guardrails suite, enabling developers to add a set of guardrail rules to Large Language Models (LLMs) designed to address the issue of users "jailbreaking" the Large Model by prompting for words that do not meet expectations. The goal is to address the issue of users "jailbreaking" the LLM by prompting words, preventing the AI from generating content that doesn't meet expectations.

NVIDIA says the corresponding AI fencing service suiteTrained on NVIDIA's Aegis content security datasetThe dataset contains 35,000 labeled data samples and is publicly available on Hugging Face at the following address (Click here to visit).

NVIDIA pointed out that the corresponding AI fencing suite is characterized by small size and high efficiency, and can run smoothly in most occasions. Enterprises can directly embed the corresponding security suite in the development of AI models, which can improve AI security deployed in healthcare, automotive, manufacturing, and other fields.

In addition, NVIDIA has provided announced a vulnerability scanning tool called Garak to test the security of models against the possibility of them outputting hallucinatory content or leaking confidential information within the organization.

statement:The content of the source of public various media platforms, if the inclusion of the content violates your rights and interests, please contact the mailbox, this site will be the first time to deal with.
Information

OpenAI Altman: Plans to launch o3 mini inference model in a few weeks

2025-1-18 11:27:24

Information

U.S. government pushes for ChatGPT product chief to testify in hopes of bolstering antitrust charges against Google

2025-1-19 13:05:06

Search