-
Anthropic's new study: Typos can 'jailbreak' GPT-4, Claude, and other AI models
December 25, 2011 - Artificial intelligence company Anthropic recently released a study revealing that security protections for large language models (LLMs) are still vulnerable, and that the process of "jailbreaking" around these protections can be automated, according to 404 Media. The study shows that simply by changing the format of a prompt, such as random case mixing, an LLM can be induced to produce content that should not be output. To validate this finding, Anthropic worked with researchers at Oxford University, Stanford University, and MATS...- 1.2k
❯
Search
Scan to open current page
Top
Checking in, please wait
Click for today's check-in bonus!
You have earned {{mission.data.mission.credit}} points today!
My Coupons
-
¥CouponsLimitation of useExpired and UnavailableLimitation of use
before
Limitation of usePermanently validCoupon ID:×Available for the following products: Available for the following products categories: Unrestricted use:Available for all products and product types
No coupons available!
Unverify
Daily tasks completed: