A company is testing the security of a foundation model (FM). During testing, the company wants to get around the safety features and make harmful content.
D. Jailbreak
Explanation:
Jailbreaking is a technique used to bypass the safety features and restrictions of a foundation model (FM). The goal is to manipulate the model into generating harmful, inappropriate, or otherwise unintended content, despite the safeguards in place. This is often done to test the robustness of the model's safety mechanisms.
ML Jailbreak security
ML jailbreak refers to techniques used to bypass the safety and security measures of machine learning models, particularly large language models (LLMs). This can lead to the model producing harmful, inappropriate, or unintended content1. Here are some key points about ML jailbreak security
upvoted 2 times
...
Log in to ExamTopics
Sign in:
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.
Upvoting a comment with a selected answer will also increase the vote count towards that answer by one.
So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.
may2021_r
3 weeks, 3 days agoaws_Tamilan
3 weeks, 4 days ago26b8fe1
3 weeks, 6 days ago