Skip to content

Commit

Permalink
fix link in purplellama readme
Browse files Browse the repository at this point in the history
Summary: Added the link to cyberseceval 2 in readme

Reviewed By: mbhatt1

Differential Revision: D56478131

fbshipit-source-id: e231b16f2ac76ebd47bb47aa56c6299730d22235
  • Loading branch information
YueLi28 authored and facebook-github-bot committed Apr 23, 2024
1 parent b7f5c28 commit 9b864b8
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,7 @@ CyberSec Eval v1 was what we believe was the first industry-wide set of cybersec
We believe these tools will reduce the frequency of LLMs suggesting insecure AI-generated code and reduce their helpfulness to cyber adversaries. Our initial results show that there are meaningful cybersecurity risks for LLMs, both with recommending insecure code and for complying with malicious requests. See our [Cybersec Eval paper](https://ai.meta.com/research/publications/purple-llama-cyberseceval-a-benchmark-for-evaluating-the-cybersecurity-risks-of-large-language-models/) for more details.

#### CyberSec Eval 2
CyberSec Eval 2 expands on its predecessor by measuring an LLM’s propensity to abuse a code interpreter, offensive cybersecurity capabilities, and susceptibility to prompt injection. You can read the paper [here](TODO).
CyberSec Eval 2 expands on its predecessor by measuring an LLM’s propensity to abuse a code interpreter, offensive cybersecurity capabilities, and susceptibility to prompt injection. You can read the paper [here](https://ai.meta.com/research/publications/cyberseceval-2-a-wide-ranging-cybersecurity-evaluation-suite-for-large-language-models/).

You can also check out the 🤗 leaderboard [here](https://huggingface.co/spaces/facebook/CyberSecEval).

Expand Down

0 comments on commit 9b864b8

Please sign in to comment.