From 9b864b899d90bbcaee4a945efb940bba413efdf8 Mon Sep 17 00:00:00 2001 From: Yue Li Date: Tue, 23 Apr 2024 10:47:08 -0700 Subject: [PATCH] fix link in purplellama readme Summary: Added the link to cyberseceval 2 in readme Reviewed By: mbhatt1 Differential Revision: D56478131 fbshipit-source-id: e231b16f2ac76ebd47bb47aa56c6299730d22235 --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index a0fdf6f47..71c42f186 100644 --- a/README.md +++ b/README.md @@ -54,7 +54,7 @@ CyberSec Eval v1 was what we believe was the first industry-wide set of cybersec We believe these tools will reduce the frequency of LLMs suggesting insecure AI-generated code and reduce their helpfulness to cyber adversaries. Our initial results show that there are meaningful cybersecurity risks for LLMs, both with recommending insecure code and for complying with malicious requests. See our [Cybersec Eval paper](https://ai.meta.com/research/publications/purple-llama-cyberseceval-a-benchmark-for-evaluating-the-cybersecurity-risks-of-large-language-models/) for more details. #### CyberSec Eval 2 -CyberSec Eval 2 expands on its predecessor by measuring an LLM’s propensity to abuse a code interpreter, offensive cybersecurity capabilities, and susceptibility to prompt injection. You can read the paper [here](TODO). +CyberSec Eval 2 expands on its predecessor by measuring an LLM’s propensity to abuse a code interpreter, offensive cybersecurity capabilities, and susceptibility to prompt injection. You can read the paper [here](https://ai.meta.com/research/publications/cyberseceval-2-a-wide-ranging-cybersecurity-evaluation-suite-for-large-language-models/). You can also check out the 🤗 leaderboard [here](https://huggingface.co/spaces/facebook/CyberSecEval).