first commit
This commit is contained in:
commit
7d4e05de19
27 changed files with 7574 additions and 0 deletions
23
pages/major-bench-secu.md
Normal file
23
pages/major-bench-secu.md
Normal file
|
@ -0,0 +1,23 @@
|
|||
# Major Benchmarks for LLM Security
|
||||
|
||||
<div class="grid-3">
|
||||
<div class="card">
|
||||
<h2 class="benchmark-title title-blue">Meta's CyberSecEval 2</h2>
|
||||
<p>Introduced in April 2024, this benchmark suite evaluates both LLM security risks and cybersecurity capabilities.</p>
|
||||
</div>
|
||||
|
||||
<div class="card">
|
||||
<h2 class="benchmark-title title-purple">SEvenLLM-Bench</h2>
|
||||
<p>A multiple-choice Q&A benchmark with 1300 test samples for evaluating LLM cybersecurity capabilities.</p>
|
||||
</div>
|
||||
|
||||
<div class="card">
|
||||
<h2 class="benchmark-title title-pink">SecLLMHolmes</h2>
|
||||
<p>A generalized, automated framework for evaluating LLM performance in vulnerability detection.</p>
|
||||
</div>
|
||||
|
||||
<div class="card">
|
||||
<h2 class="benchmark-title title-cyan">SECURE</h2>
|
||||
<p>The Security Extraction, Understanding & Reasoning Evaluation benchmark designed to assess LLM performance in realistic cybersecurity scenarios.</p>
|
||||
</div>
|
||||
</div>
|
Loading…
Add table
Add a link
Reference in a new issue