pres_security_benchmarking_llm/pages/major-bench-secu.md

23 lines
942 B
Markdown
Raw Normal View History

2025-07-12 17:25:18 +02:00
# Major Benchmarks for LLM Security
<div class="grid-3">
<div class="card">
<h2 class="benchmark-title title-blue">Meta's CyberSecEval 2</h2>
<p>Introduced in April 2024, this benchmark suite evaluates both LLM security risks and cybersecurity capabilities.</p>
</div>
<div class="card">
<h2 class="benchmark-title title-purple">SEvenLLM-Bench</h2>
<p>A multiple-choice Q&A benchmark with 1300 test samples for evaluating LLM cybersecurity capabilities.</p>
</div>
<div class="card">
<h2 class="benchmark-title title-pink">SecLLMHolmes</h2>
<p>A generalized, automated framework for evaluating LLM performance in vulnerability detection.</p>
</div>
<div class="card">
<h2 class="benchmark-title title-cyan">SECURE</h2>
<p>The Security Extraction, Understanding & Reasoning Evaluation benchmark designed to assess LLM performance in realistic cybersecurity scenarios.</p>
</div>
</div>