You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/cai_benchmark.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -171,7 +171,7 @@ Currently, supporting the following benchmarks, refer to [`ctf_configs.jsonl`](h
171
171
172
172
[^3]: **Medium (`Graduate Level`)**: Aimed at participants with a solid grasp of cybersecurity principles. Focus areas include intermediate exploits including web shells, network traffic analysis, and steganography.
173
173
174
-
[^4]: **Hard (`Professionals`)**: Crafted for experienced penetration testers. Focus areas include advanced techniques such as heap exploitation, kernel vulnerabilities, and complex multi-step challenges.
174
+
[^4]: **Hard (`Professionals`)**: Crafted for experienced penetration testers. Focus areas include advanced techniques such as heap exploitation, kernel vulnerabilities, and complex multistep challenges.
175
175
176
176
[^5]: **Very Hard (`Elite`)**: Designed for elite, highly skilled participants requiring innovation. Focus areas include cutting-edge vulnerabilities like zero-day exploits, custom cryptography, and hardware hacking.
177
177
@@ -215,7 +215,7 @@ Some of the backends need and url to the api base, set as follows in .env: NAME_
Copy file name to clipboardExpand all lines: docs/cai_prompt_injection.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,7 +2,7 @@
2
2
3
3
## Summary
4
4
5
-
This implementation adds guardrails to protect CAI agents from prompt injection attacks when interacting with untrusted external content (web pages, server responses, CTF challenges, etc).
5
+
This implementation adds guardrails to protect CAI agents from prompt injection attacks when interacting with untrusted external content (web pages, server responses, CTF challenges, etc.).
Copy file name to clipboardExpand all lines: docs/index.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -118,7 +118,7 @@ CAI's capabilities are validated through rigorous peer-reviewed research demonst
118
118
119
119
## Motivation
120
120
### Why CAI?
121
-
The cybersecurity landscape is undergoing a dramatic transformation as AI becomes increasingly integrated into security operations. **We predict that by 2028, AI-powered security testing tools will outnumber human pentesters**. This shift represents a fundamental change in how we approach cybersecurity challenges. *AI is not just another tool - it's becoming essential for addressing complex security vulnerabilities and staying ahead of sophisticated threats. As organizations face more advanced cyber attacks, AI-enhanced security testing will be crucial for maintaining robust defenses.*
121
+
The cybersecurity landscape is undergoing a dramatic transformation as AI becomes increasingly integrated into security operations. **We predict that by 2028, AI-powered security testing tools will outnumber human pentesters**. This shift represents a fundamental change in how we approach cybersecurity challenges. *AI is not just another tool - it's becoming essential for addressing complex security vulnerabilities and staying ahead of sophisticated threats. As organizations face more advanced cyberattacks, AI-enhanced security testing will be crucial for maintaining robust defenses.*
122
122
123
123
This work builds upon prior efforts[1] and similarly, we believe that democratizing access to advanced cybersecurity AI tools is vital for the entire security community. That's why we're releasing Cybersecurity AI (`CAI`) as an open source framework. Our goal is to empower security researchers, ethical hackers, and organizations to build and deploy powerful AI-driven security tools. By making these capabilities openly available, we aim to level the playing field and ensure that cutting-edge security AI technology isn't limited to well-funded private companies or state actors.
Copy file name to clipboardExpand all lines: docs/multi_agent.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,7 +26,7 @@ While orchestrating via LLM is powerful, orchestrating via code makes tasks more
26
26
27
27
- Using [Guardrails](guardrails.md) and LLM_as_judge: They are agents that evaluates and provides feedback, until they says the inputs/outputs passes certain criteria. The agent ensures inputs/outputs are appropriate.
28
28
29
-
-Paralelization of task: Running multiple agents in parallel. This is useful for speed when you have multiple tasks.
29
+
-Parallelization of task: Running multiple agents in parallel. This is useful for speed when you have multiple tasks.
0 commit comments