mirror of
https://github.com/KeygraphHQ/shannon.git
synced 2026-04-25 17:45:53 +03:00
[GH-ISSUE #38] Clarify whether benchmark reports used real customer applications or intentionally vulnerable test apps #10
Labels
No labels
pull-request
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/shannon-KeygraphHQ#10
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @saiqulhaq on GitHub (Jan 4, 2026).
Original GitHub issue: https://github.com/KeygraphHQ/shannon/issues/38
The README showcases Shannon's capabilities through sample reports in the
sample-reports/directory, specifically:While these reports demonstrate impressive vulnerability detection (20+ critical issues in Juice Shop, complete auth bypass, database exfiltration, etc.), it's not immediately clear to potential users whether these reports represent:
Why this matters:
For organizations evaluating Shannon for their security testing needs, understanding whether these impressive results come from purpose-built vulnerable apps versus real-world applications significantly affects credibility and expected performance.
Testing against deliberately insecure applications is valuable for demonstrating capabilities, but results from real customer engagements (with appropriate anonymization) would provide stronger validation of Shannon's effectiveness in production scenarios.
Suggested clarification:
Could the README explicitly state that these sample reports are from testing against intentionally vulnerable benchmark applications rather than customer engagements? Something like:
This would help potential users properly calibrate their expectations and understand that these results represent Shannon's performance against purpose-built vulnerable targets rather than real-world customer applications.
@keygraphVarun commented on GitHub (Jan 5, 2026):
Thanks for the feedback.
To clarify: the sample reports section is already prefaced with "See Shannon's capabilities in action with real penetration test results from industry-standard vulnerable applications" - which is what Juice Shop, crAPI, and c{api}tal are. Each entry also links to the GitHub repo and describes what the app is (e.g., "A notoriously insecure web application maintained by OWASP").
That said, you make a fair point that the "Benchmark Results" header could be clearer. Our actual quantitative benchmark is the XBOW benchmark as linked at the very top of the README. We'll rename that section to "Sample Reports" and add a link to the XBOW results for folks looking for standardized evaluation metrics.
Thanks for raising it.