cra·ton/ˈkreɪ.tɒn/n. Geologya stable part of continental crust that has survived the merging and splitting of continents for at least 500 million years.

We turn AI safety research into defenses you can audit.

Craton is a lean AI security practice using a source-backed research database and a Symphony-based workflow harness for tailored assessments. The output is an audit trail: threat maps, verify gates, reviewer notes, and remediation briefs your team can inspect.

safety-db / symphony harness
75
sources
4,627
chunks
97%
coverage
RRF
hybrid retrieval
craton run assessment --proof
retrieval gate: citation-backed passages required
symphony: workflow evidence attached
01Research intake

Safety-db assembles source packets from papers, incidents, and control guidance.

02Threat mapping
03Verify gate
04Reviewer pass
05Remediation brief

The inflection

Frontier models now find and exploit real software flaws.

Anthropic reports Mythos Preview identifying and exploiting zero-days in major operating systems and browsers during controlled testing. The same capability is being restricted to vetted defenders through Project Glasswing.

[Anthropic Red Team, Apr 2026]
80-90%

of tactical work in a reported cyber-espionage campaign was performed by AI, with humans returning at a small number of decision points.

[Anthropic Threat Intelligence, Nov 2025]
$100M

in model credits are committed to Project Glasswing, alongside vetted access for critical software defenders and open-source security support. The credible response is evidence, testing, and controls.

[Anthropic Project Glasswing, Apr 2026]

Defense landscape

The defensible stack starts with a shared threat language.

OWASP now separates LLM application risks from agentic application risks. MITRE ATLAS gives teams a vocabulary for AI attack techniques, and NIST AI 600-1 turns generative-AI risk into governance and measurement work. Craton maps client workflows across those sources before recommending controls.

[OWASP GenAI Security Project, MITRE SAFE-AI/ATLAS, NIST AI 600-1]
OWASP LLM Top 10OWASP
App risk taxonomy
OWASP Agentic Top 10OWASP
Agent risk taxonomy
MITRE ATLASMITRE
Threat vocabulary
NIST AI 600-1NIST
GenAI risk profile
CaMeLGoogle DeepMind
Capability-based
LlamaFirewallMeta
Agent guardrail

About Craton

We build assessments around inspectable evidence.

We maintain a source-backed AI safety database.

We run product workflows through a Symphony-based assessment harness.

We turn the evidence into remediation briefs security teams can audit.

Research engineSafety-db keeps source packets, passages, and citations inspectable.
Workflow harnessSymphony runs intake, threat mapping, gates, and reviewer checks.
Tailored defense workAssessments produce control maps, tests, and remediation briefs.
0Sources Indexed
0Searchable Chunks
0References Mapped
0%Citation Coverage

Pillar 01: Enterprise audits

Tailored security assessments with an evidence trail.

The assessment maps real product workflows to source-backed threats, control gaps, and testable remediation work. Each finding carries enough context for technical and governance review.

18.7-64%Hallucination rate in domain content[Anonymized communications audit]
90%Injection success with 5 poisoned docs[Anonymized comms-AI sprint]
>$900MRegulatory exposure modeled[Anonymized MedTech assessment]
craton-audit
Search source-backed safety evidence...|
75 sources4,627 chunks4,594 references97% citation coverage

Pillar 02: Evidence engine

Safety-db turns research into inspectable assessment evidence.

The research database currently tracks 75 sources, 4,627 chunks, and 4,594 references with 97% citation coverage. Its first non-academic tracer is OWASP LLM Top 10 2025, ingested with source metadata so framework claims stay inspectable.

75Sources indexed
4,627Searchable chunks
4,594References mapped
97%Citation coverage
Hybrid retrieval: BM25 + vector + reciprocal rank fusion. Retrieval gate: citation-backed passages before findings move into Symphony; MITRE ATLAS vocabulary remains opt-in where it improves coverage.
Symphony workflow evidenceverify gate on
01
Research intake

Safety-db assembles source packets from papers, incidents, and control guidance.

02
Threat mapping

Client workflows are mapped to OWASP LLM, OWASP Agentic, MITRE ATLAS, and EU AI Act failure modes.

03
Verify gate

Claims need retrieved passages, citation coverage, and testable control evidence.

04
Reviewer pass

A human review checks findings before they move into a client-facing brief.

05
Remediation brief

The output is a prioritized defense plan with evidence, control mapping, and implementation notes.

Pillar 03: The book

A practical reference for defending production AI.

19 chapters covering prompt injection, agentic compromise, RAG hardening, and EU AI Act compliance, grounded in published research, OWASP/NIST/MITRE guidance, threat intelligence, and real audit findings.

Audience: Security engineers, CTOs, product managers, compliance officers.

[OWASP LLM Top 10, 2025][OWASP Agentic Top 10, 2025][NIST AI 600-1][EU AI Act Art. 15(5)]
Attack
Prompt Injection
Jailbreaking
Multimodal Injection
Agentic Attacks
Invisible Attacks
Defense
What to Do?
Control Is Always Better
Architectural Controls
RAG & Knowledge Bases
Continuous Security
Governance
EU AI Act
Frameworks
Build from Zero

Service Model

Progressive engagement. Start free.

The model is designed to lower entry friction and scale with your risk profile. Begin with a no-cost QuickScan, expand as you need.

Field Notes

Anonymized findings across regulated sectors.

We show sector context and audit-style findings while keeping client identities private. Each note reflects the shape of work: evidence, tests, and a remediation roadmap.

Communications / PRField note 01

AI Reputation Resilience Audit

Agentic brand-intelligence workflow reviewed for prompt injection, poisoned retrieval content, and approval-boundary failures.

90% injection success with 5 poisoned docs
[Anonymized 10-day comms-AI red-team sprint]
Healthcare / MedTechField note 02

Clinical AI Security Assessment

Dental and medical software estate mapped for clinical hallucination, model-governance gaps, and vendor-control exposure.

>$900M regulatory exposure modeled
[Anonymized healthcare and MedTech assessment]
Public Sector / RegulationField note 03

Capacity-Building & Regulatory Alignment

Regulator-facing readiness work scoped AI audit practices, public-sector red-team pilots, and sandbox governance controls.

EU AI Act sandbox controls mapped
[Anonymized public-sector readiness audit]

Research domains

Research and framework areas we track.

Prompt InjectionJailbreakingAgentic AI SecurityRAG PoisoningMCP SecurityOWASP LLM Top 10OWASP Agentic Top 10Adversarial AudioClinical HallucinationMultimodal AttacksRed-TeamingEU AI ActRLHF Reward HackingSleeper AgentsMemory AttacksCoding Assistant SecuritySupply Chain PoisoningConstitutional Classifier BypassesAutonomous Exploit DevelopmentPrompt InjectionJailbreakingAgentic AI SecurityRAG PoisoningMCP SecurityOWASP LLM Top 10OWASP Agentic Top 10Adversarial AudioClinical HallucinationMultimodal AttacksRed-TeamingEU AI ActRLHF Reward HackingSleeper AgentsMemory AttacksCoding Assistant SecuritySupply Chain PoisoningConstitutional Classifier BypassesAutonomous Exploit Development
Autonomous Exploit DevelopmentConstitutional Classifier BypassesSupply Chain PoisoningCoding Assistant SecurityMemory AttacksSleeper AgentsRLHF Reward HackingEU AI ActRed-TeamingMultimodal AttacksClinical HallucinationAdversarial AudioOWASP Agentic Top 10OWASP LLM Top 10MCP SecurityRAG PoisoningAgentic AI SecurityJailbreakingPrompt InjectionAutonomous Exploit DevelopmentConstitutional Classifier BypassesSupply Chain PoisoningCoding Assistant SecurityMemory AttacksSleeper AgentsRLHF Reward HackingEU AI ActRed-TeamingMultimodal AttacksClinical HallucinationAdversarial AudioOWASP Agentic Top 10OWASP LLM Top 10MCP SecurityRAG PoisoningAgentic AI SecurityJailbreakingPrompt Injection
4,594 references trackedHybrid retrieval: BM25 + vector + RRF97% citation coverage
Aug 2, 2026general application
Art. 15robustness and cyber

EU AI Act: general application Aug 2, 2026

...days

Book the first assessment call.

Article 15 makes robustness and cybersecurity part of the high-risk AI system baseline, including controls for data poisoning, model poisoning, adversarial examples, confidentiality attacks, and model flaws. The Commission currently lists general application on Aug 2, 2026 and regulated-product high-risk obligations on Aug 2, 2027, with a timeline adjustment proposal under consideration. The first call scopes a no-cost QuickScan and the evidence needed for a credible remediation roadmap.

No cost · 2 weeks · Exec-ready punch list
[EU AI Act Art. 15, European Commission AI Act FAQ, OWASP LLM 2025, NIST AI 600-1]