Logo

Home | Posts | Papers | Media | About


Angles of Attack: AI Security Intelligence


Mission | Latest | Principles | Coming soon: Newsletter


Mission

→ Understand threats to AI systems, quantify risk, design mitigations, & engineer solutions at scale to the highest engineering standards, at the edge of the SOTA.

→ Go beyond the surface to grasp context, think about AI security at a systems level, anticipate AI threats, and stay ahead of the hype.

→ Use threat modeling to teach AI teams how to de-silo expertise, think like hackers, and deploy like defenders.

→ Provide AI security analysis for forward looking, systems thinking, research driven insights.


Latest


Real-Time Embedded Systems: A Quick Primer For AI

Systems that must operate in real-time–including AI–are more than just fast, they’re preictable, reliable, and robust. »

AI Systems Vs Mission-Critical Applications

One of the biggest challenges in creating mission-critical AI is baked into the nature of AI/ML systems themselves. »

When Enough Is Enough: Assurance Cases for AI System Security Design

How do we know whether our security and robustness requirements for mission-critical AI systems are good enough? »

top


Principles & Applications


AI Security Requires New Expertise

→ The AI threat model is radically different from those of traditional systems & requires fundamental paradigm shifts; de-siloing of data, dev, and security teams; training in the role and importance of data; and a lifecycle-based approach which is inseparable from Security Operations (SecOps).

→ Many data teams are still new to security concepts, and many security teams are underexposed to exactly how AI systems work–and break.

→ De-siloing expertise is an important first step, but without development of critical AI security-specific skill sets, teams will still fail to deploy AI systems securely.

Develop More Securely, Deploy More Successfully

→ The lifecycle-based, SecOps-informed approach to understanding threats to AI systems is fundamental to securing them.

→ Lifecycle-based threat modeling ensures that AI-adapted secure development lifecycle (AI-SDL) principles are baked into AI deployments from the start.

Move AI Security Principles From Theory To Practice

→ To gain a deeper contextual understanding of the true AI threat landscape, it’s necessary to both apply and go beyond the lifecycle by examining AI security applications & modeling threats through the lenses of three key themes:

AI Threat Modeling Requires A Purple Team Approach

→ Threat modeling with risk assessment for AI is a purple team activity, requiring understanding of the attack surface, organizational defense capabilities, and the regulatory policy landscape, all as applied to AI systems and their unique contexts.

→ These three thematic areas (Engineering Principles, Attack Surface & Defensive Capabilities, and Policy Application At Scale) should inform threat modeling and risk assessment for AIML applications in every organization, at every stage of development. Context matters.

Future-Proof AI Systems

→ Currency in a highly technical, ever-evolving field with mission-critical applications requires anticipating threats; anything slower means playing a hopeless game of catch-up in the AI security arms race.

→ Tightening the OODA loop in favor of defenders requires strategic, proactive threat intelligence, threat modeling, & risk assessment.

→ By studying real-world applications of key themes in context, we can better understand—and anticipate–the emergent and rapidly-shifting AI threat model.

Think Like A Security Researcher

→ Examining AI security via the context/s of Engineering Principles, Attack Surface & Defensive Capabilities, and Application-At-Scale affords savvy practitioners system-level understanding of the AI threat surface.

→ Pushing the AI security SOTA requires the cooperative, interdisciplinary de-siloing of data, security, and DevOps knowledge; applied creatively, tracked scientifically, and shared robustly.

National And Global Security Are At Stake

→ The intersections of AI security research, industrial applications, and policy represent a new frontier in international security.

→ To lead in AI innovation, nation-states will need to first lead in AI security.

Why The Frameworks We Choose Matter

→ AI security is a constantly changing field, with many facets and nearly as many ways of cataloging & understanding these aspects. The systems we choose to interpret and model these complexities shape how we see the AI ecosystem, which nuances we perceive, and ultimately, the engineering and policy philosophies we bring to the table.

→ Now, more than ever, a holistic approach to codifying & interpreting the AI threat surface affords the strategic advantage to forward-thinking practitioners.

top


🕷