Threat Modeling Insider – September 2025

Threat Modeling Insider Newsletter

47th Edition – September 2025

Welcome!

Welcome to this month’s edition of Threat Modeling Insider! In our featured guest article, LLMs are the future of Threat Modeling, right?, Felix Viktor Jedrzejewski. Davide Fucci and Oleksandr Adamov explore how large language models could support practitioners by semi-automating parts of the process. The key challenge, they argue, is defining clear metrics to measure the quality of threat models — human or AI-generated — and they invite your input to help shape this research.

Meanwhile, on the Toreon blog, Robert Hurlbut highlights the distinctions between threat modeling and threat intelligence, showing how the two practices complement each other to strengthen both tactical defenses and strategic awareness.

There’s plenty of other actionable insight ahead, so settle in and let’s get started!

Threat Modeling Insider edition

Welcome!

Welcome to this month’s edition of Threat Modeling Insider! In our featured guest article, LLMs are the future of Threat Modeling, right?, Felix Viktor Jedrzejewski. Davide Fucci and Oleksandr Adamov explore how large language models could support practitioners by semi-automating parts of the process. The key challenge, they argue, is defining clear metrics to measure the quality of threat models — human or AI-generated — and they invite your input to help shape this research.

Meanwhile, on the Toreon blog, Robert Hurlbut highlights the distinctions between threat modeling and threat intelligence, showing how the two practices complement each other to strengthen both tactical defenses and strategic awareness.

There’s plenty of other actionable insight ahead, so settle in and let’s get started!

On this edition

Tips & tricks
OWASP® Cornucopia 2.2 & Copi – A Game Engine for OWASP® Cornucopia Threat Modeling.

Training update
An update on our training sessions.

Guest article

LLMs are the future of Threat Modeling, right?

Written by Felix Viktor Jedrzejewski, Ph.D. student at Blekinge Institute of Technology (BTH) and Co-Founder of Gaetir, Davide Fucci, Docent at BTH, and Oleksandr Adamov, Senior Lecturer at BTH

Is ChatGPT—or any other Large Language Model (LLM)—a suitable partner for building a threat model? Could we see bots that eventually replace threat modeling experts? Not quite. While LLMs look promising, we are still far from such a reality.

Threat modeling remains one of the most powerful ways to systematically identify which threats a system faces and how to mitigate them. But the process is resource-intensive, requiring security and domain expertise, multiple stakeholders and significant time investment, while often struggling to deliver measurable value.

Semi-automating selected threat modeling sub-steps with LLMs could lower the barrier to entry and spread the practice more widely. To achieve this, however, we need clear metrics to evaluate the quality of both human- and machine-generated threat models.

In this article, we share insights from an ongoing research project called ThreMoLIA  building such a tool.

Moreover, we, the researchers involved in this project, ask for your input on defining quality metrics for the future.

Why Threat Modeling Needs Innovation

Threat modeling requires domain and security expertise, a high degree of manual effort, and time from a large number of stakeholders.  Moreover, threat modeling practices vary widely among companies, even in the same domain which pose a challenge when it comes to ensuring consistency and scalability. 

Industry, including SMEs with limited resources, demands a more systematic, scalable, and accessible approach to threat modeling. On the surface, LLMs and GenAI technologies appear to be the answer to fulfil that demand. But whether you let a group of experts applying STRIDE to your system architecture or rely fully on an LLM a key issue is that there is no standard way to measure the quality of the resulting threat models. We need to establish metrics to evaluate and compare (LLM-)generated or manually created threat models, which go beyond the mere “an LLM is cheaper and can do it faster.”

Without reliable ways to measure the quality of threat models, automation risks amplifying inconsistency instead of solving it. That is why defining clear, agreed-upon quality metrics is not just helpful—it is essential.

Enter LLM-Based Threat Modeling

Our idea for an LLM-based threat modeling consists of the co-production between practitioners and an LLM-based chatbot that supports the practitioner across all threat modeling phases. In our position paper[1], we presented a workflow concept of an LLM-based threat modeling tool, which is shown below.

[1] https://arxiv.org/pdf/2504.18369

Concept of an LLM-Based Threat Modeling Workflow
Concept of an LLM-Based Threat Modeling Workflow

The RAG generates context input based on a variety of system-specific data. The tool generates a threat model and simultaneously calculates a health score based on a set of empirically evaluated metrics we are currently collecting in our study.

Such metrics address one of the elephants in the room. Applying LLMs, in any capacity, pose a privacy and security concern for the input data. Generally, LLMs tend to be very context sensitive caused by the large variety of training data commercial LLMs are trained on. This also leads to potential reliability issues posed by hallucinations that can lead to wrong security decisions badly impacting the security posture of a company.

All in all, our proposed workflow and the included tool will support non-experts to accelerate the threat modeling sessions, enabling an almost continuous threat modeling.

Research Project Insights

Our research project called ThreMoLIA, Threat Modeling of LLM-Integrated Applications, financed by the Swedish Agency for Innovation Systems (Vinnova), started in 2024.[1] The goal of the project is to build a threat modeling tool together with our industry partner Ericsson AB. So far, we built a prototype according to our earlier mentioned position paper. Naturally, we aim to let practitioners test our tool on a simulated system before testing it in a full industrial context. In both cases, we require metrics to assess performance of our tool. In parallel, we started a study in which we investigate metrics evaluating threat model quality. More details about the studies and how they align in the research project will be published soon.[2]  

[1] https://www.bth.se/english/research/research-projects/thremolia—threat-modeling-for-llm-integrated-applications

[2] https://conf.researchr.org/details/esem-2025/esem-2025-research-projects-track-/3/Threat-Modeling-for-Large-Language-Model-Integrated-Applications-ThreMoLIA-

Measuring the Quality of Threat Models

The core challenge of ThreMoLIA is the lack of metrics to assess the quality of a given threat model. The goal of our ongoing study is to establish these metrics as a co-production between academia and industry.

We began by reviewing academic literature to extract metrics researchers have proposed for evaluating threat models. Next, we interviewed practitioners to test these metrics against their real-world experiences as we believe that such insights are essential to ensure our tool reflects industrial reality.

 So far, our study points to three broad categories of quality metrics:

  • Understandability – Can the entire threat model be read and comprehended with reasonable effort?
  • Coverage – Does the threat model capture all relevant threats?
  • Correctness – Are the identified threats valid and accurate?

We want your perspective: do these categories reflect your experience, or is something important missing?

Call to Action: Participate in the Survey

Your perspective will directly shape how future threat modeling tools are built and evaluated. We invite you to share your opinion in a short, anonymous, 15-minute survey:

Your input will ground academic findings in practice helping to define the how to evaluate the next generation of threat models created by hybrid human-AI teams.

Conclusion

Threat modeling is a cornerstone of secure software engineering, but it is often slow, costly, and inconsistent. By automating some of its most resource-intensive phases, we can make it faster, cheaper, and more accessible, paving the way for broader adoption across industry.

The future of threat modeling will not be defined only by new fancy AI-based tools but by we measure their effectiveness. Succeeding in building those metrics together will empower the next generation of threat modelers working efficiently with their AI counterparts.

CURATED CONTENT

Handpicked for you

Toreon Blog: Threat Modeling and Threat Intelligence: Distinct and Complementary

Threat modeling and threat intelligence are essential practices of a proactive security strategy. Occasionally, these terms can be confusing or mixed up because they sound similar (both start with “threat” – aren’t they the same?). Instead, these terms represent distinct and complementary approaches to understanding and mitigating cybersecurity risks. This blog post will explore their differences, how they complement each other, and how they can be integrated to provide a more secure posture for your organization.

The lethal trifecta for AI agents: private data, untrusted content, and external communication

Simon Willison reveals a critical security vulnerability in AI agents that can allow attackers to steal private data by exploiting the combination of three dangerous capabilities: access to private data, exposure to untrusted content, and external communication abilities.

Key takeaways are:

  • LLMs inherently follow instructions in content without reliably distinguishing their origin or importance
  • Combining tools with private data access, untrusted content, and external communication creates a severe security risk
  • Current guardrail solutions are unreliable, with most claiming only 95% attack prevention

Intent Over Tactics: A CISO's Guide to Protecting Your Crown Jewels

Many new CISOs face the overwhelming challenge of developing a comprehensive security strategy. The task is not only to secure the organization but also to create a plan that is practical, focused, and understandable for everyone, from the executive team to the operational staff. 
 
A good strategy must pass the crucial test of increasing the cost for attackers while making the company safer. With countless vulnerabilities to address, it’s essential to design a strategy that resonates with both board members and technical teams. This curated section offers a framework for creating, pitching, and implementing a security strategy that meets these high standards. For further insights, refer to the original article.

TIPS & TRICKS

OWASP® Cornucopia 2.2 & Copi - A Game Engine for OWASP® Cornucopia Threat Modeling

The pandemic drove a considerable increase in fully remote teams, which made card games quite difficult to organize. Therefore, in 2022, Grant Ongers was willing to bet a dinner at a fancy vegan restaurant that his former colleague Toby Irvine wouldn’t be able to build a fully fledged and online game engine based on the game Cornucopia (from the OWASP® Foundation) over the weekend.

There is now a new release of OWASP Cornucopia 2.2 to celebrate a new milestone in the project’s history.

Our trainings & events for 2025

Book a seat in our upcoming trainings & events

AI Whiteboard Hacking, aka Hands-On 1-Day Threat Modeling Workshop, in-person, CyberSecurity Intersection, Orlanda, USA

10 October 2025

Advanced Whiteboard Hacking a.k.a. Hands-on Threat Modeling, in-person, OWASP Global AppSec, Washington DC

4-5 November 2025

Threat Modeling Practitioner training, hybrid online, hosted by DPI 

Cohort starting on 1 December 2025

AI Whiteboard Hacking, aka Hands-On 1-Day Threat Modeling Workshop, in-person, CyberSecurity Intersection, Orlanda, USA

10 October 2025

Advanced Whiteboard Hacking a.k.a. Hands-on Threat Modeling, in-person, OWASP Global AppSec, Washington DC

4-5 November  2025

Threat Modeling Practitioner training, hybrid online, hosted by DPI

Cohort starting on 1 December

1 Day Workshop Threat Modeling with AI, in-person, Belgium, OWASP BeNeLux

Cohort starting on 1 December 2025

Advanced Whiteboard Hacking a.k.a. Hands-on Threat Modeling, in-person, Blue Team Con, Chicago, USA

1-2 December 2025

1 Day Workshop Threat Modeling with AI, in-person, Belgium, OWASP BeNeLux

Cohort starting on 1 December 2025

Advanced Whiteboard Hacking a.k.a. Hands-on Threat Modeling, in-person, Blue Team Con, Chicago, USA

1-2 December 2025

Threat Modeling Insider Newsletter

Delivering the latest Threat Modeling articles and tips straight to your mailbox.

Start typing and press Enter to search

Shopping Cart