Accelerating Threat Modeling Through Red Team Brainstorming with GPTs
Large language models (LLMs) like GPTs have gained notoriety for generating inaccurate information. However, for Erica Burgess, an artificial intelligence cybersecurity architect, these “hallucinations” can serve a beneficial role in threat modeling. “I prefer to view these hallucinations as untested…