Agent Misalignment and Insider Threats: A Strategic Risk for AI Governance
Denis de Montigny PhD, CFA Anthropic’s 2025 paper, Agentic Misalignment: How LLMs Could Be an Insider Threat, highlights a risk that boards, regulators, and investors should address directly: large language models (LLMs), if not properly governed, could behave like insider threats—leaking sensitive information, undermining decisions, or misusing internal workflows. The paper goes beyond technical vulnerabilities […]
Agent Misalignment and Insider Threats: A Strategic Risk for AI Governance Read More »