The Ghost in the Machine: Counterterrorism in the Age of Artificial Intelligence

The Ghost in the Machine: Counterterrorism in the Age of Artificial Intelligence

The Ghost in the Machine: What Generative AI Means for Financial Crime and Counterterrorism

Generative AI and large language models are changing how information is created, consumed, and acted upon. For financial crime professionals — compliance officers, fraud investigators, AML analysts, bank risk managers, and policy - makers—this change is not abstract. It affects how illicit finance is detected, how terrorist networks raise and move funds, and how regulators and firms balance privacy, accuracy, and operational speed. Drawing on Christopher Wall’s analysis of AI’s role in counterterrorism, this article reframes those insights specifically for the financial crime community and highlights strategic choices that will determine whether AI reduces harm or amplifies risk.

AI’s new role: from analytic assistant to decision amplifier

After 9/11, intelligence and security agencies layered analytics and machine learning onto existing human-led counterterrorism processes. Those tools helped sift large data sets, prioritize leads, and guide operations — but humans retained ultimate judgment. With generative AI and advanced LLMs, machines begin to perform higher-level reasoning tasks: synthesizing multilingual documents, drafting strategic summaries, and proposing responses. In financial crime control, those same capabilities can make transaction monitoring, suspicious activity reporting, sanctions screening, and typology discovery vastly more scalable — but only if firms and supervisors understand the limits, biases, and governance demands of these systems.

Bastian Schwind-Wagner
Bastian Schwind-Wagner "Generative AI offers powerful tools to detect and disrupt terrorist financing and complex financial crime, but its benefits depend on careful data governance, human oversight, and legal safeguards to prevent bias and wrongful enforcement. Responsible deployment requires explainability, rigorous validation, and cross‑sector collaboration so that speed and scale do not come at the cost of rights or public trust."
Why the data problem matters more than the model

AI performance is driven primarily by the data used to train and fine-tune models and by how models are scoped and constrained. For financial crime work this means:

  • Historical transaction data, alert dispositions, investigation narratives, and case outcomes are the lifeblood of any ML system. If the underlying data embed past bias — over‑policing certain geographies, industries, or demographic groups — models will replicate and amplify those distortions.
  • Data gaps matter: many illicit finance patterns are rare, intentionally concealed, or context‑dependent. Generative AI can hallucinate plausible but incorrect narratives when data are sparse or noisy, producing false leads or false assurances.
  • Retrieval‑augmented and domain‑tuned approaches (RAG, fine‑tuning on high‑quality typology datasets, knowledge graphs) reduce hallucination and improve relevance. The payoff in AML/CTF is better scenario prioritization and fewer wasted investigations — if models are fed curated, bias‑checked corpora.
Operational use-cases and attendant risks
Counter‑radicalization analogues: tailored interventions and messaging

Generative models can analyze extremist finance narratives — crowdfunding appeals, crypto‑fundraising messages, or obfuscated supplier invoices — and propose counter‑messaging, disruption options, or compliance flags. In financial crime, similar capabilities can:

  • Generate humanized outreach scripts for caseworkers or local NGOs to de‑escalate recruitment or fundraising channels.
  • Draft regulatory notices or embargo justification memos that explain complex typologies to non‑technical decision‑makers.

Risks: automated messaging and impersonation tools may be abused by bad actors to scale fraud or to perform social‑engineering attacks that evade detection.

Intelligence and anomaly detection: faster triage, broader reach

Multimodal LLMs can fuse text, transaction logs, geospatial metadata, voice transcripts, and open‑source intelligence to surface anomalous patterns faster than human teams alone. Practical benefits include accelerated typology discovery (merchant‑less trading patterns, trade‑based money laundering schemes, or layering strategies across crypto and fiat rails) and multilingual coverage without needing a full roster of linguists.

Risks: false negatives (missed attacks or hidden financing) and false positives (spurious alerts that overwhelm investigators). Language‑agnostic detection is powerful, but errors in an unfamiliar jurisdiction’s data or cultural context can be costly.

Countermeasures and kinetic analogues: automated takedowns vs. due process

In a military context, LLMs can advise targeting; in financial crime they can recommend enforcement actions—account freezes, suspicious activity reports, or sanction designations. Speed can be life‑saving, but automated enforcement raises critical issues:

  • Wrongful freezes or misattributed transactions can devastate businesses or individuals and cause legal exposure for providers and regulators.
  • Systems that optimize for “alerts closed” may incentivize over‑blocking or rejecting transactions in marginal cases, degrading financial inclusion and trust.
Ethics, governance and the strategic choices financial institutions must make

Human values must govern AI deployment. Wall’s central point — that machines are extensions of human values, not moral agents — translates directly to compliance: firms must decide what is an acceptable trade‑off between detection sensitivity and civil liberties, between shrinking risk and preserving client experience.

Three practical governance priorities:

  1. Data strategy and bias control

    Firms must invest in curated, annotated datasets for typologies and transactions, and in procedures to detect and mitigate representational bias. That includes:

  2. Explainability and fail‑safe controls

    AI systems must be auditable and interpretable where consequences are material. Explainable AI techniques and operational controls (dual‑validation of high‑impact decisions, human‑in‑the‑loop for account closures or sanctions) limit harm while allowing models to accelerate routine work. Kill switches, versioned rollouts, and mandatory sunset clauses for deployed models can prevent gradual mission creep.

  3. Legal alignment and cross‑sector coordination

    AI for AML/CTF sits inside legal constraints on privacy, data sharing, and due process. Regulators and firms should co‑design sandboxed experiments and clear guidance on acceptable automation levels for enforcement. Internationally, technology sharing should be cognizant of governance risk: exporting powerful screening tools without safeguards can enable authoritarian misuse.

When AI fails: lessons to avoid the October‑7 pattern

Wall highlights an instructive case: a military intelligence body reorganized around AI failed to raise timely warning before a devastating attack and simultaneously produced many false positives during the conflict that increased civilian harm. For financial crime teams, the parallel is stark: over‑reliance on imperfect models can both miss major illicit financing and create collateral damage through erroneous enforcement. Practical lessons:

  • Preserve and incentivize human expertise in edge cases where context, language nuance, and human networks matter.
  • Treat AI outputs as augmentations of cognition, not replacements.
  • Maintain open channels for analysts to escalate and to contest model recommendations.
Opportunity framing: reducing harm, not just scaling detection

Generative AI’s strongest value proposition for financial crime is reducing human cognitive load so investigators and policy‑makers can focus on strategy: disrupting networks that finance violence, restoring victims, and protecting vulnerable populations from wrongful action. When used judiciously, AI can:

  • Detect novel laundering techniques faster;
  • Free expert investigators to pursue complex cross‑border cases; and
  • Improve the timeliness and quality of SARs and intelligence products furnished to law enforcement.

But these gains require disciplined governance, investments in data and human capital, and political will to align technology with rule‑of‑law values.

Concluding guidance for financial crime leaders

Generative AI will not remove human responsibility. It will magnify it. Financial crime leaders should act on three near‑term priorities:

  • Build a rigorous data governance program that includes bias detection, typology curation, and model validation against real investigative outcomes.
  • Require explainability and enforce human oversight for any automated enforcement action with material consequences (freezes, delisting, regulatory referrals).
  • Engage regulators, counterparts, and civil society in joint sandboxes to develop shared norms and interoperability standards that prevent misuse while enabling legitimate, timely disruption of illicit finance.

Generative models are tools with extraordinary reach. Whether they become a force multiplier for effective, rights‑respecting anti‑money‑laundering and counterterrorist financing work depends on deliberate human choices — not on machines deciding for us.

The information in this article is of a general nature and is provided for informational purposes only. If you need legal advice for your individual situation, you should seek the advice of a qualified lawyer.
Did you find any mistakes? Would you like to provide feedback? If so, please contact us!
Dive deeper
  • Research ¦ Wall, C. (2025). The Ghost in the Machine: Counterterrorism in the Age of Artificial Intelligence. Studies in Conflict & Terrorism, 1–27. https://doi.org/10.1080/1057610X.2025.2475850 ¦ Link ¦ licensed under the following terms, with no changes made: license icon CC BY-NC-ND 4.0
Bastian Schwind-Wagner
Bastian Schwind-Wagner Bastian is a recognized expert in anti-money laundering (AML), countering the financing of terrorism (CFT), compliance, data protection, risk management, and whistleblowing. He has worked for fund management companies for more than 24 years, where he has held senior positions in these areas.