Matthew Toy

Thoughts, reflections and experiences

icy banner
WHY THE UK AND EU ARE LOSING THE AI INFLUENCE WAR

Why the UK and EU Are Losing the AI Influence War

Abstract

Western democracies face a new front in conflict: the cognitive battlespace, where artificial intelligence (AI) is leveraged to shape public opinion and influence behaviour. This article argues that the UK and EU are currently losing this AI-driven influence war. Authoritarian adversaries like Russia and China are deploying AI tools in sophisticated disinformation and propaganda campaigns, eroding trust in democratic institutions and fracturing social cohesion. In contrast, the UK and EU response, focused on regulation, ethical constraints, and defensive measures, has been comparatively slow and fragmented. Without a more proactive and unified strategy to employ AI in information operations and bolster societal resilience against cognitive warfare, Western nations risk strategic disadvantage. This article outlines the nature of the cognitive battle-space, examines adversarial use of AI in influence operations, evaluates UK/EU efforts and shortcomings, and suggests why urgent action is needed to regain the initiative.

Introduction

Modern conflict is no longer confined to conventional battlefields; it has expanded into the cognitive domain. The term “cognitive battlespace” refers to the arena of information and ideas, where state and non-state actors vie to influence what people think and how they behave. Today, advances in AI have supercharged this domain, enabling more sophisticated influence operations that target the hearts and minds of populations at scale. Adversaries can weaponise social media algorithms, deepfakes, and data analytics to wage psychological warfare remotely and relentlessly.

Western governments, particularly the United Kingdom and European Union member states, find themselves on the defensive. They face a deluge of AI-enhanced disinformation from authoritarian rivals but are constrained by ethical, legal, and practical challenges in responding. Early evidence suggests a troubling imbalance: Russia and China are aggressively exploiting AI for propaganda and disinformation, while the UK/EU struggle to adapt their policies and capabilities. As a result, analysts warn that Western democracies are “losing the battle of the narrative” in the context of AI (sciencebusiness.net). The stakes are high: if the UK and EU cannot secure the cognitive high ground, they risk erosion of public trust, social discord, and strategic loss of influence on the world stage.

This article explores why the UK and EU are lagging in the AI influence war. It begins by defining the cognitive battlespace and the impact of AI on information warfare. It then examines how adversaries are leveraging AI in influence operations. Next, it assesses the current UK and EU approach to cognitive warfare and highlights key shortcomings. Finally, it discusses why Western efforts are falling behind and what the implications are for future security.

The Cognitive Battlespace in the Age of AI

In cognitive warfare, the human mind becomes the battlefield. As one expert succinctly put it, the goal is to “change not only what people think, but how they think and act” (esdc.europa.eu). This form of conflict aims to shape perceptions, beliefs, and behaviours in a way that favours the aggressor’s objectives. If waged effectively over time, cognitive warfare can even fragment an entire society, gradually sapping its will to resist an adversary.

Artificial intelligence has become a force multiplier in this cognitive domain. AI algorithms can curate individualised propaganda feeds, amplify false narratives through bot networks, and create realistic fake images or videos (deepfakes) that blur the line between truth and deception. According to NATO’s Allied Command Transformation, cognitive warfare encompasses activities to affect attitudes and behaviours by influencing human cognition, effectively “modifying perceptions of reality” as a new norm of conflict (act.nato.int). In essence, AI provides powerful tools to conduct whole-of-society manipulation, turning social media platforms and information systems into weapons.

A vivid example of the cognitive battlespace in action occurred in May 2023, when an AI-generated image of a false Pentagon explosion went viral. The fake image, disseminated by bots, briefly fooled enough people that it caused a sharp but temporary dip in the U.S. stock market. Though quickly debunked, this incident demonstrated the “catastrophic potential” of AI-driven disinformation to trigger real-world consequences at machine speed (mwi.westpoint.edu) . Generative AI can manufacture convincing yet false content on a massive scale, making it increasingly difficult for populations to discern fact from fabrication.

In the cognitive battlespace, such AI-enabled tactics give malign actors a potent advantage. They can rapidly deploy influence campaigns with minimal cost or risk, while defenders struggle to identify and counter each new false narrative. As the information environment becomes saturated with AI-amplified propaganda, the traditional defenders of truth, journalists, fact-checkers, and institutions, find themselves overwhelmed. This asymmetry is at the heart of why liberal democracies are in danger of losing the cognitive war if they do not adapt quickly.

Adversaries’ AI-Driven Influence Operations

Russia and China have emerged as leading adversaries in the AI-enabled influence war, honing techniques to exploit Western vulnerabilities in the cognitive domain. Russia has a long history of information warfare against the West and has eagerly integrated AI into these efforts. Through troll farms and automated bot networks, Russia pushes AI-generated propaganda designed to destabilise societies. Moscow views cognitive warfare as a strategic tool to “destroy [the West] from within” without firing a shot. Rather than direct military confrontation with NATO (which Russia knows it would likely lose), the Kremlin invests in “cheap and highly effective” cognitive warfare to undermine Western democracies from inside (kew.org.pl) .

Russian military thinkers refer to this concept as “reflexive control,” essentially their doctrine of cognitive warfare. The idea is to manipulate an adversary’s perception and decision-making so thoroughly that the adversary “defeats themselves”. In practice, this means saturating the information space with tailored disinformation, conspiracy theories, and emotionally charged content to break the enemy’s will to resist. As one analysis describes, the battleground is the mind of the Western citizen, and the weapon is the manipulation of their understanding and cognition. By exploiting human cognitive biases, our tendencies toward emotional reaction, confirmation bias, and confusion, Russia seeks to leave citizens “unable to properly assess reality”, thus incapable of making rational decisions (for example, in elections). The goal is a weakened democratic society, rife with internal divisions and distrust, that can no longer present a united front against Russian aggression.

Concrete examples of Russia’s AI-fuelled influence operations abound. Beyond the fake Pentagon incident, Russian operatives have used generative AI to create deepfake videos of European politicians, forge fake news stories, and impersonate media outlets. Ahead of Western elections, Russian disinformation campaigns augmented with AI have aimed to sow discord and polarise public opinion. U.K. intelligence reports and independent researchers have noted that Russia’s automated bot accounts are evolving to produce more “human-like and persuasive” messages with the help of AI language models. These tactics amplify the reach and realism of propaganda, making it harder to detect and counter. Even if such interference does not always change election outcomes, it erodes public trust in information and institutions, a long-term win for the Kremlin.

China, while a newer player in European information spaces, is also investing heavily in AI for influence operations. Chinese military strategy incorporates the concept of “cognitive domain operations”, which merge AI with psychological and cyber warfare. Beijing’s aim is to shape global narratives and public opinion in its favour, deterring opposition to China’s interests. For instance, China has deployed swarms of AI-driven social media bots to spread disinformation about topics like the origins of COVID-19 and the status of Taiwan. Chinese propaganda operations use AI to generate deepfake news anchors and social media personas that promote pro-China narratives abroad. According to NATO analysts, China describes cognitive warfare as using public opinion and psychological manipulation to achieve victory, and invests in technologies (like emotion-monitoring systems for soldiers) that reveal the importance it places on the information domain. While China’s influence efforts in Europe are less overt than Russia’s, they represent a growing challenge as China seeks to project soft power and shape perceptions among European audiences, often to dissuade criticism of Beijing or divide Western unity.

The aggressive use of AI by authoritarian adversaries has put Western nations on the back foot in the information environment. Adversaries operate without the legal and ethical constraints that bind democracies. They capitalise on speed, volume, and ambiguity, launching influence campaigns faster than defenders can react. Authoritarian regimes also coordinate these efforts as part of broader hybrid warfare strategies, aligning cyber-attacks, diplomatic pressure, and economic coercion with information operations to maximise impact. In summary, Russia and China have seized the initiative in the cognitive battlespace, leaving the UK, EU, and their allies scrambling to catch up.

UK and EU Responses: Strategies and Shortcomings

Confronted with these threats, the United Kingdom and European Union have begun to recognise the urgency of the cognitive warfare challenge. In recent years, officials and strategists have taken steps to improve defences against disinformation and malign influence. However, the Western approach has so far been largely reactive and constrained, marked by cautious policy frameworks and fragmented efforts that lag the adversary’s pace of innovation.

United Kingdom: The UK government acknowledges that AI can significantly amplify information warfare. The Ministry of Defence’s Defence Artificial Intelligence Strategy (2022) warns that “AI could also be used to intensify information operations, disinformation campaigns and fake news,” for example by deploying deepfakes and bogus social media accounts. British military doctrine, including the Integrated Operating Concept (2020), emphasises that information operations are increasingly important to counter false narratives in modern conflicts (gov.uk). London’s approach has included establishing units dedicated to “strategic communications” and cyber influence and working with partners like NATO to improve information security.

The UK has also invested in research on AI and influence. For instance, the Alan Turing Institute’s research centre (CETaS) published analyses on AI-enabled influence operations in the 2024 UK elections, identifying emerging threats such as deepfake propaganda and AI-generated political smear campaigns. These studies, while finding that AI’s impact on recent elections was limited, highlighted serious concerns like AI-driven hate incitement and voter confusion (cetas.turing.ac.uk) . The implication is clear: the UK cannot be complacent. Even if traditional disinformation methods still dominate, the rapid evolution of AI means influence threats could scale up dramatically in the near future. British policymakers have started to discuss new regulations (for example, requiring transparency in AI political ads) and bolstering media literacy programs to inoculate the public against fake content.

Despite this awareness, critics argue that the UK’s response remains disjointed and under-resourced. There is no publicly articulated doctrine for cognitive warfare equivalent to adversaries’ strategies. Efforts are split among various agencies (from GCHQ handling cyber, to the Army’s 77th Brigade for information ops, to the Foreign Office for counter-disinformation), making coordination challenging. Moreover, while defensive measures (like fact-checking services and takedown of fake accounts) have improved, the UK appears reluctant to consider more assertive offensive information operations that could pre-empt adversary narratives. Legal and ethical norms, as well as fear of escalation, likely restrain such tactics. The result is that Britain often plays catch-up, reacting to disinformation waves after they have already influenced segments of the population.

European Union: The EU, as a bloc of democracies, faces additional hurdles in confronting cognitive warfare. Brussels has treated disinformation chiefly as a policy and regulatory issue tied to election security and digital platform accountability. Over the past few years, the EU implemented a Code of Practice on Disinformation (a voluntary agreement with tech companies) and stood up teams like the East StratCom Task Force (known for its EUvsDisinfo project debunking pro-Kremlin myths). Following high-profile meddling in elections and referendums, EU institutions have grown more vocal: they label Russia explicitly as the chief source of disinformation targeting Europe. The European Commission also incorporated anti-disinformation clauses into the Digital Services Act (DSA), requiring large online platforms to assess and mitigate risks from fake content.

When it comes to AI, the EU’s landmark AI Act – primarily a regulatory framework to govern AI uses – indirectly addresses some information manipulation concerns (for example, by requiring transparency for deepfakes). However, EU efforts are fundamentally defensive and norm-driven. They seek to police platforms and inform citizens, rather than actively engage in influence operations. EU leaders are wary of blurring the line between counter-propaganda and propaganda of their own, given Europe’s commitment to free expression. This creates a dilemma: open societies find it difficult to wage information war with the ruthlessness of authoritarian regimes.

European security experts are starting to grapple with this challenge. A recent EU security and defence college course underscored that cognitive warfare is an “emerging challenge” for the European Union (esdc.europa.eu) . Participants discussed the need for technological tools to detect, deter, and mitigate cognitive threats. Yet, outside of specialised circles, there is no EU-wide military command focused on cognitive warfare (unlike traditional domains such as land, sea, cyber, etc.). NATO, which includes most EU countries, has taken the lead in conceptualising cognitive warfare, but NATO’s role in offensive information activities is limited by its mandate.

A telling critique comes from a Royal United Services Institute (RUSI) commentary on disinformation and AI threats. It notes that NATO’s 2024 strategy update acknowledged the dangers of AI-enabled disinformation, using unusually strong language about the urgency of the challenge. However, the same strategy “makes no reference to how AI could be used” positively for strategic communications or to help counter disinformation (rusi.org) . In other words, Western nations are emphasising protection and defence, strengthening **governance standards, public resilience, and truth-checking mechanisms, **but they are not yet leveraging AI offensively to seize the initiative in the info sphere. This cautious approach may be ceding ground to adversaries who have no such reservations.

Why the West Is Losing the AI Influence War

Several interrelated factors explain why the UK, EU, and their allies appear to be losing ground in the cognitive domain against AI-equipped adversaries:

Reactive Posture vs. Proactive Strategy: Western responses have been largely reactive. Democracies often respond to disinformation campaigns after damage is done, issuing fact-checks or diplomatic condemnations. There is a lack of a proactive, comprehensive strategy to dominate the information environment. Adversaries, by contrast, set the narrative by deploying influence operations first and fast.

Ethical and Legal Constraints: The UK and EU operate under strict norms – adherence to truth, rule of law, and respect for civil liberties – which limit tactics in information warfare. Propaganda or deception by government is domestically unpopular and legally fraught. This makes it hard to match the scale and aggressiveness of Russian or Chinese influence operations without undermining democratic values. Authoritarians face no such constraints.

Fragmented Coordination: In Europe, tackling cognitive threats cuts across multiple jurisdictions and agencies (domestic, EU, NATO), leading to fragmentation. A unified command-and-control for information operations is lacking. Meanwhile, adversaries often orchestrate their messaging from a centralised playbook, giving them agility and consistency.

Regulatory Focus Over Capabilities: The EU’s inclination has been to regulate (AI, social media, data) to create guardrails – a necessary but slow process. However, regulation alone does not equal capability. Rules might curb some harmful content but do not stop a determined adversary. The West has invested less in developing its own AI tools for strategic communication, psyops, or rapid counter-messaging. This capability gap means ceding the technological edge to opponents.

Underestimation of Cognitive Warfare: Historically, Western security doctrine prioritised physical and cyber threats, sometimes underestimating the impact of information warfare. The concept of a sustained “cognitive war” waged in peacetime is relatively new to Western planners. Initial responses were tepid – for example, before 2016, few anticipated that online influence could significantly affect major votes. This lag in appreciation allowed adversaries to build momentum.

These factors contribute to a situation where, despite growing awareness, the UK and EU have struggled to turn rhetoric into effective countermeasures on the cognitive front. As a result, authoritarian influence campaigns continue to find fertile ground in Western societies. Each viral conspiracy theory that goes unchecked, each wedge driven between communities via disinformation, and each doubt cast on democratic institutions chips away at the West’s strategic advantage. NATO officials warn that information warfare threats “must neither be overlooked nor underestimated” in the face of the AI revolution. Yet current efforts remain a step behind the onslaught of AI-generated falsehoods.

Conclusion and Implications

If the UK, EU, and like-minded democracies do not rapidly adapt to the realities of AI-driven cognitive warfare, they risk strategic defeat in an important realm of 21st-century conflict. Losing the AI influence war doesn’t happen with a formal surrender; instead, it manifests as a gradual erosion of democratic resilience. Societies may grow deeply divided, citisens lose trust in media and governments, and adversarial narratives become entrenched. In the long run, this could weaken the political will and cohesion needed to respond to more conventional security threats. As one analysis grimly observed, the cost of inaction is high – allowing adversaries to exploit AI for malign influence can lead to a “strategic imbalance favouring adversaries”, with a flood of false narratives eroding public trust and even devastating democratic institutions if left unchecked.

Reversing this trajectory will require Western nations to elevate the priority of the cognitive battlespace in national security planning. Some broad imperatives emerge:

Develop Offensive and Defensive AI Capabilities: The UK and EU should invest in AI tools not just to detect and debunk disinformation, but also to disseminate counter-narratives that truthfully push back against authoritarian propaganda. Ethical guidelines for such operations must be established, but fear of using AI at all in information ops leaves the field open to adversaries.

Whole-of-Society Resilience: Building public resilience is crucial. Education in media literacy and critical thinking, transparency about threats, and empowering independent journalism are all part of inoculating society. A populace that can recognise manipulation is the best defence against cognitive warfare. The goal is to ensure citizens can engage with digital information sceptically, blunting the impact of fake or AI-manipulated content.

International Coordination: The transatlantic alliance and democratic partners need better coordination in the information domain. NATO’s work on cognitive warfare should be complemented by EU and UK joint initiatives to share intelligence on disinformation campaigns and align responses. A unified front can deny adversaries the ability to play divide-and-conquer with different countries.

Adaptive Governance: Western policymakers must make their regulatory frameworks more agile in the face of technological change. This might include faster mechanisms to hold platforms accountable, updated election laws regarding AI-generated content, and perhaps narrowly tailored laws against the most dangerous forms of disinformation (such as deceptive media that incites violence). The challenge is doing so without undermining free speech – a balance that requires constant calibration as AI technology evolves.

In summary, the UK and EU are at a crossroads. They can continue on the current path – risking that AI-enabled influence attacks will outpace their responses – or they can strategise anew and invest in winning the cognitive fight. The latter will demand political will and creativity: treating information space as a domain to be secured, much like land, sea, air, cyber and space. It also means confronting uncomfortable questions about using emerging technologies in ways that align with democratic values yet neutralise malign propaganda.

The cognitive battle-space is now a permanent feature of international security. Western democracies must not cede this battlefield. Maintaining an open society does not mean being defenceless. With prudent adoption of AI for good, and a staunch defence of truth, the UK, EU, and their allies can start to turn the tide in the AI influence war. Failing to do so will only embolden those who seek to “attack the democratic pillars of the West” through information manipulation. In this contest for minds and hearts, as much as in any other domain of conflict, strength and resolve will determine who prevails.

Bibliography

1. NATO Allied Command Transformation. “Cognitive Warfare.” NATO ACT, Norfolk VA.

2. Bryc, Agnieszka. “Destroy from within: Russia’s cognitive warfare on EU democracy.” Kolegium Europy Wschodniej, 27 Nov 2024.

3. European Security & Defence College (ESDC). “Cognitive warfare in the new international competition: an emerging challenge for the EU,” 28 May 2024.

4. Williams, Cameron (Modern War Institute). “Persuade, Change, and Influence with AI: Leveraging Artificial Intelligence in the Information Environment.” Modern War Institute at West Point, 14 Nov 2023.

5. UK Ministry of Defence. Defence Artificial Intelligence Strategy, June 2022. UK

6. Fitz-Gerald, Ann M., and Halyna Padalko (RUSI). “The Need for a Strategic Approach to Disinformation and AI-Driven Threats.” RUSI Commentary, 25 July 2024.

7. Science Business News. “EU is ‘losing the narrative battle’ over AI Act, says UN adviser,” 05 Dec 2024.

The Grey Mirage: Navigating Strategic Uncertainty and the Elusive Victory in Grey Zone Conflicts

Imagine a world where war is waged not with bombs and bullets, but with lines of code and viral misinformation. This is the reality of grey zone conflicts, a persistent feature of modern geopolitics characterised by cyber operations, economic coercion, and disinformation. While many initially hailed these tactics as a revolutionary new form of strategic competition, a critical examination reveals that they not only fundamentally fail to achieve strategic victory in a traditional Clausewitzian sense but also introduce profound strategic uncertainty and volatility into the international system. Extending Thomas Rid’s compelling argument that “cyber war will not take place” due to the inherent lack of decisive physical destruction, this critique applies even more broadly to the entire spectrum of grey zone conflicts.

To understand the inherent limitations of these operations, we must return to the foundational strategic thought of Carl von Clausewitz. His framework remains a lodestar: tactical successes must always serve political objectives, and the very essence of war is to impose one’s will upon the enemy. As Michael Handel succinctly summarises, Clausewitzian war aims at the destruction of enemy forces, control of vital resources, and the sway of public opinion. Grey zone tactics, however, are structurally incapable of achieving these aims in the decisive manner Clausewitz envisioned. They may sow disruption and discord, but they rarely deliver battlefield outcomes, nor can they compel political compliance in the way traditional military campaigns do. Consider, for instance, the persistent cyberattacks between nations; while disruptive and costly, they have yet to force a nation to fundamentally alter its core strategic direction.

The very nature of grey zone strategies – their calculated avoidance of outright force and immediately recognisable acts of aggression – means they cannot truly compel an adversary to accept a fundamentally new strategic order. Cyber operations, as Rid convincingly argues, rarely inflict the kind of lasting, tangible damage comparable to conventional military strikes. Disinformation campaigns, while capable of eroding trust in institutions and even mobilising populations, as seen in the Arab Spring uprisings, cannot on their own force political capitulation. Economic sanctions, though often painful and strategically useful in shaping behaviour, are notoriously slow and far from guaranteed to change a determined state’s core strategic calculations.

This inherent strategic limitation is further underscored by Colin Gray’s assertion that strategy is fundamentally about the application of force to achieve political objectives. For Gray, war is fundamentally about contesting and achieving control, and without the capacity to impose a decisive order, grey zone tactics fall drastically short of true strategic efficacy. He cautions that the absence of decisive engagement in contemporary conflicts leads not to resolution, but to a debilitating strategic paralysis. This resonates deeply with Clausewitz’s core tenet that successful war must culminate in the decisive defeat of the enemy. Grey zone conflicts, by their very nature, do not and cannot fulfil this criterion. At best, they generate protracted stalemates; at worst, they risk unintended escalation into open, conventional warfare.

Countering the Cumulative Argument and Embracing Ambiguity: Incrementalism vs. Decisiveness

It is important to acknowledge a key counterargument: that grey zone tactics, while rarely decisive alone, gain strategic effect cumulatively over time. Proponents argue that persistent cyber intrusions, disinformation, and economic pressure can erode an adversary’s strength and will. This view sees grey zone warfare as long-term shaping, not a knockout blow, exemplified by China’s “Three Warfares” doctrine.

Furthermore, the ambiguity of grey zone conflicts can be strategically useful, like nuclear deterrence. Bernard Brodie argued nuclear war’s cost shifted strategy to prevention, redefining “victory” as avoiding war. Similarly, grey zone tactics might deter and manage competition below open conflict. Incremental disruption, like cyberattacks on Iran’s nuclear program, can also shift power balances.

Hurting Without Winning and the Zero-Sum Nature of Grey Zone Competition

Thomas Schelling noted, “Victory is no longer a prerequisite for hurting the enemy.” This is key to grey zone tactics, which can aim to inflict pain and signal resolve without overt war. Even non-military gains – diplomatic wins, sanctions, legal advantages achieved through disinformation and cyber influence – become strategic victories in this zero-sum competition. This is particularly relevant as tech-savvy strategists recognise the advantages of ambiguity in these operations.

However, pursuing overwhelming military victory can backfire, escalating conflict. Grey zone tactics offer a way to avoid this, operating below the threshold of conventional war. Yet, this ambiguity breeds volatility, with miscalculation and escalation always looming.

Strategic Victory as Peace-Winning and the Challenge of Subjectivity

Rethinking “strategic victory” beyond military terms is crucial. Robert Mandel distinguishes “war-winning” from “peace-winning,” arguing true strategic victory is “peace-winning” – a multi-dimensional achievement across information, politics, economics, and diplomacy. Grey zone tactics align with this broader view, especially as public mobilisation and decentralised networks shape geopolitics.

Yet, “victory” in the grey zone remains subjective and hard to measure. Ethan Kapstein highlights the difficulty of defining metrics, gaining consensus, and obtaining reliable data in grey zone operations. Progress in one area may undermine another, increasing strategic uncertainty. Whether grey zone tactics are a “strategic win” depends on perspective and chosen metrics.

Taiwan: Strategic Uncertainty in Action

Taiwan exemplifies the inherent volatility of grey zone warfare: while hybrid strategies can pressure an opponent, they provide no clear pathway to a controlled, predictable outcome. The lack of definitive thresholds makes grey zone tactics as much a risk as an opportunity for the aggressor. Imagine China using grey zone tactics against Taiwan: cyberattacks, disinformation, and economic pressure. While this might weaken Taiwan, it’s unlikely to force capitulation without risking wider conflict. Taiwan’s reaction, U.S. responses, and the ever-present risk of miscalculation create a strategic dilemma.

While Russia has shown resilience to external grey zone pressures by controlling information, societal resilience only mitigates, not eliminates, strategic uncertainty. Even the most robust resilience strategies cannot eliminate the risk of miscalculation or escalation, underscoring the inherent volatility of grey zone conflicts. Because grey zone conflicts operate ambiguously, even careful campaigns can unexpectedly escalate, making control and predictability elusive.

Policy Implications: Actively Shaping the Grey Zone for Advantage

The inherent strategic uncertainty of grey zone conflicts demands proactive policies:

  1. Sharpen Intelligence and Active Disruption: Enhance intelligence to understand adversary intentions and develop capabilities to actively disrupt their grey zone operations.
  2. Develop Flexible and Escalatory Response Options: Create a wider range of responses, including calibrated counter-grey zone tactics and clear signalling for de-escalation and conflict management. As artificial intelligence and automation continue to reshape information warfare, states must anticipate how AI-driven disinformation, deepfake technology, and autonomous cyber operations will further complicate grey zone conflicts. Developing countermeasures that integrate AI-based detection and rapid-response systems will be critical for maintaining strategic advantage.
  3. Promote Transparency to Force Predictability: Actively expose adversary actions to force them into a more predictable strategic posture, enhancing transparency and accountability in the grey zone.
  4. Focus on Proactive Crisis Management: Develop proactive crisis management to prevent crises, including clear communication, de-escalation protocols, and persistent low-intensity engagement for stability.
  5. Re-evaluate “Victory” and Embrace Persistent Engagement: Shift from traditional victory metrics to measures of resilience, deterrence, and long-term shaping, embracing persistent engagement as the norm in grey zone competition.

Conclusion: Embracing Uncertainty, Seeking Control Through Persistent Engagement

Russia’s pre-2022 hybrid warfare campaign in Ukraine – combining cyber operations, disinformation, and economic pressure – demonstrated the limitations of grey zone tactics. Rather than coercing Ukraine into submission, these operations reinforced Ukrainian national resistance and galvanised Western military support, ultimately leading to Russia’s full-scale invasion. This case underscores the strategic volatility of grey zone competition: while these tactics can create disruption, they provide no guarantee of controlled, predictable outcomes.

This highlights how grey zone tactics, while seemingly flexible, are unlikely to deliver traditional strategic victory and introduce significant strategic uncertainty. While ambiguity and “peace-winning” are modern adaptations, they don’t guarantee predictable outcomes or escalation control. The grey zone is a volatile battlespace defined by miscalculation and instability. Navigating the grey zone requires embracing uncertainty, prioritising crisis management, and actively shaping the battlespace. In this new era of perpetual contestation, mastering the grey zone is not about winning – it is about ensuring that one’s adversaries never can.


References

  1. Brodie, Bernard. “The Absolute Weapon: Atomic Power and World Order.” The Yale Review 35, no. 3 (Spring 1946): 456-472.
  2. Gray, Colin S. The Strategy Bridge: Theory for Practice. Oxford: Oxford University Press, 2010.
  3. Handel, Michael I. Masters of War: Classical Strategic Thought. London: Frank Cass, 2001.
  4. Kania, Elsa B. “The PLA’s Latest Strategic Thinking on the Three Warfares.” The Jamestown Foundation, August 22, 2016. https://jamestown.org/program/the-plas-latest-strategic-thinking-on-the-three-warfares/.
  5. Kapstein, Ethan B. “Measuring Success in Complex Operations.” The Journal of Strategic Studies 34, no. 2 (April 2011): 267-285.
  6. Mandel, Robert. “Thinking about Victory in Strategy.” The Journal of Strategic Studies 34, no. 2 (April 2011): 199-200.
  7. Monaghan, Sean. “Twitter Revolutions? Social Media and the Arab Spring.” Whitehall Papers 69, no. 1 (2011): 21-22.
  8. Rid, Thomas. Cyber War Will Not Take Place. London: Hurst, 2013.
  9. Sanger, David E., and William J. Broad. “Obama Order Sped Up Wave of Cyberattacks Against Iran.” The New York Times, June 1, 2012. https://www.nytimes.com/2012/06/01/world/middleeast/obama-ordered-wave-of-cyberattacks-against-iran.html.
  10. Schelling, Thomas C. Arms and Influence. New Haven: Yale University Press, 1966.
  11. Simons, Greg. “Russia and information confrontation: perceptions, strategies and responses.” Journal of strategic studies 42, no. 1 (2019): 139-140.

Rethinking Warfare: Clausewitz in the Age of Cyber and Hybrid Conflict

Warfare in the age of cyber and hybrid conflict

Given the shifting sands of contemporary conflict, do we need to reassess the meaning of warfare? Clausewitz famously called war ‘a continuation of politics by other means’ (1832). But does that idea still hold up today? These days, conflicts play out on social media, in cyberspace, and even in elections—often without a single shot fired. Today’s battlespace incorporates cyber operations, climate change, mass-urbanisation, space weaponisation, and continuous strategic competition. This blurs the lines between war and peace. While classical theorists maintain that war’s fundamental nature has not changed, modern conflicts increasingly challenge traditional frameworks.

Historically, warfare was characterised by physical destruction, decisive battles, and territorial conquest. Modern conflicts, however, do not always adhere to this pattern. For instance, cyber warfare has shown that states and non-state actors can achieve strategic effects without kinetic violence. Thomas Rid (2017) contends that cyber operations can coerce, disrupt, and deceive, thereby challenging Clausewitz’s notion that war is inherently violent. The 2007 cyberattacks on Estonia and the Stuxnet virus, which incapacitated Iranian nuclear facilities, are stark reminders of strategic aggression that did not involve traditional warfare.

Clausewitz and Sun Tzu never saw Twitter battles or deepfake propaganda coming. But here we are. Rather than fighting discrete wars, we’re in a period of ongoing strategic competition. The 2018 U.S. National Defence Strategy even describes it as ‘long-term strategic competition’ (Department of Defence, 2018). This shift undermines the traditional Westphalian model, where war and peace were regarded as distinct states. Hybrid warfare thrives in ambiguity. Hoffman (2017) describes it as a mix of misinformation, economic coercion, cyberattacks, and proxy forces. The goal? Stay below the conventional threshold of war. The Russian annexation of Crimea in 2014, involving cyber operations, disinformation, and unmarked troops, serves as an exemplary case.

Despite these transformations, Clausewitz’s core concepts continue to be highly relevant. His idea of the trinity of “violence, chance, and political purpose” continues to offer a valuable framework for understanding modern conflicts. Colin Gray (1999) underscores that strategy is fundamentally about applying means to achieve political ends, irrespective of technological advancements. The risk, however, lies in excessively broadening the definition of war. If every act of geopolitical rivalry, such as economic sanctions, election interference, or cyber espionage, is termed “war,” it risks conceptual dilution. Gartzke (2013) cautions that this approach could end with unnecessary escalation, with states treating cyber incidents as casus belli when they might be closer to espionage or subversion.

So where do we go from here? Rather than discarding classical strategic theory, we should reinterpret its principles to align with current realities. Clausewitz’s trinity can evolve: “violence” can encompass non-kinetic coercion; “chance” is amplified by the unpredictability of interconnected digital systems; and “political purpose” now includes influence operations and behavioural shaping alongside territorial ambitions. Warfare may not appear as it did in Clausewitz’s era, but its essence, driven by politics and strategy, remains unchanged.

The Future of War: AI and Strategy

When looking at strategy, Clausewitz taught us that war is shaped by chance, friction, and human judgment and Colin Gray emphasised the enduring nature of strategy, despite technological change. Yet, artificial intelligence (AI) is accelerating decision-making beyond human speeds, raising a critical question: Are we entering an era where machines – not strategists – dictate the course of conflict?

The Transformative Power of AI in Strategy

AI-driven systems now process intelligence, optimise battlefield decisions, and launch cyber operations at speeds unimaginable just two decades ago. OSINT, GEOINT, and SIGINT can be ingested, analysed, and summarised into actionable insights in real time. AI-enhanced wargaming and strategic forecasting are helping policymakers anticipate threats with greater accuracy. But does this lead to better strategy, or does it introduce new vulnerabilities?

The Erosion of Traditional Strategic Advantages

Historically, military and strategic advantages were state monopolies due to the vast resources required to develop cutting-edge capabilities, but AI is breaking down these barriers. The latest open-source AI models, commercial AI applications, and dual-use technologies mean that non-state actors, corporations, and even criminal groups now have access to tools once reserved for governments.

Consider Russia’s use of AI-driven disinformation campaigns during the 2016 U.S. elections and Ukraine conflict, where AI-powered bots and deepfake technology have enabled influence operations that are difficult to counter. Similarly, China’s AI-enabled surveillance state represents a new model of strategic power – one that fuses military and civilian AI applications for geopolitical advantage.

Blurring the Lines Between War and Peace

AI does not just change warfare; it changes the very definition of conflict. The use of AI-driven cyber and information operations enables continuous engagement below the threshold of open war. Instead of clear distinctions between peace and conflict, we are witnessing an era of persistent, AI-enhanced competition.

Using China as an example again, their civil-military fusion strategy integrates AI research and applications across both sectors, allowing for rapid technological advancement with strategic implications. Will the UK and its allies struggle to counter this approach within their existing regulatory and legal frameworks?

The Impact on Deterrence and Escalation

Deterrence has traditionally relied on rational actors making calculated decisions. But what happens when autonomous systems can pre-emptively engage threats or retaliate without clear human oversight? The risk of unintended escalation grows if AI-driven platforms misinterpret data or are manipulated by adversarial AI systems.

The Pentagon’s Project Maven, which employs AI to analyse drone surveillance footage, highlights the advantages AI brings to intelligence processing. But it also raises ethical concerns – how much decision-making should be delegated to machines? And if state actors develop autonomous weapons with AI-controlled engagement protocols, does this make deterrence more fragile?

Limitations of AI in Strategy

Despite AI’s capabilities, it still struggles with unpredictability—something central to strategy. AI models are excellent at processing historical patterns but often fail in novel or asymmetric situations. This reinforces the importance of human judgment in strategic decision-making. AI-driven strategy also raises concerns about bias, such as how commercial AI models (e.g., ChatGPT, DeepSeek) reflect the interests of their creators, whether corporate or state-sponsored. If strategic decision-making increasingly relies on black-box models with unknown biases, how do we ensure accountability and transparency?

Strategic Recommendations: The Path Forward

Rather than replacing human decision-makers, I believe that AI should be seen as a force multiplier. Governments and militaries must develop frameworks for human-AI hybrid decision-making, ensuring that AI informs but does not dictate strategy.

Additionally, fail-safe mechanisms must be built into autonomous systems to prevent unintended escalation. Given the rapid development of adversarial AI defences it will be critical as states and non-state actors seek to manipulate AI-driven decision-making processes.

Finally, it is critical that military and civilian leaders must rethink strategic education in the AI era. Understanding AI’s capabilities, limitations, and strategic implications should be a core component of professional military education and policymaker training.

Are we Seeing the Automation of Strategy?

Clausewitz’s fog of war was once an unavoidable condition. If AI offers real-time clarity, does it eliminate uncertainty – or create new vulnerabilities that adversaries can exploit? As AI increasingly influences military and strategic decision-making, are we witnessing the automation of strategy itself? If so, what does that mean for accountability, escalation, and deterrence?

The future strategic environment will be defined by those who can integrate AI effectively—without surrendering human judgment to machines. The challenge ahead is ensuring that AI serves strategy, rather than strategy being dictated by AI.

My Experiences in Japan / 日本での私の経験

So here I am, about a month after having returned to London, reflecting back on my experience in Japan. The most common question I’ve had from friends is “was it worth it?” or “did you learn what you wanted to learn?” and the answer to both is a resounding yes. However it isn’t as simple as that, as there are many layers to the question which need to be unpacked a little.

The first layer is understanding what it is that I went off to Japan to learn in the first place. Of course there is the obvious “Japanese Language” side of things, but there is much more to it than that. The real learning I was hoping to take away was about myself. And if you want to learn about yourself, one of the best ways is to teach others. Which thanks to Phil’s company I was able to do exactly that with a bunch of middle school kids over 6 days of English language camps!

Class photo!

The camp is worth a whole post by itself, but the takeaway is how teaching kids made me feel about myself and reflect on what I enjoy doing professionally. It allowed me to understand what I value (honesty, enthusiasm, progress) in a much more immersive way than a work environment would.  I’ll definitely be taking this into my next role and it has helped me mature bit more as an individual.

Alongside teaching, there were many other examples of “bonus” learning opportunities I was able to take away from the experience.  One key aspect was taking myself out of a familiar environment and getting the mental/physical space to learn and reflect.  That alone was worth the proverbial price of admission.

The next layer is about “what” it is I did.  While it is obvious to some, but not always to me, it isn’t the destination that matters so much, but who you spend it with and the attitude you take with you.  I enjoy spending time by myself (quite a lot) but all my most enjoyable experiences are with friends and family.  It isn’t just quality though, it is quantity too.  People can help you overcome natural inertia (read: laziness) to get out there and do more. In fact, when Julie came to visit for a week, we crammed in more stuff than I did in the previous months!

Fun and hi-jinks!

Also, being in the right mindset (a positive one), I was able to value those times much more and care a little bit less about the latest distracting “must-own-thing”.  But it is easily forgotten and I have to remind myself often to focus on new experiences with loved ones and less on new, shiny, technology…

There is a lot more I was able to get from my travels, but that’s about my limit for for self reflection today.  I think I need to do a round up of all the beers I forgot to mention in a new post…

Nanzoin Temple / 南蔵院

With Julie’s visit this week, I have been very busy doing all sorts of fun tourist type activities! One of the best trips we did was heading off to Nanzoin Temple, which features The Reclining Buddha (more on this later).

Rather than bore you with lots of hyper-descriptive prose, I think a visual tour would be much more fun. So let’s get started!

Getting there was pretty easy, just a 30 minute train ride from Hakata station. Once we got off the train and started trekking up towards the temple grounds, we found this family of turtles with their own mini-shrine. They were kind enough to let me take way too many photos as I over-enthusiastically snapped shots from various exciting angles.

Look who came to greet us!

Clearly turtles weren’t the only photogenic fauna around, as this rather large dragonfly posed long enough for me to get a fairly zoomed in photo.

Posed for the photo and flew off as soon as I was done…

One of the first statues to be seen was of this chap with his flames and sword. We originally thought he was some bad guy, but Julie did some research and we discovered he is actually Fudoumyouou who is a Buddhist deity who protects worshippers from harm.

Later learnt this angry looking guy is a protective deity

As we spotted more of the classic Buddha style statues, we noticed they were usually wearing some sort of clothing, be that a hat, a cardigan or the like.  It turns out the reason behind this is fairly involved, so check out this post for details.

Lots of little statues, all wearing the cutest clothes

After a bit more walking, we finally made our way to the star attraction, The Reclining Buddha himself!  I must say, the statue really is impressive and quite something else to behold.  He is the world’s largest bronze statue and is the same size as the Statue of Liberty. He is so large I had to use the panorama mode of the phone to get him fully in the frame.

Even more impressive in person…

You can walk around to his feet and I think the following image really gives you a good idea of the scale of the statue…

Buddha foot with Julie touching the thousand spoke wheel symbol

Once we had seen the Buddha and enjoyed some well timed ice-creams, it was time for some exploring!

The path upwards…

After following a (very steep) trail up and out of the temple grounds, we found ourselves in some beautiful back-country, surrounded by bamboo forest. I love the sound they make as they rustle in the breeze. It was truly a magical part of the day for me.

I love bamboo forests

From the trail, we were able to visit a number of smaller temples and shrines. At one temple, we even had the Buddhist monk come out and offer us watermelon to eat! It was unexpected and very much appreciated on a very warm day. We ventured on each time and were treated to some great views through the trees.

Fabulous views over the countryside

As we wound our way back down the mountain towards the station again, we found many more statues, but the following set were really quite something else. Such a riot of colour and clothing on each of the little stone statues!

Someone likes playing dress up…

Here you can see some up close… adorable!

So dang cute…

Overall it was a brilliant day and even the photos can’t do justice to the adventure we had. If you ever get a chance, I’d highly recommend a visit!

Japanese Calligraphy / 書道

This weekend I had an opportunity to try out Japanese calligraphy at an event put on by the Fukuoka International Student Support Centre (FiSSC). Calligraphy isn’t something I’d normally be very excited about, but I’m a sucker for trying anything new so I thought I’d give it a bash anyway. It also helped that my friend Sora was coming with me, who always ready to help me avoid doing anything (too) embarrassing.

Chilling with my buddy Sora

Japanese calligraphy (shodou/書道) is an art form for writing Japanese Kanji (Chinese characters) using a brush and ink that is taught to all Japanese students in primary school, but it is rarely used after that. As such the event had a good mix of Japanese and non-Japanese people attending, with quite the range of skills on display, from very basic through to really artful. The sensei (teacher) for the lesson was truly an artist and was able to create evocative works with little apparent effort. Practice definitely makes perfect.

We were given our instructions, ink, brush and a stack of paper to practise with, after which we were told to get on with practising! My own works were somewhat less than perfect, but I learnt a trick or two over the period and I really felt that I’d improved at least a little bit over the duration of the class.

The sensei watches on as a student tries to do “光” or “light”

I’m glad I did manage to improve a little as I was “asked” to try out writing a kanji out on one of the larger sheets (the joys of being very-obviously-foreign). I choose the character for “joy” or “喜” as it is simple and represents something I think we could all do with a little more of in our lives. The end result was middle of the road enough to neither attract laughter nor praise. Phew, embarrassment avoided!

My “masterpiece”. On the left is my name in katakana – マシュートイ

My own experiments aside, the work produced by masters of the art truly is something else though. I’m impressed that so much feeling and emotion can be conveyed through a simple black and white medium.

All said, I had a lot of fun and learnt the basics of a new skill. Not a bad result for a Saturday afternoon!

My other creations. One of them isn’t a real kanji, can you guess which one? 🙂

With Friends / 友達と

The past few weeks have been a social whirlwind and I’ve had an opportunity to try out all sorts of new and interesting activities .  I’ll make sure I do a proper set of updates, but this time I’d like to briefly tell you about an adventure into Thai cuisine.

As a bit of background, there is a regular “language lounge”, where we swap English and Japanese practice with learners from both languages, which is really good fun. I get to meet all sorts of awesome people and learn about Japanese language and culture from them. After one particular lounge a group of 6 of us decided to head off to grab some dinner, opting to try out a Thai restaurant in Tenjin.

All smiles 🙂

This was the first time I’ve tried Thai food while in Japan and I have to say, I found it more “authentic” than what you would normally get in London. Not sure why I think it is more authentic, apart from it not feeling particularly Western in its flavours and presentation. That aside, the food itself was really tasty! Lots of spice though, so those with a sensitive tongue will need to be extra careful.

I feel that this style of setting makes it really easy to communicate effectively with people. The casual atmosphere allows you to relax and banter back and forth, using a mix of both native and learnt languages. I really enjoy being able to try out new words, expressions and such without anyone to judge you on technical performance. In fact the only real judge is whether you are able to effectively get your ideas across to the people you are talking to (eventually).

I made a commitment to use my time in Japan as effectively as possible and I’m super happy I’ve managed to get into more of this style activity, in addition to the classroom education and personal study.

Views of Fukuoka / 福岡の景色

One of the things many locals of Fukuoka like about their city is that it has all of the city conveniences but also, not too far away, it has lots of nature and scenery. I haven’t had an opportunity yet to venture outside of the city properly (something I intend to remedy shortly) but I have found many lovely places while out running.

My favourite so far has to be Heiwaminami Ryokuchi Park (say that 3 times quickly) which is only a few kilometre jog from my apartment. Not only is it 100% trails with wonderful green foliage, but it is super-duper undulating. This means I can get all the hill training I could possibly want, all in one handy-dandy park!

Let’s take a quick photo tour…

Heiwaminami Ryokuchi Park

Photos never do justice to how steep a trail really is. But this was super steep, honest!  That aside, I love how green and verdant everything looks. Cannot get enough of this sort of trail.

Peeking out through the trees

Most of the views are obstructed by all the beautiful green foliage I was just mentioning, but occasionally you can peek out and see the city below.

View from the observation tower

One unique aspect of the park is the tower you can walk up to get 360 degree views of Fukuoka.

Cemeteries get all the best views

Right next to the park is a large cemetery. Japanese culture includes a lot of respect for the dead, so you can just imagine how this is both immaculately maintained and presented.

I’m a sucker for attractive looking clouds…

On the way back from the park now, waiting at an intersection. Not too much exciting happening, but I do like how the building in front is telling me something is “Good” with a solid thumbs up included. Thank you random building, things are indeed “Good”.

Filled with fishies!

What you can’t see very well in this final photo is all the fishies in the canal. The water is very clear, so you can see all the big buggers and little guys all swimming around.

Beer Review Round-up / ビールレビューまとめ

I thought it about time to do a round-up of the various beers I’ve tried over the past few weeks. I must admit to quite enjoying this series, as I really don’t know what to expect with each drink I try.

To start things off, we have two beers which are from the same brewery. Yaho Brewing is proving to be an excellent source of tasty beers, I must say. The first one shares the same name with the very enjoyable “A beer for you and me” that I got not long after arriving. Whether it is actually the same beer is up for debate, but it shares the funky looking frog motif and is equally tasty. Really enjoyed this one, it is equal parts easy drinking, smoothness and tasty flavour.

If you see this frog on a beer, it is likely to be a tasty

The second from this brewery is the Yona Yona Ale. Not much to say apart from it being a very solid and enjoyable pale ale. They are nearly double the price of their cheapest rivals (~£1.80 to 90p) but totally worth the extra spend in my books.

Another winner!

From Suntory we have 金麦 or “Gold Wheat” beer. Apart from really liking the design of the can itself, there isn’t a lot I can write about this middle-of-the-road beer. Nothing wrong with it, nothing all that right. In a word: Inoffensive.

Average

Speaking of refreshing looking can designs, next we have a Lemon Beer from Grand Kirin. The name is something like ひこうき雲と私 or literally “the aeroplane cloud and me”. I imagine it loses something in the translation.

On the beer itself, I found that the lemon flavour is there, but it is quite muted.Not a bad idea, I like the idea of a refreshing summer beer, but I don’t think this is it for me. Overall, it wasn’t too bad really, decently drinkable, but doesn’t really commit to the concept – it isn’t beery enough and not lemony enough either.

Was expecting something, but not quite this

The Suntory Beer Recipe (a blueberry flavoured beer) wasn’t ever going to be great, but I at least thought it would be interesting. It does have a bit of a fruity flavour, but it was underwhelming and left me wishing for something more exciting.

Similar to the lemon beer above, it just doesn’t commit to the concept to make it worth not having a less weird and better tasting normal beer.

I really need to stop experimenting…

Finally, I had much higher hopes for this Kirin “Prime Brew” beer, but sadly it left me thinking… “Meh, okay”. It was drinkable and if pushed I could have a few, but it really wasn’t a “prime” anything and left me wanting for something more flavoursome.

Prime Brew huh? Not that impressed

My adventure continues…

Page 2 of 4

Powered by WordPress & Theme by Anders Norén