Matthew Toy

Thoughts, reflections and experiences

icy banner
Europe's Leadership Vacuum in the Shadow of Russia and America

Europe’s Leadership Vacuum in the Shadow of Russia and America

The concept of ‘strategic culture’ as critiqued in Hew Strachan’s “The Direction of War: Contemporary Strategy in Historical Perspective” emphasises continuity and a nation’s resistance to change, shaped historically and geographically. Strategic culture includes historical memory, institutional norms, core national values, and collective threat perceptions, all contributing to a nation’s strategic posture. This comprehensive framework is valuable when examining Europe’s contemporary security challenges, specifically the strategic vacuum highlighted by the ongoing war in Ukraine and America’s ongoing withdrawal from global leadership.

Europe’s Strategic Culture

European strategic culture, forged during the Cold War, assumed stability through American military and diplomatic leadership. Strachan argues convincingly that such cultural assumptions hinder strategic flexibility, creating vulnerabilities when geopolitical realities shift dramatically, as they have since Russia’s invasion of Ukraine in 2022.

NATO-centric thinking, predicated on the guarantee of American power projection, has revealed problematic inertia… European states, notably the UK and the EU members, have found themselves scrambling to define a coherent, autonomous response.

America’s Strategic Shift from Protector to Competitor

America’s strategic withdrawal from Europe, evidenced by Obama’s pivot to Asia, that accelerated by Trump V1.0’s transactional approach, Biden’s reticence and culminating with Trump 2.0’s recent dramatic geopolitical hand grenades. This reflects not merely a change in policy but a radical break from previous expectations. This withdrawal is a revolutionary, not evolutionary, shift in global strategy, shattering Europe’s assumption of guaranteed U.S. engagement.

Strategically, this creates immediate tensions:

  • The U.S. increasingly frames its engagement with Europe as transactional and conditional upon shared responsibilities, as demonstrated by U.S. ambivalence toward NATO under Trump and Biden’s conditional engagement in Ukraine.
  • Simultaneously, Russia’s aggression has starkly shown that the belief in a diminished threat from inter-state warfare, fashionable among policymakers since the Cold War’s end, is dangerously misplaced. Strachan’s scepticism about overly optimistic predictions of war’s obsolescence resonates strongly here, given recent events.

This combination reveals Europe’s strategic culture as critically unprepared for the harsh geopolitical realities of the 21st century.

Europe’s Strategic Awakening

Europe has not been entirely inactive. The EU’s Strategic Compass, adopted in 2022, and the UK’s Integrated Review Refresh in 2023 demonstrate genuine acknowledgment of new realities. These documents move beyond purely reactive policies and represent Europe’s incremental shift towards strategic autonomy:

  • Increased defence expenditure: Germany’s Zeitenwende is a prime example.
  • Increased EU defence coordination, exemplified by the European Peace Facility funding Ukraine’s defence.
  • Renewed commitment to territorial defence and enhanced military deployments in Eastern Europe.

Yet, despite these efforts, the doctrinal and strategic mindset change has been incomplete. European policies continue to implicitly rely on the assumption of sustained U.S. involvement, despite public and political statements affirming Europe’s need for self-sufficiency.

Russia and America as Mirrors

The actions of Russia and the retreat of America each independently expose the inadequacies of Europe’s current strategic posture:

Russia’s Actions: Highlighted Europe’s continuing strategic vulnerability, emphasising weaknesses in rapid military deployment, critical capability gaps (such as long-range precision munitions and air defence), and dependence on U.S. logistical, intelligence, and strategic capabilities.

America’s Pivot Away: Underscores that strategic autonomy isn’t merely desirable but imperative. Starting with Biden administration’s reluctance to escalate beyond certain lines in Ukraine and Washington’s growing Indo-Pacific focus expose a stark misalignment between European expectations and American strategy. The most recent signals from Trump are an unequivocal message to Europe: unless there is something in it for America, you are on your own.

The Limits of Integration and NATO

While deeper European integration and renewed commitment to NATO might appear sufficient, these solutions alone are inadequate. Integration without clear autonomous capabilities risks perpetual dependency, and NATO’s structure, inherently reliant on American leadership, cannot compensate for America’s strategic reorientation. As Strachan underscores, relying purely on continuity without adaptability is strategically naive.

From Reactive Culture to Proactive Realism

Europe’s security doctrine requires nuanced recalibration rather than wholesale abandonment. The gap is not merely military, it is doctrinal, conceptual, and philosophical. A robust European strategic doctrine should:

  1. Recognise NATO’s Limitations: Explicitly acknowledge NATO’s limitations without undermining its centrality to European defence.
  2. Embed Strategic Autonomy: Clearly outline Europe’s independent capabilities and strategic objectives, moving beyond rhetoric to practical operational frameworks. Europe must realistically assess scenarios in which it may need to act without guaranteed American backing.
  3. Rethink Strategic Culture: Move beyond traditional assumptions of continuity—what previously seemed unthinkable, such as large-scale inter-state conflict, must become integral to planning and preparedness again.

Engaging Broader Perspectives

Drawing briefly from constructivist insights, strategic culture is not immutable but socially constructed, implying that European nations have the agency to reshape it consciously. Additionally, realist thinkers like John Mearsheimer caution against complacency in alliance politics, reinforcing the need for independent European capabilities.

Rethinking Doctrine for Strategic Resilience

The UK’s Integrated Review and the EU’s Strategic Compass represent valuable first steps toward a more strategic and independent Europe. However, they still fall short of addressing the fundamental gap that Russia’s aggression and America’s strategic recalibration have exposed.

Addressing Europe’s leadership vacuum demands overcoming historical and cultural inertia. It requires strategic humility: recognising that the stability provided by Cold War-era assumptions no longer applies, that threats are tangible, and that peace through strength must be anchored not in external assurances, but in Europe’s credible, independently sustainable power.

Europe must confront this reality head-on, accepting change not merely rhetorically but operationally, doctrinally, and culturally. Only then will Europe secure genuine strategic autonomy, prepared not just for today’s threats but also for tomorrow’s inevitable uncertainties.

Bibliography

  • Strachan, Hew. The Direction of War: Contemporary Strategy in Historical Perspective. Cambridge University Press, 2013.
  • European Union. “Strategic Compass for Security and Defence.” 2022.
  • United Kingdom Government. “Integrated Review Refresh.” 2023.
  • Mearsheimer, John J. The Tragedy of Great Power Politics. W. W. Norton & Company, 2001.
  • Smith, Rupert. The Utility of Force: The Art of War in the Modern World. Penguin, 2005.

[Video] UK and EU AI Influence

Artificial intelligence isn’t just reshaping industries—it’s reshaping reality. While the UK and EU focus on regulating AI and combating misinformation, adversarial states like Russia and China are weaponizing it for influence warfare. The AI-driven disinformation battle isn’t coming; it’s already here.

In my latest article, “Why the UK and EU Are Losing the AI Influence War”, I explore how Europe’s slow response, defensive posture, and reliance on outdated regulatory approaches are leaving it vulnerable to AI-enhanced propaganda campaigns.

To bring these ideas to life, I’ve created a video that visualises the scale of the challenge and why urgent action is needed. Watch it below:

The AI influence war is no longer a hypothetical—it’s unfolding in real-time. Europe’s current strategies are reactive and insufficient, while adversaries leverage AI to manipulate narratives at unprecedented speed. Without a cognitive security unit, AI-powered countermeasures, and a national security-driven approach, the UK and EU risk losing control of their own information space.

The question isn’t whether AI will reshape public perception, it’s who will be in control of that perception. Will Europe rise to the challenge, or will it remain a passive battleground for AI-driven narratives?

What do you think? Should the UK and EU take a more aggressive stance in countering AI-enhanced disinformation? Feel free to discuss in the comments.

WHY THE UK AND EU ARE LOSING THE AI INFLUENCE WAR

Why the UK and EU Are Losing the AI Influence War

Abstract

Western democracies face a new front in conflict: the cognitive battlespace, where artificial intelligence (AI) is leveraged to shape public opinion and influence behaviour. This article argues that the UK and EU are currently losing this AI-driven influence war. Authoritarian adversaries like Russia and China are deploying AI tools in sophisticated disinformation and propaganda campaigns, eroding trust in democratic institutions and fracturing social cohesion. In contrast, the UK and EU response, focused on regulation, ethical constraints, and defensive measures, has been comparatively slow and fragmented. Without a more proactive and unified strategy to employ AI in information operations and bolster societal resilience against cognitive warfare, Western nations risk strategic disadvantage. This article outlines the nature of the cognitive battle-space, examines adversarial use of AI in influence operations, evaluates UK/EU efforts and shortcomings, and suggests why urgent action is needed to regain the initiative.

Introduction

Modern conflict is no longer confined to conventional battlefields; it has expanded into the cognitive domain. The term “cognitive battlespace” refers to the arena of information and ideas, where state and non-state actors vie to influence what people think and how they behave. Today, advances in AI have supercharged this domain, enabling more sophisticated influence operations that target the hearts and minds of populations at scale. Adversaries can weaponise social media algorithms, deepfakes, and data analytics to wage psychological warfare remotely and relentlessly.

Western governments, particularly the United Kingdom and European Union member states, find themselves on the defensive. They face a deluge of AI-enhanced disinformation from authoritarian rivals but are constrained by ethical, legal, and practical challenges in responding. Early evidence suggests a troubling imbalance: Russia and China are aggressively exploiting AI for propaganda and disinformation, while the UK/EU struggle to adapt their policies and capabilities. As a result, analysts warn that Western democracies are “losing the battle of the narrative” in the context of AI (sciencebusiness.net). The stakes are high: if the UK and EU cannot secure the cognitive high ground, they risk erosion of public trust, social discord, and strategic loss of influence on the world stage.

This article explores why the UK and EU are lagging in the AI influence war. It begins by defining the cognitive battlespace and the impact of AI on information warfare. It then examines how adversaries are leveraging AI in influence operations. Next, it assesses the current UK and EU approach to cognitive warfare and highlights key shortcomings. Finally, it discusses why Western efforts are falling behind and what the implications are for future security.

The Cognitive Battlespace in the Age of AI

In cognitive warfare, the human mind becomes the battlefield. As one expert succinctly put it, the goal is to “change not only what people think, but how they think and act” (esdc.europa.eu). This form of conflict aims to shape perceptions, beliefs, and behaviours in a way that favours the aggressor’s objectives. If waged effectively over time, cognitive warfare can even fragment an entire society, gradually sapping its will to resist an adversary.

Artificial intelligence has become a force multiplier in this cognitive domain. AI algorithms can curate individualised propaganda feeds, amplify false narratives through bot networks, and create realistic fake images or videos (deepfakes) that blur the line between truth and deception. According to NATO’s Allied Command Transformation, cognitive warfare encompasses activities to affect attitudes and behaviours by influencing human cognition, effectively “modifying perceptions of reality” as a new norm of conflict (act.nato.int). In essence, AI provides powerful tools to conduct whole-of-society manipulation, turning social media platforms and information systems into weapons.

A vivid example of the cognitive battlespace in action occurred in May 2023, when an AI-generated image of a false Pentagon explosion went viral. The fake image, disseminated by bots, briefly fooled enough people that it caused a sharp but temporary dip in the U.S. stock market. Though quickly debunked, this incident demonstrated the “catastrophic potential” of AI-driven disinformation to trigger real-world consequences at machine speed (mwi.westpoint.edu) . Generative AI can manufacture convincing yet false content on a massive scale, making it increasingly difficult for populations to discern fact from fabrication.

In the cognitive battlespace, such AI-enabled tactics give malign actors a potent advantage. They can rapidly deploy influence campaigns with minimal cost or risk, while defenders struggle to identify and counter each new false narrative. As the information environment becomes saturated with AI-amplified propaganda, the traditional defenders of truth, journalists, fact-checkers, and institutions, find themselves overwhelmed. This asymmetry is at the heart of why liberal democracies are in danger of losing the cognitive war if they do not adapt quickly.

Adversaries’ AI-Driven Influence Operations

Russia and China have emerged as leading adversaries in the AI-enabled influence war, honing techniques to exploit Western vulnerabilities in the cognitive domain. Russia has a long history of information warfare against the West and has eagerly integrated AI into these efforts. Through troll farms and automated bot networks, Russia pushes AI-generated propaganda designed to destabilise societies. Moscow views cognitive warfare as a strategic tool to “destroy [the West] from within” without firing a shot. Rather than direct military confrontation with NATO (which Russia knows it would likely lose), the Kremlin invests in “cheap and highly effective” cognitive warfare to undermine Western democracies from inside (kew.org.pl) .

Russian military thinkers refer to this concept as “reflexive control,” essentially their doctrine of cognitive warfare. The idea is to manipulate an adversary’s perception and decision-making so thoroughly that the adversary “defeats themselves”. In practice, this means saturating the information space with tailored disinformation, conspiracy theories, and emotionally charged content to break the enemy’s will to resist. As one analysis describes, the battleground is the mind of the Western citizen, and the weapon is the manipulation of their understanding and cognition. By exploiting human cognitive biases, our tendencies toward emotional reaction, confirmation bias, and confusion, Russia seeks to leave citizens “unable to properly assess reality”, thus incapable of making rational decisions (for example, in elections). The goal is a weakened democratic society, rife with internal divisions and distrust, that can no longer present a united front against Russian aggression.

Concrete examples of Russia’s AI-fuelled influence operations abound. Beyond the fake Pentagon incident, Russian operatives have used generative AI to create deepfake videos of European politicians, forge fake news stories, and impersonate media outlets. Ahead of Western elections, Russian disinformation campaigns augmented with AI have aimed to sow discord and polarise public opinion. U.K. intelligence reports and independent researchers have noted that Russia’s automated bot accounts are evolving to produce more “human-like and persuasive” messages with the help of AI language models. These tactics amplify the reach and realism of propaganda, making it harder to detect and counter. Even if such interference does not always change election outcomes, it erodes public trust in information and institutions, a long-term win for the Kremlin.

China, while a newer player in European information spaces, is also investing heavily in AI for influence operations. Chinese military strategy incorporates the concept of “cognitive domain operations”, which merge AI with psychological and cyber warfare. Beijing’s aim is to shape global narratives and public opinion in its favour, deterring opposition to China’s interests. For instance, China has deployed swarms of AI-driven social media bots to spread disinformation about topics like the origins of COVID-19 and the status of Taiwan. Chinese propaganda operations use AI to generate deepfake news anchors and social media personas that promote pro-China narratives abroad. According to NATO analysts, China describes cognitive warfare as using public opinion and psychological manipulation to achieve victory, and invests in technologies (like emotion-monitoring systems for soldiers) that reveal the importance it places on the information domain. While China’s influence efforts in Europe are less overt than Russia’s, they represent a growing challenge as China seeks to project soft power and shape perceptions among European audiences, often to dissuade criticism of Beijing or divide Western unity.

The aggressive use of AI by authoritarian adversaries has put Western nations on the back foot in the information environment. Adversaries operate without the legal and ethical constraints that bind democracies. They capitalise on speed, volume, and ambiguity, launching influence campaigns faster than defenders can react. Authoritarian regimes also coordinate these efforts as part of broader hybrid warfare strategies, aligning cyber-attacks, diplomatic pressure, and economic coercion with information operations to maximise impact. In summary, Russia and China have seized the initiative in the cognitive battlespace, leaving the UK, EU, and their allies scrambling to catch up.

UK and EU Responses: Strategies and Shortcomings

Confronted with these threats, the United Kingdom and European Union have begun to recognise the urgency of the cognitive warfare challenge. In recent years, officials and strategists have taken steps to improve defences against disinformation and malign influence. However, the Western approach has so far been largely reactive and constrained, marked by cautious policy frameworks and fragmented efforts that lag the adversary’s pace of innovation.

United Kingdom: The UK government acknowledges that AI can significantly amplify information warfare. The Ministry of Defence’s Defence Artificial Intelligence Strategy (2022) warns that “AI could also be used to intensify information operations, disinformation campaigns and fake news,” for example by deploying deepfakes and bogus social media accounts. British military doctrine, including the Integrated Operating Concept (2020), emphasises that information operations are increasingly important to counter false narratives in modern conflicts (gov.uk). London’s approach has included establishing units dedicated to “strategic communications” and cyber influence and working with partners like NATO to improve information security.

The UK has also invested in research on AI and influence. For instance, the Alan Turing Institute’s research centre (CETaS) published analyses on AI-enabled influence operations in the 2024 UK elections, identifying emerging threats such as deepfake propaganda and AI-generated political smear campaigns. These studies, while finding that AI’s impact on recent elections was limited, highlighted serious concerns like AI-driven hate incitement and voter confusion (cetas.turing.ac.uk) . The implication is clear: the UK cannot be complacent. Even if traditional disinformation methods still dominate, the rapid evolution of AI means influence threats could scale up dramatically in the near future. British policymakers have started to discuss new regulations (for example, requiring transparency in AI political ads) and bolstering media literacy programs to inoculate the public against fake content.

Despite this awareness, critics argue that the UK’s response remains disjointed and under-resourced. There is no publicly articulated doctrine for cognitive warfare equivalent to adversaries’ strategies. Efforts are split among various agencies (from GCHQ handling cyber, to the Army’s 77th Brigade for information ops, to the Foreign Office for counter-disinformation), making coordination challenging. Moreover, while defensive measures (like fact-checking services and takedown of fake accounts) have improved, the UK appears reluctant to consider more assertive offensive information operations that could pre-empt adversary narratives. Legal and ethical norms, as well as fear of escalation, likely restrain such tactics. The result is that Britain often plays catch-up, reacting to disinformation waves after they have already influenced segments of the population.

European Union: The EU, as a bloc of democracies, faces additional hurdles in confronting cognitive warfare. Brussels has treated disinformation chiefly as a policy and regulatory issue tied to election security and digital platform accountability. Over the past few years, the EU implemented a Code of Practice on Disinformation (a voluntary agreement with tech companies) and stood up teams like the East StratCom Task Force (known for its EUvsDisinfo project debunking pro-Kremlin myths). Following high-profile meddling in elections and referendums, EU institutions have grown more vocal: they label Russia explicitly as the chief source of disinformation targeting Europe. The European Commission also incorporated anti-disinformation clauses into the Digital Services Act (DSA), requiring large online platforms to assess and mitigate risks from fake content.

When it comes to AI, the EU’s landmark AI Act – primarily a regulatory framework to govern AI uses – indirectly addresses some information manipulation concerns (for example, by requiring transparency for deepfakes). However, EU efforts are fundamentally defensive and norm-driven. They seek to police platforms and inform citizens, rather than actively engage in influence operations. EU leaders are wary of blurring the line between counter-propaganda and propaganda of their own, given Europe’s commitment to free expression. This creates a dilemma: open societies find it difficult to wage information war with the ruthlessness of authoritarian regimes.

European security experts are starting to grapple with this challenge. A recent EU security and defence college course underscored that cognitive warfare is an “emerging challenge” for the European Union (esdc.europa.eu) . Participants discussed the need for technological tools to detect, deter, and mitigate cognitive threats. Yet, outside of specialised circles, there is no EU-wide military command focused on cognitive warfare (unlike traditional domains such as land, sea, cyber, etc.). NATO, which includes most EU countries, has taken the lead in conceptualising cognitive warfare, but NATO’s role in offensive information activities is limited by its mandate.

A telling critique comes from a Royal United Services Institute (RUSI) commentary on disinformation and AI threats. It notes that NATO’s 2024 strategy update acknowledged the dangers of AI-enabled disinformation, using unusually strong language about the urgency of the challenge. However, the same strategy “makes no reference to how AI could be used” positively for strategic communications or to help counter disinformation (rusi.org) . In other words, Western nations are emphasising protection and defence, strengthening **governance standards, public resilience, and truth-checking mechanisms, **but they are not yet leveraging AI offensively to seize the initiative in the info sphere. This cautious approach may be ceding ground to adversaries who have no such reservations.

Why the West Is Losing the AI Influence War

Several interrelated factors explain why the UK, EU, and their allies appear to be losing ground in the cognitive domain against AI-equipped adversaries:

Reactive Posture vs. Proactive Strategy: Western responses have been largely reactive. Democracies often respond to disinformation campaigns after damage is done, issuing fact-checks or diplomatic condemnations. There is a lack of a proactive, comprehensive strategy to dominate the information environment. Adversaries, by contrast, set the narrative by deploying influence operations first and fast.

Ethical and Legal Constraints: The UK and EU operate under strict norms – adherence to truth, rule of law, and respect for civil liberties – which limit tactics in information warfare. Propaganda or deception by government is domestically unpopular and legally fraught. This makes it hard to match the scale and aggressiveness of Russian or Chinese influence operations without undermining democratic values. Authoritarians face no such constraints.

Fragmented Coordination: In Europe, tackling cognitive threats cuts across multiple jurisdictions and agencies (domestic, EU, NATO), leading to fragmentation. A unified command-and-control for information operations is lacking. Meanwhile, adversaries often orchestrate their messaging from a centralised playbook, giving them agility and consistency.

Regulatory Focus Over Capabilities: The EU’s inclination has been to regulate (AI, social media, data) to create guardrails – a necessary but slow process. However, regulation alone does not equal capability. Rules might curb some harmful content but do not stop a determined adversary. The West has invested less in developing its own AI tools for strategic communication, psyops, or rapid counter-messaging. This capability gap means ceding the technological edge to opponents.

Underestimation of Cognitive Warfare: Historically, Western security doctrine prioritised physical and cyber threats, sometimes underestimating the impact of information warfare. The concept of a sustained “cognitive war” waged in peacetime is relatively new to Western planners. Initial responses were tepid – for example, before 2016, few anticipated that online influence could significantly affect major votes. This lag in appreciation allowed adversaries to build momentum.

These factors contribute to a situation where, despite growing awareness, the UK and EU have struggled to turn rhetoric into effective countermeasures on the cognitive front. As a result, authoritarian influence campaigns continue to find fertile ground in Western societies. Each viral conspiracy theory that goes unchecked, each wedge driven between communities via disinformation, and each doubt cast on democratic institutions chips away at the West’s strategic advantage. NATO officials warn that information warfare threats “must neither be overlooked nor underestimated” in the face of the AI revolution. Yet current efforts remain a step behind the onslaught of AI-generated falsehoods.

Conclusion and Implications

If the UK, EU, and like-minded democracies do not rapidly adapt to the realities of AI-driven cognitive warfare, they risk strategic defeat in an important realm of 21st-century conflict. Losing the AI influence war doesn’t happen with a formal surrender; instead, it manifests as a gradual erosion of democratic resilience. Societies may grow deeply divided, citisens lose trust in media and governments, and adversarial narratives become entrenched. In the long run, this could weaken the political will and cohesion needed to respond to more conventional security threats. As one analysis grimly observed, the cost of inaction is high – allowing adversaries to exploit AI for malign influence can lead to a “strategic imbalance favouring adversaries”, with a flood of false narratives eroding public trust and even devastating democratic institutions if left unchecked.

Reversing this trajectory will require Western nations to elevate the priority of the cognitive battlespace in national security planning. Some broad imperatives emerge:

Develop Offensive and Defensive AI Capabilities: The UK and EU should invest in AI tools not just to detect and debunk disinformation, but also to disseminate counter-narratives that truthfully push back against authoritarian propaganda. Ethical guidelines for such operations must be established, but fear of using AI at all in information ops leaves the field open to adversaries.

Whole-of-Society Resilience: Building public resilience is crucial. Education in media literacy and critical thinking, transparency about threats, and empowering independent journalism are all part of inoculating society. A populace that can recognise manipulation is the best defence against cognitive warfare. The goal is to ensure citizens can engage with digital information sceptically, blunting the impact of fake or AI-manipulated content.

International Coordination: The transatlantic alliance and democratic partners need better coordination in the information domain. NATO’s work on cognitive warfare should be complemented by EU and UK joint initiatives to share intelligence on disinformation campaigns and align responses. A unified front can deny adversaries the ability to play divide-and-conquer with different countries.

Adaptive Governance: Western policymakers must make their regulatory frameworks more agile in the face of technological change. This might include faster mechanisms to hold platforms accountable, updated election laws regarding AI-generated content, and perhaps narrowly tailored laws against the most dangerous forms of disinformation (such as deceptive media that incites violence). The challenge is doing so without undermining free speech – a balance that requires constant calibration as AI technology evolves.

In summary, the UK and EU are at a crossroads. They can continue on the current path – risking that AI-enabled influence attacks will outpace their responses – or they can strategise anew and invest in winning the cognitive fight. The latter will demand political will and creativity: treating information space as a domain to be secured, much like land, sea, air, cyber and space. It also means confronting uncomfortable questions about using emerging technologies in ways that align with democratic values yet neutralise malign propaganda.

The cognitive battle-space is now a permanent feature of international security. Western democracies must not cede this battlefield. Maintaining an open society does not mean being defenceless. With prudent adoption of AI for good, and a staunch defence of truth, the UK, EU, and their allies can start to turn the tide in the AI influence war. Failing to do so will only embolden those who seek to “attack the democratic pillars of the West” through information manipulation. In this contest for minds and hearts, as much as in any other domain of conflict, strength and resolve will determine who prevails.

Bibliography

1. NATO Allied Command Transformation. “Cognitive Warfare.” NATO ACT, Norfolk VA.

2. Bryc, Agnieszka. “Destroy from within: Russia’s cognitive warfare on EU democracy.” Kolegium Europy Wschodniej, 27 Nov 2024.

3. European Security & Defence College (ESDC). “Cognitive warfare in the new international competition: an emerging challenge for the EU,” 28 May 2024.

4. Williams, Cameron (Modern War Institute). “Persuade, Change, and Influence with AI: Leveraging Artificial Intelligence in the Information Environment.” Modern War Institute at West Point, 14 Nov 2023.

5. UK Ministry of Defence. Defence Artificial Intelligence Strategy, June 2022. UK

6. Fitz-Gerald, Ann M., and Halyna Padalko (RUSI). “The Need for a Strategic Approach to Disinformation and AI-Driven Threats.” RUSI Commentary, 25 July 2024.

7. Science Business News. “EU is ‘losing the narrative battle’ over AI Act, says UN adviser,” 05 Dec 2024.

The Grey Mirage: Navigating Strategic Uncertainty and the Elusive Victory in Grey Zone Conflicts

Imagine a world where war is waged not with bombs and bullets, but with lines of code and viral misinformation. This is the reality of grey zone conflicts, a persistent feature of modern geopolitics characterised by cyber operations, economic coercion, and disinformation. While many initially hailed these tactics as a revolutionary new form of strategic competition, a critical examination reveals that they not only fundamentally fail to achieve strategic victory in a traditional Clausewitzian sense but also introduce profound strategic uncertainty and volatility into the international system. Extending Thomas Rid’s compelling argument that “cyber war will not take place” due to the inherent lack of decisive physical destruction, this critique applies even more broadly to the entire spectrum of grey zone conflicts.

To understand the inherent limitations of these operations, we must return to the foundational strategic thought of Carl von Clausewitz. His framework remains a lodestar: tactical successes must always serve political objectives, and the very essence of war is to impose one’s will upon the enemy. As Michael Handel succinctly summarises, Clausewitzian war aims at the destruction of enemy forces, control of vital resources, and the sway of public opinion. Grey zone tactics, however, are structurally incapable of achieving these aims in the decisive manner Clausewitz envisioned. They may sow disruption and discord, but they rarely deliver battlefield outcomes, nor can they compel political compliance in the way traditional military campaigns do. Consider, for instance, the persistent cyberattacks between nations; while disruptive and costly, they have yet to force a nation to fundamentally alter its core strategic direction.

The very nature of grey zone strategies – their calculated avoidance of outright force and immediately recognisable acts of aggression – means they cannot truly compel an adversary to accept a fundamentally new strategic order. Cyber operations, as Rid convincingly argues, rarely inflict the kind of lasting, tangible damage comparable to conventional military strikes. Disinformation campaigns, while capable of eroding trust in institutions and even mobilising populations, as seen in the Arab Spring uprisings, cannot on their own force political capitulation. Economic sanctions, though often painful and strategically useful in shaping behaviour, are notoriously slow and far from guaranteed to change a determined state’s core strategic calculations.

This inherent strategic limitation is further underscored by Colin Gray’s assertion that strategy is fundamentally about the application of force to achieve political objectives. For Gray, war is fundamentally about contesting and achieving control, and without the capacity to impose a decisive order, grey zone tactics fall drastically short of true strategic efficacy. He cautions that the absence of decisive engagement in contemporary conflicts leads not to resolution, but to a debilitating strategic paralysis. This resonates deeply with Clausewitz’s core tenet that successful war must culminate in the decisive defeat of the enemy. Grey zone conflicts, by their very nature, do not and cannot fulfil this criterion. At best, they generate protracted stalemates; at worst, they risk unintended escalation into open, conventional warfare.

Countering the Cumulative Argument and Embracing Ambiguity: Incrementalism vs. Decisiveness

It is important to acknowledge a key counterargument: that grey zone tactics, while rarely decisive alone, gain strategic effect cumulatively over time. Proponents argue that persistent cyber intrusions, disinformation, and economic pressure can erode an adversary’s strength and will. This view sees grey zone warfare as long-term shaping, not a knockout blow, exemplified by China’s “Three Warfares” doctrine.

Furthermore, the ambiguity of grey zone conflicts can be strategically useful, like nuclear deterrence. Bernard Brodie argued nuclear war’s cost shifted strategy to prevention, redefining “victory” as avoiding war. Similarly, grey zone tactics might deter and manage competition below open conflict. Incremental disruption, like cyberattacks on Iran’s nuclear program, can also shift power balances.

Hurting Without Winning and the Zero-Sum Nature of Grey Zone Competition

Thomas Schelling noted, “Victory is no longer a prerequisite for hurting the enemy.” This is key to grey zone tactics, which can aim to inflict pain and signal resolve without overt war. Even non-military gains – diplomatic wins, sanctions, legal advantages achieved through disinformation and cyber influence – become strategic victories in this zero-sum competition. This is particularly relevant as tech-savvy strategists recognise the advantages of ambiguity in these operations.

However, pursuing overwhelming military victory can backfire, escalating conflict. Grey zone tactics offer a way to avoid this, operating below the threshold of conventional war. Yet, this ambiguity breeds volatility, with miscalculation and escalation always looming.

Strategic Victory as Peace-Winning and the Challenge of Subjectivity

Rethinking “strategic victory” beyond military terms is crucial. Robert Mandel distinguishes “war-winning” from “peace-winning,” arguing true strategic victory is “peace-winning” – a multi-dimensional achievement across information, politics, economics, and diplomacy. Grey zone tactics align with this broader view, especially as public mobilisation and decentralised networks shape geopolitics.

Yet, “victory” in the grey zone remains subjective and hard to measure. Ethan Kapstein highlights the difficulty of defining metrics, gaining consensus, and obtaining reliable data in grey zone operations. Progress in one area may undermine another, increasing strategic uncertainty. Whether grey zone tactics are a “strategic win” depends on perspective and chosen metrics.

Taiwan: Strategic Uncertainty in Action

Taiwan exemplifies the inherent volatility of grey zone warfare: while hybrid strategies can pressure an opponent, they provide no clear pathway to a controlled, predictable outcome. The lack of definitive thresholds makes grey zone tactics as much a risk as an opportunity for the aggressor. Imagine China using grey zone tactics against Taiwan: cyberattacks, disinformation, and economic pressure. While this might weaken Taiwan, it’s unlikely to force capitulation without risking wider conflict. Taiwan’s reaction, U.S. responses, and the ever-present risk of miscalculation create a strategic dilemma.

While Russia has shown resilience to external grey zone pressures by controlling information, societal resilience only mitigates, not eliminates, strategic uncertainty. Even the most robust resilience strategies cannot eliminate the risk of miscalculation or escalation, underscoring the inherent volatility of grey zone conflicts. Because grey zone conflicts operate ambiguously, even careful campaigns can unexpectedly escalate, making control and predictability elusive.

Policy Implications: Actively Shaping the Grey Zone for Advantage

The inherent strategic uncertainty of grey zone conflicts demands proactive policies:

  1. Sharpen Intelligence and Active Disruption: Enhance intelligence to understand adversary intentions and develop capabilities to actively disrupt their grey zone operations.
  2. Develop Flexible and Escalatory Response Options: Create a wider range of responses, including calibrated counter-grey zone tactics and clear signalling for de-escalation and conflict management. As artificial intelligence and automation continue to reshape information warfare, states must anticipate how AI-driven disinformation, deepfake technology, and autonomous cyber operations will further complicate grey zone conflicts. Developing countermeasures that integrate AI-based detection and rapid-response systems will be critical for maintaining strategic advantage.
  3. Promote Transparency to Force Predictability: Actively expose adversary actions to force them into a more predictable strategic posture, enhancing transparency and accountability in the grey zone.
  4. Focus on Proactive Crisis Management: Develop proactive crisis management to prevent crises, including clear communication, de-escalation protocols, and persistent low-intensity engagement for stability.
  5. Re-evaluate “Victory” and Embrace Persistent Engagement: Shift from traditional victory metrics to measures of resilience, deterrence, and long-term shaping, embracing persistent engagement as the norm in grey zone competition.

Conclusion: Embracing Uncertainty, Seeking Control Through Persistent Engagement

Russia’s pre-2022 hybrid warfare campaign in Ukraine – combining cyber operations, disinformation, and economic pressure – demonstrated the limitations of grey zone tactics. Rather than coercing Ukraine into submission, these operations reinforced Ukrainian national resistance and galvanised Western military support, ultimately leading to Russia’s full-scale invasion. This case underscores the strategic volatility of grey zone competition: while these tactics can create disruption, they provide no guarantee of controlled, predictable outcomes.

This highlights how grey zone tactics, while seemingly flexible, are unlikely to deliver traditional strategic victory and introduce significant strategic uncertainty. While ambiguity and “peace-winning” are modern adaptations, they don’t guarantee predictable outcomes or escalation control. The grey zone is a volatile battlespace defined by miscalculation and instability. Navigating the grey zone requires embracing uncertainty, prioritising crisis management, and actively shaping the battlespace. In this new era of perpetual contestation, mastering the grey zone is not about winning – it is about ensuring that one’s adversaries never can.


References

  1. Brodie, Bernard. “The Absolute Weapon: Atomic Power and World Order.” The Yale Review 35, no. 3 (Spring 1946): 456-472.
  2. Gray, Colin S. The Strategy Bridge: Theory for Practice. Oxford: Oxford University Press, 2010.
  3. Handel, Michael I. Masters of War: Classical Strategic Thought. London: Frank Cass, 2001.
  4. Kania, Elsa B. “The PLA’s Latest Strategic Thinking on the Three Warfares.” The Jamestown Foundation, August 22, 2016. https://jamestown.org/program/the-plas-latest-strategic-thinking-on-the-three-warfares/.
  5. Kapstein, Ethan B. “Measuring Success in Complex Operations.” The Journal of Strategic Studies 34, no. 2 (April 2011): 267-285.
  6. Mandel, Robert. “Thinking about Victory in Strategy.” The Journal of Strategic Studies 34, no. 2 (April 2011): 199-200.
  7. Monaghan, Sean. “Twitter Revolutions? Social Media and the Arab Spring.” Whitehall Papers 69, no. 1 (2011): 21-22.
  8. Rid, Thomas. Cyber War Will Not Take Place. London: Hurst, 2013.
  9. Sanger, David E., and William J. Broad. “Obama Order Sped Up Wave of Cyberattacks Against Iran.” The New York Times, June 1, 2012. https://www.nytimes.com/2012/06/01/world/middleeast/obama-ordered-wave-of-cyberattacks-against-iran.html.
  10. Schelling, Thomas C. Arms and Influence. New Haven: Yale University Press, 1966.
  11. Simons, Greg. “Russia and information confrontation: perceptions, strategies and responses.” Journal of strategic studies 42, no. 1 (2019): 139-140.
Past meets Future

Rethinking Warfare: Clausewitz in the Age of Cyber and Hybrid Conflict

Carl von Clausewitz’s claim that war is “a continuation of politics by other means” has survived railways, radio and nuclear weapons.  Today the “other means” range from data-wiping malware that bricks ventilators to viral deep-fakes that never fired a shot.  The central puzzle is whether these novelties merely change the character of war (the tools, tempo and terrain) or whether they erode its immutable nature of violence, chance and political purpose (Echevarria 2002; Strachan 2013). 

International lawyers behind the Tallinn Manual 2.0 accept that a non-international armed conflict may now consist solely of cyber operations if the effects rival kinetic force (Schmitt 2017).  Thomas Rid counters that almost all cyber activity is better classed as espionage, sabotage or subversion. Such attacks may be potent, but not war in the Clausewitzian sense (Rid 2017).  The Russian “AcidRain” attack of February 2022 sits precisely on that fault-line: a single wiper disabled thousands of Ukrainian satellite modems and 5,800 German wind-turbines, yet no bombs fell (SentinelLabs 2022; Greenberg 2023).  If violence is judged by effect on human life rather than by the immediate mechanics of injury, Clausewitz still works; if it is judged by physical harm alone, he wobbles. 

The 2022 US National Defense Strategy elevates “integrated deterrence”, urging day-to-day campaigning below the armed-attack threshold (US DoD 2022).  US Cyber Command’s doctrine of persistent engagement pushes the same logic into practice, contesting adversaries continually rather than waiting for crises (USCYBERCOM 2022).  Fischerkeller and Harknett argue that such calibrated friction stabilises the domain; Lynch casts it as a new “power-sinew contest” in which outright war is the exception, not the rule (Fischerkeller & Harknett 2019; Lynch 2024).  The danger is conceptual inflation: call every malicious packet “war” and escalation thresholds blur, yet forcing every new tactic into Clausewitz’s vocabulary risks missing genuine novelty. 

Frank Hoffman’s once-handy term “hybrid warfare” now covers almost any sub-threshold activity.  NATO’s recent work on cognitive warfare goes further, treating perception itself as decisive terrain and calling for a fresh taxonomy of “acts of cognitive war” (NATO Innovation Hub 2023).  Clausewitz, writing in an age of limited literacy, rarely considered the deliberate collapse of an adversary’s shared reality as a line of operation.  The gap is undeniable – but it need not be fatal if his categories can stretch. 

Clausewitzian elementDigital-age inflectionIllustrative case
ViolencePhysical harm or systemic disruption that produces downstream human sufferingAcidRain modem wipe, 2022
ChanceAmplified by tightly coupled networks where small code changes trigger cascading failuresLog4j exploit cascade, 2021
Political purposeTerritorial control plus cognitive or behavioural manipulation2016 US election interference

The table shows how old categories bend.  Violence migrates into infrastructure; chance spikes in opaque systems; political purpose colonises the infosphere.  None of these shifts removes politics from the centre – precisely why the trinity still maps the ground.

There are 3 key areas where Clausewitz’s wisdom holds strongly:

  1. Politics first.  Colin Gray insists that strategy is the orchestration of means to political ends; replacing artillery with algorithms does not move that lodestar (Gray 1999).
  2. Escalation logic.  Even in cyberspace, deterrence depends on adversaries reading tacit red lines.  Clausewitz’s emphasis on uncertainty and friction remains apt.
  3. Human cost.  Cyber operations hurt indirectly – frozen hospital wards, confused electorates – but the harm is felt by bodies in time and space, not by circuits.

There are however, a number of places where the strain shows, namely where:

  • Systemic cyber harm approaches “force” while sidestepping bodily violence.
  • Persistent, below-threshold campaigning blurs the war–peace boundary Clausewitz assumed.
  • The trinity was never meant to classify acts aimed at belief rather than battalions.

For now, Rid’s scepticism still holds true – most cyber operations do not meet Clausewitz’s threshold of war.  Yet as societies entangle their critical functions ever more tightly with code, the line between systemic disruption and physical violence narrows.  Clausewitz’s trinity of violence, chance, political purpose – still offers the clearest compass, because politics, not technology, remains the centre of gravity of strategy.  The compass, however, is being asked to steer across novel terrain.  Should a future campaign achieve political aims through cyber-enabled systemic coercion alone, the Prussian might finally need more than a tune-up.  Until then, his core logic endures, and while needing adaptation, it has not been eclipsed.

Bibliography

Clausewitz, C. v. (1832) On War.  Berlin: Ferdinand Dümmler.

Echevarria, A. J. (2002) ‘Clausewitz’s Center of Gravity: Changing Our Warfighting Doctrine – Again!’.  Carlisle, PA: US Army Strategic Studies Institute. 

Fischerkeller, M. P. and Harknett, R. J. (2019) ‘Persistent Engagement, Agreed Competition, and Cyberspace Interaction Dynamics’. The Cyber Defense Review

Gray, C. S. (1999) Modern Strategy.  Oxford: Oxford University Press. 

Greenberg, A. (2023) ‘Ukraine Suffered More Wiper Malware in 2022 Than Anywhere, Ever’. WIRED, 22 February. 

Lynch, T. F. III (2024) ‘Forward Persistence in Great Power Cyber Competition’.  Washington, DC: National Defense University. 

NATO Innovation Hub (2023) The Cognitive Warfare Concept.  Norfolk, VA: NATO ACT. 

Rid, T. (2017) Cyber War Will Not Take Place.  Oxford: Oxford University Press. 

Schmitt, M. N. (ed.) (2017) Tallinn Manual 2.0 on the International Law Applicable to Cyber Operations.  Cambridge: Cambridge University Press. 

SentinelLabs (2022) ‘AcidRain: A Modem Wiper Rains Down on Europe’.  SentinelOne Labs Blog, 31 March. 

US Cyber Command (2022) ‘CYBER 101 – Defend Forward and Persistent Engagement’.  Press release, 25 October. 

US Department of Defense (2022) National Defense Strategy of the United States of America.  Washington, DC. 

The Future of War: AI and Strategy

When Carl von Clausewitz wrote that war is “a continuation of politics by other means,” he centred conflict on purpose rather than technology. Colin Gray later warned that strategic constants outlive every gadget. Artificial intelligence now accelerates observation-decision loops from minutes to milliseconds, but whether that shift dethrones human strategy is still contested.

Speed Meets Friction

Ukrainian drone teams run machine-vision updates at the front line every fortnight, turning quadcopters into near-autonomous kamikaze platforms (Bondar 2025). Yet the same coders struggle with false positives – such as bears flagged as enemy sentries – and with mesh-network bottlenecks once EW jammers blanket the spectrum. AI compresses time, but it also multiplies friction, the very element Clausewitz thought ineradicable. We have to be conscious that false positives do not just waste munitions; when the same image-detection stack mis-tags an ambulance as a supply truck, the result is shrapnel in a paediatric ward, not an algorithmic hiccup. In 2025, the World Health Organization stated that hospitals reported 205 deaths from strike-related service loss in Ukraine.

Open-source models still give insurgents propaganda reach, but the sharper edge of algorithmic warfare sits with states.  Israel’s Lavender system, revealed in 2024, generated a list of roughly 37,000 potential Hamas targets and was used even when commanders expected up to twenty civilian deaths per strike—a machine-driven tempo that unsettled some of the intelligence officers involved (McKernan & Davies 2024).  Cutting-edge autonomy, however, still demands high-end GPUs, abundant power and proprietary data.  That keeps strategic dominance gated by infrastructure, mirroring geopolitical power.  Yet, as Kuner (2024) notes, Brussels carved a national-security escape hatch into the AI Act precisely to preserve state leverage over the biggest models.

Washington’s Replicator initiative aims to field “thousands of attriable autonomous systems” within two years (DoD 2024). Beijing answers through civil-military fusion; Moscow improvises with AI-augmented loitering munitions. These programmes underpin an operating concept of continuous, sub-threshold contest, paralleling U.S. Cyber Command’s “persistent engagement”. Strategic deterrence thus rests on the hope that algorithmic agents still read tacit red lines the way humans do. In Stanford’s 2024 crisis simulations, LLM agents recommended first-strike escalation in seventy-nine per cent of runs, providing evidence that algorithmic ‘rationality’ may be anything but.

If LLM advisers escalate crises in simulation nine times out of ten, the locus of judgement drifts from commander to code.  The next question is whether that drift merely speeds execution or begins to automate strategy itself.

Promoters once claimed AI would dissolve uncertainty; real battlefields say different. Sensor glut, spoofed tracks and synthetic “ghost columns” now drown analysts in contradictory feeds (Collazzo 2025). AI redistributes fog rather than lifting it – accelerating some judgements while blinding others through overload or deception (Askew and Salinas 2025). 

The Pentagon updated Directive 3000.09 on autonomous weapons in late 2024, tightening human-in-the-loop requirements. At the multilateral level, UN talks in May 2025 once again failed to agree binding rules, though Secretary-General Guterres set a 2026 deadline (Le Poidevin 2025). Norms lag well behind code, keeping accountability – and escalation liability – firmly in human hands. 

Strategic implications

The transformative impact of AI on strategic paradigms can be distilled into a few key considerations:

  • Advantage remains political. AI is a means; objectives still emanate from human intent. Strategy therefore keeps its Clausewitzian anchor in politics.
  • Automation magnifies misperception. Faster loops leave less time for reflection; and black-box models hide their own failure modes.  Bias and data poisoning risk strategic self-harm.
  • Deterrence becomes brittle. Autonomous systems may over-react to spoofed inputs; adversaries may test thresholds in micro-seconds rather than hours, shortening the ladder of de-escalation.

Conclusions

AI does not automate strategy; it amplifies both its promise and its pathologies. Machines accelerate tactics, generate options and even draft operational plans, but they do not choose political ends – and they continue to manifest friction, chance and uncertainty. Thomas Rid remains broadly right that most cyber and AI activity falls short of war (Rid, 2017), yet as energy grids, logistics chains and battlefield kill cycles digitise, the gap between systemic disruption and physical violence narrows. For the moment, Clausewitz’s compass still points true – but the ground beneath it is starting to slide.

Select bibliography

Askew, M. and Salinas, A. (2025) ‘AI Will Make the Mind Games of War More Risky’, Business Insider, 18 Apr.

Bondar, K. (2025) Ukraine’s Future Vision and Current Capabilities for Waging AI-Enabled Autonomous Warfare.  CSIS.

Collazzo, A. (2025) ‘Warfare at the Speed of Thought’, Modern War Institute, 21 Feb.

Department of Defense (2024) ‘Replicator Initiative Progress Update’.

Kuner, C. (2024) ‘The AI Act National Security Exception’, Verfassungsblog, 15 Dec.

Le Poidevin, O. (2025) ‘Nations Meet at UN for “Killer Robot” Talks’, Reuters, 12 May.

McKernan, B., & Davies, H. (2024). The Machine Did It Coldly. The Guardian, 3 April. https://www.theguardian.com/world/2024/apr/03/israel-gaza-ai-database-hamas-airstrikes  

Rid, T. (2017) Cyber War Will Not Take Place.  Oxford: Oxford University Press.

Stanford HAI (2024) Escalation Risks from LLMs in Military and Diplomatic Contexts.

World Health Organization (2025) WHO’s Health Emergency Appeal 2025. Geneva: World Health Organization.

My Experiences in Japan / 日本での私の経験

So here I am, about a month after having returned to London, reflecting back on my experience in Japan. The most common question I’ve had from friends is “was it worth it?” or “did you learn what you wanted to learn?” and the answer to both is a resounding yes. However it isn’t as simple as that, as there are many layers to the question which need to be unpacked a little.

The first layer is understanding what it is that I went off to Japan to learn in the first place. Of course there is the obvious “Japanese Language” side of things, but there is much more to it than that. The real learning I was hoping to take away was about myself. And if you want to learn about yourself, one of the best ways is to teach others. Which thanks to Phil’s company I was able to do exactly that with a bunch of middle school kids over 6 days of English language camps!

Class photo!

The camp is worth a whole post by itself, but the takeaway is how teaching kids made me feel about myself and reflect on what I enjoy doing professionally. It allowed me to understand what I value (honesty, enthusiasm, progress) in a much more immersive way than a work environment would.  I’ll definitely be taking this into my next role and it has helped me mature bit more as an individual.

Alongside teaching, there were many other examples of “bonus” learning opportunities I was able to take away from the experience.  One key aspect was taking myself out of a familiar environment and getting the mental/physical space to learn and reflect.  That alone was worth the proverbial price of admission.

The next layer is about “what” it is I did.  While it is obvious to some, but not always to me, it isn’t the destination that matters so much, but who you spend it with and the attitude you take with you.  I enjoy spending time by myself (quite a lot) but all my most enjoyable experiences are with friends and family.  It isn’t just quality though, it is quantity too.  People can help you overcome natural inertia (read: laziness) to get out there and do more. In fact, when Julie came to visit for a week, we crammed in more stuff than I did in the previous months!

Fun and hi-jinks!

Also, being in the right mindset (a positive one), I was able to value those times much more and care a little bit less about the latest distracting “must-own-thing”.  But it is easily forgotten and I have to remind myself often to focus on new experiences with loved ones and less on new, shiny, technology…

There is a lot more I was able to get from my travels, but that’s about my limit for for self reflection today.  I think I need to do a round up of all the beers I forgot to mention in a new post…

Nanzoin Temple / 南蔵院

With Julie’s visit this week, I have been very busy doing all sorts of fun tourist type activities! One of the best trips we did was heading off to Nanzoin Temple, which features The Reclining Buddha (more on this later).

Rather than bore you with lots of hyper-descriptive prose, I think a visual tour would be much more fun. So let’s get started!

Getting there was pretty easy, just a 30 minute train ride from Hakata station. Once we got off the train and started trekking up towards the temple grounds, we found this family of turtles with their own mini-shrine. They were kind enough to let me take way too many photos as I over-enthusiastically snapped shots from various exciting angles.

Look who came to greet us!

Clearly turtles weren’t the only photogenic fauna around, as this rather large dragonfly posed long enough for me to get a fairly zoomed in photo.

Posed for the photo and flew off as soon as I was done…

One of the first statues to be seen was of this chap with his flames and sword. We originally thought he was some bad guy, but Julie did some research and we discovered he is actually Fudoumyouou who is a Buddhist deity who protects worshippers from harm.

Later learnt this angry looking guy is a protective deity

As we spotted more of the classic Buddha style statues, we noticed they were usually wearing some sort of clothing, be that a hat, a cardigan or the like.  It turns out the reason behind this is fairly involved, so check out this post for details.

Lots of little statues, all wearing the cutest clothes

After a bit more walking, we finally made our way to the star attraction, The Reclining Buddha himself!  I must say, the statue really is impressive and quite something else to behold.  He is the world’s largest bronze statue and is the same size as the Statue of Liberty. He is so large I had to use the panorama mode of the phone to get him fully in the frame.

Even more impressive in person…

You can walk around to his feet and I think the following image really gives you a good idea of the scale of the statue…

Buddha foot with Julie touching the thousand spoke wheel symbol

Once we had seen the Buddha and enjoyed some well timed ice-creams, it was time for some exploring!

The path upwards…

After following a (very steep) trail up and out of the temple grounds, we found ourselves in some beautiful back-country, surrounded by bamboo forest. I love the sound they make as they rustle in the breeze. It was truly a magical part of the day for me.

I love bamboo forests

From the trail, we were able to visit a number of smaller temples and shrines. At one temple, we even had the Buddhist monk come out and offer us watermelon to eat! It was unexpected and very much appreciated on a very warm day. We ventured on each time and were treated to some great views through the trees.

Fabulous views over the countryside

As we wound our way back down the mountain towards the station again, we found many more statues, but the following set were really quite something else. Such a riot of colour and clothing on each of the little stone statues!

Someone likes playing dress up…

Here you can see some up close… adorable!

So dang cute…

Overall it was a brilliant day and even the photos can’t do justice to the adventure we had. If you ever get a chance, I’d highly recommend a visit!

Japanese Calligraphy / 書道

This weekend I had an opportunity to try out Japanese calligraphy at an event put on by the Fukuoka International Student Support Centre (FiSSC). Calligraphy isn’t something I’d normally be very excited about, but I’m a sucker for trying anything new so I thought I’d give it a bash anyway. It also helped that my friend Sora was coming with me, who always ready to help me avoid doing anything (too) embarrassing.

Chilling with my buddy Sora

Japanese calligraphy (shodou/書道) is an art form for writing Japanese Kanji (Chinese characters) using a brush and ink that is taught to all Japanese students in primary school, but it is rarely used after that. As such the event had a good mix of Japanese and non-Japanese people attending, with quite the range of skills on display, from very basic through to really artful. The sensei (teacher) for the lesson was truly an artist and was able to create evocative works with little apparent effort. Practice definitely makes perfect.

We were given our instructions, ink, brush and a stack of paper to practise with, after which we were told to get on with practising! My own works were somewhat less than perfect, but I learnt a trick or two over the period and I really felt that I’d improved at least a little bit over the duration of the class.

The sensei watches on as a student tries to do “光” or “light”

I’m glad I did manage to improve a little as I was “asked” to try out writing a kanji out on one of the larger sheets (the joys of being very-obviously-foreign). I choose the character for “joy” or “喜” as it is simple and represents something I think we could all do with a little more of in our lives. The end result was middle of the road enough to neither attract laughter nor praise. Phew, embarrassment avoided!

My “masterpiece”. On the left is my name in katakana – マシュートイ

My own experiments aside, the work produced by masters of the art truly is something else though. I’m impressed that so much feeling and emotion can be conveyed through a simple black and white medium.

All said, I had a lot of fun and learnt the basics of a new skill. Not a bad result for a Saturday afternoon!

My other creations. One of them isn’t a real kanji, can you guess which one? 🙂

With Friends / 友達と

The past few weeks have been a social whirlwind and I’ve had an opportunity to try out all sorts of new and interesting activities .  I’ll make sure I do a proper set of updates, but this time I’d like to briefly tell you about an adventure into Thai cuisine.

As a bit of background, there is a regular “language lounge”, where we swap English and Japanese practice with learners from both languages, which is really good fun. I get to meet all sorts of awesome people and learn about Japanese language and culture from them. After one particular lounge a group of 6 of us decided to head off to grab some dinner, opting to try out a Thai restaurant in Tenjin.

All smiles 🙂

This was the first time I’ve tried Thai food while in Japan and I have to say, I found it more “authentic” than what you would normally get in London. Not sure why I think it is more authentic, apart from it not feeling particularly Western in its flavours and presentation. That aside, the food itself was really tasty! Lots of spice though, so those with a sensitive tongue will need to be extra careful.

I feel that this style of setting makes it really easy to communicate effectively with people. The casual atmosphere allows you to relax and banter back and forth, using a mix of both native and learnt languages. I really enjoy being able to try out new words, expressions and such without anyone to judge you on technical performance. In fact the only real judge is whether you are able to effectively get your ideas across to the people you are talking to (eventually).

I made a commitment to use my time in Japan as effectively as possible and I’m super happy I’ve managed to get into more of this style activity, in addition to the classroom education and personal study.

Page 2 of 4

Powered by WordPress & Theme by Anders Norén