Double Pendulum - Why Meaning Is Chaotic
- 1 day ago
- 5 min read
What Physics Can Teach Organizations About Context Manipulation
This post addresses one decision: whether your current monitoring approach can detect campaigns that manipulate meaning through context rather than lies. The urgency is straightforward — the tools that caught last decade's disinformation are the wrong tools for what is being deployed today.
Simulation - Turn the "Chaos On"!
Why Do Identical Words Produce Opposite Meanings Across Campaigns?
TL;DR: Meaning is not in the words — it is in the contextual frame surrounding them, and that frame is the primary target of modern information operations.
Evidence:
The word "Modernisierung" carries positive connotations in mainstream policy discourse. In campaigns monitored by evai.ai, the same term was consistently framed as linking it to financial burden and coercion; the content was factually accurate, but the context inverted the valence.
A 2023 Reuters Institute report found that the majority of viral mis- and disinformation episodes studied did not involve fabricated facts, but selectively framed or decontextualised true information.
Any statement is a vector whose direction is determined by the surrounding context, not by its literal content.
So what: Keyword and sentiment tools that ignore contextual embedding will consistently produce false negatives for the most sophisticated campaigns.
What Is Semantic Drift, and Why Is It Harder to Detect Than Classic Misinformation?
TL;DR: Semantic drift is the gradual, coordinated shift in the associations surrounding a term or narrative — it is slow, deniable, and nearly invisible to snapshot monitoring.
Evidence:
Classic misinformation (a fabricated quote, a false statistic) is detectable with fact-checking. Semantic drift requires longitudinal tracking: the same term must be compared across time, platform, and speaker network.
In evai.ai's analysis of the AfD municipal-level communication strategy, drift was not in explicit claims but in consistent co-occurrence of specific terms with threat-coded vocabulary — a pattern only visible across thousands of posts over months.
Platform algorithms amplify semantic drift because engagement correlates with emotional valence, not factual accuracy.
Detection requires vector-space modelling over time (semantic change detection), not classification of individual statements as true or false.
So what: Near-real-time semantic baseline monitoring is the minimum viable capability.
How Does Context Function as an "Initial Condition" for Meaning?
TL;DR: Like a double pendulum, a small change in contextual initial conditions produces exponentially diverging meaning trajectories — and the divergence is deterministic, not random.
Evidence:
In physics, deterministic chaos describes systems that are fully rule-governed yet practically unpredictable because of sensitivity to initial conditions. Change the starting angle of a double pendulum by 0.001 degrees, and within seconds, the paths bear no resemblance to each other.
The context window preceding any statement functions as its initial condition. A sentence about energy costs, framed after economic-anxiety messaging, takes an entirely different interpretive trajectory than the same sentence framed after a policy reform announcement.
This is not a loose metaphor — it describes how large language models work. Prompt context determines model output with the same sensitive dependency: tiny changes in framing reliably produce large changes in interpretation.
Adversarial actors who understand this do not need to fabricate content. They need only to control what comes before the message, which is cheaper and more durable.
So what: Monitoring systems that evaluate content in isolation are blind to the most scalable attack vector in information operations today.
What Are the Six Criteria for a Context-Aware Narrative Monitoring System?
TL;DR: Effectiveness depends less on data volume than on whether the system tracks contextual embedding, temporal drift, and cross-platform amplification simultaneously.
Ranked by diagnostic value:
Temporal semantic baseline: Can the system compare how a term's associative network has shifted over 30, 90, 180 days?
Cross-platform context aggregation: A narrative seeded on Telegram often completes its contextual frame on mainstream platforms within two to four days.
Speaker-network attribution: Who is doing the reframing, and are accounts exhibiting coordinated behaviour?
Sentiment disaggregation by sub-audience: Aggregate scores mask campaigns that are neutral for one audience while toxic for another.
Velocity and acceleration metrics: Drift that accelerates is more actionable than drift that is stable.
Human-readable narrative summaries: Analysts cannot act on raw vector distances.
Build vs. Buy: Can Off-the-Shelf Tools Detect Context Manipulation?
TL;DR: General-purpose social listening tools were designed for brand sentiment, not information operations. The gap is architectural, not a feature request.
Evidence:
Standard social listening platforms aggregate keyword mentions and classify sentiment. They were built for marketing use cases: volume, reach, and share of voice. None of these metrics captures contextual drift.
Building in-house requires a fine-tuned embedding model per language and domain, a longitudinal vector store, a coordination detection layer, and analyst tooling. Engineering cost typically exceeds €500k before the first production deployment.
Purpose-built disinformation monitoring systems are designed around the specific forensics of information operations. The cost-benefit calculus tilts heavily toward buy for organizations without a dedicated ML research function.
Key risk of building: the threat landscape is constantly evolving. A proprietary system built to detect last year's techniques requires ongoing retraining investment that compounds annually.
How Will LLMs Change the Scale of Context Manipulation?
TL;DR: LLMs industrialise context manipulation by making it trivial to generate thousands of contextually varied versions of the same message. Monitoring must now operate at generation speed, not publication speed.
Evidence:
Generating a contextually manipulated version of a statement previously required a skilled human operator. Current LLMs require a prompt. The cost of producing 10,000 contextual variants of a narrative has dropped to near zero.
Detection is not symmetric: generating synthetic contextual frames is cheap; detecting that they are synthetic and coordinated is computationally expensive.
evai.ai's analysis of synthetic persona activity on financial topics (Swatch campaign) showed that the most effective campaigns used LLM-generated contextual framing for niche sub-audiences — not mass-broadcast false content.
Organizations that monitor publication counts are measuring the wrong variable. The relevant variable is contextual surface area: how many distinct contextual frames does a narrative appear in simultaneously?
So what: Monitoring infrastructure needs to be re-budgeted for the LLM era. The detection problem is an order of magnitude larger than in 2020.
FAQ
Isn't this just a restatement of 'framing effects' from media studies?
Framing theory describes the phenomenon; the chaos model explains the mechanism and predicts where divergence accelerates, which determines the optimal intervention point.
We already track sentiment. Is that not sufficient?
Aggregate sentiment tells you how people feel about a topic; it does not tell you what contextual frame is producing that feeling or whether it is the product of coordinated activity.
How much historical data is needed before baseline comparison becomes meaningful?
In practice, 90 days of continuous embedding data per platform provides a workable baseline for most domains.
Can this approach work for smaller organizations without a dedicated analyst?
A well-designed system surfaces pre-interpreted findings rather than raw data. The analyst time requirement is proportional to response complexity, not detection complexity.
How do you distinguish coordinated inauthentic behaviour from genuine grassroots usage?
Network-level signals: account age, posting velocity, linguistic diversity, inter-account timing — distinguish coordination from organic drift. No single signal is conclusive; the combination is diagnostic.
Is there a false positive risk?
Yes. Systems should be tuned for high precision at the cost of recall for automated alerts, with human review as the gate before any public response.
What is the biggest misconception about disinformation monitoring?
The problem is finding false content. The harder problem is tracking true content being weaponised through context.
Methods & Data Appendix
Available Evidence
evai.ai case analysis: Swatch disinformation campaign — coordinated fake account network, sentiment spike pattern, measurable stock price correlation.
evai.ai case analysis: AfD municipal communication strategy document — narrative pattern analysis, vocabulary co-occurrence mapping.
Reuters Institute Digital News Report 2023 — referenced for base rate of context manipulation vs. fabrication in viral episodes.

