2016 Election

The concept of “prebunking” emerges as a proactive strategy in the fight against disinformation, an ever-present challenge in the digital era where information spreads at unprecedented speed and scale. In essence, prebunking involves the preemptive education of the public about the techniques and potential contents of disinformation campaigns before they encounter them. This method seeks not only to forewarn but also to forearm individuals, making them more resilient to the effects of misleading information.

Understanding disinformation

Disinformation, by definition, is false information that is deliberately spread with the intent to deceive or mislead. It’s a subset of misinformation, which encompasses all false information regardless of intent.

In our current “information age,” the rapid dissemination of information through social media, news outlets, and other digital platforms has amplified the reach and impact of disinformation campaigns. These campaigns can have various motives, including political manipulation, financial gain, or social disruption — and at times, all of the above; particularly in the case of information warfare.

The mechanism of prebunking

Prebunking works on the principle of “inoculation theory,” a concept borrowed from virology. Much like a vaccine introduces a weakened form of a virus to stimulate the immune system’s response to it, prebunking introduces individuals to a weakened form of an argument or disinformation tactic, thereby enabling them to recognize and resist such tactics in the future.

The process typically involves several key elements:

  • Exposure to Techniques: Educating people on the common techniques used in disinformation campaigns, such as emotional manipulation, conspiracy theories, fake experts, and misleading statistics.
  • Content Examples: Providing specific examples of disinformation can help individuals recognize similar patterns in future encounters.
  • Critical Thinking: Encouraging critical thinking and healthy skepticism, particularly regarding information sources and their motives. Helping people identify trustworthy media sources and discern credible sources in general.
  • Engagement: Interactive and engaging educational methods, such as games or interactive modules, have been found to be particularly effective in prebunking efforts.

The effectiveness of prebunking

Research into the effectiveness of prebunking is promising. Studies have shown that when individuals are forewarned about specific misleading strategies or the general prevalence of disinformation, they are better able to identify false information and less likely to be influenced by it. Prebunking can also increase resilience against disinformation across various subjects, from health misinformation such as the anti-vaccine movement to political propaganda.

However, the effectiveness of prebunking can vary based on several factors:

  • Timing: For prebunking to be most effective, it needs to occur before exposure to disinformation. Once false beliefs have taken root, they are much harder to correct — due to the backfire effect and other psychological, cognitive, and social factors.
  • Relevance: The prebunking content must be relevant to the audience’s experiences and the types of disinformation they are likely to encounter.
  • Repetition: Like many educational interventions, the effects of prebunking can diminish over time, suggesting that periodic refreshers may be necessary.

Challenges and considerations

While promising, prebunking is not a panacea for the disinformation dilemma. It faces several challenges:

  • Scalability: Effectively deploying prebunking campaigns at scale, particularly in a rapidly changing information environment, is difficult.
  • Targeting: Identifying and reaching the most vulnerable or targeted groups before they encounter disinformation requires sophisticated understanding and resources.
  • Adaptation by Disinformers: As prebunking strategies become more widespread, those who spread disinformation may adapt their tactics to circumvent these defenses.

Moreover, there is the ethical consideration of how to prebunk without inadvertently suppressing legitimate debate or dissent, ensuring that the fight against disinformation does not become a vector for censorship.

The role of technology and media

Given the digital nature of contemporary disinformation campaigns, technology companies and media organizations play a crucial role in prebunking efforts. Algorithms that prioritize transparency, the promotion of factual content, and the demotion of known disinformation sources can aid in prebunking. Media literacy campaigns, undertaken by educational institutions and NGOs, can also equip the public with the tools they need to navigate the information landscape critically.

Prebunking represents a proactive and promising approach to mitigating the effects of disinformation. By educating the public about the tactics used in disinformation campaigns and fostering critical engagement with media, it’s possible to build a more informed and resilient society.

However, the dynamic and complex nature of digital disinformation means that prebunking must be part of a broader strategy that includes technology solutions, regulatory measures, and ongoing research. As we navigate this challenge, the goal remains clear: to cultivate an information ecosystem where truth prevails, and public discourse thrives on accuracy and integrity.

Read more

conspiracy theories, disinformation, and fake news

Conspiracy Theory Dictionary: From QAnon to Gnostics

In half a decade we’ve gone from Jeb Bush making a serious run for president to Marjorie Taylor Greene running unopposed and winning a House seat in Georgia. QAnon came seemingly out of nowhere, but taps into a much deeper and older series of conspiracy theories that have surfaced, resurfaced, and been remixed throughout time.

Essentially, QAnon is a recycling of the Protocols of the Elders of Zion conspiracy theory that drove the Nazi ideology and led to the genocide of over 6 million Jews, gypsies, gays, and others who made Hitler mad. It’s a derivative of the global cabal conspiracy theory, and is riddled with the kind of conspiratorial paranoia that led to the deaths of over 75 million people in World War II.

The spread of the QAnon conspiracy theory greatly benefits from historical memory, getting a generous marketing boost from sheer familiarity. It also benefits from an authoritarian mentality growing louder in America, with a predilection for magical thinking and a susceptibility to conspiratorial thinking.

conspiracy theories, by midjourney

Tales as old as time

Conspiracy theories have been around much longer even than the Protocols — stretching back about as long as recorded history itself. Why do people believe in conspiracy theories? In an increasingly complex world brimming with real-time communication capabilities, the cognitive appeal of easy answers may simply be stronger than ever before.

Anthropologists believe that conspiracy theory has been around for about as long as human beings have been able to communicate. Historians describe one of the earliest conspiracy theories as originating in ancient Mesopotamia, involving a god named Marduk and a goddess called Tiamat — both figures in Babylonian creation mythology.

According to the myth, Marduk defeated Tiamat in battle and created the world from her body — but some ancient Mesopotamians at the time thought that the story was not actually a mere myth, but a political cover-up of a real-life conspiracy in which the followers of Marduk secretly plotted to overthrow Tiamat to seize power.

This “original conspiracy theory” was likely driven by political tensions between city-states in ancient Mesopotamia, although there are very few written records still around to corroborate the origin of the theory or perception of the story at the time. Nevertheless, the Marduk-Tiamat myth is regarded as one of the earliest known examples of widespread belief in conspiracy theories, and it points to the relative commonality and frequency of false narratives throughout history.

Whether deployed purposefully to deceive a population for political advantage, created to exploit people economically, or invented “naturally” as a simple yet satisfying explanation for otherwise complicated and overwhelming phenomena, conspiracy theories are undoubtedly here to stay in culture more broadly for some time to come. We had best get the lay of the land, and understand the language we might use to describe and talk about them.

conspiracy theories: old men around the world map, by midjourney

Conspiracy Theory Dictionary

4chanA notorious internet message board with an unruly culture capable of trolling, pranks, and crimes.
8chanIf 4chan wasn’t raw and lawless enough for you, you could try the even more right-wing “free speech”-haven 8chan while it still stood (now 8kun). Described by its founder Frederick Bennan as “if 4chan and reddit had a baby,” the site is notorious for incubating Gamergate, which morphed into PizzaGate, which morphed into QAnon — and for generally being a cesspool of humanity’s worst stuff.
9/11 truthersPeople who believe the attacks on the Twin Towers in New York City in 2001 were either known about ahead of time and allowed to happen, or were intentionally planned by the US government.
alien abductionPeople who claim to have been captured by intelligent life from another planet, taken to a spaceship or other plane of existence, and brought back — as well as the folks who believe them.
American carnageEvocative of “immense loss” in the Nazi mythology
AntifaAntifa is anti-fascism, so the anti-anti-fascists are just fascists wrapped in a double negative. They are the real cancel culture — and a dangerous one (book burning and everything!).
Anti-SemitismOne of history’s oldest hatreds, stretching back to early biblical times
Biblical inerrancyBiblical inerrancy is the doctrine that the Bible, in its original manuscripts, is without error or fault in all its teachings. 
birtherismOne of Donald Trump‘s original Big Lies — that President Barack Obama wasn’t born in the U.S. and therefore, wasn’t a “legitimate” president.
Black Lives MatterA social justice movement advocating for non-violent civil disobedience in protest against incidents of police brutality and all racially motivated violence against black people.
blood libelA false accusation or myth that Jewish people used the blood of Christians, especially children, in religious rituals, historically used to justify persecution of Jews.
child traffickingThe illegal practice of procuring or trading children for the purpose of exploitation, such as forced labor, sexual exploitation, or illegal adoption.
Christian IdentityA religious belief system that asserts that white people of European descent are God’s chosen people, often associated with white supremacist and extremist groups.
climate change denialThe rejection or dismissal of the scientific consensus that the climate is changing and that human activity is a significant contributing factor. Part of a broader cultural trend of science denialism.
The ConfederacyRefers to the Confederate States of America, a group of 11 southern states that seceded from the United States in 1861, leading to the American Civil War, primarily over the issue of slavery.
contaminationThe presence of an unwanted substance or impurity in another substance, making it unsafe or unsuitable for use.
cosmopolitanismAnother term for globalist or internationalist, which are all dog whistles for Jewish people (see also: global cabal, blood libel)
Crossing the RubiconA phrase that signifies passing a point of no return, derived from Julius Caesar’s irreversible crossing of the Rubicon River in 49 BC, leading to the Roman Civil War.
cultural MarxismAnti-semitic conspiracy theory alleging that Jewish intellectuals who fled the Hitler regime were responsible for infecting American culture with their communist takeover plans and that this holy war is the war the right-wing fights each day.
deep stateThe idea of a body within the government and military that operates independently of elected officials, often believed to manipulate government policy and direction.
DVE(Domestic Violent Extremism): Refers to violent acts committed within a country’s borders by individuals motivated by domestic political, religious, racial, or social ideologies.
fake newsInformation that is false or misleading, created and disseminated with the intent to deceive the public or sway public opinion.
GamerGateA controversy that started in 2014 involving the harassment of women in the video game industry, under the guise of advocating for ethics in gaming journalism.
George SorosA Hungarian-American billionaire investor and philanthropist, often the subject of unfounded conspiracy theories alleging he manipulates global politics and economies.
HollywoodThe historic center of the United States film industry, often used to refer broadly to American cinema and its cultural influence.
IlluminatiA term often associated with various conspiracy theories that allege a secret society controlling world affairs, originally referring to the Bavarian Illuminati, an Enlightenment-era secret society.
InfoWarsA controversial far-right media platform known for promoting conspiracy theories, disinformation, and misinformation, hosted by clinical narcissist Alex Jones.
JFK assassinationThe assassination of President John F. Kennedy on November 22, 1963, in Dallas, Texas, an event surrounded by numerous conspiracy theories regarding the motives and identities of the assassins.
John Birch SocietyThe QAnon of its day (circa 1960s), this extreme right-wing group was theoretically about anti-communist ideals but espoused a host of conspiracy theories and outlandish beliefs.
lamestream mediaDerogatory term for any media that isn’t right-wing media.
leftist apocalypseA hyperbolic term used by some critics to describe a scenario where leftist or progressive policies lead to societal collapse or significant negative consequences.
Makers and TakersA right-wing economic dichotomy used to describe individuals or groups who contribute to society or the economy (makers) versus those who are perceived to take from it without contributing (takers). See also: Mudsill Theory, trickle down economics, supply side economics, Reaganomics, Libertarianism
micro-propaganda machineMPM: Refers to the use of targeted, small-scale dissemination of propaganda, often through social media and other digital platforms, to influence public opinion or behavior.
motivated reasoningThe cognitive process where individuals form conclusions that are more favorable to their preexisting beliefs or desires, rather than based on objective evidence.
New World OrderA conspiracy theory that posits a secretly emerging totalitarian world government, often associated with fears of loss of sovereignty and individual freedoms. (see also, OWG, ZOG)
nullificationA constitutional “theory” put forth by southern states before the Civil War that they have the power to invalidate any federal laws or judicial decisions they consider unconstitutional. It’s never been upheld by the federal courts.
One World GovernmentThe concept of a single government authority that would govern the entire world, often discussed in the context of global cooperation or, conversely, as a dystopian threat in conspiracy theories. (see also: NWO, ZOG)
PizzaGateA debunked and baseless conspiracy theory alleging the involvement of certain U.S. political figures in a child sex trafficking ring, supposedly operated out of a Washington, D.C., pizzeria.
post-truthRefers to a cultural and political context in which debate is framed largely by appeals to emotion disconnected from the details of policy, and by the repeated assertion of talking points to which factual rebuttals are ignored.
PRpublic relations
propagandaInformation, especially of a biased or misleading nature, used to promote a political cause or point of view.
Protocols of the Elders of ZionForged anti-semitic document alleging a secret Jewish child murder conspiracy used by Hitler to gin up support for his regime.
PsyOpsPsychological operations: Operations intended to convey selected information and indicators to audiences to influence their emotions, motives, objective reasoning, and ultimately the behavior of governments, organizations, groups, and individuals. Used as part of hybrid warfare and information warfare tactics in geopolitical (and, sadly, domestic) arenas.
QAnonA baseless conspiracy theory alleging that a secret cabal of Satan-worshipping pedophiles is running a global child sex-trafficking ring and plotting against former U.S. President Donald Trump.
Q DropsMessages or “drops” posted on internet forums by “Q,” the anonymous figure at the center of the QAnon conspiracy theory, often cryptic and claiming to reveal secret information about a supposed deep state conspiracy.
reactionary modernismA term that describes the combination of modern technological development with traditionalist or reactionary political and cultural beliefs, often seen in fascist ideologies.
Reichstag fireAn arson attack on the Reichstag building (home of the German parliament) in Berlin on February 27, 1933, which the Nazi regime used as a pretext to claim that Communists were plotting against the German government.
RothschildsA wealthy Jewish family of bankers, often subject to various unfounded conspiracy theories alleging they control global financial systems and world events.
sock puppetsOnline identities used for purposes of deception, such as to praise, defend, or support a person or organization while appearing to be an independent party.
“Stand back and stand by”A phrase used by former U.S. President Donald Trump during a presidential debate, which was interpreted as a call to readiness by the Proud Boys, a far-right and neo-fascist organization that seemed to answer his calling during the riot and coup attempt at the Capitol on January 6, 2021.
The StormWithin the context of QAnon, a prophesied event in which members of the supposed deep state cabal will be arrested and punished for their crimes.
WikiLeaksWikiLeaks is a controversial platform known for publishing classified and secret documents from anonymous sources, gaining international attention for its major leaks. While it has played a significant role in exposing hidden information, its release of selectively edited materials has also contributed to the spread of conspiracy theories related to American and Russian politics.
ZOGZOG (Zionist Occupation Government): A conspiracy theory claiming that Jewish people secretly control a country, particularly the United States, while the term itself is antisemitic and unfounded.
Read more

The concept of a “confirmation loop” in psychology is a critical element to understand in the contexts of media literacy, disinformation, and political ideologies. It operates on the basic human tendency to seek out, interpret, favor, and recall information in a way that confirms one’s preexisting beliefs or hypotheses, known as confirmation bias. This bias is a type of cognitive bias and a systematic error of inductive reasoning that affects the decisions and judgments that people make.

Understanding the confirmation loop

A confirmation loop occurs when confirmation bias is reinforced in a cyclical manner, often exacerbated by the selective exposure to information that aligns with one’s existing beliefs. In the digital age, this is particularly prevalent due to the echo chambers created by online social networks and personalized content algorithms.

These technologies tend to present us with information that aligns with our existing views, thus creating a loop where our beliefs are constantly confirmed, and alternative viewpoints are rarely encountered. This can solidify and deepen convictions, making individuals more susceptible to disinformation and conspiracy theories, and less tolerant of opposing viewpoints.

Media literacy and disinformation

Media literacy is the ability to identify different types of media and understand the messages they’re sending. It’s crucial in breaking the confirmation loop as it involves critically evaluating sources of information, their purposes, and their impacts on our thoughts and beliefs.

With the rise of digital media, individuals are bombarded with an overwhelming amount of information, making it challenging to distinguish between credible information and disinformation. It is paramount to find your own set of credible sources, and verify the ethics and integrity of new sources you come across.

Disinformation, or false information deliberately spread to deceive people, thrives in an environment where confirmation loops are strong. Individuals trapped in confirmation loops are more likely to accept information that aligns with their preexisting beliefs without scrutinizing its credibility. This makes disinformation a powerful tool in manipulating public opinion, especially in politically charged environments.

Political ideologies

The impact of confirmation loops on political ideologies cannot be overstated. Political beliefs are deeply held and can significantly influence how information is perceived and processed.

When individuals only consume media that aligns with their political beliefs, they’re in a confirmation loop that can reinforce partisan views and deepen divides. This is particularly concerning in democratic societies where informed and diverse opinions are essential for healthy political discourse.

Operation of the confirmation loop

The operation of the confirmation loop can be seen in various everyday situations. For instance, a person might exclusively watch news channels that reflect their political leanings, follow like-minded individuals on social media, and participate in online forums that share their viewpoints.

Algorithms on many platforms like Facebook and Twitter (X) detect these preferences and continue to feed similar content, thus reinforcing the loop. Over time, this can result in a narrowed perspective, where alternative viewpoints are not just ignored but may also be actively discredited or mocked.

Becoming more aware and breaking the loop

Becoming more aware of confirmation loops and working to break them is essential for fostering open-mindedness and reducing susceptibility to disinformation. Here are several strategies to achieve this:

  1. Diversify Information Sources: Actively seek out and engage with credible sources of information that offer differing viewpoints. This can help broaden your perspective and challenge your preconceived notions.
  2. Critical Thinking: Develop critical thinking skills to analyze and question the information you encounter. Look for evidence, check sources, and consider the purpose and potential biases behind the information.
  3. Media Literacy Education: Invest time in learning about media literacy. Understanding how media is created, its various forms, and its impact can help you navigate information more effectively.
  4. Reflect on Biases: Regularly reflect on your own biases and consider how they might be affecting your interpretation of information. Self-awareness is a crucial step in mitigating the impact of confirmation loops.
  5. Engage in Constructive Dialogue: Engage in respectful and constructive dialogues with individuals who hold different viewpoints. This can expose you to new perspectives and reduce the polarization exacerbated by confirmation loops.

The confirmation loop is a powerful psychological phenomenon that plays a significant role in shaping our beliefs and perceptions, especially in the context of media literacy, disinformation, and political ideologies. By understanding how it operates and actively working to mitigate its effects, individuals can become more informed, open-minded, and resilient against disinformation.

The path toward breaking the confirmation loop involves a conscious effort to engage with diverse information sources, practice critical thinking, and foster an environment of open and respectful discourse.

Read more

Fact-checking is a critical process used in journalism to verify the factual accuracy of information before it’s published or broadcast. This practice is key to maintaining the credibility and ethical standards of journalism and media as reliable information sources. It involves checking statements, claims, and data in various media forms for accuracy and context.

Ethical standards in fact-checking

The ethical backbone of fact-checking lies in journalistic integrity, emphasizing accuracy, fairness, and impartiality. Accuracy ensures information is cross-checked with credible sources. Fairness mandates balanced presentation, and impartiality requires fact-checkers to remain as unbiased in their evaluations as humanly possible.

To evaluate a media source’s credibility, look for a masthead, mission statement, about page, or ethics statement that explains the publication’s approach to journalism. Without a stated commitment to journalistic ethics and standards, it’s entirely possible the website or outlet is publishing opinion and/or unverified claims.

Fact-checking in the U.S.: A historical perspective

Fact-checking in the U.S. has evolved alongside journalism. The rise of investigative journalism in the early 20th century highlighted the need for thorough research and factual accuracy. However, recent developments in digital and social media have introduced significant challenges.

Challenges from disinformation and propaganda

The digital era has seen an explosion of disinformation and propaganda, particularly on social media. ‘Fake news‘, a term now synonymous with fabricated or distorted stories, poses a significant hurdle for fact-checkers. The difficulty lies not only in the volume of information but also in the sophisticated methods used to spread falsehoods, such as deepfakes and doctored media.

Bias and trust issues in fact-checking

The subjectivity of fact-checkers has been scrutinized, with some suggesting that personal or organizational biases might influence their work. This perception has led to a trust deficit in certain circles, where fact-checking itself is viewed as potentially politically or ideologically motivated.

Despite challenges, fact-checking remains crucial for journalism. Future efforts may involve leveraging technology like AI for assistance, though human judgment is still essential. The ongoing battle against disinformation will require innovation, collaboration with tech platforms, transparency in the fact-checking process, and public education in media literacy.

Fact-checking stands as a vital element of journalistic integrity and a bulwark against disinformation and propaganda. In the U.S., and globally, the commitment to factual accuracy is fundamental for a functioning democracy and an informed society. Upholding these standards helps protect the credibility of the media and trusted authorities, and supports the fundamental role of journalism in maintaining an informed public and a healthy democracy.

Read more

The concept of cherry-picking refers to the practice of selectively choosing data or facts that support one’s argument while ignoring those that may contradict it. This method is widely recognized not just as a logical fallacy but also as a technique commonly employed in the dissemination of disinformation. Cherry-picking can significantly impact the way information is understood and can influence political ideology, public opinion, and policy making.

Cherry-picking and disinformation

Disinformation, broadly defined, is false or misleading information that is spread deliberately, often to deceive or mislead the public. Cherry-picking plays a crucial role in the creation and propagation of disinformation.

By focusing only on certain pieces of evidence while excluding others, individuals or entities can create a skewed or entirely false narrative. This manipulation of facts is particularly effective because the information presented can be entirely true in isolation, making the deceit harder to detect. In the realm of disinformation, cherry-picking is a tool to shape perceptions, create false equivalencies, and undermine credible sources of information.

The role of cherry-picking in political ideology

Political ideologies are comprehensive sets of ethical ideals, principles, doctrines, myths, or symbols of a social movement, institution, class, or large group that explains how society should work. Cherry-picking can significantly influence political ideologies by providing a biased view of facts that aligns with specific beliefs or policies.

This biased information can reinforce existing beliefs, creating echo chambers where individuals are exposed only to viewpoints similar to their own. The practice can deepen political divisions, making it more challenging for individuals with differing viewpoints to find common ground or engage in constructive dialogue.

Counteracting cherry-picking

Identifying and countering cherry-picking requires a critical approach to information consumption and sharing. Here are several strategies:

  1. Diversify Information Sources: One of the most effective ways to recognize cherry-picking is by consuming information from a wide range of sources. This diversity of trustworthy sources helps in comparing different viewpoints and identifying when certain facts are being omitted or overly emphasized.
  2. Fact-Checking and Research: Before accepting or sharing information, it’s essential to verify the facts. Use reputable fact-checking organizations and consult multiple sources to get a fuller picture of the issue at hand.
  3. Critical Thinking: Develop the habit of critically assessing the information you come across. Ask yourself whether the evidence supports the conclusion, what might be missing, and whether the sources are credible.
  4. Educate About Logical Fallacies: Understanding and educating others about logical fallacies, like cherry-picking, can help people recognize when they’re being manipulated. This knowledge can foster healthier public discourse and empower individuals to demand more from their information sources.
  5. Promote Media Literacy: Advocating for media literacy education can equip people with the skills needed to critically evaluate information sources, understand media messages, and recognize bias and manipulation, including cherry-picking.
  6. Encourage Open Dialogue: Encouraging open, respectful dialogue between individuals with differing viewpoints can help combat the effects of cherry-picking. By engaging in conversations that consider multiple perspectives, individuals can bridge the gap between divergent ideologies and find common ground.
  7. Support Transparent Reporting: Advocating for and supporting media outlets that prioritize transparency, accountability, and comprehensive reporting can help reduce the impact of cherry-picking. Encourage media consumers to support organizations that make their sources and methodologies clear.

Cherry-picking is a powerful tool in the dissemination of disinformation and in shaping political ideologies. Its ability to subtly manipulate perceptions makes it a significant challenge to open, informed public discourse.

By promoting critical thinking, media literacy, and the consumption of a diverse range of information, individuals can become more adept at identifying and countering cherry-picked information. The fight against disinformation and the promotion of a well-informed public require vigilance, education, and a commitment to truth and transparency.

Read more

Microtargeting is a marketing and political strategy that leverages data analytics to deliver customized messages to specific groups within a larger population. This approach has become increasingly prevalent in the realms of digital media and advertising, and its influence on political campaigns has grown significantly.

Understanding microtargeting

Microtargeting begins with the collection and analysis of vast amounts of data about individuals. This data can include demographics (age, gender, income), psychographics (interests, habits, values), and behaviors (purchase history, online activity). By analyzing this data, organizations can identify small, specific groups of people who share common characteristics or interests. The next step involves crafting tailored messages that resonate with these groups, significantly increasing the likelihood of engagement compared to broad, one-size-fits-all communications.

Microtargeting and digital media

Digital media platforms, with their treasure troves of user data, have become the primary arenas for microtargeting. Social media networks, search engines, and websites collect extensive information on user behavior, preferences, and interactions. This data enables advertisers and organizations to identify and segment their audiences with remarkable precision.

Microtargeting, by Midjourney

Digital platforms offer sophisticated tools that allow for the delivery of customized content directly to individuals or narrowly defined groups, ensuring that the message is relevant and appealing to each recipient. The interactive nature of digital media also provides immediate feedback, allowing for the refinement of targeting strategies in real time.

Application in advertising

In the advertising domain, microtargeting has revolutionized how brands connect with consumers. Rather than casting a wide net with generic advertisements, companies can now send personalized messages that speak directly to the needs and desires of their target audience. This approach can improve the effectiveness of advertising campaigns — but comes with a tradeoff in terms of user data privacy.

Microtargeted ads can appear on social media feeds, as search engine results, within mobile apps, or as personalized email campaigns, making them a versatile tool for marketers. Thanks to growing awareness of the data privacy implications — including the passage of regulations including the GDPR, CCPA, DMA and others — users are beginning to have more control over what data is collected about them and how it is used.

Expanding role in political campaigns

The impact of microtargeting reaches its zenith in the realm of political campaigns. Political parties and candidates use microtargeting to understand voter preferences, concerns, and motivations at an unprecedented level of detail. This intelligence allows campaigns to tailor their communications, focusing on issues that resonate with specific voter segments.

For example, a campaign might send messages about environmental policies to voters identified as being concerned about climate change, while emphasizing tax reform to those worried about economic issues. A campaign might target swing voters with characteristics that match their party’s more consistent voting base, hoping to influence their decision to vote for the “right” candidate.

Microtargeting in politics also extends to voter mobilization efforts. Campaigns can identify individuals who are supportive but historically less likely to vote and target them with messages designed to motivate them to get to the polls. Similarly, microtargeting can help in shaping campaign strategies, determining where to hold rallies, whom to engage for endorsements, and what issues to highlight in speeches.

Ethical considerations and challenges

The rise of microtargeting raises significant ethical and moral questions and challenges. Concerns about privacy, data protection, and the potential for manipulation are at the forefront. The use of personal information for targeting purposes has sparked debates on the need for stricter regulation and transparency. In politics, there’s apprehension that microtargeting might deepen societal divisions by enabling campaigns to exploit sensitive issues or disseminate misleading information — or even disinformation — to susceptible groups.

Furthermore, the effectiveness of microtargeting in influencing consumer behavior and voter decisions has led to calls for more responsible use of data analytics. Critics argue for the development of ethical guidelines that balance the benefits of personalized communication with the imperative to protect individual privacy and maintain democratic integrity.

Microtargeting represents a significant evolution in the way organizations communicate with individuals, driven by advances in data analytics and digital technology. Its application across advertising and, more notably, political campaigns, has demonstrated its power to influence behavior and decision-making.

However, as microtargeting continues to evolve, it will be crucial for society to address the ethical and regulatory challenges it presents. Ensuring transparency, protecting privacy, and promoting responsible use will be essential in harnessing the benefits of microtargeting while mitigating its potential risks. As we move forward, the dialogue between technology, ethics, and regulation will shape the future of microtargeting in our increasingly digital world.

Read more

Power is the water we swim in; the air we breathe. So pervasive we forget, ignore, or never even become fully aware of it.

It’s the ability to offload risk and responsibility if things go wrong, while accruing credit and authority if things go right. Blame flows downward, and reward flows upwards.

Where does power derive from? In a just world, it comes from knowledge, mastery, excellence, and merit. In an unjust world, it comes from itself — power qua power.

What is power?

This is a work in progress:

  • Power is a “strong force” acting on us everywhere, in every interaction.
  • Power is having the luxury of meeting a lack of resistance to one’s directives, desires, or assumptions. Friction tends to “get out of the way” of those in power.
  • It’s the ability to get what you want to have done, done.
  • The ability to influence others to your way of thinking.
  • Having enough gravity to warp reality — a reality-distortion field. Controlling reality itself.
  • It is the ability to hack your inner People Pleaser.
  • The ability to ignore feedback loops.
  • The ability to accrue accolades and offload risk: Offloading risk and responsibility
  • The ability to appear confident without any supporting factual basis.
  • The ability to make claims and have them taken seriously, at face value.
  • The ability to control reality.
  • The ability to say something enough times that, despite insufficient evidence, you can make it become β€œtrue.”
  • The ability to make other people feel crazy about your own lack of memory.
  • The expectation that others will read your mind and give you what you want before you even have to say it.
  • The assumption that anything anyone else has to do takes them only 5 minutes and can be done by anyone without special expertise. The assumption that things you have to do are arduous, time-consuming, and require excessive skill.
  • Now, in the age of instant 24/7 media orgy gratification, we are sitting ducks β€” waiting to be hijacked every day over some deliberately manufactured crisis.
  • These days everyone seems to want Power without responsibility.
  • Power is a FEELING.
  • It flows from social dominance games.
  • It’s the raison d’Γͺtre of those who believe in a hierarchical worldview.
  • It’s being able to export DISCOMFORT to others.
  • It’s when you keep the fun parts & β€œdelegate” the gruntwork to underlings.

About power:

  • Who is the master and who is the servant?
  • Who gets to do what they want, and who has to do what others say?
  • Tricks of power:
    • vagueness
    • citing invisible data you aren’t able to see
    • controlling the agenda
    • being “too busy” to have time for you
    • unreasonable criticism
    • refusing to be satisfied by any reasonable means
  • Interrelation of power and control
    • To avoid destroying ourselves, men will have to learn how to give up control and be more gentle; self-reflexive. Yet I don’t see this happening — instead they prefer their power games and petty squabbles
    • Privilege is about being groomed to control
  • Having material needs met is not enough — American culture trains us to believe that we need psychological domination to “be happy.” We confuse the rush of power with happiness — and for some, it’s all they know.
  • Power is a story the powerful made up, and got everyone to believe.
    • Some are True Believers
    • Some just recognize it as useful
    • Others become True Believers along the way
  • Power is a very large lensing effect. It’s seen as “right” and sees itself as “right” no matter what, or how clueless
  • 1% power: the theater of friendly coercion. Get everything leveraged up just below the threshold of detectable overt tyranny.
  • The rich and powerful are so rich and powerful that their main obstacle left is Mother Nature herself taking their life away. They hate the messy organic reality of life and all its flaws and limitations.
  • Celebrity culture = denial of death. Leveraging fame as a way to live on, and have the kind of wealth and power that insulates one from many dangers
Read more

“Source amnesia” is a psychological phenomenon that occurs when an individual can remember information but cannot recall where the information came from. In the context of media and disinformation, source amnesia plays a crucial role in how misinformation spreads and becomes entrenched in people’s beliefs. This overview will delve into the nature of source amnesia, its implications for media consumption, and strategies for addressing it.

Understanding source amnesia

Source amnesia is part of the broader category of memory errors where the content of a memory is dissociated from its source. This dissociation can lead to a situation where individuals accept information as true without remembering or critically evaluating where they learned it. The human brain tends to remember facts or narratives more readily than it does the context or source of those facts, especially if the information aligns with pre-existing beliefs or emotions. This bias can lead to the uncritical acceptance of misinformation if the original source was unreliable but the content is memorable.

Source amnesia in the media landscape

The role of source amnesia in media consumption has become increasingly significant in the digital age. The vast amount of information available online and the speed at which it spreads mean that individuals are often exposed to news, facts, and narratives from myriad sources, many of which might be dubious or outright false. Social media platforms, in particular, exacerbate this problem by presenting information in a context where source credibility is often obscured or secondary to engagement.

Disinformation campaigns deliberately exploit source amnesia. They spread misleading or false information, knowing that once the information is detached from its dubious origins, it is more likely to be believed and shared. This effect is amplified by confirmation bias, where individuals are more likely to remember and agree with information that confirms their pre-existing beliefs, regardless of the source’s credibility.

Implications of source amnesia

The implications of source amnesia in the context of media and disinformation are profound. It can lead to the widespread acceptance of false narratives, undermining public discourse and trust in legitimate information sources. Elections, public health initiatives, and social cohesion can be adversely affected when disinformation is accepted as truth due to source amnesia.

The phenomenon also poses challenges for fact-checkers and educators, as debunking misinformation requires not just presenting the facts but also overcoming the emotional resonance and simplicity of the original, misleading narratives.

Addressing source amnesia

Combating source amnesia and its implications for disinformation requires a multi-pronged approach, focusing on education, media literacy, and critical thinking. Here are some strategies:

  1. Media Literacy Education: Teaching people to critically evaluate sources and the context of the information they consume can help mitigate source amnesia. This includes understanding the bias and reliability of different media outlets, recognizing the hallmarks of credible journalism, and checking multiple sources before accepting information as true.
  2. Critical Thinking Skills: Encouraging critical thinking can help individuals question the information they encounter, making them less likely to accept it uncritically. This involves skepticism about information that aligns too neatly with pre-existing beliefs or seems designed to elicit an emotional response.
  3. Source Citing: Encouraging the practice of citing sources in media reports and social media posts can help readers trace the origin of information. This practice can aid in evaluating the credibility of the information and combat the spread of disinformation.
  4. Digital Platforms’ Responsibility: Social media platforms and search engines play a crucial role in addressing source amnesia by improving algorithms to prioritize reliable sources and by providing clear indicators of source credibility. These platforms can also implement features that encourage users to evaluate the source before sharing information.
  5. Public Awareness Campaigns: Governments and NGOs can run public awareness campaigns highlighting the importance of source evaluation. These campaigns can include guidelines for identifying credible sources and the risks of spreading unverified information.

Source amnesia is a significant challenge in the fight against disinformation, making it easy for false narratives to spread unchecked. By understanding this phenomenon and implementing strategies to address it, society can better safeguard against the corrosive effects of misinformation.

It requires a concerted effort from individuals, educators, media outlets, and digital platforms to ensure that the public remains informed and critical in their consumption of information. This collective action can foster a more informed public, resilient against the pitfalls of source amnesia and the spread of disinformation.

Read more

The backfire effect is a cognitive phenomenon that occurs when individuals are presented with information that contradicts their existing beliefs, leading them not only to reject the challenging information but also to further entrench themselves in their original beliefs.

This effect is counterintuitive, as one might expect that presenting factual information would correct misconceptions. However, due to various psychological mechanisms, the opposite can occur, complicating efforts to counter misinformation, disinformation, and the spread of conspiracy theories.

Origin and mechanism

The term “backfire effect” was popularized by researchers Brendan Nyhan and Jason Reifler, who in 2010 conducted studies demonstrating that corrections to false political information could actually deepen an individual’s commitment to their initial misconception. This effect is thought to stem from a combination of cognitive dissonance (the discomfort experienced when holding two conflicting beliefs) and identity-protective cognition (wherein individuals process information in a way that protects their sense of identity and group belonging).

Relation to media, disinformation, echo chambers, and media bubbles

In the context of media and disinformation, the backfire effect is particularly relevant. The proliferation of digital media platforms has made it easier than ever for individuals to encounter information that contradicts their beliefs — but paradoxically, it has also made it easier for them to insulate themselves in echo chambers and media bubblesβ€”environments where their existing beliefs are constantly reinforced and rarely challenged.

Echo chambers refer to situations where individuals are exposed only to opinions and information that reinforce their existing beliefs, limiting their exposure to diverse perspectives. Media bubbles are similar, often facilitated by algorithms on social media platforms that curate content to match users’ interests and past behaviors, inadvertently reinforcing their existing beliefs and psychological biases.

Disinformation campaigns can exploit these dynamics by deliberately spreading misleading or false information, knowing that it is likely to be uncritically accepted and amplified within certain echo chambers or media bubbles. This can exacerbate the backfire effect, as attempts to correct the misinformation can lead to individuals further entrenching themselves in the false beliefs, especially if those beliefs are tied to their identity or worldview.

How the backfire effect happens

The backfire effect happens through a few key psychological processes:

  1. Cognitive Dissonance: When confronted with evidence that contradicts their beliefs, individuals experience discomfort. To alleviate this discomfort, they often reject the new information in favor of their pre-existing beliefs.
  2. Confirmation Bias: Individuals tend to favor information that confirms their existing beliefs and disregard information that contradicts them. This tendency towards bias can lead them to misinterpret or dismiss corrective information.
  3. Identity Defense: For many, beliefs are tied to their identity and social groups. Challenging these beliefs can feel like a personal attack, leading individuals to double down on their beliefs as a form of identity defense.

Prevention and mitigation

Preventing the backfire effect and its impact on public discourse and belief systems requires a multifaceted approach:

  1. Promote Media Literacy: Educating the public on how to critically evaluate sources and understand the mechanisms behind the spread of misinformation can empower individuals to think critically and assess the information they encounter.
  2. Encourage Exposure to Diverse Viewpoints: Breaking out of media bubbles and echo chambers by intentionally seeking out and engaging with a variety of perspectives can reduce the likelihood of the backfire effect by making conflicting information less threatening and more normal.
  3. Emphasize Shared Values: Framing challenging information in the context of shared values or goals can make it less threatening to an individual’s identity, reducing the defensive reaction.
  4. Use Fact-Checking and Corrections Carefully: Presenting corrections in a way that is non-confrontational and, when possible, aligns with the individual’s worldview or values can make the correction more acceptable. Visual aids and narratives that resonate with the individual’s experiences or beliefs can also be more effective than plain factual corrections.
  5. Foster Open Dialogue: Encouraging open, respectful conversations about contentious issues can help to humanize opposing viewpoints and reduce the instinctive defensive reactions to conflicting information.

The backfire effect presents a significant challenge in the fight against misinformation and disinformation, particularly in the context of digital media. Understanding the psychological underpinnings of this effect is crucial for developing strategies to promote a more informed and less polarized public discourse. By fostering critical thinking, encouraging exposure to diverse viewpoints, and promoting respectful dialogue, it may be possible to mitigate the impact of the backfire effect and create a healthier information ecosystem.

Read more

A “filter bubble” is a concept in the realm of digital publishing, media, and web technology, particularly significant in understanding the dynamics of disinformation and political polarization. At its core, a filter bubble is a state of intellectual isolation that can occur when algorithms selectively guess what information a user would like to see based on past behavior and preferences. This concept is crucial in the digital age, where much of our information comes from the internet and online sources.

Origins and mechanics

The term was popularized by internet activist Eli Pariser around 2011. It describes how personalization algorithms in search engines and social media platforms can isolate users in cultural or ideological bubbles. These algorithms, driven by AI and machine learning, curate content – be it news, search results, or social media posts – based on individual user preferences, search histories, and previous interactions.

filter bubble, by DALL-E 3

The intended purpose is to enhance user experience by providing relevant and tailored content. However, this leads to a situation where users are less likely to encounter information that challenges or broadens their worldview.

Filter bubbles in the context of disinformation

In the sphere of media and information, filter bubbles can exacerbate the spread of disinformation and propaganda. When users are consistently exposed to a certain type of content, especially if it’s sensational or aligns with their pre-existing beliefs, they become more susceptible to misinformation. This effect is compounded on platforms where sensational content is more likely to be shared and become viral, often irrespective of its accuracy.

Disinformation campaigns, aware of these dynamics, often exploit filter bubbles to spread misleading narratives. By tailoring content to specific groups, they can effectively reinforce existing beliefs or sow discord, making it a significant challenge in the fight against fake news and propaganda.

Impact on political beliefs and US politics

The role of filter bubbles in shaping political beliefs is profound, particularly in the polarized landscape of recent US politics. These bubbles create echo chambers where one-sided political views are amplified without exposure to opposing viewpoints. This can intensify partisanship, as individuals within these bubbles are more likely to develop extreme views and less likely to understand or empathize with the other side.

Recent years in the US have seen a stark divide in political beliefs, influenced heavily by the media sources individuals consume. For instance, the right and left wings of the political spectrum often inhabit separate media ecosystems, with their own preferred news sources and social media platforms. This separation contributes to a lack of shared reality, where even basic facts can be subject to dispute, complicating political discourse and decision-making.

Filter bubbles in elections and political campaigns

Political campaigns have increasingly utilized data analytics and targeted advertising to reach potential voters within these filter bubbles. While this can be an effective campaign strategy, it also means that voters receive highly personalized messages that can reinforce their existing beliefs and psychological biases, rather than presenting a diverse range of perspectives.

Breaking out of filter bubbles

Addressing the challenges posed by filter bubbles involves both individual and systemic actions. On the individual level, it requires awareness and a conscious effort to seek out diverse sources of information. On a systemic level, it calls for responsibility from tech companies to modify their algorithms to expose users to a broader range of content and viewpoints.

Filter bubbles play a significant role in the dissemination and reception of information in today’s digital age. Their impact on political beliefs and the democratic process — indeed, on democracy itself — in the United States cannot be overstated. Understanding and mitigating the effects of filter bubbles is crucial in fostering a well-informed public, capable of critical thinking and engaging in healthy democratic discourse.

Read more

Dark money refers to political spending by organizations that are not required to disclose their donors or how much money they spend. This allows wealthy individuals and special interest groups to secretly fund political campaigns and influence elections without transparency or accountability.

The term “dark money” gained prominence after the 2010 Supreme Court decision in Citizens United v. Federal Election Commission. In that case, the Court ruled that corporations and unions could spend unlimited amounts of money on political campaigns, as long as the spending was not coordinated with a candidate’s campaign.

This decision opened the floodgates for massive amounts of dark money to flow into political campaigns, often with no way for the public to know who was behind it. Dark money can come from a variety of sources, including wealthy individuals, corporations, trade associations, and non-profit organizations.

Hidden donors

Non-profit organizations, in particular, have become a popular way for donors to hide their political contributions. These organizations can operate under section 501(c)(4) of the tax code, which allows them to engage in some political activity as long as it is not their primary purpose. These groups are not required to disclose their donors, which means that wealthy individuals and corporations can funnel unlimited amounts of money into political campaigns without anyone knowing where the money came from.

Another way that dark money is used in politics is through “shell corporations.” These are companies that exist solely to make political donations and are often set up specifically to hide the identity of the true donor. For example, a wealthy individual could set up a shell corporation and then use that corporation to donate to a political campaign. Because the corporation is listed as the donor, the individual’s name does not appear on any public disclosure forms.

The money can be used to run ads, create content and propaganda, fund opposition research, pay armadas of PR people, send direct mail, lobby Congress, hire social media influencers, and many other powerful marketing strategies to reach and court voters.

These practices erode at the foundations of representative democracy, and the kind of government the Founders had in mind. One is free to vote for who one wishes, and to advocate for who ones wishes to hold power, but one has no Constitutional right to anonymity when doing so. It infringes on others peoples’ rights as well — the right to representative and transparent government.

Dark money impact

Dark money can have a significant impact on elections and public policy. Because the source of the money is not known, candidates and elected officials may be influenced by the interests of the donors rather than the needs of their constituents. This can lead to policies that benefit wealthy donors and special interest groups rather than the broader public.

There have been some efforts to increase transparency around dark money. For example, the DISCLOSE Act, which has been introduced in Congress several times since 2010, would require organizations that spend money on political campaigns to disclose their donors (the acronym stands for “Democracy Is Strengthened by Casting Light On Spending in Elections”). However, these efforts have been met with resistance from groups that benefit from the lack of transparency — who, somewhat ironically, have been using their influence with the Republican Party to make sure the GOP opposes the bill and prevents it from passing, or even coming up for a vote at all.

In addition to the impact on elections and policy, dark money can also undermine public trust in government. When voters feel that their voices are being drowned out by the interests of wealthy donors and special interest groups, they may become disillusioned with the political process and less likely to participate.

Overall, dark money is a significant problem in American politics. The lack of transparency and accountability around political spending allows wealthy individuals and special interest groups to wield undue influence over elections and policy. To address this problem, it will be important to increase transparency around political spending and reduce the influence of money in politics.

Dark Money: Learn more

Read more

The term “hoax” is derived from “hocus,” a term that has been in use since the late 18th century. It originally referred to a trick or deception, often of a playful or harmless nature. The essence of a hoax was its capacity to deceive, typically for entertainment or to prove a point without malicious intent. Over time, the scope and implications of a hoax have broadened significantly. What was once a term denoting jest or trickery has morphed into a label for deliberate falsehoods intended to mislead or manipulate public perception.

From playful deception to malicious misinformation

As society entered the age of mass communication, the potential reach and impact of hoaxes expanded dramatically. The advent of newspapers, radio, television, and eventually the internet and social media platforms, transformed the way informationβ€”and misinformationβ€”circulated. Hoaxes began to be used not just for amusement but for more nefarious purposes, including political manipulation, financial fraud, and social engineering. The line between a harmless prank and damaging disinformation and misinformation became increasingly blurred.

The political weaponization of “hoax”

In the contemporary political landscape, particularly within the US, the term “hoax” has been co-opted as a tool for disinformation and propaganda. This strategic appropriation has been most visible among certain factions of the right-wing, where it is used to discredit damaging information, undermine factual reporting, and challenge the legitimacy of institutional findings or scientific consensus. This application of “hoax” serves multiple purposes: it seeks to sow doubt, rally political bases, and divert attention from substantive issues.

the politicization of hoaxes, via fake scandals that tie up the media unwittingly in bullshit for years, by DALL-E 3

This tactic involves labeling genuine concerns, credible investigations, and verified facts as “hoaxes” to delegitimize opponents and minimize the impact of damaging revelations. It is a form of gaslighting on a mass scale, where the goal is not just to deny wrongdoing but to erode the very foundations of truth and consensus. By branding something as a “hoax,” these actors attempt to preemptively dismiss any criticism or negative information, regardless of its veracity.

Case Studies: The “Hoax” label in action

High-profile instances of this strategy include the dismissal of climate change data, the denial of election results, and the rejection of public health advice during the COVID-19 pandemic. In each case, the term “hoax” has been employed not as a description of a specific act of deception, but as a blanket term intended to cast doubt on the legitimacy of scientifically or empirically supported conclusions. This usage represents a significant departure from the term’s origins, emphasizing denial and division over dialogue and discovery.

The impact on public discourse and trust

The strategic labeling of inconvenient truths as “hoaxes” has profound implications for public discourse and trust in institutions. It creates an environment where facts are fungible, and truth is contingent on political allegiance rather than empirical evidence. This erosion of shared reality undermines democratic processes, hampers effective governance, and polarizes society.

Moreover, the frequent use of “hoax” in political discourse dilutes the term’s meaning and impact, making it more difficult to identify and respond to genuine instances of deception. When everything can be dismissed as a hoax, the capacity for critical engagement and informed decision-making is significantly compromised.

Moving Forward: Navigating a “post-hoax” landscape

The challenge moving forward is to reclaim the narrative space that has been distorted by the misuse of “hoax” and similar terms. This involves promoting media literacy, encouraging critical thinking, and fostering a public culture that values truth and accountability over partisanship. It also requires the media, educators, and public figures to be vigilant in their language, carefully distinguishing between genuine skepticism and disingenuous dismissal.

The evolution of “hoax” from a term denoting playful deception to a tool for political disinformation reflects broader shifts in how information, truth, and reality are contested in the public sphere. Understanding this transformation is crucial for navigating the complexities of the modern informational landscape and for fostering a more informed, resilient, and cohesive society.

Read more

Malware, short for “malicious software,” is any software intentionally designed to cause damage to a computer, server, client, or computer network. This cybersecurity threat encompasses a variety of software types, including viruses, worms, trojan horses, ransomware, spyware, adware, and more. Each type has a different method of infection and damage.

Who uses malware and what for

Malware is utilized by a wide range of actors, from amateur hackers to sophisticated cybercriminals, and even nation-states. The motives can vary greatly:

  • Cybercriminals often deploy malware to steal personal, financial, or business information, which can be used for financial gain through fraud or direct theft.
  • Hacktivists use malware to disrupt services or bring attention to political or social causes.
  • Nation-states and state-sponsored actors might deploy sophisticated malware for espionage and intelligence, to gain strategic advantage, sabotage, or influence geopolitical dynamics.
Malware, illustrated by DALL-E 3

Role in disinformation and geopolitical espionage

Malware plays a significant role in disinformation campaigns and geopolitical espionage. State-sponsored actors might use malware to infiltrate the networks of other nations, steal sensitive information (hacked emails perhaps?), and manipulate or disrupt critical infrastructure. In terms of disinformation, malware can be used to gain unauthorized access to media outlets or social media accounts, spreading false information to influence public opinion or destabilize political situations.

Preventing malware

Preventing malware involves multiple layers of security measures:

  • Educate Users: The first line of defense is often the users themselves. Educating them about the dangers of phishing emails, not to click on suspicious links, and the importance of not downloading or opening files from unknown sources can significantly reduce the risk of malware infections.
  • Regular Software Updates: Keeping all software up to date, including operating systems and antivirus programs, can protect against known vulnerabilities that malware exploits.
  • Use Antivirus Software: A robust antivirus program can detect and remove many types of malware. Regular scans and real-time protection features are crucial.
  • Firewalls: Both hardware and software firewalls can block unauthorized access to your network, which can help prevent malware from spreading.
  • Backups: Regularly backing up important data ensures that, in the event of a malware attack, the lost data can be recovered without paying ransoms or losing critical information.

Famous malware incidents in foreign affairs

Several high-profile malware incidents have had significant implications in the realm of foreign affairs:

  • Stuxnet: Discovered in 2010, Stuxnet was a highly sophisticated worm that targeted supervisory control and data acquisition (SCADA) systems and was believed to be designed to damage Iran’s nuclear program. It is widely thought to be a cyberweapon developed by the United States and Israel, though neither has confirmed involvement.
  • WannaCry: In May 2017, the WannaCry ransomware attack affected over 200,000 computers across 150 countries, with the UK’s National Health Service, Spain’s TelefΓ³nica, FedEx, and Deutsche Bahn among those impacted. The attack exploited a vulnerability in Microsoft Windows, and North Korea was widely blamed for the attack.
  • NotPetya: Initially thought to be ransomware, NotPetya emerged in 2017 and caused extensive damage, particularly in Ukraine. It later spread globally, affecting businesses and causing billions of dollars in damages. It is believed to have been a state-sponsored attack originating from Russia, designed as a geopolitical tool under the guise of ransomware.
  • SolarWinds: Uncovered in December 2020, the SolarWinds hack was a sophisticated supply chain attack that compromised the Orion software suite used by numerous US government agencies and thousands of private companies. It allowed the attackers, believed to be Russian state-sponsored, to spy on the internal communications of affected organizations for months.

In conclusion, malware is a versatile and dangerous tool in the hands of cybercriminals and state actors alike, used for everything from financial theft to sophisticated geopolitical maneuvers. The proliferation of malware in global affairs underscores the need for robust cybersecurity practices at all levels, from individual users to national governments. Awareness, education, and the implementation of comprehensive security measures are key to defending against the threats posed by malware.

Read more

Wealth Cult -- rich men behaving badly, by Midjourney

A network of exceedingly wealthy individuals and organizations have channeled their vast fortunes into influencing American politics, policy, and public opinion — they’ve formed a wealth cult. And they’ve leveraged that cult and its considerable fortune to influence and in many ways dramatically transform American politics.

The term “dark money” refers to political spending meant to influence the decision-making and critical thinking of the public and lawmakers where the source of the money is not disclosed. This lack of transparency makes it challenging to trace the influence back to its origins, hence the term “dark.”

And, it is dark indeed.

Wealth cult anchors the trench coat

The Wealth Cult is one of 3 primary groups or clusters supporting the right-wing and generally, the Republican Party. It anchors the trench coat by funding the 2 cults above it: the Christian Cult, and the White Cult.

Its story is stealthy and significant.

A bunch of billionaires toast themselves to themselves, by Midjourney

The wealth cult has funded disinformation campaigns, the spread of conspiracy theories, created fake social movements through astroturfing, enabled violent extremists to attack their country’s capitol, cruelly deprived vulnerable people (especially immigrants, poor people, and women) of the kind of state aid granted generously throughout the developed world, bribed regulators, rigged elections, crashed economies, and on and on in service of their extremist free market ideology beliefs.

They believe in “makers and takers,” or Mudsill Theory, as it was once called by pedophile and racist Senator and slavery enthusiast James Henry Hammond. Some people were born to serve others, they say. Hierarchies are natural, they claim. Wealthy men should make all the decisions — because that’s what’s best for everyone, they say in paternalistic tones.

I don’t buy it. I believe all men are created equal. So did a certain Founder of our country.

Continue reading Wealth Cult: The oligarchs influencing American politics from the shadows
Read more

Below is a list of the covert gang of folks trying to take down the US government — the anti-government oligarchs who think they run the place. The Koch network of megarich political operatives has been anointing itself the true (shadowy) leaders of American politics for several decades.

Spearheaded by Charles Koch, the billionaire fossil fuel magnate who inherited his father Fred Koch’s oil business, the highly active and secretive Koch network — aka the “Kochtopus” — features a sprawling network of donors, think tanks, non-profits, political operatives, PR hacks, and other fellow travelers who have come to believe that democracy is incompatible with their ability to amass infinite amounts of wealth.

Despite their obvious and profligate success as some of the world’s richest people, they whine that the system of US government is very unfair to them and their ability to do whatever they want to keep making a buck — the environment, the people, and even the whole planet be damned. Part of an ever larger wealth cult of individuals spending unprecedented amounts of cash to kneecap the US government from any ability to regulate business or create a social safety net for those exploited by concentrated (and to a large extent inherited) wealth, the Koch network is the largest and most formidable group within the larger project of US oligarchy.

The Kochtopus

By 2016 the Koch network of private political groups had a paid staff of 1600 people in 35 states — a payroll larger than that of the Republican National Committee (RNC) itself. They managed a pool of funds from about 400 or so of the richest people in the United States, whose goal was to capture the government and run it according to their extremist views of economic and social policy. They found convenient alignment with the GOP, which has been the party of Big Business ever since it succeeded in first being the party of the Common Man in the 1850s and 60s.

Are we to be just a wholly-owned subsidiary of Koch Industries? Who will help stand and fight for our independence from oligarchy?

  • Philip Anschutz — Founder of Qwest Communications. Colorado oil and entertainment magnate and billionaire dubbed the world’s “greediest executive” by Fortune Magazine in 2002.
  • American Energy Alliance — Koch-funded tax-exempt nonprofit lobbying for corporate-friendly energy policies
  • American Enterprise Institute — The American Enterprise Institute (AEI) is a public policy think tank based in Washington, D.C. Established in 1938, it is one of the oldest and most influential think tanks in the United States. AEI is primarily known for its conservative and free-market-oriented policy research and advocacy.
  • Americans for Prosperity
  • Harry and Lynde Bradley — midwestern defense contractors and Koch donors
  • Michael Catanzaro
  • Cato Institute
  • Center to Protect Patient Rights — The Koch network’s fake front group for fighting against Obama‘s Affordable Care Act.
  • CGCN Group — right-wing lobbying group
  • Citizens for a Sound Economy
  • Club for Growth
  • Competitive Enterprise Institute — Right-wing think tank funded by the Kochs and other oil and gas barons
  • Continental Resources — Harold Hamm’s shale-oil company
  • Joseph Coors — Colorado beer magnate
  • Betsy and Dick DeVos — founders of the Amway MLM empire, and one of the richest families in Michigan
  • Myron Ebell — Outspoken client change denier picked to head Trump’s EPA transition team who previously worked at the Koch-funded Competitive Enterprise Institute.
  • Richard Farmer — Chairman of the Cintas Corporation in Cincinnati, the nation’s largest uniform supply company. Legal problems against him included an employee’s gruesome death thanks to violating safety laws.
  • Freedom Partners — the Koch donor group
  • Freedom School — the all-white CO private school funded by Charles Koch in the 1960s
  • FreedomWorks
  • Richard Gilliam — Head of Virginia coal mining company Cumberland Resources, and Koch network donor.
  • Harold Hamm — Oklahoma fracking king and charter member of the Koch donors’ circle, Hamm became a billionaire founding the Continental Resources shale-oil company
  • Diane Hendricks — $3.6 billion building supply company owner and Trump inaugural committee donor, and the wealthiest woman in Wisconsin.
  • Charles Koch — CEO of Koch Industries and patriarch of the Koch empire following his father and brother’s death, and estrangement from his other younger brother. Former member of the John Birch Society, a group so far to the right that even arch-conservative William F. Buckley excommunicated them from the mainstream party in the 1950s.
  • The Charles Koch Foundation
  • (David Koch) — deceased twin brother of Bill Koch and younger brother to Charles who ran a failed campaign in 1980 as the vice presidential nominee of the Libertarian Party — netting 1% of the popular vote. In 2011 he echoed spurious claims from conservative pundit Dinesh D’Souza that Obama got his “radical” political outlook from his African father.
  • The Leadership Institute
  • Michael McKenna — president of the lobbying firm MWR Strategies, whose clients include Koch Industries, picked by Trump to serve on the Department of Energy transition team
  • Rebekah Mercer — daughter of hedge fund billionaire and right-wing Koch donor Robert Mercer, she worked with Steve Bannon on several projects including Breitbart News, Cambridge Analytica, and Gab.
  • Robert Mercer — billionaire NY hedge fund manager and next largest donor after the Kochs themselves, sometimes even surpassing them
  • MWR Strategies — lobbying firm for the energy industry whose clients include Koch Industries, whose president Michael McKenna served on the Trump energy transition team
  • John M. Olin — chemical and munitions magnate and Koch donor
  • George Pearson — Former head of the Koch Foundation
  • Mike Pence — Charles Koch’s number one pick for president in 2012.
  • Mike Pompeo — former Republican Kansas Congressman who got picked first to lead the CIA, then later as Secretary of State under Trump. He was the single largest recipient of Koch money in Congress as of 2017. The Kochs had been investors and partners in Pompeo’s business ventures before he got into politics.
  • The Reason Foundation
  • Richard Mellon Scaife — heir to the Mellon banking and Gulf Oil fortunes
  • David Schnare — self-described “free-market environmentalist” on Trump’s EPA transition team
  • Marc Short — ran the Kochs’ secretive donor club, Freedom Partners, before becoming a senior advisor to vice president Mike Pence during the Trump transition
  • State Policy Network
  • The Tax Foundation
  • Tea Party

Koch Network Mind Map

This mind map shows the intersections between the Koch network and the larger network of GOP donors, reactionaries, and evil billionaires who feel entitled to control American politics via the fortunes they’ve made or acquired.

Read more