Next Article in Journal
A Bibliometric Analysis of the Scientific Outcomes of European Projects on the Digital Transformation of SMEs
Previous Article in Journal
Citation Rate Challenges for a Small Journal Indexed in Scopus and WoS—Case Study from Central Europe (Croatia), Editorial View
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

What the Neuroscience and Psychology of Magic Reveal about Misinformation

by
Robert G. Alexander
*,
Stephen L. Macknik
and
Susana Martinez-Conde
Departments of Ophthalmology, Neurology, and Physiology & Pharmacology, State University of New York Downstate Health Sciences University, Brooklyn, NY 11203, USA
*
Author to whom correspondence should be addressed.
Publications 2022, 10(4), 33; https://doi.org/10.3390/publications10040033
Submission received: 25 May 2022 / Revised: 12 August 2022 / Accepted: 25 August 2022 / Published: 23 September 2022
(This article belongs to the Special Issue Science Information, Media Misinformation, and Public Trust)

Abstract

:
When we believe misinformation, we have succumbed to an illusion: our perception or interpretation of the world does not match reality. We often trust misinformation for reasons that are unrelated to an objective, critical interpretation of the available data: Key facts go unnoticed or unreported. Overwhelming information prevents the formulation of alternative explanations. Statements become more believable every time they are repeated. Events are reframed or given “spin” to mislead audiences. In magic shows, illusionists apply similar techniques to convince spectators that false and even seemingly impossible events have happened. Yet, many magicians are “honest liars”, asking audiences to suspend their disbelief only during the performance, for the sole purpose of entertainment. Magic misdirection has been studied in the lab for over a century. Psychological research has sought to understand magic from a scientific perspective and to apply the tools of magic to the understanding of cognitive and perceptual processes. More recently, neuroscientific investigations have also explored the relationship between magic illusions and their underlying brain mechanisms. We propose that the insights gained from such studies can be applied to understanding the prevalence and success of misinformation. Here, we review some of the common factors in how people experience magic during a performance and are subject to misinformation in their daily lives. Considering these factors will be important in reducing misinformation and encouraging critical thinking in society.

1. Introduction

Discrediting false reports is crucial for reinforcing and maintaining society’s trust in science. Unfortunately, members of the public are often barraged with incomplete, out-of-date, and biased information. Perhaps worse still is the spread of “disinformation”, or misinformation that is actively disseminated with the intent to deceive others. False news, whether intentionally malicious or not, can distort people’s views about important issues [1] and encourage them to engage in risky, unproductive, or harmful behaviors—such as taking unproven medications, and overreacting or underreacting to situations (e.g., ignoring safety recommendations [2]). Beyond their impact on health and safety, false news can facilitate terrorist propaganda [3], stock market manipulations [4], and a host of other malicious actions. Why do people believe false information, and—in many cases—maintain a belief in misinformation even after a story is retracted, disputed, or debunked?
Belief in misinformation might best be thought of as an illusion: a mismatch between factual reality and what is perceived and understood. Biases in information processing—what we see as appealing, intuitive, or believable—can alter our understanding of news headlines and social media posts. For example, cognitive biases can lead to implausible claims being interpreted as plausible or even proven (e.g., the Pizzagate conspiracy theory), for reasons unrelated to a story’s content or its objective accuracy [5,6].
Here, we explore what another type of life experience, in which illusions and inaccurate beliefs are prevalent, can teach us. In a magic show, one of the performer’s main objectives is to make spectators experience seemingly impossible things. To accomplish this, magicians mislead their audiences, preventing them from accurately determining or understanding the reality of the methods behind the magical effects. Psychology researchers—and more recently, neuroscientists—have studied magic for over a century, aiming to provide a cognitive and perceptual understanding of how magic works in the minds of spectators [7,8,9,10,11,12,13,14,15,16,17]. As a result, a great deal is known about the psychology of magic, including how observers come to question whether impossible events have occurred. This review relates those fundamental findings to how and why people mistakenly believe that misinformation is accurate.

2. Attention Illusions

We cannot possibly process all the information that the world presents to us. Beyond our sensory (visual, auditory, etc.) constraints, we are largely limited to noticing only those things that we attend to. Objects or events that are new, unusual, moving, or otherwise salient strongly draw our attention in a bottom-up manner (a process called “attentional capture” in psychology). When more than one event is visible, our attention tends to follow the more salient event or the one that begins first [16,18,19,20,21]. We can also direct our attention to specific objects and events in a top-down manner, as guided by our goals and intentions. Magicians manipulate the audience’s attention using both bottom-up and top-down processes to control what is noticed [16,19,22].
Psychologist and magicians often think of attention as a “spotlight” that enhances the items or parts of a scene we are focusing on—see Figure 1. The attentional spotlight is not an accurate metaphor from a neuroscientific perspective, however: rather than making our object of interest more salient, our attentional system actively suppresses (via inhibitory neurons) how salient everything else is. That is, rather that illuminating whatever draws our interest, our attention gives us tunnel vision: it obscures all the rest [23,24,25,26,27,28,29,30,31,32,33]. Suppressed crucial information passes us by without ever entering our awareness.

2.1. Information Overwhelm

Magicians use well-practiced sleight-of-hand techniques, storytelling, gimmicks, special effects, and surprising events to overwhelm the audience’s sensory, cognitive, and even emotional processes. The magician’s hand and body movements, narration, and timing can appear innocent, but they influence the facts and events that spectators perceive—and the ones they ignore (see Figure 2). For example, by asking the audience a simple question, a magician can turn their attention inward, generating an internal dialogue that prevents observers from focusing on the actions that take place onstage [35,36]. Similarly, a magician might use humor or other emotive content to misdirect the audience [16,35]. Or a magician might perform more than one gesture at the same time (i.e., “a big motion covers a small motion”) to hide a specific action or to split the audience’s attention across several spatial locations [7,11,16].
The human attentional system gets overwhelmed more easily than we might think. Objects and events that we usually recognize without delay become much more difficult to identify in certain contexts [38,39,40,41]. For example, items or patterns can blend perceptually with adjacent ones and become unrecognizable, a phenomenon known as “visual crowding” [42,43]. Similarly, sounds can become unrecognizable when preceded and followed by nonoverlapping noise, a phenomenon known as “auditory masking” [41,44].
Many magic tricks involve multiple cards, coins, etc., changing position, and often rely on the spectators’ powerlessness to multitask—that is, their inability to track the locations of such objects and the magician’s hands simultaneously. As a result, the audience is effectively prevented from focusing on the elements that may be key to the magical effect. One example is the classic “Cups and Balls” trick, in which the magician uses sleight-of-hand to make balls appear and disappear inside of upside-down cups [7]. The magic duo Penn & Teller perform this trick with transparent (rather than opaque) cups. Thus, when Penn & Teller place balls secretly inside of the cups, they should be fully visible to audience members. Instead, the balls typically go unnoticed at first, because the spectators’ attention is directed elsewhere [11].
The failure to notice the balls as they are secretly loaded into the transparent cups is a kind of change blindness. With change blindness, observers fail to notice that something has become different from how it was before, even if the change is dramatic or if they are looking directly at the location of the changes [45,46,47,48,49]. Moreover, even when such changes are initially noticed, memory limitations can prevent viewers from remembering the specific alterations that took place over a period of time. Magicians sometimes exploit change blindness when swapping out one card for another, to later reveal that the suit or number of the card has changed. Certain objects can be placed in conspicuous locations or removed from sight—all without spectators noticing.
Social media has dramatically increased the volume of data that we must filter, process, and comprehend on an unremitting basis, which may facilitate a comparable kind of deception. Even within a very limited time window, social media displays an overwhelming amount of data to the user—in the form of image, video, and text. All of this takes place at a rapid rate on a small screen, while individuals continue to navigate their daily life, processing events around them, and responding to personal and work demands. The swiftly switching news cycle may similarly prevent viewers from noticing shifts in the framing of a story. In other words, information can come at viewers too quickly for them to stay on top of important details or even track what the latest crisis is. Forced to multitask, people are rendered unable to separate the grain from the straw (in information processing terms, the signal from the noise). As a result, they fail to notice important changes or vital elements in a story.
Visualizing data in animated or video form—showing multiple moving data points—might make people even more vulnerable to misinformation than when they view static images. That is, dynamic images can cause people to miss important changes in the data, such as when attending to one set of data points causes changes in other datapoints to go unnoticed [50]. Research has shown that observers can only track accurately the motion of up to roughly four simultaneously moving objects [51,52].

Insufficient Time to Verify Facts

Magic tricks require that the secret method remains hidden behind the magical effect, creating the illusion that there is no plausible explanation other than that the impossible has happened [53,54]. As such, magicians aim to prevent audiences from carefully evaluating alternative (i.e., veridical) explanations. One way to achieve this is by moving the show at a fast pace, so that there is no time for spectators to deeply analyze the events occurring onstage. Many magical effects would fail if audiences could stop and carefully check for the accuracy of their perceptions (in fact, viewers of magic tricks on YouTube often slow down, pause, and rewind videos with the goal of unveiling magicians’ secret methods).
Fact-checking techniques could help counter misinformation, but just like spectators of magic shows, social media users may find themselves unable to conduct in-depth examinations of the content they encounter. Thus, users tend to spend little time evaluating possible explanations, and they rarely interrupt their actions to fact-check online stories [55,56]. This is especially true for groups of people with relatively little practice in online information use, such as older, low-income adults [57,58]. But even those who are fluent in verification practices do not always take the time to authenticate messages and news received on social media. This is all the more difficult when multitasking: indeed, habitual media multitaskers can be worse at judging the accuracy of news. In the end, rather than taking the time to compare data and opinions across reliable sources, social media users tend to simply decide, in the moment, whether a statement is fact or falsehood.

2.2. Unattended Events Are Ignored

Events and viewpoints that are presented but are not the focus of attention frequently go unnoticed, a phenomenon known as “inattentional blindness” [59]. Inattentional blindness can happen even for seemingly obvious and salient events: in the classic example, observers watching a video of people passing a basketball failed to see a gorilla walk into the scene and thump its chest [59]. Even viewers who looked directly at the gorilla often did not see it, so long as their attention was directed towards other ongoing events in the video (i.e., counting the number of times the basketball was passed between the players) [60].
Inattentional blindness can be intentionally created by “misdirecting” observers, purposely diverting their attention away from certain events which would be otherwise apparent (i.e., the intruding gorilla in the example above) and towards competing concurrent events (i.e., the players passing the ball to one another) [16,19,61,62,63,64,65]. In magic, misdirection can be so effective that spectators fail to notice events even when looking directly at them [66]. The sophisticated technique by which magicians can draw spectators’ attention away from a secret action without redirecting their gaze has been described as “covert misdirection” [16].
Magicians also use narrative and storytelling (i.e., “patter”) to distract audiences at crucial points in the performance. Narrative can moreover hold an audience’s attention on specifically intended information [61] and thus serve to turn their attention away from other content which may be more important. Just as emotion can drive attention during magic shows [16], compelling and emotional news stories can pull attention towards a narrow slice of information and away from other facts or perspectives.
To sum up, news stories can use misdirection to focus attention toward particular viewpoints or aspects of a report—and therefore away from others—thereby causing potentially important information to be missed. In science reporting, misleading uses of narrative can take a variety of different forms. Narrative can misdirect attention away from questions of whether or not a concept is legitimate, a tactic sometimes referred to as “diversionary reframing” [67]. For example, someone might avoid mentioning allegations made against them, while blaming their past actions and positions on problematic policies or regulations (or pointing to the opposition from another party to avoid taking blame for a political failure). Diversionary reframing tactics include diverting attention away from a controversial statement by undermining the legitimacy of critics [67]. Other narratives reframe discussions or “put spin” on a story by focusing audiences’ attention on its positive aspects. For instance, discussions surrounding controversial immigrant detention centers have sometimes focused on the education support provided to detained children to discourage questions about the legitimacy and appropriateness of holding children in such facilities [68].
Even in original scientific articles where discoveries are first announced, scientists frequently report findings in ways that fail to accurately reflect them or their significance [69]. Storytelling is an important part of writing an academic article, as scientists must select the data points, analyses, and all other aspects of a study to present to the reader. Namely, the title, abstract, and conclusions of a research report can all be framed to highlight a particular point, which is acceptable practice in scientific communication. However, storytelling becomes problematic when scientists craft a misleading narrative, with unwarranted interpretations or conclusions that extend beyond the data. Examples include reports that focus on the clinical or societal importance of the findings, even when such results are statistically insignificant [70,71,72]. In clinical trials, a report with spin might be framed to convey that a treatment is beneficial, despite little or no objective evidence in support of that conclusion [73]. Such spin is prevalent in scientific publishing and reporting [71,74,75], leading to a substantial distortion of research results. This is particularly concerning for clinical trials, where misled patients might proceed with therapies or interventions that are unsupported by science.
Discoveries can also be given “spin” through misleading press releases or the ensuing news reports. Press releases often contain exaggerated causal claims and unsupported advice [76], while failing to incorporate key limitations or important facts [77,78,79]. Spin of this kind tends to spread from source to source: when the abstract of the original published article has spin, usually the press release has it too [80]. Similarly, exaggerated claims in press releases tend to be repeated in news reports [76,79,80,81]. Even if there is no spin in the original article, reporters and editors have incentives to report findings in misleading ways. For example, the pressure to drive clicks to a news website, or to encourage potential readers to subscribe, can lead journalists to exaggerate scientific findings.
On social media, infographics—which are more about storytelling than about visualizing data [82]–and other data visualizations can also facilitate deceptive narratives. Data visualizations can be misleading (intentionally or otherwise) even when they include all the relevant information: for example, misleading labels or distorted axes can cause people to misinterpret data [83]. Whether in news reports or in social media posts, data plots and figures must be read and actively interpreted before their patterns can be understood. Unfortunately, people often form quick initial impressions from visualizations before they carefully examine specific aspects of the images. As a result, an individual that is focusing on the wrong elements of a visualization can miss even highly salient information. In one dramatic example, 93% of people failed to notice a conspicuous image of a dinosaur embedded in visually presented data because they were focusing their attention on another aspect of the plot (the position of the Xs, rather than the Os that formed the dinosaur; [84])—see Figure 3. Misleading captions or content in a Twitter thread or other social media posts could produce comparable attentional failures by directing the viewer’s attention towards some patterns and away from others.
Readers can also be misled into incorrect conclusions by confusingly complicated plots, or by peculiar choices in how data is presented (such as by showing the number of COVID-19 cases over time in a pattern of decreasing values, rather than chronologically cumulative case numbers) [50,83,85,86]. People often incorrectly interpret complex information or events as though they were their typical, more simple versions [87]: in magic, for example, spectators often fail to realize that the magician may not cut a deck of cards in the normal way, but using instead a modified maneuver that allows them to control which card ends up on top [88].
Removing or replacing the original narrative—for instance, by posting images without their proper context (i.e., the title, caption, and data labels that belong to an image) can also mislead viewers. In 2020, a data visualization of the fires in Australia was shared online with the inaccurate caption “This is a NASA photograph”. In truth, the image was an artistic visualization, using composite data compiled over time (meaning that the depicted fires were not all simultaneously burning), and somewhat exaggerating the scale of the fires due to glow effects from the way the hotspots were rendered [89]—see Figure 4. Such artistic visualizations are easily mistaken for photographs when reposted in misleading contexts: news media often showed this particular image alongside actual photographs, providing a context that may have further led readers to conclude the image was a photograph too [90].

2.3. We See What We Expect to See

Even when there is no intended deception in a report, readers can miss important information. One well-studied reason for these failures is “confirmation bias”, in which people irrationally consider only information supporting their expectations [91,92]. Magicians often perform movements that resemble typical everyday actions, relying on confirmation bias to prevent the audience from realizing that something different is happening. The audience may believe that a hand gesture is a common, ordinary action, but that movement may conceal a palmed coin [9,10,16].
People are typically unaware that confirmation bias influences their judgments [93]. Instead, when presented with information that contradicts our beliefs, we often react in a defensive manner. Conversely, we feel more inclined to trust information that is consistent with our prior beliefs than we do other information [91,92,94]. We also assign more relevance to stories that speak from our own perspective, rather than from a different viewpoint. At least in part, confirmation bias may result from familiar information being easier to process than unfamiliar information [95]. Additionally, when people feel motivated to confirm their existing opinions, they choose to interpret information in ways that allow them to feel validated [96,97], relying on the cognitive processes that are most likely to lead to their preferred conclusion [98].
Because social media feeds keep users engaged by tailoring content algorithmically to their preferences, people are relatively unlikely to be exposed to differing viewpoints online [99,100]. Thus, Facebook users tend to see information that supports their existing beliefs [101,102], which may reinforce prior biases and increase the spread of disinformation [103,104]. The result can be an “echo chamber”, where mutual followers tend to have similar opinions that reinforce existing views [105].
Lack of exposure to differing viewpoints cannot be attributed solely to algorithms, however: even when the social media posts challenging users’ opinions are visible, that information is often ignored [106] and therefore not shared with other users. In fact, when confronted with too much new or uncomfortable information on their social media feeds, people can feel overwhelmed, reaching a state of “social media fatigue” that may lead them to disengage or temporarily stop using social media [107].

3. Memory Illusions

The way we remember the past can substantially deviate from reality. When biographical or historical events are described in a certain way, our recollection of such (biased) description can overshadow our perception of what happened right in front of us.
Studies have shown that memories are most vulnerable to alteration at the time of retrieval. Although the underlying mechanisms are not well understood, the research evidence indicates that retrieving a memory can enhance the subsequent learning of new information [108,109], and change the original memory [110,111]. That is, whenever we relive a memory, we re-store it as a new memory for later access—essentially wiping off the old memory in the process.
Our memory’s malleability provides a fertile ground for misinformation. Magicians sometimes take advantage of this pliability by recasting events that took place onstage in a manner that will bias spectators’ memories of the performance [19] and prevent them from figuring out the secret method when they try to puzzle it out later [7]. This same process may explain the enhanced susceptibility of eye witnesses in criminal proceedings to misleading suggestions [109].

The Illusory Truth Effect: Repeated Lies Become Plausible

Events and stories seem truer when repeated multiple times. For example, if a magician tosses a coin from hand to hand a few times, and then fakes a throw from right hand to left hand (while secretly concealing the coin in the right hand), it can look as though the magician catches the tossed coin in their left hand. Then, when the magician opens that hand and reveals it to be empty, the coin appears to have vanished [9]. Similar tricks can be performed with balls [62,112] or other items: the object is truly thrown a few times and then the magician merely pretends to throw the object. The initial repetitions create the expectation and false perception that the same event is happening again. This illusion results in part from “priming”, in which the presentation of a stimulus makes subsequent appearances of the stimulus more salient [113].
Repetition can also prime people to believe false information. News headlines that are seen only once are more likely to be believed if they are true [114,115]. But news headlines that are seen multiple times are more likely to be believed irrespective of their accuracy. That is, a repeated message becomes increasingly plausible each time around, even if people initially believed it to be false [115,116]. One possible explanation is that repeated exposure makes information easier to process. Thus, people can interpret their feeling of ease with a message as evidence of its veracity [117]. This phenomenon, called the “illusory truth effect”, can last months after exposure [118].
Although highly implausible statements (such as “a single elephant weighs less than a single ant”) continue to be considered false in the face of repetition [5,6], research has shown that even such implausible statements seem somewhat more plausible when repeated [115]. The illusory truth effect persists when individuals are warned about its existence [119], and when the message contradicts prior knowledge [120,121]. In other words, even if people are able to identify the truth, repetition can boost the perceived accuracy of false statements [121]. Provided enough repetition, obvious lies may begin to seem credible.
Labels identifying material as potentially false can reduce the size of the illusory truth effect in some cases [122], but they do not eliminate it. With repetition, inaccurate statements become more credible even when valid alternatives are available [123], and even when the misinformation comes from a source that is known to be untrustworthy and unreliable [124]. Thus, an expert in a field repudiating a piece of news as false may not be enough to counter repeated misinformation.
In the past decade, there was some concern that repeating a story to refute it could backfire. The reasoning was that (1) people would dismiss any new information that they disagreed with, due to confirmation bias, and (2) additional exposure to false information (even if it was immediately refuted) might strengthen its hold on people’s minds [101,125,126,127]. Nevertheless, such “backfire effects” have proven to be rare, and there is evidence that direct refutations can be successful in correcting misinformation [128,129,130,131].

4. Illusions of Choice

The way we think about and remember events can permanently alter our understanding of them [132,133,134]. Many factors in our environment can alter our behavior, and we tend to assume that what we see is as it appears—that there are no concealed events secretly changing the outcome of an action. The result can be an “illusion of choice”, in which we believe we are making decisions freely, but specific results are inevitable and out of our control.

4.1. Our Decisions Can Be Influenced

If a magician asks you to choose a card, it may seem as though you are free to make any choice you like. However, magicians have a wide variety of “forcing” techniques that constrain spectators’ choices, making them much less free than they appear to be [135]. That is, the magician influences or limits an individual’s choice without their awareness [136,137]. Depending on the technique used, the spectator’s choice may be completely predetermined, or the magician might merely enhance the probability that they will make a particular decision. In one simple example, the magician might discretely or overtly discard any card chosen by an audience member—possibly with an entertaining flourish—until the “correct” card is selected. Alternatively, the magician may use gestural and conversational primes to encourage the spectator to make a certain choice (such as drawing the number three in the air and making a diamond shape with their hands to increase the chance that a volunteer will pick the three of diamonds [138]). Time pressure to produce an action quickly (such as picking a card) can urge audiences to act without thinking deeply, and thus compel them to perform the most obvious action [54]. Therefore, a magician can force a spectator’s card selection by having them quickly choose a card at a time when only one card is visible [54]. Or a magician might ask the spectator to draw the card from the top of a shuffled deck—a card whose placement they have controlled with sleight of hand. The result is that audiences believe they are freely making their own decisions when that is not the case. Alternatively, a spectator’s choice could be free but irrelevant to the outcome [139], as when either selection will lead to the same end result (a magic technique known as “magician’s choice” [140]). The conclusion is predetermined, and the spectator’s involvement in the outcome is an illusion.
Social media users may choose to like some messages and not others, but the platform may be showing them a small subset of posts selected by an algorithm. If users are presented only with information that supports their existing attitudes, they will never have the actual option to accept or reject contrary opinions.
The phenomenon known as “choice blindness” also shows that people do not only misremember manipulated choices as their own, but they also vigorously defend choices that they did not make (though they believe they did) [141].
One example in which a magician can exploit this kind of vulnerability is by covertly changing the card chosen by a volunteer and then asking them to “go ahead, take your card” [88]. Comparable bait-and-switch tactics can be found in online advertising: some ads are designed to look like real news articles, opinion polls, or online search results. The nature of such ads is misleading in itself, with many users unable to reliably identify the content as advertisements [142]. Consequently, people click on ads believing that they are moving on to another trustworthy story on the same website [143], but instead they link to external sites that gather their email addresses, or help to financially support misinformation websites [143,144]. Because these ads often use unsubstantiated language or frame information in misleading ways, they can also spread misinformation directly [144,145]. Bait-and-switch tactics are also common when using search engines: Google, for example, has increasingly dedicated space to advertisements that look like search results [146,147].

4.2. Misleading or Unbalanced Viewpoints

Magicians also manipulate audiences’ interpretation of events by controlling their viewpoint. Magical effects that seem wondrous from the vantage point of spectators can appear mundane when viewed from another angle. Thus, in stage performances the perspective is often controlled by the context: the audience can only view the trick from the front of the stage. Movements, objects, and events that are solely visible from backstage are concealed from the audience’s view.
In a news story, how different kinds of information or viewpoints are presented can similarly lead audiences to accept a single interpretation of facts as truthful, without ever knowing that alternative perspectives existed.
Even when several perspectives are represented, news can give imbalanced treatments to different viewpoints. Journalists typically value balanced reporting, and strive to present information accurately [148], but in practice there is no widely recognized and agreed upon definition of balanced reporting [149]. That is, balanced reporting might refer to seeking out and presenting all different sides or angles in every story, or to approaching information in a detached and neutral way. Regardless of the chosen approach, finding an appropriate balance is more difficult than one might assume.
Other times, stories simply omit information and thereby prevent readers from considering it. For instance, a story might mention a change to a number (e.g., “5% fewer COVID-related deaths”) rather than the actual number (which, although lower than before, might still be strikingly high).
Just as spectators in a magic show lack the information or the time to figure out the secret method behind a magical effect, reporters may be unable to determine fact from fiction for comparable reasons: a scarcity of time, scientific expertise, or other resources. Local news sources have especially struggled in the past decades [150,151,152,153] and may need to reduce resources to cut costs, possibly including cutting fact-checking verification practices.
Unlike magic audiences, journalists are expected to extract (and provide) veridical information from potentially deceitful appearances. Yet, with a limited ability to determine the accuracy of two positions, some journalists seeking to avoid perceived bias will present both sides of a dispute with equal weight, irrespective of their respective merits [154,155]. Unfortunately, if one of such viewpoints is based on untruths and misinformation, portraying both sides as equally valid is inherently misleading. Likewise, giving domain experts and nonexperts an equal voice can distort the truth. For example, journalists might misrepresent the current scientific consensus on climate change by quoting both environmental scientists and climate change deniers [155,156,157,158].
If 1% of scientists hold a particular viewpoint but they receive an equal treatment to the other 99%, such reporting would give the incorrect impression of a 50/50 split in opinion. Because audiences do not know the actual distribution of opinions among scientists, equal balance can produce a form of sampling bias, where minority opinion is given too much weight and misinformation appears to be legitimized. Equal balance practices may also increase public distrust of science: when scientific findings are presented alongside misinformed criticisms, the scientific conclusions can appear less certain than they really are. This situation becomes even more problematic when there are more than two sides in a story: given time and space constraints in reporting, journalists can experience difficulties when selecting which viewpoints to include, omit, or discuss in detail.
Overall, the above kind of “balanced” reporting can undermine the interests of audiences who try to draw meaningful conclusions, both by highlighting misinformation and by limiting the time given to in-depth discussions of accurate information [159,160]. One study found that stories with an equitable tone (toward two political candidates) were less likely to focus in-depth on substance or issues, as well as less likely to include sourced content. They were also generally shorter in length than “slanted” stories that favored one candidate over the other [161]. Whereas “balanced” reporting practices may be expeditious to journalists, giving equal importance to truth and to misinformation clouds matters and leads to information biases [148].

4.3. Why Warnings Fail to Prevent Misperceptions

The sense of wonder that accompanies a magic show requires that the performance seems impossible. But being warned that the magical effects are “just tricks” may do nothing to diminish the emotions felt by the audience. The warning is, in fact, redundant: audiences typically know that magic shows are not accurate portrayals of physical reality.
In false news reports, misinformation similarly triggers emotional responses even in contexts where it should not. When misinformation is labelled as potentially inaccurate and contested (e.g., Facebook stating on a page that independent fact-checkers have said the information is false, or Twitter warning users that claims about election fraud are disputed), readers often still believe it and share the misleading claims [115,162,163,164,165,166]. In fact, even when readers notice the warnings, they often ignore them, so long as the warning does not interrupt the readers’ actions [163,167,168]. Even if warnings are heeded, the fact that only some messages include them can make people think that the other messages that do not include such warnings must be accurate—when they may contain other misinformation that the platform failed to detect [163].

5. Potential Answers and Areas for Future Research

Many of the cognitive biases discussed above are ingrained and therefore very difficult to circumvent or prevent. However, being aware of these vulnerabilities may help individuals to critically evaluate information and ameliorate bias, as well as assist developers and legislators in the creation of more effective tools for combatting misinformation. Topical experts can be encouraged to vocally dispute misinformation, and scientists supported to engage in communication and outreach practices, to help journalists and the public to accurately understand discoveries. Financial assistance can be provided to local news outlets to provide them with fact-checking resources.
Some direct steps can also be taken to decrease the flow of misinformation, including the implementation of warnings on disputed messages. Warnings are often, but not always ignored, and therefore can work to slow the spread of misinformation [163,167,168]. Further, warnings can be made more effective by interrupting a user’s actions (e.g., requiring a mouse click to continue) and by periodically changing so that users do not dismiss them or bypass them out of habit. At the extreme, users who spread large amounts of misinformation can be deplatformed, removing their messaging from a social media platform. An important caveat is that as measures against misinformation grow in severity, so will the likelihood that they will be viewed as paternalistic, imposing, punitive, and restrictive of free speech [169,170].
Some spectators of magic shows enjoy focusing their efforts on dissecting tricks and identifying the methods behind magical effects. Similarly, some social media users take effortful steps to authenticate dubious claims, helping to slow the spread of misinformation with fact-checking behaviors. Users who share their findings on social media can encourage others to remove misleading messages or to reconsider their behavior when posting in the future. These efforts by individual users should be supported, and their fact-checking encouraged, by providing easy-to-use and freely available verification tools. Some fact-checking resources have become widely available and used, such as the website Snopes.com. Similar initiatives should be supported, with the caveat that users should also be educated about which tools are reliable, given that sites masquerading as fact-checking services have also proliferated. Fact checking, verification, and authentication processes that are already in place should also be revisited regularly to consider new misinformation techniques and tools, particularly when dealing with topics where misinformation can create substantial harm.
Engineers and other developers should take misuse-related considerations seriously when determining research priorities and when creating or modifying tools such as social media platforms. Developers working with such tools should proactively act to minimize foreseeable harmful applications. Domain experts in neuroscience and psychology should be included—and the factors discussed above considered—in discussions of these issues and in the development of legislation dealing with misinformation.
False news has existed for a long time, and it will continue to exist despite efforts to counter it. As communication becomes progressively faster and easier, misinformation will also increase in scale. With technological advances, intentional disinformation attempts will likely become more effective, and new methods and approaches for deceiving the public will appear. For example, artificial intelligence (AI) may enable new kinds of automated disinformation attempts at ever-larger scales. Misinformation will also expand on smaller, personal levels, as technologies enable greater numbers of users to create authentic-seeming fake images and videos. These images can be employed for image-based abuse (also called “revenge porn”), bullying, blackmail, political sabotage, and other malicious purposes. Fake images might even be used as false evidence of a crime, potentially leading to wrongful convictions [171].
Individuals can now use AI tools to produce “deepfake” videos that appear convincingly real, in which new words are digitally inserted into audio tracks, or the face or body of a person is digitally altered to match someone else’s appearance [172,173]. Deepfake videos will likely become increasingly realistic as software tools improve. Recently developed text-to-image synthesis techniques can also generate new—and often realistic—images from visual descriptions provided by the user as text [174]. As a result, users can very quickly and easily create new images to include in misleading social media messages. There is no equivalent in magic to date: although magicians often incorporate cutting-edge technologies into their acts, no existing automated tool has yet been used to create magical effects from simple text prompts. Though research on magic does not yet illuminate the impact of—or what tools may combat—AI-facilitated deceptions, we expect that determining how people perceive the accuracy or falsehood of audio and video recordings will be a necessary direction for future research on misinformation.

6. Conclusions

Magical effects are incredibly robust: they work even though audiences know that they are being tricked. Similarly, people often accept and disperse misinformation despite warnings that the facts are disputed and potentially false [115,162,163,164,165,166]. Thus, increasing the awareness of scientific facts has proven ineffective in countering the flow of misinformation [175,176]. In news stories, as in magic performances, people do not view information in a detached and neutral way [177]. Psychological and emotional appeal can be much more important than factual accuracy. In this paper, we reviewed some factors that help explain how people react to misinformation, grounding the discussion around the ways in which magicians exploit audiences’ vulnerabilities and biases in perception and cognition.
Audiences are often fooled by the concealment of key information in magic shows: no spectator can accurately understand data that are absent. Likewise, due to reframing and spin, and even outright data falsification, news reports and social media posts are often impossible to evaluate accurately by readers.
Even when information is present—and highly salient—observers can fail to process it accurately. Attention can be misdirected away from central facts and data. Confirmation bias and other cognitive biases can cause people to miss or disregard accurate information. Sampling biases can mislead news story consumers and social media users by granting legitimacy to fringe opinions. Repetition causes false information to gain plausibility with each iteration.
Research studies on perception and cognition have been—and continue to be—conducted to determine the situations in which magic tricks succeed or fail. Scientists may draw on this growing literature to help identify the scenarios that withstand misinformation, and to devise more fruitful methods for countering its spread. Reducing the prevalence and societal impact of misinformation will require multi-layered approaches that consider the range of biases we discussed here.

Author Contributions

Conceptualization, R.G.A. and S.M.-C.; Writing—Original Draft Preparation, R.G.A.; Writing—Review and Editing, R.G.A., S.L.M. and S.M.-C.; Funding Acquisition, S.L.M. and S.M.-C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the New York State Empire Innovator Program, by the National Science Foundation (Award 1734887 to S.M.-C. and S.L.M.; Award 1523614 to S.L.M.), and by the National Institute of Health (Awards R01EY031971 and R01CA258021 to S.M.-C. and S.L.M.).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest. The funding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Flynn, D.; Nyhan, B.; Reifler, J. The Nature and Origins of Misperceptions: Understanding False and Unsupported Beliefs about Politics. Political Psychol. 2017, 38, 127–150. [Google Scholar] [CrossRef]
  2. Pennycook, G.; McPhetres, J.; Zhang, Y.; Lu, J.G. David G Rand Fighting COVID-19 misinformation on social media: Experimental evidence for a scalable ac-curacy-nudge intervention. Psychol. Sci. 2020, 31, 770–780. [Google Scholar] [CrossRef]
  3. Chiluwa, I.E. Deception in Online Terrorist Propaganda: A Study of ISIS and Boko Haram. In Handbook of Research on Deception, Fake News, and Misinformation Online; IGI Global: Hershey, PA, USA, 2019; pp. 520–537. [Google Scholar]
  4. Mantere, M. Stock Market Manipulation Using Cyberattacks Together with Misinformation Disseminated through Social Media. In Proceedings of the 2013 International Conference on Social Computing, Alexandria, VA, USA, 8–14 September 2013; pp. 950–954. [Google Scholar] [CrossRef]
  5. Lacassagne, D.; Béna, J.; Corneille, O. Is Earth a perfect square? Repetition increases the perceived truth of highly implausible statements. Cognition 2022, 223, 105052. [Google Scholar] [CrossRef]
  6. Fazio, L.K.; Rand, D.G.; Pennycook, G. Repetition increases perceived truth equally for plausible and implausible statements. Psychon. Bull. Rev. 2019, 26, 1705–1710. [Google Scholar] [CrossRef]
  7. Macknik, S.; Martinez-Conde, S.; Blakeslee, S. Sleights of Mind: What the Neuroscience of Magic Reveals about Our Everyday Deceptions; Henry Holt & Co.: New York, NY, USA, 2011. [Google Scholar]
  8. Martinez-Conde, S.; Macknik, S.L. Magic and the brain. Sci. Am. 2008, 299, 72–79. [Google Scholar] [CrossRef]
  9. Cui, J.; Otero-Millan, J.; Macknik, S.L.; King, M.; Martinez-Conde, S. Social Misdirection Fails to Enhance a Magic Illusion. Front. Hum. Neurosci. 2011, 5, 103. [Google Scholar] [CrossRef]
  10. Otero-Millan, J.; Macknik, S.L.; Robbins, A.; Martinez-Conde, S. Stronger Misdirection in Curved than in Straight Motion. Front. Hum. Neurosci. 2011, 5, 133. [Google Scholar] [CrossRef]
  11. Rieiro, H.; Martinez-Conde, S.; Macknik, S.L. Perceptual elements in Penn & Teller’s “Cups and Balls” magic trick. PeerJ 2013, 1, e19. [Google Scholar]
  12. Barnhart, A.S.; Costela, F.M.; Martinez-Conde, S.; Macknik, S.L.; Goldinger, S.D. Microsaccades reflect the dynamics of misdirected attention in magic. J. Eye Mov. Res. 2019, 12. [Google Scholar] [CrossRef] [PubMed]
  13. Kuhn, G.; Olson, J.; Raz, A. Editorial: The Psychology of Magic and the Magic of Psychology. Front. Psychol. 2016, 7, 1358. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Rensink, R.A.; Kuhn, G. A framework for using magic to study the mind. Front. Psychol. 2015, 5, 1508. [Google Scholar] [CrossRef]
  15. Thomas, C.; Didierjean, A.; Maquestiaux, F.; Gygax, P. Does magic offer a cryptozoology ground for psychology? Rev. Gen. Psychol. 2015, 19, 117–128. [Google Scholar] [CrossRef]
  16. Macknik, S.L.; King, M.; Randi, J.; Robbins, A.; Teller; Thompson, J.; Martinez-Conde, S. Attention and awareness in stage magic: Turning tricks into research. Nat. Rev. Neurosci. 2008, 9, 871–879. [Google Scholar] [CrossRef] [PubMed]
  17. Demacheva, I.; Ladouceur, M.; Steinberg, E.; Pogossova, G.; Raz, A. The Applied Cognitive Psychology of Attention: A Step Closer to Understanding Magic Tricks. Appl. Cogn. Psychol. 2012, 26, 541–549. [Google Scholar] [CrossRef]
  18. Folk, C.L.; Remington, R.W.; Wright, J.H. The structure of attentional control: Contingent attentional capture by apparent motion, abrupt onset, and color. J. Exp. Psychol. Hum. Percept. Perform 1994, 20, 317–329. [Google Scholar] [CrossRef]
  19. Lamont, P.; Wiseman, R. Magic in Theory; Hermetic Press: Seattle, WA, USA, 1999. [Google Scholar]
  20. Ascanio, A. The psychology of palming. In The Magic of Ascanio—The Structural Conception of Magic; Páginas: Madrid, Spain, 1982; pp. 77–86. [Google Scholar]
  21. Alexander, R.G.; Waite, S.; Macknik, S.L.; Martinez-Conde, S. What do radiologists look for? Advances and limitations of perceptual learning in radiologic search. J. Vis. 2020, 20, 1–13. [Google Scholar] [CrossRef] [PubMed]
  22. Sharpe, S. Conjurers Psychological Secrets; Hades Publications: Calgary, AB, Cannada, 1988. [Google Scholar]
  23. Maunsell, J.H.R.; Cook, E.P. The role of attention in visual processing. Philos. Trans. R. Soc. B Biol. Sci. 2002, 357, 1063–1072. [Google Scholar] [CrossRef]
  24. Reynolds, J.H.; Chelazzi, L. Attentional Modulation of Visual Processing. Annu. Rev. Neurosci. 2004, 27, 611–647. [Google Scholar] [CrossRef] [PubMed]
  25. McAdams, C.J.; Reid, R.C. Attention Modulates the Responses of Simple Cells in Monkey Primary Visual Cortex. J. Neurosci. 2005, 25, 11023–11033. [Google Scholar] [CrossRef] [PubMed]
  26. Motter, B.C. Focal attention produces spatially selective processing in visual cortical areas V1, V2, and V4 in the presence of competing stimuli. J. Neurophysiol. 1993, 70, 909–919. [Google Scholar] [CrossRef] [PubMed]
  27. Lavie, N.; Tsal, Y. Perceptual load as a major determinant of the locus of selection in visual attention. Percept. Psychophys. 1994, 56, 183–197. [Google Scholar] [CrossRef] [PubMed]
  28. Silver, M.A.; Ress, D.; Heeger, D.J. Neural Correlates of Sustained Spatial Attention in Human Early Visual Cortex. J. Neurophysiol. 2007, 97, 229–237. [Google Scholar] [CrossRef] [PubMed]
  29. Ardid, S.; Wang, X.-J.; Compte, A. An Integrated Microcircuit Model of Attentional Processing in the Neocortex. J. Neurosci. 2007, 27, 8486–8495. [Google Scholar] [CrossRef] [PubMed]
  30. Treue, S.; Trujillo, J.C.M. Feature-based attention influences motion processing gain in macaque visual cortex. Nature 1999, 399, 575–579. [Google Scholar] [CrossRef]
  31. Connor, C.E.; Gallant, J.L.; Preddie, D.C.; Van Essen, D.C. Responses in area V4 depend on the spatial relationship between stimulus and attention. J. Neurophysiol. 1996, 75, 1306–1308. [Google Scholar] [CrossRef] [PubMed]
  32. Moore, T.; Armstrong, K.M. Selective gating of visual signals by microstimulation of frontal cortex. Nature 2003, 421, 370–373. [Google Scholar] [CrossRef] [PubMed]
  33. Chen, Y.; Martinez-Conde, S.; Macknik, S.L.; Bereshpolova, Y.; Swadlow, H.A.; Alonso, J.-M. Task difficulty modulates the activity of specific neuronal populations in primary visual cortex. Nat. Neurosci. 2008, 11, 974–982. [Google Scholar] [CrossRef]
  34. Tse, P.U. Voluntary attention modulates the brightness of overlapping transparent surfaces. Vis. Res. 2005, 45, 1095–1098. [Google Scholar] [CrossRef]
  35. Macknik, S.; Martinez-Conde, S. How Neural Constraints Set Us Free to Create. Am. Sci. 2020, 108, 196–199. [Google Scholar] [CrossRef]
  36. Quiroga, R.Q. Magic and cognitive neuroscience. Curr. Biol. 2016, 26, R390–R394. [Google Scholar] [CrossRef] [Green Version]
  37. Martinez-Conde, S.; Macknik, S.L. Mind tricks. Nature 2007, 448, 414. [Google Scholar] [CrossRef]
  38. Scott, S.K.; Rosen, S.; Beaman, C.P.; Davis, J.P.; Wise, R.J.S. The neural processing of masked speech: Evidence for different mechanisms in the left and right temporal lobes. J. Acoust. Soc. Am. 2009, 125, 1737–1743. [Google Scholar] [CrossRef] [PubMed]
  39. Gutschalk, A.; Micheyl, C.; Oxenham, A.J. Neural correlates of auditory perceptual awareness under informational masking. PLoS Biol. 2008, 6, e138. [Google Scholar] [CrossRef] [PubMed]
  40. Pelli, D.G. Crowding: A cortical constraint on object recognition. Curr. Opin. Neurobiol. 2008, 18, 445–451. [Google Scholar] [CrossRef] [PubMed]
  41. Zhang, M.; Denison, R.N.; Pelli, D.G.; Le, T.T.C.; Ihlefeld, A. An auditory-visual tradeoff in susceptibility to clutter. Sci. Rep. 2021, 11, 1–8. [Google Scholar] [CrossRef]
  42. Pelli, D.G.; Tillman, K.A. The uncrowded window of object recognition. Nat. Neurosci. 2008, 11, 1129–1135. [Google Scholar] [CrossRef]
  43. Louie, E.G.; Bressler, D.W.; Whitney, D. Holistic crowding: Selective interference between configural represen-tations of faces in crowded scenes. J. Vis. 2007, 7, 24. [Google Scholar] [CrossRef]
  44. Goldsworthy, R.L.; Greenberg, J.E. Analysis of speech-based Speech Transmission Index methods with implications for nonlinear operations. J. Acoust. Soc. Am. 2004, 116, 3679–3689. [Google Scholar] [CrossRef]
  45. Simons, D.J.; Franconeri, S.L.; Reimer, R.L. Change Blindness in the Absence of a Visual Disruption. Perception 2000, 29, 1143–1154. [Google Scholar] [CrossRef]
  46. Rensink, R.A. Change detection. Annu. Rev. Psychol. 2002, 53, 245–277. [Google Scholar] [CrossRef]
  47. O’Regan, J.K.; Deubel, H.; Clark, J.J.; Rensink, R.A. Picture Changes during Blinks: Looking without Seeing and Seeing without Looking. Vis. Cogn. 2000, 7, 191–211. [Google Scholar] [CrossRef]
  48. Alexander, R.G.; Schmidt, J.; Zelinsky, G.J. Are summary statistics enough? Evidence for the importance of shape in guiding visual search. Vis. Cogn. 2014, 22, 595–609. [Google Scholar] [CrossRef]
  49. Alexander, R.G.; Mintz, R.J.; Custodio, P.J.; Macknik, S.L.; Vaziri, A.; Venkatakrishnan, A.; Gindina, S.; Martinez-Conde, S. Gaze mechanisms enabling the detection of faint stars in the night sky. Eur. J. Neurosci. 2021, 54, 5357–5367. [Google Scholar] [CrossRef]
  50. Szafir, D.A. The good, the bad, and the biased. Interactions 2018, 25, 26–33. [Google Scholar] [CrossRef]
  51. Zelinsky, G.J.; Neider, M.B. An eye movement analysis of multiple object tracking in a realistic environment. Vis. Cogn. 2008, 16, 553–566. [Google Scholar] [CrossRef]
  52. Pylyshyn, Z.W.; Storm, R.W. Tracking multiple independent targets: Evidence for a parallel tracking mechanism*. Spat. Vis. 1988, 3, 179–197. [Google Scholar] [CrossRef]
  53. Tamariz, J. The Magic Way; Frakson Books: Madrid, Spain, 1988. [Google Scholar]
  54. Kuhn, G.; Amlani, A.A.; Rensink, R.A. Towards a science of magic. Trends Cogn. Sci. 2008, 12, 349–354. [Google Scholar] [CrossRef]
  55. McGrew, S.; Breakstone, J.; Ortega, T.; Smith, M.; Wineburg, S. Can Students Evaluate Online Sources? Learning from Assessments of Civic Online Reasoning. Theory Res. Soc. Educ. 2017, 46, 165–193. [Google Scholar] [CrossRef]
  56. Wineburg, S.; McGrew, S. Lateral Reading: Reading Less and Learning More When Evaluating Digital Information. Teach. Coll. Rec. 2017. [Google Scholar] [CrossRef]
  57. Horrigan, J.B. How People Approach Facts and Information: People Deal in Varying Ways with Tensions about What Information to Trust and how Much They Want to Learn. 2017. Available online: https://www.pewresearch.org/internet/2017/09/11/how-people-approach-facts-and-information/ (accessed on 24 May 2022).
  58. Riley, N.S. America’s real digital divide. The New York Times, 2 February 2018; A23. [Google Scholar]
  59. Simons, D.J.; Chabris, C.F. Gorillas in our midst: Sustained inattentional blindness for dynamic events. Perception 1999, 28, 1059–1074. [Google Scholar] [CrossRef]
  60. Memmert, D. The effects of eye movements, age, and expertise on inattentional blindness. Conscious. Cogn. 2006, 15, 620–627. [Google Scholar] [CrossRef] [PubMed]
  61. Martinez-Conde, S.; Alexander, R.G.; Blum, D.; Britton, N.; Lipska, B.K.; Quirk, G.J.; Swiss, J.I.; Willems, R.M.; Macknik, S.L. The Storytelling Brain: How Neuroscience Stories Help Bridge the Gap between Research and Society. J. Neurosci. 2019, 39, 8285–8290. [Google Scholar] [CrossRef] [PubMed]
  62. Kuhn, G.; Land, M.F. There’s more to magic than meets the eye. Curr. Biol. 2006, 16, R950–R951. [Google Scholar] [CrossRef] [PubMed]
  63. Kuhn, G.; Tatler, B. Magic and Fixation: Now You Don’t See it, Now You Do. Perception 2005, 34, 1155–1161. [Google Scholar] [CrossRef]
  64. Kuhn, G.; Tatler, B.; Findlay, J.M.; Cole, G.G. Misdirection in magic: Implications for the relationship between eye gaze and attention. Vis. Cogn. 2008, 16, 391–405. [Google Scholar] [CrossRef]
  65. Most, S.B. What’s “inattentional” about inattentional blindness? Conscious. Cogn. 2010, 19, 1102–1104; discussion 1107–1109. [Google Scholar] [CrossRef]
  66. Tamariz, J.; Lehn, D.B. The Five Points in Magic; Hermetic Press Incorporated: Seattle, WA, USA, 2007. [Google Scholar]
  67. Freudenburg, W.R.; Alario, M. Weapons of Mass Distraction: Magicianship, Misdirection, and the Dark Side of Legitimation. Sociol. Forum. 2007, 22, 146–173. [Google Scholar] [CrossRef]
  68. Dittmann, L.; Gerber, J. Framing Legitimacy: A Qualitative Analysis Examining Local Print-Media Portrayals of an Immigrant Family Detention Center in Texas. J. Qual. Crim. Justice Criminol. 2016. [Google Scholar] [CrossRef]
  69. Boutron, I.; Ravaud, P. Misrepresentation and distortion of research in biomedical literature. Proc. Natl. Acad. Sci. USA 2018, 115, 2613–2619. [Google Scholar] [CrossRef]
  70. Premji, S.; Messing, K.; Lippel, K. Would a “one-handed” scientist lack rigor? How scientists discuss the work-relatedness of musculoskeletal disorders in formal and informal communications. Am. J. Ind. Med. 2008, 51, 173–185. [Google Scholar] [CrossRef]
  71. Chiu, K.; Grundy, Q.; Bero, L. ‘Spin’ in published biomedical literature: A methodological systematic review. PLoS Biol. 2017, 15, e2002173. [Google Scholar] [CrossRef] [PubMed]
  72. Cofield, S.S.; Corona, R.V.; Allison, D. Use of Causal Language in Observational Studies of Obesity and Nutrition. Obes. Facts 2010, 3, 353–356. [Google Scholar] [CrossRef] [PubMed]
  73. Boutron, I.; Altman, D.G.; Hopewell, S.; Vera-Badillo, F.; Tannock, I.; Ravaud, P. Impact of Spin in the Abstracts of Articles Reporting Results of Randomized Controlled Trials in the Field of Cancer: The SPIIN Randomized Controlled Trial. J. Clin. Oncol. 2014, 32, 4120–4126. [Google Scholar] [CrossRef] [PubMed]
  74. Boutron, I.; Dutton, S.; Ravaud, P.; Altman, D.G. Reporting and Interpretation of Randomized Controlled Trials with Statistically Nonsignificant Results for Primary Outcomes. JAMA 2010, 303, 2058–2064. [Google Scholar] [CrossRef]
  75. Lazarus, C.; Haneef, R.; Ravaud, P.; Hopewell, S.; Altman, D.G.; Boutron, I. Peer reviewers identified spin in manuscripts of nonrandomized studies assessing therapeutic interventions, but their impact on spin in abstract conclusions was limited. J. Clin. Epidemiol. 2016, 77, 44–51. [Google Scholar] [CrossRef] [PubMed]
  76. Sumner, P.; Vivian-Griffiths, S.; Boivin, J.; Williams, A.; Venetis, C.; Davies, A.; Ogden, J.; Whelan, L.; Hughes, B.; Dalton, B.; et al. The association between exaggeration in health related science news and academic press releases: Retrospective observational study. BMJ 2014, 349, g7015. [Google Scholar] [CrossRef]
  77. Kuriya, B.; Schneid, E.C.; Bell, C.M. Quality of Pharmaceutical Industry Press Releases Based on Original Research. PLoS ONE 2008, 3, e2828. [Google Scholar] [CrossRef]
  78. Woloshin, S.; Schwartz, L.M. Press releases: Translating research into news. JAMA 2002, 287, 2856–2858. [Google Scholar] [CrossRef]
  79. Woloshin, S.; Schwartz, L.M.; Casella, S.L.; Kennedy, A.T.; Larson, R.J. Press releases by academic medical centers: Not so academic? Ann. Intern. Med. 2009, 150, 613–618. [Google Scholar] [CrossRef]
  80. Yavchitz, A.; Boutron, I.; Bafeta, A.; Marroun, I.; Charles, P.; Mantz, J.; Ravaud, P. Misrepresentation of Randomized Controlled Trials in Press Releases and News Coverage: A Cohort Study. PLoS Med. 2012, 9, e1001308. [Google Scholar] [CrossRef] [Green Version]
  81. Sumner, P.; Vivian-Griffiths, S.; Boivin, J.; Williams, A.; Bott, L.; Adams, R.; Venetis, C.A.; Whelan, L.; Hughes, B.; Chambers, C.D. Exaggerations and Caveats in Press Releases and Health-Related Science News. PLoS ONE 2016, 11, e0168217. [Google Scholar] [CrossRef] [PubMed]
  82. Krum, R. Cool Infographics: Effective Communication with Data Visualization and Design; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  83. Pandey, A.V.; Rall, K.; Satterthwaite, M.L.; Nov, O.; Bertini, E. How deceptive are deceptive visualizations? An empirical analysis of common distortion tech-niques. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Online, 18 April 2015. [Google Scholar]
  84. Boger, T.; Most, S.B.; Franconeri, S.L. Jurassic Mark: Inattentional Blindness for a Datasaurus Reveals that Visu-alizations are Explored, not Seen. arXiv 2021, arXiv:2108.05182v2. [Google Scholar]
  85. Engledowl, C.; Weiland, T. Data (Mis)representation and COVID-19: Leveraging Misleading Data Visualizations for Developing Statistical Literacy across Grades V 16. J. Stat. Data Sci. Educ. 2021, 29, 160–164. [Google Scholar] [CrossRef]
  86. Doan, S. Misrepresenting COVID-19: Lying with Charts during the Second Golden Age of Data Design. J. Bus. Tech. Commun. 2020, 35, 73–79. [Google Scholar] [CrossRef]
  87. Kahneman, D.; Frederick, S. Representativeness Revisited: Attribute Substitution in Intuitive Judgment. In Heuristics and Biases; Cambrige University Press: Cambridge, UK, 2002; pp. 49–81. [Google Scholar]
  88. Pailhès, A.; Kuhn, G. The apparent action causation: Using a magician forcing technique to investigate our illu-sory sense of agency over the outcome of our choices. Q. J. Exp. Psychol. 2020, 73, 1784–1795. [Google Scholar] [CrossRef] [PubMed]
  89. Hearsey, A. Australia is Burning/A 3D Visualization. 2019. Available online: https://anthonyhearsey.com/australia-is-burning-a-3d-visualisation (accessed on 24 May 2022).
  90. Suhardjono, L.; Oscario, A.; Luzar, L.; Sriherlambang, B. Infographie, data visualization, and the danger of reality distortion: The case study of Australia fire 2020. IOP Conf. Ser. Earth Environ. Sci. 2021, 729. [Google Scholar] [CrossRef]
  91. Koriat, A.; Lichtenstein, S.; Fischhoff, B. Reasons for confidence. J. Exp. Psychol. Hum. Learn. Mem. 1980, 6, 107. [Google Scholar] [CrossRef]
  92. Nickerson, R.S. Confirmation bias: A ubiquitous phenomenon in many guises. Rev. Gen. Psychol. 1998, 2, 175–220. [Google Scholar] [CrossRef]
  93. Tversky, A.; Kahneman, D. Judgment under Uncertainty: Heuristics and Biases. Science 1974, 185, 1124–1131. [Google Scholar] [CrossRef] [PubMed]
  94. Devine, P.G.; Hirt, E.R.; Gehrke, E.M. Diagnostic and confirmation strategies in trait hypothesis testing. J. Personal. Soc. Psychol. 1990, 58, 952. [Google Scholar] [CrossRef]
  95. Kahneman, D. Maps of Bounded Rationality: Psychology for Behavioral Economics. Am. Econ. Rev. 2003, 93, 1449–1475. [Google Scholar] [CrossRef] [Green Version]
  96. Camaj, L. From Selective Exposure to Selective Information Processing: A Motivated Reasoning Approach. Media Commun. 2019, 7, 8–11. [Google Scholar] [CrossRef]
  97. Bolsen, T.; Druckman, J.N.; Cook, F.L. The Influence of Partisan Motivated Reasoning on Public Opinion. Politi-Behav. 2013, 36, 235–262. [Google Scholar] [CrossRef]
  98. Kunda, Z. The case for motivated reasoning. Psychol. Bull. 1990, 108, 480–498. [Google Scholar] [CrossRef] [PubMed]
  99. Chou, W.-Y.S.; Oh, A.; Klein, W.M.P. Addressing Health-Related Misinformation on Social Media. JAMA 2018, 320, 2417–2418. [Google Scholar] [CrossRef] [PubMed]
  100. Guilbeault, D.; Becker, J.; Centola, D. Social learning and partisan bias in the interpretation of climate trends. Proc. Natl. Acad. Sci. USA 2018, 115, 9714–9719. [Google Scholar] [CrossRef]
  101. Zollo, F. Dealing with digital misinformation: A polarised context of narratives and tribes. EFSA J. 2019, 17, e170720. [Google Scholar] [CrossRef]
  102. Zollo, F.; Quattrociocchi, W. Misinformation spreading on Facebook. In Complex Spreading Phenomena in Social Systems; Springer: Berlin/Heidelberg, Germany, 2018; pp. 177–196. [Google Scholar]
  103. Wardle, C.; Derakhshan, H. Information Disorder: Toward an Interdisciplinary Framework for Research and Policy Making; Council of Europe: Strasbourg, France, 2017; p. 27. [Google Scholar]
  104. Benlian, A. Web Personalization Cues and Their Differential Effects on User Assessments of Website Value. J. Manag. Inf. Syst. 2015, 32, 225–260. [Google Scholar] [CrossRef]
  105. Colleoni, E.; Rozza, A.; Arvidsson, A. Echo Chamber or Public Sphere? Predicting Political Orientation and Measuring Political Homophily in Twitter Using Big Data. J. Commun. 2014, 64, 317–332. [Google Scholar] [CrossRef]
  106. Moravec, P.; Minas, R.; Dennis, A.R. Fake news on social media: People believe what they want to believe when it makes no sense at all. Kelley Sch. Bus. Res. Pap. 2018, 18–87. [Google Scholar] [CrossRef]
  107. Islam, A.K.M.N.; Laato, S.; Talukder, S.; Sutinen, E. Misinformation sharing and social media fatigue during COVID-19: An affordance and cognitive load perspective. Technol. Forecast. Soc. Chang. 2020, 159, 120201. [Google Scholar] [CrossRef]
  108. Chan, J.C.K.; Meissner, C.A.; Davis, S.D. Retrieval potentiates new learning: A theoretical and meta-analytic review. Psychol. Bull. 2018, 144, 1111–1146. [Google Scholar] [CrossRef]
  109. Chan, J.C.K.; Manley, K.D.; Lang, K. Retrieval-enhanced suggestibility: A retrospective and a new investigation. J. Appl. Res. Mem. Cogn. 2017, 6, 213–229. [Google Scholar] [CrossRef]
  110. Merckelbach, H.; Jelicic, M.; Pieters, M. Misinformation increases symptom reporting: A test–retest study. JRSM Short Rep. 2011, 2, 1–6. [Google Scholar] [CrossRef]
  111. Stille, L.; Norin, E.; Sikström, S. Self-delivered misinformation-Merging the choice blindness and misinformation effect paradigms. PLoS ONE 2017, 12, e0173606. [Google Scholar] [CrossRef]
  112. Triplett, N. The Psychology of Conjuring Deceptions. Am. J. Psychol. 1900, 11, 439. [Google Scholar] [CrossRef]
  113. Kolb, B.; Whishaw, I.Q. Fundamentals of Human Neuropsychology, 6th ed.; Worth Publishers: New York, NY, USA, 2008. [Google Scholar]
  114. Rosenzweig, L.R.; Bago, B.; Berinsky, A.J.; Rand, D.G. Happiness and surprise are associated with worse truth discernment of COVID-19 headlines among social media users in Nigeria. Harv. Kennedy Sch. Misinf. Rev. 2021. [Google Scholar] [CrossRef]
  115. Pennycook, G.; Cannon, T.D.; Rand, D.G. Prior exposure increases perceived accuracy of fake news. J. Exp. Psychol. Gen. 2018, 147, 1865–1880. [Google Scholar] [CrossRef]
  116. Begg, I.M.; Anas, A.P.; Farinacci, S. Dissociation of processes in belief: Source recollection, statement familiarity, and the illusion of truth. J. Exp. Psychol. Gen. 1992, 121, 446–458. [Google Scholar] [CrossRef]
  117. Wang, W.-C.; Brashier, N.M.; Wing, E.A.; Marsh, E.J.; Cabeza, R. On Known Unknowns: Fluency and the Neural Mechanisms of Illusory Truth. J. Cogn. Neurosci. 2016, 28, 739–746. [Google Scholar] [CrossRef]
  118. Brown, A.S.; Nix, L.A. Turning lies into truths: Referential validation of falsehoods. J. Exp. Psychol. Learn. Mem. Cogn. 1996, 22, 1088–1100. [Google Scholar] [CrossRef]
  119. Nadarevic, L.; Aßfalg, A. Unveiling the truth: Warnings reduce the repetition-based truth effect. Psychol. Res. 2016, 81, 814–826. [Google Scholar] [CrossRef]
  120. Fazio, L.K.; Perfors, A.; Ecker, U. Repetition increases perceived truth even for known falsehoods. Collabra Psychol. 2020, 6, 38. [Google Scholar] [CrossRef]
  121. Fazio, L.K.; Marsh, E.J. Prior knowledge does not protect against illusory truth effects. J. Exp. Psychol. Gen. 2009, 144, 993–1002. [Google Scholar] [CrossRef]
  122. Jalbert, M.; Schwarz, N.; Newman, E. Only half of what i’ll tell you is true: Expecting to encounter falsehoods reduces illusory truth. J. Appl. Res. Mem. Cogn. 2020, 9, 602–613. [Google Scholar] [CrossRef]
  123. Unkelbach, C.; Greifeneder, R. Experiential fluency and declarative advice jointly inform judgments of truth. J. Exp. Soc. Psychol. 2018, 79, 78–86. [Google Scholar] [CrossRef]
  124. Henkel, L.A.; Mattson, M.E. Reading is believing: The truth effect and source credibility. Conscious. Cogn. 2011, 20, 1705–1721. [Google Scholar] [CrossRef]
  125. Nyhan, B.; Reifler, J.; Richey, S.; Freed, G.L. Effective Messages in Vaccine Promotion: A Randomized Trial. Pediatrics 2014, 133, e835–e842. [Google Scholar] [CrossRef]
  126. Peter, C.; Koch, T. When debunking scientific myths fails (and when it does not) the backfire effect in the context of journalistic coverage and immediate judgments as prevention strategy. Sci. Commun. 2016, 38, 3–25. [Google Scholar] [CrossRef]
  127. Lewandowsky, S.; Ecker, U.K.; Seifert, C.M.; Schwarz, N.; Cook, J. Misinformation and its correction: Continued influence and successful debiasing. Psycho-Log. Sci. Public Interest 2012, 13, 106–131. [Google Scholar] [CrossRef]
  128. Wood, T.; Porter, E. The Elusive Backfire Effect: Mass Attitudes’ Steadfast Factual Adherence. Politiical Behav. 2019, 41, 135–163. [Google Scholar] [CrossRef]
  129. Ecker, U.K.H.; Lewandowsky, S.; Chadwick, M. Can corrections spread misinformation to new audiences? Testing for the elusive familiarity backfire effect. Cogn. Res. Princ. Implic. 2020, 5, 41. [Google Scholar] [CrossRef]
  130. Swire-Thompson, B.; DeGutis, J.; Lazer, D. Searching for the backfire effect: Measurement and design consider-ations. J. Appl. Res. Mem. Cogn. 2020, 9, 286–299. [Google Scholar] [CrossRef]
  131. Schmid, P.; Betsch, C. Effective strategies for rebutting science denialism in public discussions. Nat. Hum. Behav. 2019, 3, 931–939. [Google Scholar] [CrossRef]
  132. Nadel, L.; Hupbach, A.; Gomez, R.; Newman-Smith, K. Memory formation, consolidation and transformation. Neurosci. Biobehav. Rev. 2012, 36, 1640–1645. [Google Scholar] [CrossRef]
  133. Hupbach, A.; Gomez, R.; Hardt, O.; Nadel, L. Reconsolidation of episodic memories: A subtle reminder triggers integration of new information. Learn. Mem. 2007, 14, 47–53. [Google Scholar] [CrossRef]
  134. Haubrich, J.; Nader, K. Memory reconsolidation. Behavioral Neuroscience of Learning and Memory; Springer: Cham, Switzerland, 2016; pp. 151–176. [Google Scholar]
  135. Pailhes, A.; Rensink, R.A.; Kuhn, G. A psychologically based taxonomy of magicians’ forcing techniques: How magicians influence our choices, and how to use this to study psychological mechanisms. Conscious. Cogn. 2020, 86, 103038. [Google Scholar] [CrossRef] [PubMed]
  136. Olson, J.A.; Amlani, A.A.; Raz, A.; Rensink, R.A. Influencing choice without awareness. Conscious. Cogn. 2015, 37, 225–236. [Google Scholar] [CrossRef]
  137. Shalom, D.E.; Serro, M.G.D.S.; Giaconia, M.; Martinez, L.M.; Rieznik, A.; Sigman, M. Choosing in Freedom or Forced to Choose? Introspective Blindness to Psychological Forcing in Stage-Magic. PLoS ONE 2013, 8, e58254. [Google Scholar] [CrossRef]
  138. Pailhès, A.; Kuhn, G. Influencing choices with conversational primes: How a magic trick unconsciously influences card choices. Proc. Natl. Acad. Sci. USA 2020, 117, 17675–17679. [Google Scholar] [CrossRef]
  139. Cole, G.G. Forcing the issue: Little psychological influence in a magician’s paradigm. Conscious. Cogn. 2020, 84, 103002. [Google Scholar] [CrossRef] [PubMed]
  140. Pailhès, A.; Kumari, S.; Kuhn, G. The magician’s choice: Providing illusory choice and sense of agency with the equivoque forcing technique. J. Exp. Psychol. Gen. 2020, 150, 1358. [Google Scholar] [CrossRef] [PubMed]
  141. Johansson, P.; Hall, L.; Sikström, S. From Change Blindness to Choice Blindness. Psychologia 2008, 51, 142–155. [Google Scholar] [CrossRef]
  142. Hyman, A.D.; Franklyn, D.; Yee, C.; Rahmati, M. Can consumers recognize native advertising: Does it matter. Yale JL Tech. 2017, 19, 77. [Google Scholar]
  143. Zeng, E.; Kohno, T.; Roesner, F. Bad news: Clickbait and deceptive ads on news and misinformation websites. In Proceedings of the Workshop on Technology and Consumer Protection (ConPro), Online, 21 May 2020; IEEE: New York, NY, USA, 2020. [Google Scholar]
  144. Zeng, E.; Wei, M.; Gregersen, T.; Kohno, T.; Roesner, F. Polls, clickbait, and commemorative $2 bills: Problematic political advertising on news and media websites around the 2020 US elections. In Proceedings of the 21st ACM Internet Measurement Conference, Online, 2 November 2021. [Google Scholar]
  145. Keating, D.; Schaul, K.; Shapiro, L. The Facebook Ads Russians Targeted at Different Groups; Washington Post: Washington, DC, USA, 2017. [Google Scholar]
  146. Hazan, J.G. Stop being evil: A proposal for unbiased Google search. Mich. Law Rev. 2012, 111, 789. [Google Scholar]
  147. Fowler, G.A. How does Google’s monopoly hurt you? In Try These Searches; The Washington Post: Washington, DC, USA, 2020. [Google Scholar]
  148. Bennett, W.L. News: The Politics of Illusion; University of Chicago Press: Chicago, IL, USA, 2016. [Google Scholar]
  149. Benham, J. Best Practices for Journalistic Balance: Gatekeeping, Imbalance and the Fake News Era. J. Pract. 2019, 14, 791–811. [Google Scholar] [CrossRef]
  150. Pickard, V. America’s Battle for Media Democracy: The Triumph of Corporate Libertarianism and the Future of Media Reform. Triplec Commun. Capital. Crit. Open Access J. A Glob. Sustain. Inf. Soc. 2014, 12. [Google Scholar] [CrossRef]
  151. Pickard, V. Media failures in the age of Trump. Political Econ. Commun. 2017, 4, 118–122. [Google Scholar]
  152. Stites, T. Layoffs and Cutbacks Lead to a New World of News Deserts. Nieman Lab. 2011. Available online: https://www.niemanlab.org/2011/12/tom-stites-layoffs-and-cutbacks-lead-to-a-new-world-of-news-deserts/ (accessed on 24 May 2022).
  153. Sullivan, M. Ghosting the News: Local Journalism and the Crisis of American Democracyl. Columbia Glob. Rep. 2020. [Google Scholar] [CrossRef]
  154. Carpini, M.X.D.; Entman, R.M. Democracy without Citizens: Media and the Decay of American Politics. Politi-Sci. Q. 1990, 105, 139. [Google Scholar] [CrossRef]
  155. Boykoff, M.T. Flogging a dead norm? Newspaper coverage of anthropogenic climate change in the United States and United Kingdom from 2003 to 2016. Area 2007, 39, 470–481. [Google Scholar] [CrossRef]
  156. Corbett, J.B.; Durfee, J.L. Testing public (un) certainty of science: Media representations of global warming. Sci. Commun. 2004, 26, 129–151. [Google Scholar] [CrossRef]
  157. Cunningham, B. Re-thinking objectivity. Columbia J. Rev. 2003, 42, 24–32. [Google Scholar]
  158. Boykoff, M.T.; Boykoff, J.M. Climate change and journalistic norms: A case-study of US mass-media coverage. Geoforum 2007, 38, 1190–1204. [Google Scholar] [CrossRef]
  159. Patterson, T.E. Informing the News: The need for Knowledge-Based Journalism; Vintage: New York, NY, USA, 2013. [Google Scholar]
  160. Rosenstiel, T. Political Polling and the New Media Culture: A Case of More Being Less. Public Opin. Q. 2005, 69, 698–715. [Google Scholar] [CrossRef]
  161. Dunaway, J.L.; Davis, N.T.; Padgett, J.; Scholl, R.M. Objectivity and Information Bias in Campaign News. J. Commun. 2015, 65, 770–792. [Google Scholar] [CrossRef]
  162. Ross, B.; Jung, A.; Heisel, J.; Stieglitz, S. Fake news on social media: The (in) effectiveness of warning messages. In Proceedings of the 39th International Conference on Information Systems (ICIS 2018), San Francisco, CA, USA, 13–16 December 2018. [Google Scholar]
  163. Hua, Y.; Namavari, A.; Cheng, K.; Naaman, M.; Ristenpart, T. Adapting Security Warnings to Counter Online Disinformation. In Proceedings of the 30th {USENIX} Security Symposium ({USENIX} Security 21), Online, 11–13 August 2021. [Google Scholar]
  164. Gao, M.; Xiao, Z.; Karahalios, K.; Fu, W.T. To label or not to label: The effect of stance and credibility labels on readers’ selection and perception of news articles. Proc. ACM Hum.-Comput. Interact. 2018, 2, 1–16. [Google Scholar] [CrossRef]
  165. Pennycook, G.; Bear, A.; Collins, E.T.; Rand, D.G. The Implied Truth Effect: Attaching Warnings to a Subset of Fake News Headlines Increases Perceived Accuracy of Headlines without Warnings. Manag. Sci. 2020, 66, 4944–4957. [Google Scholar] [CrossRef]
  166. Seo, H.; Xiong, A.; Lee, D. Trust It or Not: Effects of Machine-Learning Warnings in Helping Individuals Mitigate Misinformation. In Proceedings of the 10th ACM Conference on Web Science, Online, 26 June 2019. [Google Scholar]
  167. Wu, M.; Miller, R.C.; Garfinkel, S.L. Do security toolbars actually prevent phishing attacks? In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. Association for Computing Machinery, Montréal, QC, Canada, 22–27 April 2006; pp. 601–610. [Google Scholar]
  168. Egelman, S.; Cranor, L.F.; Hong, J. You’ve been warned: An empirical study of the effectiveness of web browser phishing warnings. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Online, 6 April 2008; Association for Computing Machinery: Florence, Italy, 2008; pp. 1065–1074. [Google Scholar]
  169. Jhaver, S.; Boylston, C.; Yang, D.; Bruckman, A. Evaluating the Effectiveness of Deplatforming as a Moderation Strategy on Twitter. Proc. ACM Human-Comput. Interact. 2021, 5, 1–30. [Google Scholar] [CrossRef]
  170. Saltz, E.; Leibowicz, C.R.; Wardle, C. Encounters with visual misinformation and labels across platforms: An interview and diary study to inform ecosystem approaches to misinformation interventions. In Proceedings of the Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems, Online, 8 May 2021. [Google Scholar]
  171. Maras, M.-H.; Alexandrou, A. Determining authenticity of video evidence in the age of artificial intelligence and in the wake of Deepfake videos. Int. J. Évid. Proof 2018, 23, 255–262. [Google Scholar] [CrossRef]
  172. Kaplan, A. Artificial Intelligence, Social Media, and Fake News: Is This the End of Democracy? Istanbul University Press: Istanbul, Turkey, 2020; p. 149. [Google Scholar]
  173. Fletcher, J. Deepfakes, artificial intelligence, and some kind of dystopia: The new faces of online post-fact perfor-mance. Theatre J. 2018, 70, 455–471. [Google Scholar] [CrossRef]
  174. Frolov, S.; Hinz, T.; Raue, F.; Hees, J.; Dengel, A. Adversarial text-to-image synthesis: A review. Neural Netw. 2021, 144, 187–209. [Google Scholar] [CrossRef]
  175. National Academies of Sciences, Engineering, and Medicine; Division of Behavioral and Social Sciences and Education; Committee on the Science of Science Communication. Communicating Science Effectively: A Research Agenda; The National Academies Press: Washington, DC, USA, 2017; p. 152. [Google Scholar]
  176. Attanasio, R. Communicating environmental sciences: Public discourse and policy development. Integr. Environ. Assess. Manag. 2018, 14, 167–168. [Google Scholar] [CrossRef]
  177. Martinez-Conde, S.; Macknik, S.L. Opinion: Finding the plot in science storytelling in hopes of enhancing science communication. Proc. Natl. Acad. Sci. USA 2017, 114, 8127–8129. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Attentional allocation modulates the perceived saliency and brightness of visual objects. Fixate your gaze on the central blue dot while directing your attention to one of the three overlapping circles. Whichever circle you focus your attention on will appear brighter than the other two (unattended) circles, and it will also seem to be positioned in front of them. Modified from [34].
Figure 1. Attentional allocation modulates the perceived saliency and brightness of visual objects. Fixate your gaze on the central blue dot while directing your attention to one of the three overlapping circles. Whichever circle you focus your attention on will appear brighter than the other two (unattended) circles, and it will also seem to be positioned in front of them. Modified from [34].
Publications 10 00033 g001
Figure 2. Magician Teller’s performance of The Miser’s Dream trick at the 2007 “Magic of Consciousness” Symposium [37]. The Miser’s Dream is a classic magic routine in which myriad coins seem to appear where previously there were none. The magician uses auditory cues with precise timing to make audiences believe that many coins have dropped into the metal bucket when that was not the case. Teller relies on the audience’s multi-sensory integration processes, using his body language, gaze direction, and sleight of hand techniques to misdirect spectators. From [16].
Figure 2. Magician Teller’s performance of The Miser’s Dream trick at the 2007 “Magic of Consciousness” Symposium [37]. The Miser’s Dream is a classic magic routine in which myriad coins seem to appear where previously there were none. The magician uses auditory cues with precise timing to make audiences believe that many coins have dropped into the metal bucket when that was not the case. Teller relies on the audience’s multi-sensory integration processes, using his body language, gaze direction, and sleight of hand techniques to misdirect spectators. From [16].
Publications 10 00033 g002
Figure 3. In Boger et al. [84], participants viewed a series of plots while they performed a task that required them to focus on one aspect of each display. In one example, participants were tasked with identifying which quadrants of the plots contained blue Xs (i.e., upper left and lower right in the left panel). In some instances, other symbols on the plot formed the contour of a dinosaur (right panel). 93% of participants failed to notice any unusual patterns in the data. When explicitly asked if they had noticed a dinosaur, they reported that they had not. From [84].
Figure 3. In Boger et al. [84], participants viewed a series of plots while they performed a task that required them to focus on one aspect of each display. In one example, participants were tasked with identifying which quadrants of the plots contained blue Xs (i.e., upper left and lower right in the left panel). In some instances, other symbols on the plot formed the contour of a dinosaur (right panel). 93% of participants failed to notice any unusual patterns in the data. When explicitly asked if they had noticed a dinosaur, they reported that they had not. From [84].
Publications 10 00033 g003
Figure 4. This visualization of fire hotspots in Australia may appear lifelike, but it is not a photograph and does not accurately reflect the size or number of fires burning at any one time. Australia is Burning/A 3D Visualization (2019) (available at https://anthonyhearsey.com/australia-is-burning-a-3d-visualisation, accessed on 24 August 2022) by Anthony Hearsey. Reproduction courtesy of the artist.
Figure 4. This visualization of fire hotspots in Australia may appear lifelike, but it is not a photograph and does not accurately reflect the size or number of fires burning at any one time. Australia is Burning/A 3D Visualization (2019) (available at https://anthonyhearsey.com/australia-is-burning-a-3d-visualisation, accessed on 24 August 2022) by Anthony Hearsey. Reproduction courtesy of the artist.
Publications 10 00033 g004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Alexander, R.G.; Macknik, S.L.; Martinez-Conde, S. What the Neuroscience and Psychology of Magic Reveal about Misinformation. Publications 2022, 10, 33. https://doi.org/10.3390/publications10040033

AMA Style

Alexander RG, Macknik SL, Martinez-Conde S. What the Neuroscience and Psychology of Magic Reveal about Misinformation. Publications. 2022; 10(4):33. https://doi.org/10.3390/publications10040033

Chicago/Turabian Style

Alexander, Robert G., Stephen L. Macknik, and Susana Martinez-Conde. 2022. "What the Neuroscience and Psychology of Magic Reveal about Misinformation" Publications 10, no. 4: 33. https://doi.org/10.3390/publications10040033

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop