The Internet has revolutionised lives — whether in terms of working, finding information or entertainment, connecting with others, or shopping. The online world has made things easier and opened up new opportunities. Similarly, the internet presents individuals and societies with major challenges: The underlying technologies do not necessarily serve users’ best interests.
“We’re interested in questions such as: How can we create online environments that respect human autonomy and promote truth? And what can people themselves do to avoid being misled and manipulated?”, says Anastasia Kozyreva, lead author and researcher at the Center for Adaptive Rationality at the Max Planck Institute for Human Development.
The research team began by examining the differences between the online and offline worlds, and identified four challenges:
- User behaviour is influenced by manipulative choice architectures. These “dark patterns” steer users toward unintended behaviours, including advertising that blends into the content or navigation of a page to generate more clicks, or confusing privacy settings that prompt people to share more information than they really want to.
- The information presented by AI-powered information architectures is not neutral. It is personalised on the basis of the data collected from users. This means that two people who enter the same term into a search engine will probably be shown different results. That can be helpful when, for example, looking up a restaurant and the search engine displays hits in local neighbourhood at the top of the list, rather than a restaurant with the same name on the other side of the world. However, when shown news or political content solely on the basis of preferences, there is a risk of users being in a filter bubble where there is no exposure to other opinions.
- The research team sees false and misleading information as another challenge for people online. Videos and posts propagating conspiracy theories and unverified rumours can spread rapidly through social media, causing real harm. For example, people may decide not to get vaccinated due to misinformation about vaccines, putting themselves and others at risk.
- Distracting online environments constantly seek to attract users’ attention — whether by means of push notifications, flashing displays, pop-up ads, or constantly updated content. The aim is to capture and hold users’ attention for as long as possible. That is the very basis of the business models of internet platforms. Users may find themselves spending more time on screens than intended — with no real benefit and at the cost of attention for other things.
Taking a behavioural science perspective, the researchers propose specific interventions to address the four challenges. They suggest that “boosting tools” can be used to train new competencies and enable better, more autonomous decisions in the online world.
Self-nudging is one of the cognitive tools that people can use to create “healthier” choice and information environments for themselves. Self-nudging empowers people to set up their digital environment in the way that works best for them.
This might involve turning off notifications from apps or rearranging one’s smartphone home screen so that only useful apps are displayed — such as the calendar, camera, and maps — along with meditation and weather apps. Everything that is overly distracting, such as social media and games, is better tucked away in folders. The researchers also recommend that users consciously set time limits on their social media use.
“The digital world is full of traps,” says Ralph Hertwig, Director of the Center for Adaptive Rationality at the Max Planck Institute for Human Development. “But we can take steps to avoid falling into them. In the same way as we might hide our chocolate stash at the back of the cupboard and put a bowl of apples on the table, we can turn off notifications from apps that permanently demand our attention. Out of sight is out of mind – whether in real life or in the digital world.”
In the same habit as looking right and left before crossing a street, users should make a habit of asking certain questions to evaluate the content we encounter online.
Questions such as: What is the origin of the information? Which sources are cited? Can I find similar content on reputable websites? This approach can boost users’ competence in evaluating the reliability of online information. Internet platforms care also capable of helping users assess content — for example, by displaying decision trees that remind users to check the source and the facts before sharing content.
Generally, policymakers also need to consider putting in place stronger regulatory measures to ensure that internet users retain control over the digital environment and their personal data — for example, through default privacy settings. Lastly, the smart and self-determined use of digital technologies needs to be taught in both school and adult education. The earlier, the better.
The researchers emphasize that none of the interventions they propose can singlehandedly counter online manipulation or prevent the spread of misinformation.
“It will take a combination of smart cognitive tools, early media literacy education, and a regulatory framework that limits the power of commercial interests to hijack people’s attention to make the online world a more democratic and truthful place,” says Stephan Lewandowsky, professor of cognitive psychology at the University of Bristol.