Wikipedia bots more like humans than expected: Report
Bots behave differently in culturally distinct online environments, say researchers
Latest News »
- Deals Buzz: Softbank plans to invest in Uber
- Market live: Sensex, Nifty open higher, shares of Axis Bank, Airtel fall 1% each
- CPM flags down 3rd Rajya Sabha term for Sitaram Yechury
- Rupee weakens marginally against US dollar in early trade
- US House votes to impose sanctions on Russia, Iran and North Korea
London: ‘Benevolent bots’ or software robots designed to improve articles on Wikipedia sometimes have online ‘fights’ over content that can continue for years, say scientists who warn that artificial intelligence systems may behave more like humans than expected.
Editing bots on Wikipedia undo vandalism, enforce bans, check spelling, create links and import content automatically, whereas other bots (which are non-editing) can mine data, identify data or identify copyright infringements.
Researchers from the University of Oxford and the Alan Turing Institute in the UK analysed how much they disrupted Wikipedia, observing how they interacted on 13 different language editions over 10 years (from 2001 to 2010).
They found that bots interacted with one another, whether or not this was by design, and it led to unpredictable consequences.
Researchers said that bots are more like humans than you might expect. Bots appear to behave differently in culturally distinct online environments.
The findings are a warning to those using artificial intelligence for building autonomous vehicles, cyber security systems or for managing social media. We may have to devote more attention to bots’ diverse social life and their different cultures, researchers said.
The research found that although the online world has become an ecosystem of bots, our knowledge of how they interact with each other is still rather poor.
Although bots are automatons that do not have the capacity for emotions, bot to bot interactions are unpredictable and act in distinctive ways.
Researchers found that German editions of Wikipedia had fewest conflicts between bots, with each undoing another’s edits 24 times, on average, over 10 years.
This shows relative efficiency, when compared with bots on the Portuguese Wikipedia edition, which undid another bot’s edits 185 times, on average, over ten years, researchers said. Bots on English Wikipedia undid another bot’s work 105 times, on average, over ten years, three times the rate of human reverts, they said.
The findings show that even simple autonomous algorithms can produce complex interactions that result in unintended consequences—‘sterile fights’ that may continue for years, or reach deadlock in some cases. While bots constitute a tiny proportion (0.1%) of Wikipedia editors, they stand behind a significant proportion of all edits.
Although such conflicts represent a small proportion of bots’ overall editorial activity, the findings are significant in highlighting their unpredictability and complexity.
“We find that bots behave differently in different cultural environments and their conflicts are also very different to the ones between human editors,” said Milena Tsvetkova, from the Oxford Internet Institute. “This has implications not only for how we design artificial agents but also for how we study them. We need more research into the sociology of bots,” said Tsvetkova.