As long as there is news, there is fake news and as long as there is fake news, you can guarantee that theyâll be conspiracy theories. The anguish and uncertainty of crises, especially societal crises, create an incentive to understand whatâs happening, amplifying the spread of conspiracy theories. Many times, these theories are based on small shreds of truth, centred on the dichotomy between those in power and the rest of us in the infinite battle of good versus evil.
2020, a tale of fact versus fiction
Since the COVID-19 pandemic and the Black Lives Matter protests that erupted after the murder of George Floyd, conspiracy theorists are having a field day, as for most, the world right now feels just too dystopic to be the result of a series of coincidences. Whenever something huge happens, think 9/11, the 1969 moon landing or Princess Dianaâs 1997 death, thereâs a trail of conspiracy theories materialising after them fuelled on the idea that âthere must be something more to this.â The internet and particularly social media have made these theories inescapable and to some, appear scarily legit. What tech is being developed to combat this?
Fake News Guard are using tech to debunk fake news. Stating how âdisinformation can be produced faster than it is fact-checked,â the company use tools such as âa search engineâ and âa track record feature, to show which newspapers and publications have put out fake stories in the pastâ as well as âan algorithm to help detect fake news in an unsupervised mannerâ for both individuals and organisations.
Similarly, Untrue News is an open-source search engine that only displays fake news results when topics are searched. It does this using âboth automated and semi-automated natural language processing techniques to crawl and identify false and misleading news articlesâ in multiple languages. The startup relies on the International Fact-Checking Network from the nonprofit Poynter Institute to separate accurate news from fake news. By solely identifying false news, the company âfills a gap left by multipurpose toolsâ and âuses both automated and semi-automated natural language processing techniques to crawl and identify false and misleading news articles.â
Factmata uses tech to âextract claims being made by actors online [in the form of] journalists, Twitter users or Facebook profiles and highlights the common themes or narratives evolving,â says CEO of the company, Dhruv Ghulati. âWe also have separate algorithms which answer the question, âare these likely to be threatening, hateful, sexist, or propaganda content,â and should they be removed from a network?â Factmataâs tech is also designed for both businesses and organisations with a focus on selling the tech to businesses.
Speaking on the current requirement for this technology, Dhruv states âwe are going into a world where information will be the new battlefield where warfare takes place, from spreading lies, confusing populations, altering voter behaviour to changing stock prices.â
Faheem Nasir, Head of Social Media at Logically explains how the company uses AI and algorithms âdesigned to identify the accuracy and credibility of content using a three-pronged approach.â This approach utilises networks, metadata and content. The companyâs tech is âbuilt on a set of modular processes, capable of analysing boundless amounts of dataâ that helps turn data into insight through machine learning algorithms.
âOnce weâve ingested and analysed the data, our clustering algorithms highlight similarities and disparities between similar data sets, such as articles from different publishers covering the same subject,â Faheem explains. They also offer a free app that âgathers credible news stories from across the political spectrum so that users can get all the facts, evaluate biases within news content and come to their own conclusions based on a plurality of sources.â
But what about freedom of speech?
It gets complicated though, because as much as conspiracy theories seem crazy, many people spreading these theories online actually believe them. While actors attempting to infiltrate peopleâs thinking and peddle political propaganda for ulterior motives are awful, there are also fears over the censoring of any information online. If online speech begins to be censored, how far will it go and who decides what is and what isnât allowed to be said online?
Just before the pandemic started to kick off in the West, Time Magazine wrote an article stating that âfreedom [of speech] is often abused, but â truly â whose fault is that? Is it Twitterâs fault if I lie about the news? It is my responsibility to exercise my rights responsibly.â But we canât trust people to act responsibly, thatâs the whole point. Treading the line between censoring dangerous and offensive content and allowing free speech is a dangerous road to walk on and one that is turning into an increasingly loaded debate.
Dhruv emphasizes how the onus is on the individual to make sure they are only sharing fact-checked, truthful information. âSpreading lies or not doing your research on facts has negative externalities on others, as does being abusive and sexist. This discourse needs to be monitored if it is in the public sphere communication and not a private channel.â
Mattia, CEO of Fake News Guard mentions how asking big tech companies to âintroduce editorial control is a blow to the very core of their business model.â He goes on to say how âthey are now trying to do this with algorithms and various techniques, but the technology is not fully there yet.â His start-up offers âtools that can help people to stay informedâ targeted at regular internet users. They are also creating âa data-driven hubâ that users can come to when they have doubts about a news story.
Likewise, Faheem mentions how âfor many social media platforms, effectively fighting fake news is counterproductive to their business models. Whether it be through paid ads or someone sharing a suspicious post, social media and algorithm-based platforms are not only susceptible to the propagation of misinformation, they often profit from it.â
âThe idea for Logically came after seeing the political polarisation that surrounded the Brexit and US Presidential campaigns, with both campaigns using highly targeted social media advertising to aid them. It was clear that the public both needed and wanted access to credible information they could reliably use to inform their decisions.â He continues. âFreedom of speech and expression are fundamental rights which should always be respected, but when it is to the detriment of public wellbeing, it is paramount that we develop comprehensive solutions to these complex problems.â
âIt is known that there are vested-interest groups whose jobs are to produce fake news in order to push their agenda, be it political, environmental or else. It becomes excruciatingly difficult to tell truths from falsehoods,â says Dr. Vinicius Wolosyzn from Untrue News.
âLook at the Brazilian election in 2018 that elected far-right Jair Bolsonaro as president. The far-right had huge teams (being investigated to this day) producing fake news non-stop, some of which was too ridiculous to be believed by most people. But it worked! And now Brazil (and the world if you consider the deforestation of the Amazon) is paying the price for not having acted early enough to avoid the spread of fake news,â he continues.
âFreedom of speech is not freedom to be racist, a bigot, misogynist; or freedom to spread lies and falsehoods. Untrue.news aims to empower people who don't want to be deceived. We believe that only in an environment of truth, true freedom of speech can be fully exercised,â he states.
How is big tech responding to this?
Recently, Mark Zuckerberg stated that he doesnât think that âFacebook should be the arbiter of truth of everything that people say online.â Whatsapp, owned by Facebook, set up a COVID-19 fact-checking bot in March to fight fake news being spread about the pandemic. The messaging service also imposed restrictions on forwarding fake news, with users only being allowed to forward a message once if that message has been previously forwarded over 5 times.
Instagram, also owned by Facebook, has recently made efforts to remove COVID-19 accounts from recommendations and the explore page as well as downranking false information from stories unless itâs posted from credible health organisations.
Twitter announced that it will remove tweets that promote harmful information relating to COVID-19 and are trying to stop users tweeting articles they have not read. Last year, Twitter acquired Fabula AI, a startup using deep learning to identify online fake news and disinformation.
Recently, Google announced that they would be spending $6.5M on fighting fake news relating to COVID-19. The money will be spent on utilising global fact-checkers and nonprofit organisations dedicated to combating misinformation online.
Finally, YouTube which is owned by Google and arguably the home of conspiracy theories has one in four videos posted that contain fake news. They have also been accused of âactively promoting misinformationâ by a senior UK politician. Earlier this year, YouTube announced that they would be implementing machine learning to recommend content from trusted news sources to users and use human moderators to tackle fake video content. Susan Wojcicki, CEO of YouTube mentioned that removing certain creators from the platform would be going too far and is instead opting to provide links under videos displaying accurate information on the topic at hand.
The battle for a free, fair and safe internetÂ
Striking a balance between keeping the internet the intriguing and opinionated hotbed that it is while tackling fake and harmful misinformation will be one of the greatest challenges for big tech over the next decade.Â
Scandals such as Cambridge Analytica highlight that this issue requires action as it fundamentally affects how we think and shapes the world that we live in. Though, we should strive to retain the essence of social media which is self-expression while holding big companies and advertisers spreading false stories accountable.
Speaking on the urgency for services such as Logically, Faheem states âwe rely on information to make meaningful decisions that affect our lives, but the nature of the internet means that flawed news reaches more people faster than ever before. The effects of misinformation have shaken democracies, provoked public health epidemics and induced lethal violence.â
âWe strive to nurture a fair and free internet, to take people out of their divisive filter bubbles and echo chambers and help make better sense of the world around us.â