A week before race riots broke out across England and Northern Ireland, entrepreneurial trolls operating from Africa launched a Facebook page called "Patriots of Britain".
They started publishing anti-immigrant misinformation from the moment it was created in late July and boosted its content with fake followers from Vietnam and India, with the probable aim of building an audience that they could then drive to websites featuring ads.
After the fatal stabbing of three young girls in the seaside town of Southport on July 29, false online posts wrongly identifying the suspected killer as an Islamist migrant triggered widespread violence.
It was here that the page's administrators saw an opportunity.
The "Patriots of Britain" page began spreading rumours of "masked gangs hunting down people in Britain" and sharing call-outs for people to attend protests organised by anti-immigrant groups.
On August 2, it posted a picture of a smiling cartoon man with a knife protruding from his chest with the caption "New logo of Southport".
When ABC NEWS Verify flagged the unauthentic behaviour on this page with Meta, the parent company of Facebook, it was quickly taken down.
The short life span of "Patriots of Britain" meant its impact was minimal, but similar pages spreading potentially dangerous falsehoods might soon fall through the cracks.
ABC NEWS Verify was able to detect the page with a Meta-owned product called CrowdTangle. Today, after years of the tool bringing it public humiliation, Facebook will be killing it off for good.
What is CrowdTangle and why is it being closed?
Meta purchased CrowdTangle in 2016 when it was mostly used as a marketing and social media management tool.
The dashboard — effectively a super-charged social media search engine — offered a free and rare way to peek inside Facebook and Instagram and track what was trending and who was responsible. It was information that was not readily available on YouTube and X (formerly Twitter).
Most of Facebook's 3 billion monthly active users have never used it and will not even notice at first when it is gone. However, its imminent closure has sent academics, journalists, and civic groups who rely on it to expose disinformation into a panic.
Reporting from the US suggested Meta had slowly been starving CrowdTangle of resources since 2021 and had planned to deactivate it in 2022, but was forced to keep it operational due to the EU Digital Services Act requiring social media companies to allow researchers access to platform data.
Meta is rolling out a replacement called the Meta Content Library (MCL) but it is restricting access to approved academics and non-profit researchers.
The tech giant has defied high-level calls, including from the US Congress and the European Commission, to keep CrowdTangle alive and will officially deactivate it on August 14.
Australian eSafety Commissioner Julie Inman Grant, who has been pushing big tech for more transparency, lamented the loss of a near real-time social monitoring tool ahead of the US presidential race in November and the next Australian federal election.
"They're referring it to it as the deepfake election, so we may see totally new vectors of harm and spread of mis- and disinformation," the commissioner said.
"If we're killing social media monitoring tools that are giving us just a degree of transparency, then it's going to limit our ability to be able to intercede and stop those harmful activities from undermining democratic processes."
The transparency tool has been a thorn in Meta's side for years.
CrowdTangle's data has repeatedly been used by reporters and academics to illuminate how the top trending content on Facebook is often hyper-political or rife with misinformation.
In 2020, The New York Times relied on the tool to reveal how Plandemic — a conspiracy theory film featuring testimony from a discredited anti-vaccine activist — was one of the top-performing posts throughout COVID-19.
The ABC has relied on CrowdTangle to uncover a coordinated network of African-run Facebook pages attempting to influence the outcome of this year's UK general election. We also used it to show how misinformation and racist conspiracy theories dominated last year's Voice to Parliament referendum.
ABC NEWS Verify understands that some members of the UK government have also used CrowdTangle. When it became clear that incendiary posts were fanning the flames of the race riots, the government contacted the major social media companies.
"I have been clear it is unacceptable that people are using social media to cause damage, distress and destruction in our communities," UK technology secretary Peter Kyle told ABC NEWS Verify.
"I've had useful meetings with TikTok, Meta, Google and X, to make clear their responsibility to continue to work with us to stop the spread of hateful misinformation and incitement.
"There is a significant amount of content circulating that platforms need to be dealing with at pace."
A standard set by the 'Musk effect'
Hundreds of people have been arrested for their involvement in the riots and some offenders have already been handed jail terms.
Although UK police named 17-year-old Cardiff-born Axel Rudakubana as the suspect behind the Southport stabbings, social media posts attacking Muslim migrants have continued to proliferate online.
British Prime Minister Keir Starmer has also been under pressure to launch an inquiry into social media platforms' role in turbo-charging the riots.
Despite pressure from governments worldwide for transparency, the major tech companies are sliding into further obfuscation.
"I call this the Musk effect," said Marc Owen Jones, author and digital disinformation expert from Qatar's Hamad Bin Khalifa University.
Elon Musk, the owner of X, has been accused of personally agitating the unrest after publishing on the platform that "civil war is inevitable" in response to a post blaming the violence on mass migration and "open borders".
"He closed Twitter's (now X) trust and safety team, which was responsible for combating disinformation, hate speech and platform manipulation," Dr Jones said.
"He fired the whole team to save money and because of his principles … if Musk does something and gets away with it, then other companies can follow suit."
AI Jesus is taking over Facebook
It's not just pages spouting violent rhetoric that could be missed without CrowdTangle. The recent AI boom has caused Facebook and Instagram to be flooded with bizarre and disturbing content.
Over the past year, AI-generated images have become one of the most popular and most shared types of content on Meta platforms.
Searchable phrases in the CrowdTangle dashboard such as "why don't pictures like this ever trend" or "#photographychallenge" have revealed a surreal religious subculture.
AI-generated images of Jesus Christ appearing in white fluffy clouds or performing heroic feats are garnering up to 1.6 million reactions and shares, putting them on par with the most popular posts by Taylor Swift.
Some of the content generated in this AI-generated blitz is exploitative and deeply rooted in racial stereotypes, including those showing sickly and deformed children from developing countries writhing in pain.
Much of this content, like with the "Patriots of Britain" page, is boosted through unauthentic activity and likely operated by spammers and scammers, according to the Stanford Internet Observatory.
Tech publisher 404 media found the mass production of these images was in part being fuelled by Meta's own content creator program, which pays users it invites to the program for posts that perform well.
Meta did not respond directly to ABC questions about why it was shutting down CrowdTangle, but instead directed us to a blog post highlighting its other research tool, Meta Content Library (MCL).
The tech giant bills its CrowdTangle replacement as an improved dashboard that it will continue to support. However, MCL will only be available to vetted academics and non-profit institutions — a limitation that excludes almost all news organisations and reporters from access.
Responses to MCL's functionality from those who have used it have been mixed.
Queensland University of Technology disinformation expert Axel Bruns said MCL was a "fairly solid product" that addressed some of the gaps in CrowdTangle and allowed researchers to examine content at a granular level.
However, Professor Bruns told ABC NEWS Verify that restricting access to the replacement puts the onus of flagging disinformation and harmful content almost entirely on academics.
"We need more, not fewer, organisations that do this sort of work, because clearly there is a lot of mis- and disinformation or influence operations out there," he said.
Dr Jones said Meta did better than its competitors at responding to bad actors manipulating crises to spread hate speech, but the closure of a publicly accessible transparency tool meant "we're going to lose a pressure point that keeps them honest".
"Without CrowdTangle, disinformation and misinformation is definitely going to get worse," Dr Jones said.
"Without us being able to provide the evidence saying … this looks like manipulation, Facebook can just ignore us, deny that there's any problem and we don't have the ability to question that."