The law has made it a crime for social media platforms not to remove "abhorrent violent material" quickly. The crime would be punishable by up to three years in prison and a fine of $10.5 million Australian (around $10 million Cdn), or 10 per cent of the platform's annual turnover — whichever is larger.
Abhorrent violent material is defined as acts of terrorism, murder, attempted murder, torture, rape and kidnapping. The material must be recorded by the perpetrator or an accomplice for the law to apply. Platforms anywhere in the world would face fines of up to $840,000 Australian (around $799,000 Cdn) if they fail to notify Australian Federal Police if they are aware their service was streaming "abhorrent violent conduct" occurring in Australia.
'Clumsy and flawed'
Dreyfus described the bill as "clumsy and flawed," and the timetable to pass it as "ridiculous." Labor first saw the legislation late Monday.
The bill could potentially undermine Australia's security co-operation with the United States by requiring U.S. internet providers to share content data with Australian Federal Police, in breach of U.S. law, Dreyfus said.
The Digital Industry Group Inc. — an association representing the digital industry in Australia, including Facebook, Google and Twitter — said taking down abhorrent content was a "highly complex problem" that required consultation with a range of experts, which the government had not done.
"This law, which was conceived and passed in five days without any meaningful consultation, does nothing to address hate speech, which was the fundamental motivation for the tragic Christchurch terrorist attacks," the group's managing director Sunita Bose said in a statement.
"This creates a strict internet intermediary liability regime that is out of step with the notice-and-takedown regimes in Europe and the United States, and is therefore bad for internet users as it encourages companies to proactively surveil the vast volumes of user-generated content being uploaded at any given minute," Bose added.
Arthur Moses, president of the Australian Law Council, the nation's top lawyers group, said the law could lead to media censorship and prevent whistleblowers from using social media to shine a light on atrocities because of social media companies' fear of prosecution.
"Media freedom and whistleblowing of atrocities here and overseas have been put at risk by the ill-informed live stream laws passed by the federal Parliament," Moses said.
The penalties would be "bad for certainty and bad for business," which could scare off online business investment in Australia, Moses said.
Australian Industry Group chief executive Innes Willox, a leading business advocate, said more time was required to ensure the law did not unnecessarily impinge on existing fundamental media rights and freedoms.
Scott Farquhar, co-founder of the Sydney-based software company Atlassian, predicted job losses in the technology industry.
"As of today, any person working at any company [globally] that allows users to upload videos or images could go to jail," Farquhar tweeted. "Guilty until proven innocent."
Fergus Hanson, head of the International Cyber Policy Centre at the Australian Strategic Policy Institute, saw problems in the legislation's definitions, including how much time a company had to "expeditiously" remove offensive material.
Facebook live streamed the Christchurch massacre for 17 minutes without interruption before reacting. Facebook said it removed 1.5 million videos of the shootings during the first 24 hours afterward.
It was allegedly filmed by Brenton Harrison Tarrant, 28, the suspected gunman who faces 50 charges of murder and 39 charges of attempted murder and is scheduled to appear in court Friday.
Executives of Facebook, Google, Twitter, internet service providers and Australian phone companies met Prime Minister Scott Morrison and three ministers last week to discuss social media regulation. Communications Minister Mitch Fifield said Facebook "did not present any immediate solutions to the issues arising out of the horror that occurred in Christchurch."
Facebook did not immediately respond to a request for comment on Thursday. CEO Mark Zuckerberg used an op-ed in the Washington Post last week to invite a more active role by governments and regulators to deal with harmful online content.
"The rules governing the internet allowed a generation of entrepreneurs to build services that changed the world and created a lot of value in people's lives," Zuckerberg wrote. "It's time to update these rules to define clear responsibilities for people, companies and governments going forward."
Morrison wants to take the Australian law to a Group of 20 countries forum as a model for holding social media companies to account.
New Zealand's Justice Minister Andrew Little said his government has also made a commitment to review the role of social media and the obligations of the companies that provide the platforms. He said he has asked officials to look at the effectiveness of current hate speech laws and whether there were gaps that need to be filled.
Little said he didn't see any irony in that people were watching hearings into a bill that would place new restrictions on guns in real time on Facebook, the same platform the shooter used to broadcast the massacre.
"There's a world of difference, I think, between the exercise of a democratic function and a democratic institution like a national parliament, and some of the more toxic stuff that you see put out by individuals," he said.