By Kelvin Chan, Business Writer
LONDON (AP) --Proposed British rules aimed at cracking down on harmful online content should be beefed up with tougher measures like making it illegal to send unsolicited graphic images, requiring porn sites to ensure children can't gain access and moving faster to hold tech executives criminally liable for failing to uphold the regulations, lawmakers said in a new report.
The committee of lawmakers recommended a series of major changes to the U.K. government's draft online safety bill early Tuesday that would make digital and social media companies more responsible for protecting users from child abuse, racist content and other harmful material found on their platforms.
The proposals by Prime Minister Boris Johnson's government and similar rules that the European Union is working on underline how Europe is at the vanguard of the global movement to rein in the power of digital giants like Google and Facebook parent Meta Platforms.
The committee of British lawmakers is scrutinizing the draft bill to offer recommendations on how the government can improve it before it's presented to Parliament next year for approval.
"The era of self-regulation for big tech has come to an end," committee chairman Damian Collins said. "The companies are clearly responsible for services they have designed and profit from, and need to be held to account for the decisions they make."
The committee has been hearing evidence from tech executives, researchers and whistleblowers such as former Facebook data scientist Frances Haugen, whose revelations alleging that the company put profits ahead of safety have galvanized legislative and regulatory efforts around the world to clamp down on online hate speech and misinformation.
Under the proposed rules, the U.K.'s communications regulator, Ofcom, would be appointed to investigate internet companies for any breaches, with the power to impose fines of up 18 million pounds ($24 million) or 10% of a company's annual revenue, whichever is higher.
Criminal offenses should be drawn up to cover new online crimes such as cyberflashing — sending someone unsolicited graphic images — or sending flashing images to someone with epilepsy with the intention of causing a seizure, the report said.
Internet companies would have to follow codes of conduct on areas like child exploitation and terrorism and would be required to carry out risk assessments for algorithms recommending videos and other content to users that could lead them down "rabbit holes" filled with false information.
To protect children from accessing online porn, websites should use "age assurance" technology to ensure users are old enough to be on the site, the report said. To ease concerns it would result in a data grab by big tech companies dominating the verification systems, the report advised Ofcom to set standards to protect privacy and restrict the amount of information they can collect.
One of the more controversial parts of the British government's proposal is a provision holding senior managers at internet companies criminally liable for failing to give regulators information needed to assess whether they're complying with the rules, which would kick in after a two-year review period.
That's raised concerns tech companies would use the time to stall instead of complying, spurring digital minister Nadine Dorries to vow to make it happen "under a much shorter timeframe" of up to six months — something the committee backed.
However, digital rights groups and tech companies have warned that the provision could have unintended consequences. Google said it would be an incentive for companies to use automated systems to take down content "at scale" to protect themselves from prosecution.
Authoritarian countries could look to the U.K. rules as justification for threatening staff to get political speech and journalistic content taken down that their leaders don't like, Twitter's senior director of global public policy strategy, Nick Pickles, told the committee in October.
The government has two months to respond to the committee's report before presenting it to Parliament for approval.
South Korea fines Meta $15 million for illegally collecting information on Facebook users
South Korea's privacy watchdog on Tuesday fined social media company Meta 21.6 billion won ($15 million) for illegally collecting sensitive personal information from Facebook users, including data about their political views and sexual orientation, and sharing it with thousands of advertisers.
It was the latest in a series of penalties against Meta by South Korean authorities in recent years as they increase their scrutiny of how the company, which also owns Instagram and WhatsApp, handles private information.
Following a four-year investigation, South Korea's Personal Information Protection Commission concluded that Meta unlawfully collected sensitive information about around 980,000 Facebook users, including their religion, political views and whether they were in same-sex unions, from July 2018 to March 2022.
It said the company shared the data with around 4,000 advertisers.
South Korea's privacy law provides strict protection for information related to personal beliefs, political views and sexual behavior, and bars companies from processing or using such data without the specific consent of the person involved.
The commission said Meta amassed sensitive information by analyzing the pages the Facebook users liked or the advertisements they clicked on.
The company categorized ads to identify users interested in themes such as specific religions, same-sex and transgender issues, and issues related to North Korean escapees, said Lee Eun Jung, a director at the commission who led the investigation on Meta.
"While Meta collected this sensitive information and used it for individualized services, they made only vague mentions of this use in their data policy and did not obtain specific consent," Lee said.
Lee... Read More