The European Union’s executive arm proposed a bill that would limit police use of facial-recognition software in public and ban the marketing or use of certain kinds of AI systems, in one of the broadest efforts yet to regulate high-stakes applications of artificial intelligence.

The bill proposed on Wednesday would also create a list of so-called high-risk uses of AI that would be subject to new supervision and standards for their development and use, such as critical infrastructure, college admissions and loan applications. Regulators could fine a company up to 6% of its annual world-wide revenue for the most severe violations, though in practice EU officials rarely if ever mete out their maximum fines.

The bill is one of the broadest of its kind to be proposed by a Western government, and part of the EU’s expansion of its role as a global tech enforcer.

In recent years, the EU has sought to take a global lead in drafting and enforcing new regulations aimed at taming the alleged excesses of big tech companies and curbing potential dangers of new technologies, in areas ranging from digital competition to online-content moderation. The bloc’s new privacy law, the General Data Protection Regulation, helped set a template for broadly applied rules backed by stiff fines that has been followed in some ways by other countries—and some U.S. states.

SHARE YOUR THOUGHTS

How should governments balance privacy and innovation? Join the conversation below.

Wednesday’s proposal faces a long road—and potential changes—before it becomes law. In the EU, such laws must be approved by both the European Council, representing the bloc’s 27 national governments, and the directly elected European Parliament, which can take years.

Some digital-rights activists, while applauding parts of the proposed legislation, said other elements appear too vague and offer too many loopholes. Some others aligned with industry, argued that the EU’s proposed rules would give an advantage to companies in China, which wouldn’t face them.

“It’s going to make it prohibitively expensive or even technologically infeasible to build AI in Europe,” said Benjamin Mueller, a senior policy analyst at the Center for Data Innovation, part of a tech-aligned think tank. “The U.S. and China are going to look on with amusement as the EU kneecaps its own startups.”

Some tech-industry lobbyists, however, said they were relieved the draft wasn’t more draconian, and applauded the approach of imposing strict oversight on only some types of so-called high-risk uses of AI, such as software for critical infrastructure and algorithms that police use to predict crimes.

“It’s positive that the commission has taken this risk-based approach,” said Christian Borggreen, vice president and head of the Brussels office at the Computer & Communications Industry Association, which represents a number of large technology companies including Amazon, Facebook and Google.

There are a handful of specific practices that face outright bans in the bill. In addition to social credit systems, such as those used by the Chinese government, it also would ban AI systems that use “subliminal techniques” or take advantage of people with disabilities to “materially distort a person’s behavior” in a way that could cause physical or psychological harm.

The bill would create new oversight for high-risk uses of AI.

Photo: focke strangmann/Shutterstock

While police would be generally blocked from using what is described as “remote biometric identification systems”—such as facial recognition—in public places in real time, judges can approve exemptions that include finding abducted children, stopping imminent terrorist threats and locating suspects of certain crimes, ranging from fraud to murder.

“The list of exemptions is incredibly wide,” said Sarah Chander, a senior policy adviser at European Digital Rights, a network of nongovernmental organizations. Such a list “kind of defeats the purpose for claiming something is a ban.”

More on Tech Regulation in Europe

Large banks have pioneered the work of unpicking their artificial intelligence algorithms to regulators, as part of government efforts to prevent another global credit crisis. That makes them a test case for how a broader range of companies will eventually have to do the same, according to Andre Franca, a former director at Goldman Sachs’ model risk management team, and current data science director at AI startup causaLens.

In the past decade, for instance, banks have had to hire teams of people to help present regulators with the mathematical code underlying their AI models, in some cases comprising more than 100 pages per model, Dr. Franca said. Technology companies and advertising agencies that use AI may have to produce similar documentation for regulators.

The new bill would allow the European Commission, the EU’s executive arm, to identify new uses of artificial intelligence to add to the high-risk list subjecting them to stricter oversight. That includes ongoing risk assessments, quality requirements for data used to train AI software and obligations for humans to oversee algorithms’ operation.

The EU could also send teams of regulators to companies to scrutinize algorithms in person if they fall into the high-risk categories laid out in the regulations, Dr. Franca said. That includes systems that identify people’s biometric information—a person’s face or fingerprints—and algorithms that could impact a person’s safety. Regulators from the ECB often personally scrutinize the computer code of banks over several days of workshops and meetings, he added.

While some lower-risk AI systems, such as chatbots, would face obligations to inform users they are not real people, the EU says most uses of AI, including videogames or spam filters, would have no new rules under the bill.

The new regulations might not necessarily have the same impact as GDPR, simply because AI is so broadly defined, according to Julien Cornebise, an honorary associate professor in computer science at University College London and a former research scientist at Google.

“AI is a moving goal post,” he said. “Our phones are doing things everyday that would have been considered ‘AI’ 20 years ago. There’s a risk that could cause the regulation to be either lost in definition or obsolete quickly.”

Write to Sam Schechner at [email protected] and Parmy Olson at [email protected]

Copyright ©2020 Dow Jones & Company, Inc. All Rights Reserved. 87990cbe856818d5eddac44c7b1cdeb8

This post first appeared on wsj.com

Leave a Reply

Your email address will not be published. Required fields are marked *

You May Also Like

We could be living in a croissant-shaped bubble in space, say scientists as Nasa funds work

A TEAM of astronomers led by Boston University has revealed that the…

How to Turn Your Favorite Web Apps Into Desktop Apps

As you may have noticed, we do most of our computing through…

Inside America’s School Internet Censorship Machine

APS, which installed Blocksi in May, stopped using the filter on most…

10 Best October Lego Prime Day Deals (2023)

“What?” you ask, slowly blinking. Didn’t we already have Prime Day? Have…