I interpreted an interview in Madrid for “The Fake Paralympians” a BBC podcast produced by Simon Maybin and presented by Paralympian swimmer Dan Pepper . The interview is from around 12:00 to 17:25 in episode 5, Court.
Streaming link here.
As part of its emerging role as a global regulatory watchdog, the European Commission published a proposal on 21 April for regulations to govern artificial intelligence use in the European Union.
The economic stakes are high: the Commission predicts European public and private investment in AI reaching €20 billion a year this decade, and that was before the additional earmark of up to €134 billion earmarked for digital transitions in Europe’s Covid-19 pandemic recovery fund, some of which the Commission presumes will fund AI, too. Add to that counting investments in AI outside the EU but which target EU residents, since these rules will apply to any use of AI in the EU, not just by EU-based companies or governments.
Things aren’t going to change overnight: the EU’s AI rules proposal is the result of three years of work by bureaucrats, industry experts, and public consultations and must go through the European Parliament—which requested it—before it can become law. EU member states then often take years to transpose EU-level regulations into their national legal codes.
The proposal defines four tiers for AI-related activity and differing levels of oversight for each. The first tier is unacceptable risk: some AI uses would be banned outright in public spaces, with specific exceptions granted by national laws and subject to additional oversight and stricter logging and human oversight. The to-be-banned AI activity that has probably garnered the most attention is real-time remote biometric identification, i.e. facial recognition. The proposal also bans subliminal behavior modification and social scoring applications. The proposal suggests fines of up to 6 percent of commercial violators’ global annual revenue.
The proposal next defines a high-risk category, determined by the purpose of the system and the potential and probability of harm. Examples listed in the proposal include job recruiting, credit checks, and the justice system. The rules would require such AI applications to use high-quality datasets, document their traceability, share information with users, and account for human oversight. The EU would create a central registry of such systems under the proposed rules and require approval before deployment.
Limited-risk activities, such as the use of chatbots or deepfakes on a website, will have less oversight but will require a warning label, to allow users to opt in or out. Then finally there is a tier for applications judged to present minimal risk.
As often happens when governments propose dense new rulebooks (this one is 108 pages), the initial reactions from industry and civil society groups seem to be more about the existence and reach of industry oversight than the specific content of the rules. One tech-funded think tank told the Wall Street Journal that it could become “infeasible to build AI in Europe.” In turn, privacy-focused civil society groups such as European Digital Rights (EDRi) said in a statement that the “regulation allows too wide a scope for self-regulation by companies.”
“I think one of the ideas behind this piece of regulation was trying to balance risk and get people excited about AI and regain trust,” says Lisa-Maria Neudert, AI governance researcher at the University of Oxford, England, and the Weizenbaum Institut in Berlin, Germany. A 2019 Lloyds Register Foundation poll found that the global public is about evenly split between fear and excitement about AI.
“I can imagine it might help if you have an experienced large legal team,” to help with compliance, Neudert says, and it may be “a difficult balance to strike” between rules that remain startup-friendly and succeed in reining in mega-corporations.
AI researchers Mona Sloane and Andrea Renda write in VentureBeat that the rules are weaker on monitoring of how AI plays out after approval and launch, neglecting “a crucial feature of AI-related risk: that it is pervasive, and it is emergent, often evolving in unpredictable ways after it has been developed and deployed.”
Europe has already been learning from the impact its sweeping 2018 General Data Protection Regulation (GDPR) had on global tech and privacy. Yes, some outside websites still serve Europeans a page telling them the website owners can’t be bothered to comply with GDPR, so Europeans can’t see any content. But most have found a way to adapt in order to reach this unified market of 448 million people.
“I don’t think we should generalize [from GDPR to the proposed AI rules], but it’s fair to assume that such a big piece of legislation will have effects beyond the EU,” Neudert says. It will be easier for legislators in other places to follow a template than to replicate the EU’s heavy investment in research, community engagement, and rule-writing.
While tech companies and their industry groups may grumble about the need to comply with the incipient AI rules, Register columnist Rupert Goodwin suggests they’d be better off focusing on forming the industry groups that will shape the implementation and enforcement of the rules in the future: “You may already be in one of the industry organizations for AI ethics or assessment; if not, then consider them the seeds from which influence will grow.”
Andrés Colao speaks from his own experience as a patient who has seen the COVID-19 pandemic cripple an already weak healthcare system. He is the spokesperson for AFESA, a Spanish charity of people with mental illness and their relatives. For those who had a disorder diagnosed before the COVID-19 pandemic, the crisis has left them in limbo.
Jorge Daniel Castilla, who was undergoing treatment for a mental health condition, says, “I have had a couple of calls since March, the last one was in June to ask how I was doing. My therapy has been left up in the air.”
The crisis has been especially difficult for people seeking psychiatric and psychological services. “There are patients who have suffered a lot,” Colao says.
COVID-19 has caused a tsunami in mental health. During the first wave, 93% of countries surveyed by the World Health Organization (WHO) suffered paralysis in one or more services for patients with mental, neurological and substance abuse problems. Almost 40% of participating European countries reported worse conditions: they had stopped three out of four health services. “The stricter the lockdown, the more severe the impact,” says Marcin Rodzinka, spokesperson for Mental Health Europe, a network of mental health service users and professionals. This happened in Spain, for example, which shut its mental health outpatient centres.Continue reading Translated Story: No appointments for mental health patients during the COVID-19 pandemic
Online conferences can be easier than in-person conferences to integrate into a busy schedule, but they still require some advance planning and thoughtful behavior throughout. If you’re attending the ACS Fall 2020 Virtual Conference & Expo or any professional meeting online for the first time, we’ve got advice to help you be a good virtual citizen and boost your chemistry career.Continue reading Getting Value out of Virtual Conferences