RISKS-LIST: Risks-Forum Digest Thursday 5 August 2021 Volume 32 : Issue 80
ACM FORUM ON RISKS TO THE PUBLIC IN COMPUTERS AND RELATED SYSTEMS (comp.risks) Peter G. Neumann, founder and still moderator ***** See last item for further information, disclaimers, caveats, etc. ***** This issue is archived at <http://www.risks.org> as <http://catless.ncl.ac.uk/Risks/32.80> The current issue can also be found at <http://www.csl.sri.com/users/risko/risks.txt> Contents: The World Is Suffering from Champlain Towers South Syndrome (WiReD) Beware using telemedicine for voice and speech therapy (techxplore) In China, Big Brother is a service (Facebook) AI flunks COVID test (Will Douglas Heaven vis Henry Baker) YouTube suspends Sky News Australia uploads over COVID-19 misinformation (Engadget) Abridged info on RISKS (comp.risks) ---------------------------------------------------------------------- Date: Sat, 31 Jul 2021 00:36:53 -0400 From: "Gabe Goldberg" <g...@gabegold.com> Subject: The World Is Suffering from Champlain Towers South Syndrome (WiReD) This behavior is far from limited to condo residents in South Florida. Even the alleged geniuses running big tech companies—at least they’re rewarded like geniuses—seem to suffer from an intentional or self-delusional refusal to face inconvenient truths. For years, as I documented in my book and others reported as well, Facebook had warnings that its inability to police destructive posts in foreign nations would come back to bite it. It was years before it took significant action. More recently, as The Wall Street Journal reported, a team at Facebook (nicknamed “Eat Your Veggies”) tested a less toxic version of the News Feed. The new algorithm appeared to work, but Mark Zuckerberg decided, presumably because he preferred the status quo, to water it down considerably, and asked the team to “not bring him something like that again.” I’d argue that bringing comity to the News Feed is vital to Facebook in the long run, even if engagement were to temporarily drop, and that it should be prioritized. Or look at Google. The company has succeeded by catering to the whims of its largely male engineering workforce. Changing that philosophy might be painful, and possibly even risky. But incident after incident of unacceptable responses to sexual harassment will only erode the company’s reputation, possibly leading those wonderful engineers to go elsewhere. It’s only one of a number of festering problems at Google, which seemingly -- but not really -- has the luxury of huge profits to keep its immediate outlook sunny. Google and Facebook are joined by its cousins in dominance -- Amazon, Apple, and Microsoft -- which have all stuck to anticompetitive behaviors that pay off handsomely in the short term but have alienated the media and government. Now they are targets of regulation and lawsuits. https://www.wired.com/story/plaintext-champlain-towers-south-syndrome/ ------------------------------ Date: Sat, 31 Jul 2021 12:56:30 +0800 From: "Richard Stein" <rmst...@ieee.org> Subject: Beware using telemedicine for voice and speech therapy (techxplore) https://techxplore.com/news/2021-07-beware-telemedicine-voice-speech-therapy.html "Because many of the voice metrics collected from virtual platforms had clinically significant differences from those collected in person, Weerathunge and the team urge caution for voice and speech therapists using telepractice." Risk: Analog-to-digital speech capture and reproduction algorithms ------------------------------ Date: Sat, 31 Jul 2021 10:58:15 +0300 From: "Amos Shapir" <amos...@gmail.com> Subject: In China, Big Brother is a service (Facebook) This video was posted to China's International Radio FB page. It boasts the latest innovation: just stand in front of an airport info booth, and BB will show you who you are, where you are going to, and when. https://www.facebook.com/watch/?v=340566300891396 ------------------------------ Date: Sun, 01 Aug 2021 17:28:04 -0700 From: "Henry Baker" <hbak...@pipeline.com> Subject: AI flunks COVID test (Will Douglas Heaven) It's a good thing that AI is doing so much better in other areas -- e.g., prison sentences -- because AI got a pretty big black eye for COVID. :-) "In the end, many hundreds of predictive tools were developed. None of them made a real difference, and some were potentially harmful." "This pandemic was a big test for AI and medicine," says Driggs ... "But I don't think we passed that test." "Because patients scanned while lying down were more likely to be seriously ill, the AI learned wrongly to predict serious covid risk from a person's *position*." "Some AIs were found to be picking up on the *text font* that certain hospitals used to label the scans. As a result, fonts from hospitals with more serious caseloads became predictors of covid risk." https://www.technologyreview.com/2021/07/30/1030329/machine-learning-ai-failed-covid-hospital-diagnosis-pandemic/ Hundreds of AI tools have been built to catch covid. None of them helped. Some have been used in hospitals, despite not being properly tested. But the pandemic could help make medical AI better. Will Douglas Heaven, 30 Jul 2021 When covid-19 struck Europe in March 2020, hospitals were plunged into a health crisis that was still badly understood. "Doctors really didn't have a clue how to manage these patients," says Laure Wynants, an epidemiologist at Maastricht University in the Netherlands, who studies predictive tools. But there was data coming out of China, which had a four-month head start in the race to beat the pandemic. If machine-learning algorithms could be trained on that data to help doctors understand what they were seeing and make decisions, it just might save lives. "I thought, 'If there's any time that AI could prove its usefulness, it's now,'" says Wynants. "I had my hopes up." It never happened -- but not for lack of effort. Research teams around the world stepped up to help. The AI community, in particular, rushed to develop software that many believed would allow hospitals to diagnose or triage patients faster, bringing much-needed support to the front lines -- in theory. In the end, many hundreds of predictive tools were developed. None of them made a real difference, and some were potentially harmful. That's the damning conclusion of multiple studies published in the last few months. In June, the Turing Institute, the UK's national center for data science and AI, put out a report summing up discussions at a series of workshops it held in late 2020. The clear consensus was that AI tools had made little, if any, impact in the fight against covid. Not fit for clinical use This echoes the results of two major studies that assessed hundreds of predictive tools developed last year. Wynants is lead author of one of them, a review in the British Medical Journal that is still being updated as new tools are released and existing ones tested. She and her colleagues have looked at 232 algorithms for diagnosing patients or predicting how sick those with the disease might get. They found that none of them were fit for clinical use. Just two have been singled out as being promising enough for future testing. "It's shocking," says Wynants. "I went into it with some worries, but this exceeded my fears." Wynants's study is backed up by another large review carried out by Derek Driggs, a machine-learning researcher at the University of Cambridge, and his colleagues, and published in Nature Machine Intelligence. This team zoomed in on deep-learning models for diagnosing covid and predicting patient risk from medical images, such as chest x-rays and chest computer tomography (CT) scans. They looked at 415 published tools and, like Wynants and her colleagues, concluded that none were fit for clinical use. "This pandemic was a big test for AI and medicine," says Driggs, who is himself working on a machine-learning tool to help doctors during the pandemic. "It would have gone a long way to getting the public on our side," he says. "But I don't think we passed that test." Both teams found that researchers repeated the same basic errors in the way they trained or tested their tools. Incorrect assumptions about the data often meant that the trained models did not work as claimed. Wynants and Driggs still believe AI has the potential to help. But they are concerned that it could be harmful if built in the wrong way because they could miss diagnoses or underestimate risk for vulnerable patients. "There is a lot of hype about machine-learning models and what they can do today," says Driggs. Unrealistic expectations encourage the use of these tools before they are ready. Wynants and Driggs both say that a few of the algorithms they looked at have already been used in hospitals, and some are being marketed by private developers. "I fear that they may have harmed patients," says Wynants. So what went wrong? And how do we bridge that gap? If there's an upside, it is that the pandemic has made it clear to many researchers that the way AI tools are built needs to change. "The pandemic has put problems in the spotlight that we've been dragging along for some time," says Wynants. What went wrong Many of the problems that were uncovered are linked to the poor quality of the data that researchers used to develop their tools. Information about covid patients, including medical scans, was collected and shared in the middle of a global pandemic, often by the doctors struggling to treat those patients. Researchers wanted to help quickly, and these were the only public data sets available. But this meant that many tools were built using mislabeled data or data from unknown sources. Driggs highlights the problem of what he calls Frankenstein data sets, which are spliced together from multiple sources and can contain duplicates. This means that some tools end up being tested on the same data they were trained on, making them appear more accurate than they are. It also muddies the origin of certain data sets. This can mean that researchers miss important features that skew the training of their models. Many unwittingly used a data set that contained chest scans of children who did not have covid as their examples of what non-covid cases looked like. But as a result, the AIs learned to identify kids, not covid. Driggs's group trained its own model using a data set that contained a mix of scans taken when patients were lying down and standing up. Because patients scanned while lying down were more likely to be seriously ill, the AI learned wrongly to predict serious covid risk from a person's position. In yet other cases, some AIs were found to be picking up on the text font that certain hospitals used to label the scans. As a result, fonts from hospitals with more serious caseloads became predictors of covid risk. Errors like these seem obvious in hindsight. They can also be fixed by adjusting the models, if researchers are aware of them. It is possible to acknowledge the shortcomings and release a less accurate, but less misleading model. But many tools were developed either by AI researchers who lacked the medical expertise to spot flaws in the data or by medical researchers who lacked the mathematical skills to compensate for those flaws. A more subtle problem Driggs highlights is incorporation bias, or bias introduced at the point a data set is labeled. For example, many medical scans were labeled according to whether the radiologists who created them said they showed covid. But that embeds, or incorporates, any biases of that particular doctor into the ground truth of a data set. It would be much better to label a medical scan with the result of a PCR test rather than one doctor's opinion, says Driggs. But there isn't always time for statistical niceties in busy hospitals. That hasn't stopped some of these tools from being rushed into clinical practice. Wynants says it isn't clear which ones are being used or how. Hospitals will sometimes say that they are using a tool only for research purposes, which makes it hard to assess how much doctors are relying on them. "There's a lot of secrecy," she says. Wynants asked one company that was marketing deep-learning algorithms to share information about its approach but did not hear back. She later found several published models from researchers tied to this company, all of them with a high risk of bias. "We don't actually know what the company implemented," she says. According to Wynants, some hospitals are even signing nondisclosure agreements with medical AI vendors. When she asked doctors what algorithms or software they were using, they sometimes told her they weren't allowed to say. How to fix it What's the fix? Better data would help, but in times of crisis that's a big ask. It's more important to make the most of the data sets we have. The simplest move would be for AI teams to collaborate more with clinicians, says Driggs. Researchers also need to share their models and disclose how they were trained so that others can test them and build on them. "Those are two things we could do today," he says. "And they would solve maybe 50% of the issues that we identified." Getting hold of data would also be easier if formats were standardized, says Bilal Mateen, a doctor who leads the clinical technology team at the Wellcome Trust, a global health research charity based in London. Another problem Wynants, Driggs, and Mateen all identify is that most researchers rushed to develop their own models, rather than working together or improving existing ones. The result was that the collective effort of researchers around the world produced hundreds of mediocre tools, rather than a handful of properly trained and tested ones. "The models are so similar -- they almost all use the same techniques with minor tweaks, the same inputs -- and they all make the same mistakes," says Wynants. "If all these people making new models instead tested models that were already available, maybe we'd have something that could really help in the clinic by now." In a sense, this is an old problem with research. Academic researchers have few career incentives to share work or validate existing results. There's no reward for pushing through the last mile that takes tech from "lab bench to bedside," says Mateen. To address this issue, the World Health Organization is considering an emergency data-sharing contract that would kick in during international health crises. It would let researchers move data across borders more easily, says Mateen. Before the G7 summit in the UK in June, leading scientific groups from participating nations also called for "data readiness" in preparation for future health emergencies. Such initiatives sound a little vague, and calls for change always have a whiff of wishful thinking about them. But Mateen has what he calls a "naïvely optimistic" view. Before the pandemic, momentum for such initiatives had stalled. "It felt like it was too high of a mountain to hike and the view wasn't worth it," he says. "Covid has put a lot of this back on the agenda." "Until we buy into the idea that we need to sort out the unsexy problems before the sexy ones, we're doomed to repeat the same mistakes," says Mateen. "It's unacceptable if it doesn't happen. To forget the lessons of this pandemic is disrespectful to those who passed away." ------------------------------ Date: Tue, 3 Aug 2021 08:06:05 +0900 From: "ファーバーデイビッド J" <far...@keio.jp> Subject: YouTube suspends Sky News Australia uploads over COVID-19 misinformation (Engadget) https://www.engadget.com/youtube-sky-news-australia-ban-covid-misinformation-213810128.html ------------------------------ Date: Mon, 1 Aug 2020 11:11:11 -0800 From: risks-requ...@csl.sri.com Subject: Abridged info on RISKS (comp.risks) The ACM RISKS Forum is a MODERATED digest. Its Usenet manifestation is comp.risks, the feed for which is donated by panix.com as of June 2011. => SUBSCRIPTIONS: The mailman Web interface can be used directly to subscribe and unsubscribe: http://mls.csl.sri.com/mailman/listinfo/risks => SUBMISSIONS: to risks@CSL.sri.com with meaningful SUBJECT: line that includes the string `notsp'. Otherwise your message may not be read. *** This attention-string has never changed, but might if spammers use it. => SPAM challenge-responses will not be honored. Instead, use an alternative address from which you never send mail where the address becomes public! => The complete INFO file (submissions, default disclaimers, archive sites, copyright policy, etc.) is online. <http://www.CSL.sri.com/risksinfo.html> *** Contributors are assumed to have read the full info file for guidelines! => OFFICIAL ARCHIVES: http://www.risks.org takes you to Lindsay Marshall's searchable html archive at newcastle: http://catless.ncl.ac.uk/Risks/VL.IS --> VoLume, ISsue. Also, ftp://ftp.sri.com/risks for the current volume/previous directories or ftp://ftp.sri.com/VL/risks-VL.IS for previous VoLume If none of those work for you, the most recent issue is always at http://www.csl.sri.com/users/risko/risks.txt, and index at /risks-32.00 ALTERNATIVE ARCHIVES: http://seclists.org/risks/ (only since mid-2001) *** NOTE: If a cited URL fails, we do not try to update them. Try browsing on the keywords in the subject line or cited article leads. Apologies for what Office365 and SafeLinks may have done to URLs. ==> Special Offer to Join ACM for readers of the ACM RISKS Forum: <http://www.acm.org/joinacm1> ------------------------------ End of RISKS-FORUM Digest 32.80 ************************