Ha! "There's a fun sub-result, which is, if you have a very deviant concept ... if you have a very weirdo concept that other people don't share, you're actually much more likely to be aware that you have a deviant concept."
At least I *know* I'm a deviant. On 12/29/19 8:43 AM, Roger Critchlow wrote: > I thought she was arguing that very mechanisms that google, facebook, > twitter, etc. are using right now to engage people's interest online are > already engendering and entrenching all sorts of weird beliefs. 6-9 minutes > of activated charcoal advocacy videos and you're probably certain that black > smoothies are okay, maybe even good for you. There are no neutral platforms, > because the order in which content is presented is never neutral, and it is > especially biased if its goal is to keep you clicking. Whether this allows > focused election manipulation seems dubious, but it does allow for thousands > of bizarre theories to be injected into the public consciousness at low cost, > and some of them even make money. Hey, some of them, bizarre as they are, > might turn out to be correct, not that the platforms have any interest in > that aspect, because that wouldn't be neutral. > [...] > > On Sat, Dec 28, 2019 at 10:23 AM Steven A Smith <sasm...@swcp.com > <mailto:sasm...@swcp.com>> wrote: > > REC - > > Good find! > > I am not closely following the development and results of GAN work, but > it seems like this kind of study explicates at least ONE GOOD REASON for > worrying about AI changing the nature of the world as we know it (even if it > isn't a precise existential threat). Convolved with Carl's offering around > "weaponizing complexity", it feels more and more believable (recursion > unintended) that the wielders of strong AI/ML will have the upper hand in any > tactical and possibly strategic domain (warfare, public opinion, markets, > etc.). > [...] > > On 12/27/19 8:21 PM, Roger Critchlow wrote: >> This talk was mentioned on hacker news this week and inspired my >> babbling at Saveur this morning. >> https://slideslive.com/38921495/how-to-know. The talk was delivered at >> Neural IPS on December 9 and discusses recent research on how people come to >> believe they know something. >> >> This paper >> https://www.mitpressjournals.org/doi/full/10.1162/opmi_a_00017 describes the >> Amazon Mechanical Turk experiment on people becoming certain they understood >> the boolean rule they were being taught by examples. -- ☣ uǝlƃ ============================================================ FRIAM Applied Complexity Group listserv Meets Fridays 9a-11:30 at cafe at St. John's College to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com archives back to 2003: http://friam.471366.n2.nabble.com/ FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove