by aston on 6/22/16, 12:32 AM with 94 comments
by lacker on 6/22/16, 1:12 AM
IMO, a better approach to AI safety research is to focus on securing the first channels that a malicious AI would be likely to exploit. Like spam, and security. Can you make communications spam-resistant? Can you make an unhackable internet service?
Those seem hard, but more plausible than the "Watch out for paperclip optimizers" approach to AI safety. It just feels like inventing a way to build a nuclear weapon that can't actually explode, and then hoping the problem of nuclear war is solved.
by colah3 on 6/22/16, 12:54 AM
Google Post: https://research.googleblog.com/2016/06/bringing-precision-t...
It was a pleasure for us to work on this with OpenAI and others. John/Paul/Jacob are good friends, and wonderful colleagues! :)
by fiatmoney on 6/22/16, 4:59 AM
The hard part is actually figuring out what you care about, particularly in the context of a truly universal optimizer that can decide to trade off anything in the pursuit of its objectives.
This has been a core problem of philosophy for 3000 years - that is, putting some amount of rigorous codification behind human preferences. You could think of it as a branch of deontology, or maybe aesthetics. It is extremely unlikely that a group sponsored by Sam Altman, whose brilliant idea was "let's put the government in charge of it" [1], will make a breakthrough there.
I don't actually doubt that AIs would lead to philosophical implications, and philosophers like Nick Land have actually explored some of that area. But I severely doubt the ability of AI researchers to do serious philosophy and simultaneously build an AI that reifies those concepts.
by arcanus on 6/22/16, 1:24 AM
In particular, I'm talking about verification and validation testing. I'm curious why generally these approaches are not being leveraged to ensure quality of output here.
I suspect this is because of the persistent belief that AI will annihilate humanity with one mishap, but I'm suggesting that we approach this much more like traditional engineering problems, such as building a bridge or flying a plane, whereby rigorous standards of are continually applied to ensure the system behaves as designed.
The resulting system will look much more like continuous integration with robust regression testing and high line coverage than it will be the sexy research ideas presented here, but I can't help but think it will be more robust. These systems are too complicated to treat them as anything but a black box, at least from a quality assurance standpoint.
by Animats on 6/22/16, 4:12 AM
Yes. That's why I was critical of an academic AI effort which attempts automatic driving by training a supervised learning system by observing human drivers. That's going to work OK for a while, and then do something really stupid, because it has no model of catastrophic actions.
by chrisfosterelli on 6/22/16, 12:51 AM
by pizza on 6/22/16, 1:11 AM
[0] http://www.wireheading.com/ - David Pearce's ideas are.. interesting.. to say the least ;)
by DrNuke on 6/22/16, 11:02 AM
by Mendenhall on 6/22/16, 1:33 AM
by glaberficken on 6/22/16, 11:42 AM
I assume this probably already worked into the current prototypes. Does anyone have references to discussions about this in current gen self driving car prototypes?
by fitzwatermellow on 6/22/16, 1:03 PM
To me this is the toughest nut in the lot. Training a Pac-man agent to avoid ghosts and eat pellets, in a world of infinite hazards and cautions! Any strategies?
by w_t_payne on 6/22/16, 11:07 AM
by kordless on 6/22/16, 5:38 AM
Oh brother. Avoiding negative side effects is a wasteful proposition. Learning from those side effects, however, is priceless.
by mountaineer22 on 6/22/16, 2:31 AM
by JoeAltmaier on 6/22/16, 1:27 PM
by daveguy on 6/22/16, 2:27 AM
http://xkcd.com/1696/ (the current xkcd)
by yarou on 6/22/16, 1:36 AM
by logicallee on 6/22/16, 3:14 PM