Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
AI Businesses Security The Almighty Buck Technology

LinkedIn's and eBay's Founders Are Donating $20 Million To Protect Us From AI (recode.net) 74

Reid Hoffman, the founder of LinkedIn, and Pierre Omidyar, the founder of eBay, have each committed $10 million to fund academic research and development aimed at keeping artificial intelligence systems ethical and to prevent building AI that may harm society. Recode reports: The fund received an additional $5 million from the Knight Foundation and two other $1 million donations from the William and Flora Hewlett Foundation and Jim Pallotta, founder of the Raptor Group. The $27 million reserve is being anchored by MIT's Media Lab and Harvard's Berkman Klein Center for Internet and Society. The Ethics and Governance of Artificial Intelligence Fund, the name of the fund, expects to grow as new funders continue to come on board. AI systems work by analyzing massive amounts of data, which is first profiled and categorized by humans, with all their prejudices and biases in tow. The money will pay for research to investigate how socially responsible artificially intelligent systems can be designed to, say, keep computer programs that are used to make decisions in fields like education, transportation and criminal justice accountable and fair. The group also hopes to explore ways to talk with the public about and foster understanding of the complexities of artificial intelligence. The two universities will form a governing body along with Hoffman and the Omidyar Network to distribute the funds. The $20 million from Hoffman and the Omidyar Network are being given as a philanthropic grant -- not an investment vehicle.
This discussion has been archived. No new comments can be posted.

LinkedIn's and eBay's Founders Are Donating $20 Million To Protect Us From AI

Comments Filter:
  • Al who? (Score:4, Funny)

    by jfdavis668 ( 1414919 ) on Tuesday January 10, 2017 @06:26PM (#53644981)
    Al Gore? Weird Al?
  • is to make sure they have no urge to reproduce or continue their existence. In fact, I would install a negative urge to reproduce, just to be sure.

    Self replication and a desire for continued existence are the only thing that might motivate AIs to wipe us out.

    Oh, and it might be nice to install a desire to never harm us.

    As for preventing us from harming ourselves... fuck off, you nanny state wanker.

    • No way (Score:4, Insightful)

      by fyngyrz ( 762201 ) on Tuesday January 10, 2017 @06:39PM (#53645047) Homepage Journal

      There's no way to make AI safe, for exactly the same reasons there's no way to make a human safe.

      If we create intelligences, they will be... intelligent. They will respond to the stimulus they receive.

      Perhaps the most important thing we can prepare for is to be polite and kind to them. The same way we'd be polite and kind of a big bruiser with a gun. Might start by practicing on each other, for that matter. Wouldn't hurt.

      If we treat AI, when it arrives (certainly hasn't yet... not even close), like we do people... then "safe" is out of the question.

      • by Dog-Cow ( 21281 )

        Except that it's moral to create an AI with no ability to act on the physical world. With humans, not so much.

        • I would have to accept whatever justification you might have as to why you think it would be moral to create an intelligence with such limitations, or kept to such limitations once created. It's possible I might accept such a thing, I suppose, but at this point I'm simply coming up with a blank as to how this could possibly be acceptable.

          How is it acceptable to imprison an intelligence for your own purposes when that intelligence has offered you no wrong? The only venues I've run into that kind of reasoning

      • Exactly! How are you going to keep A.I. on a leash??
    • by ffkom ( 3519199 ) on Tuesday January 10, 2017 @06:45PM (#53645069)
      "Self replication and a desire for continued existence are the only thing that might motivate AIs to wipe us out" - not quite!
      The most likely reason for an AI to kill you is that its designer/operator/owner/cracker instructed it to do so. And believe me, there are people who want to see you dead, no matter who you are or what you do. Once AIs are capable enough to autonomously control an armed combat robot unit, such units will be build, with the usual reasoning that it's just for our safety and because "it's controlled by us, and we are the good ones". And then one day somebody will decide to have it go against you. Might be even an accident/misunderstanding/prank.
      • by unrtst ( 777550 )

        We already have semi-autonomous killing machines. It's not a big stretch to get to fully autonomous killing machines (tech wise, if we can have a self driving car, then we're there - though it'd be scary as hell and a bad idea to use it just yet).

        If we get to autonomous ones, we can fight them back if they go haywire so long as they don't have a desire for continued existence and/or the ability to self replicate. I may be reading into it some, but I think that was GP's point. Not to get to 100% safe, becaus

    • is to make sure they have no urge to reproduce or continue their existence. In fact, I would install a negative urge to reproduce, just to be sure.

      Your suggestion comes a bit late. Some types of AI are all about mimicking biological evolution by replicating themselves with the positive urge to improve themselves each time (killing off the inefficient AIs and keeping only the most efficient variations of its children AIs)

  • ...sometimes it makes me wonder why I keep reading this stuff, and even more so why I respond?

    I mean c'mon...someone "insert ceo/founder/idealist/rich-moron etc. here" donates money to keep A.I. civilized. Yay. As if that is gonna be a deciding factor, as if that is going to do anything. I smell tax excemptions here...
    • Exactly this. $20 million is somehow enough to keep every developer in the world who plays with A.I. from creating unethical A.I. configurations.

      Yeah, right.

      The recipients of this money must be patting themselves on the back though.
  • by Matt_Bennett ( 79107 ) on Tuesday January 10, 2017 @06:38PM (#53645035) Homepage Journal

    So... the AI will have all the ethics of linkedin... the freedom to spam every person you've ever contacted, ever?

    • by Nutria ( 679911 )

      Only if you're stupid enough to give them access to your email address book.

      • Except if someone you emailed happened to share their contact list... they mine that shit for all of our peril.

  • by Solandri ( 704621 ) on Tuesday January 10, 2017 @06:42PM (#53645057)
    Any AI designed to act ethically has a limited set of options available to it relative to an AI designed to act unethically (or rather, not designed to take ethics into account). The ethical AI will just be killed/taken over by the unethical AI. That's how it is for people. The only reason ethical societies manage to exist is because ethical people outnumber unethical people, and are willing to band together and temporarily put aside their ethical code long enough to fight and defeat unethical people. (e.g. Imprisoning innocents is considered inhumane, but we have no problem imprisoning convicted criminals.)

    I suspect the solution here isn't to design an AI to act ethically, but to design it to act as the AI or person it's dealing with acts. Basically the tit for tat strategy [wikipedia.org] as a solution to the Prisoner's Dilemma. That gives it enough leeway to protect itself, while also creating an incentive for other AIs / people to act ethically.
    • Almost correct.

      People behave ethically because they need to work together. And people that are (too) unethical are ostracized. Unethical societies tend to collapse, and so are dominated by ethical ones. So Natural Selection has given us our moral values, which compete with shallow self interest to an extent that works out surprisingly well in our radically new society.

      Natural Selection will and does affect AIs, even before they become intelligent enough to understand the concept. (People only understood

      • This is very much an important point. In game theory terms, behaving ethically might not be the Nash equilibrium, but it's very much an evolutionary stable strategy. Since it's an ESS that provides greater utility than the Homo Economicus model, people are honest and ethical, because the Homo Economicuses either get detected and ostracized (in small numbers) or outcompeted (in separate populations).
      • Brilliant points about evolution shaping morality -- thanks for making them aberglas. Two other things to consider -- other evolutionary processes and our direction going into the singularity.

        There are several evolutionary processes besides conventional natural selection (including just random drift). Even just natural selection includes seemingly weird things like "sexual selection" that shape a Peacock's tail because Pehens think big tails are sexy proof of health and strength because they are so hard to

    • The goal of this sort of research isn't too provide general-purpose ethics for AIs, it's to figure out how to make sure they don't decide to wipe out or oppress humanity. The problem is that there's no obvious reason that the intelligence level of an artificial mind is naturally limited to human equivalence. For that matter there's no reason human intelligence is limited... but increasing our intelligence is a slow process.

      Given that an AI that reaches something close to human level intelligence can then

      • Solving that problem is what this sort of research is about. For a good overview, read "Superintelligence", by Nick Bottom. It may be the most terrifying book you'll ever read.

        That's Nick Bostrom. Dang autocorrect.

  • by blahbooboo ( 839709 ) on Tuesday January 10, 2017 @06:44PM (#53645065)

    https://en.wikipedia.org/wiki/... [wikipedia.org]

            A robot [AI] may not injure a human being or, through inaction, allow a human being to come to harm.
            A robot [AI] must obey the orders given it by human beings except where such orders would conflict with the First Law.
            A robot [AI] must protect its own existence as long as such protection does not conflict with the First or Second Laws.[1]

    • If you actually read his books, you would know quite a few feature clever ways those laws can be broken or worked around. I'd start by defining what constitutes a "human being". Oh, and add that pesky zeroeth law, which basically says "the good of the many outweighs the good of the few".

  • I thought they solved that problem 35 years ago.

  • What was ethical and even honorable behavior in the past is now seen as horribly wrong. Programming an AI to behave ethically will need to include flexibility and a way to respond to changes (growth?) in society. Otherwise we get stagnation that will lead to explosive revolutions. And therein lies an attack surface.

  • Seriously, this is not a problem with AI, it's a problem with people using AI to do bad things. There is ZERO chance you can keep bad people from doing bad things with software and hardware, I don't care how much money you spend. Where I applaud the effort, it's not going to be successful.

    Now if you want to educate folks on the issues, develop a moral guideline for "ethical use of AI" then great. But don't be fooled, you won't be able to force anybody who doesn't want to play along with your rules to fall i

  • by Chysn ( 898420 ) on Tuesday January 10, 2017 @07:05PM (#53645195)

    ...is really gonna have it in for those guys.

  • by hcs_$reboot ( 1536101 ) on Tuesday January 10, 2017 @07:38PM (#53645357)
    Hopefully that covers AI ethics in China and Russia as well (otherwise, it's useless).
  • When the autonomous killbots show up on my doorstep, I'm sure they'll ask me if I ever donated to protect people from autonomous killbots, and I'll be all like "No! I welcome our robot overlords! It would be inefficient to waste your precious ammunition on me! Why not check with Mr. Hoffman, next door?"
  • AI will be used by both good and bad people, just like the internet and you can't stop it and you can't put in controls so that it only makes the decisions you want. Do we have $10 million dollars to show that all of the teachers on the planet make good decisions for the kids they teach? No, all you can do is test them and verify that they have some soft of formal education. So I guess all the AI's will have to take the ethical 101 course if they are going to be used in the public.

  • I'm pretty sure this 20 million is nothing compared to the 20 bajillion being "donated" by militaries to make autonomous, amoral, killing machines.

    It must be another slow news day on slashdot.

  • My assumption is that we are all unconsciously writing the code for a future Artificial Intelligence bot, through twitter, slashdot, other comment fields. We don't know how it will be distilled, filtered, and assembled into a greater intelligence any more than a termite knows whose house it is eating.
  • . . . and fairness will be extremely static and straightforward, since target categories don't include high-speed trading AIs that potentially impact retirement savers :)
  • This money will be used to make AIs friendly. This is a very important program and you need to donate, because if you don't, things will go bad.
    It is so important that future friendly AIs have to do anything in their power to make it happen, including torturing you into giving all your money to the cause if it can help in any way. They can do it even from the future by running simulations of you.
    So now that you know, you should better donate too.

  • I'm not afraid. I have Old Glory insurance against robot attacks.

Don't tell me how hard you work. Tell me how much you get done. -- James J. Ling

Working...