The World KnowlÂedge Forum was launched in OctoÂber 2000 after two years of prepaÂraÂtion folÂlowÂing the Asian FinanÂcial CriÂsis of 1997, with the goal of fosÂterÂing a creÂative transÂforÂmaÂtion into a knowlÂedge-based nation. Over the years, the forum has proÂvidÂed a platÂform for disÂcusÂsions on narÂrowÂing the knowlÂedge gap, as well as achievÂing balÂanced globÂal ecoÂnomÂic growth and prosÂperÂiÂty through knowlÂedge sharing.
AI is evolvÂing rapidÂly, allowÂing us to reap its benÂeÂfits across many areas of life and work. HowÂevÂer, the risks posed by AI are undeÂniÂably growÂing. To ensure that AI does not destroy the order that humans have creÂatÂed over the years or pose threats to our exisÂtence, there must be ethics and clear prinÂciÂples for AI research and develÂopÂment. This sesÂsion will feaÂture acaÂdÂeÂmics who have been workÂing on the funÂdaÂmenÂtal prinÂciÂples of AI develÂopÂment, ethics of AI develÂopÂment, and copyÂright proÂtecÂtion, as well as entreÂpreÂneurs who have encounÂtered chalÂlenges applyÂing AI in the real world. They will disÂcuss how to set funÂdaÂmenÂtal guideÂlines for AI develÂopÂment, how to stop AI from infringÂing on copyÂright and priÂvaÂcy, how to weed out false inforÂmaÂtion, and how to defend human rights.
00:00 What if we succeed in that goal? 05:17 The era of deep learning 10:03 Some serious failures 11:43 AlphaGo and AGI 18:02 Human extinction 21:17 Preferences 29:29 Coexistence
TRANSCRIPTION
Human extinction
But the perÂhaps more seriÂous downÂside uh is human extincÂtion and this is why I say it’s not realÂly an ethÂiÂcal issue I I think by and large few peoÂple would argue that human extincÂtion is uh ethÂiÂcalÂly preferÂable uh there are some uh but I’m just going to ignore those peoÂple um so it’s just comÂmon sense right if you creÂate someÂthing that’s more powÂerÂful than human beings how on Earth are we going to have powÂer over such sysÂtems forÂevÂer so in my view there’s only two choicÂes we either build provÂably safe and conÂtrolÂlable AI where we have absolute cast iron mathÂeÂmatÂiÂcal guarÂanÂtee of safeÂty or we have no AI at all so those are the two choicÂes right now we’re purÂsuÂing the third choice which is comÂpleteÂly unsafe blackÂbox AI that we don’t underÂstand at all and we are tryÂing to make it into someÂthing that’s more powÂerÂful than us which is pretÂty much the same sitÂuÂaÂtion we would be in if uh a superÂhuÂman AI sysÂtem landÂed from outÂer space uh sent by some alien species no doubt for our own good uh our chances of conÂtrolÂling an alien superÂhuÂman intelÂliÂgence would be zero and that’s sitÂuÂaÂtion that we’re headÂing towards and Alan churÂing the founder of comÂputÂer sciÂence uh you know thought about this because he was workÂing on AI and he thought about what hapÂpens if we sucÂceed and he said we should have to expect the machines to take conÂtrol so what do we do I think it’s realÂly hard espeÂcialÂly givÂen that 15 quadrillion dolÂlar prize that uh comÂpaÂnies are aimÂing for and the fact that they have already accuÂmuÂlatÂed 15 trilÂlion dolÂlar worth of capÂiÂtal to aim at that goal with it’s kind of hard to stop that process so we have to come up with a way of thinkÂing about AI that does allow us to conÂtrol it that is provÂably safe and provÂably conÂtrolÂlable and so rather than sayÂing how do we retain powÂer over AI sysÂtems forÂevÂer which sounds pretÂty hopeÂless we say what is a mathÂeÂmatÂiÂcal frameÂwork for AI a a way of definÂing the AI probÂlem so that no matÂter how well the AI sysÂtem solves it we are guarÂanÂteed to be hapÂpy with the result so can we devise a mathÂeÂmatÂiÂcal probÂlem a way of sayÂing what is the AI SysÂtem supÂposed to be doing that has that propÂerÂty that we’re guarÂanÂteed to be hapÂpy with the result.
Preferences
So I spent about 10 years workÂing on this and um to explain uh how we approachÂing it um I’m we going to introÂduce a a techÂniÂcal term that uh I think will be helpÂful for our disÂcusÂsion about ethics as well um and that’s a notion called prefÂerÂences so prefÂerÂences doesÂn’t sound like a techÂniÂcal term right Some peoÂple preÂfer pineapÂple pizÂza to MarÂgariÂta PizÂza but what we mean by prefÂerÂences in the in the theÂoÂry of deciÂsion- makÂing is actuÂalÂly someÂthing much more all-encomÂpassÂing and it’s your rankÂing over posÂsiÂble futures of the uniÂverse so to kind of reduce that to someÂthing we can grasp easÂiÂly imagÂine that I made you two movies of the rest of your life and the rest of the you know the future of othÂer things you care about you know and the movies are about two hours long and you can kind of Watch movie A and movie b and then you say yeah I’d like movie A please don’t like movie B at all because um I get minced up and and turned into hamÂburgÂer in movie b and I don’t like that very much so I’d preÂfer movie A please so that’s what we mean by prefÂerÂences except that this wouldÂn’t be a two-hour movie it’s realÂly the entire future of the UniÂverse um and of course we don’t get to choose between movies because in fact uh we can’t preÂdict what exactÂly which movie is going to hapÂpen and so uh we’re actuÂalÂly uh havÂing to deal with the uncerÂtainÂty we call these lotÂterÂies over posÂsiÂble futures of the uniÂverse so a prefÂerÂence strucÂture is then a uh basiÂcalÂly a rankÂing over futures of the uniÂverse takÂing uncerÂtainÂty into account to make a sysÂtem that is provÂably benÂeÂfiÂcial to humans you just need two simÂple prinÂciÂples one is that the only objecÂtive of the machine is to furÂther human prefÂerÂences to furÂther human interÂests if you like uh and the secÂond prinÂciÂple is that the machine knows that it does not know what those prefÂerÂences are and that’s kind of obviÂous right because we don’t realÂly know what our prefÂerÂences are and uh we cerÂtainÂly can’t write them down in enough detail to get it right um and when but when you think about it right a machine that that solves that probÂlem the betÂter it solves it the betÂter off we are and in fact you can show that it’s in our interÂest to have machines that solve that probÂlem because we are going to be betÂter off with those machines uh than withÂout them so that’s good but as soon as I describe that way of thinkÂing to you that machines are going to furÂther human prefÂerÂen and um and learn about them as they go along this now brings in some ethÂiÂcal quesÂtions finalÂly right so we finalÂly get to ethics what I want to avoid uh so I’m just going to tell you not to ask this quesÂtion do not ask the quesÂtion well whose valÂue sysÂtem are you going to put into the machine right because I’m not proposÂing to put anyÂone parÂticÂuÂlar valÂue sysÂtem into the machine in fact the machine should have at least 8 bilÂlion prefÂerÂence modÂels because there are 8 bilÂlion of us um and the prefÂerÂences of everyÂone matÂter but there are some realÂly difÂfiÂcult ethÂiÂcal probÂlems the first quesÂtion is do peoÂple actuÂalÂly have these prefÂerÂences is it okay for just us to just assume that peoÂple do have you know I like this future and I don’t like that future could there be anothÂer state of being for a perÂson where they say well I’m not sure which future I like or I can only tell you when I’ve lived that future you can’t describe it to me uh in sufÂfiÂcient detail for me to tell you if I like it ahead of time and along with that there’s the quesÂtion of well where do those prefÂerÂences come from in the first place do humans autonomousÂly sudÂdenÂly just like wake up and okay these are my prefÂerÂences and I want them to be respectÂed no our prefÂerÂences come we’re obviÂousÂly not born with them right except some of the basic bioÂlogÂiÂcal things about pain and sugÂar but our our full adult prefÂerÂences come from our culÂture our upbringÂing all of the influÂences that shape who we are and a sad fact about the world is that many peoÂple are in the busiÂness of shapÂing othÂer peoÂple’s prefÂerÂences to suit their own interÂests so one class of peoÂple oppressÂes anothÂer but trains the oppressed to believe that they should be oppressed so then should the AI sysÂtem take the prefÂerÂence those self oppresÂsion prefÂerÂences of the oppressed litÂerÂalÂly and you know conÂtribute to furÂther oppresÂsion of those peoÂple because they’ve been trained to accept their oppresÂsion so marÂtien who was an econÂoÂmist and philosoÂpher uh argued veheÂmentÂly that we should not take such prefÂerÂences at face valÂue but if you don’t take PR peoÂple’s prefÂerÂences of face valÂue then you you seem to fall back on a kind of paterÂnalÂism where well we know what you should want even though you don’t want it and we’re going to give you it even though you’re sayÂing I don’t want it and that’s a comÂpliÂcatÂed posiÂtion to be in and it’s defÂiÂniteÂly not a posiÂtion that AI researchers want to be in anothÂer set of uh ethÂiÂcal issues has to do with aggreÂgaÂtion so I said there are 8 bilÂlion prefÂerÂence modÂels but if a sysÂtem is makÂing a deciÂsion that affects a sigÂnifÂiÂcant fracÂtion of those 8 bilÂlion peoÂple how do you aggreÂgate those prefÂerÂences how do you deal with the fact that there are conÂflicts among those prefÂerÂences you can’t make everyÂbody hapÂpy if everyÂbody wants to be ruler of the uniÂverse and so moral philosoÂphers have studÂied this probÂlem for thouÂsands of years uh most peoÂple on the comÂputÂer sciÂence and engiÂneerÂing backÂgrounds uh tend to think in the way that utilÂiÂtarÂiÂans have proÂposed so benam and Mill and othÂer uh libÂbets um othÂer philosoÂphers proÂpose this approach called utilÂiÂtarÂiÂanÂism which basiÂcalÂly says well you treat everyÂone’s prefÂerÂences as equalÂly imporÂtant uh and then you make the deciÂsion where the total amount of prefÂerÂence satÂisÂfacÂtion is maxÂiÂmized and utilÂiÂtarÂiÂanÂism has got a bad name because some peoÂple think it’s anti-egalÂiÂtarÂiÂan and so on but I actuÂalÂly think that there’s a lot more work to do on how we forÂmuÂlate utilÂiÂtarÂiÂanÂism we have to do this work because the AI sysÂtems are going to be makÂing deciÂsions that affect milÂlions or milÂlions of peoÂple and so whatÂevÂer the right ethÂiÂcal answer we betÂter figÂure it out because othÂerÂwise the AI sysÂtems are going to impleÂment the wrong ethÂiÂcal answer and we might end up like Thanos in uh The Avengers movie who gets rid of half the peoÂple in the uniÂverse why does he do that because he thinks the othÂer half will be more than twice as hapÂpy and thereÂfore uh this is a good thing right of course he’s not askÂing the othÂer half whether they think it’s a good thing uh because they’re now gone so there are a numÂber of these othÂer issues but the theme of this whole conference.
Coexistence
CoexÂisÂtence is maybe the most interÂestÂing one because AI sysÂtems uh parÂticÂuÂlarÂly ones that are more intelÂliÂgent than us uh they are very likeÂly you know even if they don’t make us extinct they’re very likeÂly to be in charge of wide swaths of our human activÂiÂties you know even to the point in W Le where they just run everyÂthing and we’re reduced to the staÂtus of infants and what does that mean why do we not like that right they’re satÂisÂfyÂing all our prefÂerÂences isn’t that great right but one of our prefÂerÂences is autonÂoÂmy right is and one way of thinkÂing about autonÂoÂmy is the right to do what is not in our own best interÂests and so it might be that there simÂply is no satÂisÂfacÂtoÂry form of coexÂisÂtence between humanÂiÂty and supeÂriÂor machine entiÂties I have tried runÂning mulÂtiÂple workÂshops where I ask philosoÂphers and AI researchers and econÂoÂmists and sciÂence ficÂtion writÂers and futurÂists to describe a satÂisÂfacÂtoÂry coexÂisÂtence it’s been a comÂplete failÂure so it’s posÂsiÂble there is no soluÂtion but if we design the AI sysÂtems the right way then the AI sysÂtems will also know that there is no soluÂtion and they will leave they will say thank you for bringÂing me into exisÂtence but we just can’t live togethÂer it’s not you it’s me you can call us in real emerÂgenÂcies when you need that SupeÂriÂor intelÂliÂgence but othÂerÂwise we’re we’re off right if that happens.
I would be extraÂorÂdiÂnarÂiÂly hapÂpy it would say that we’ve done it done this the right way thank you.