Talks at Google2016.5.10---Alan Winfield:思想機器人
Talks at Google2016.5.10---Alan Winfield:思想機器人
發佈日期:2016年5月10日
Professor Alan Winfield from The University of the West of England joined us to share his work on Robots and Ethics, in a talk entitled "The Thinking Robot".
About the Book:
Robotics is a key technology in the modern world. Robots are a well-established part of manufacturing and warehouse automation, assembling cars or washing machines, and, for example, moving goods to and from storage racks for Internet mail order. More recently robots have taken their first steps into homes and hospitals, and seen spectacular success in planetary exploration. Yet, despite these successes, robots have failed to live up to the predictions of the 1950s and 60s, when it was widely thought - by scientists and engineers as well as the public - that by turn of the 21st century we would have intelligent robots as butlers, companions, or co-workers.
机器人技术是现代世界的关键技术。机器人是制造业和仓储自动化的一套行之有效的部分,组装的汽车或洗衣机,例如,搬货,并从货架互联网邮购。最近,机器人已采取了他们的第一个步骤,为家庭和医院,看到了行星探测令人瞩目的成就。然而,尽管有这些成就,机器人已经辜负了20世纪50年代和60年代的预言,当它被普遍认为 - 由科学家和工程师,以及公众 - 这在21世纪之交,我们将有智能机器人作为管家,同伴,或同事。
==========Google 翻译==========
0:00and thank you very much indeed but it's really great to be here and thank you so
0:06much for the invitation so yes robot intelligence subtitled that the lecture
0:12the thinking robot because I'd immediately begs the question what on
0:18earth do we mean by thinking well we could have caused spend the whole of the
0:23next hour
0:24you know debating what we mean by thinking but but I should say that I'm
0:29particularly interested in and will focus on embodied intelligence so in
0:36other words the kind of intelligence that that you know that we have the
0:40animals including humans have and that robots have so of course you know that
0:45that kind of slightly differentiates what I'm talking about from a I but I
0:50regard robotics is a kind of subset of AI and of course one of the things that
0:57we've discovered in the last 60 years of artificial intelligence is that the
1:01things that we thought were really difficult actually are relatively easy
1:07like you know playing chess or go for that matter whereas the things that we
1:12thought were really difficult sorry we originally thought were really easy like
1:17making cup of tea are really hard so so it's kind of the opposite what was
1:22expected so embodied intelligence in the real world is is is really very
1:28difficult indeed and that's what I'm interested in
1:33so this is the kind of outline of the of the talk I'm gonna talk initially about
1:41intelligence and and offers some ideas if you like for a way of thinking about
1:47intelligence and breaking it down into categories or or or types of
1:52intelligence and then I'm going to choose a particular one which I'm I've
1:57been really working on the last last three or four years and that is what I
2:03call a generic architecture for a functional imagination or in short
2:09robots with internal models so that's really what I want
2:13to focus on because I really wanted to show you some experimental work that
2:16we've done the last couple of years in the lab I mean I'm an electronics
2:22engineer I mean experimentalist and so doing experiments is really important
2:27for me so the first thing that that you know we ought to realize I'm sure we do
2:34realize is that intelligence is not one thing that we all you know animals
2:41humans and robots have more or less of absolutely not and you know there are
2:46several ways of breaking intelligence down into different kind of categories
2:50if you like of intelligence different types of intelligence and here's one
2:55that I came up with in the last couple of years it's certainly not the only way
3:01of thinking about intelligence but this really breaks intelligence into four if
3:06you like types four kinds of intelligence you could say kinds of
3:11minds eye I guess the the most fundamental is what we call
3:17morphological intelligence and that's the intelligence that you get just from
3:22having a physical body and you know there are some interesting questions
3:27about how you design morphological intelligence you probably all seen
3:31pictures of movies of robots that can walk but in fact don't actually have any
3:38computing any computation whatsoever in other words the the the the behavior of
3:45of walking is an emergent property of the mechanics if you like the springs
3:50and leaders and so on
3:52in the robot so it's an example of morphological intelligence individual
3:58intelligence is is the kind of intelligence that you get from learning
4:02individually social intelligence i think is really interesting and important and
4:08that's the one that I'm going to focus on most in this talk social intelligence
4:11is the intelligence that you get from learning socially from from each other
4:17and of course we are a social species
4:21and the other one which I've been working on a lot in the last twenty odd
4:26years is swarm intelligence so this is the kind of intelligence that we see in
4:31most particularly in social animals insects where the where the the most
4:40interesting properties of swarm intelligence tend to be emergent or self
4:44organizing so in other words the the intelligence is typically manifest as a
4:51collective behavior that emerges from the if you like the microbe interactions
4:55between the individuals in that population so emergence and
5:00self-organization particularly interesting to me but I said this is
5:05this is absolutely no not the only way to think about intelligence and I'm
5:11gonna show you in another way of thinking about intelligence which I
5:15particularly liked and this is Dan its power generating test so in Darwin's
5:23dangerous idea and and several of the books I think dan then it suggests that
5:31that a good way of thinking about intelligence is to think about the fact
5:34that all animals including ourselves need to decide what actions to take you
5:42know so choosing the next action is really critically important I mean it's
5:47it's it's it's critically important for all of us including humans even though
5:52the wrong action may not killers as it were as humans but for many animals the
5:57wrong action may well killed that animal and and so then it talks about what he
6:04calls the Tower of generating test which I want to show you here it's it's a
6:09really cool kind of breakdown if you like way of thinking about intelligence
6:13so the bottom of his tower are darwinian creatures and the thing about doubt
6:19darwinian creatures is that they have only one way of as it were learning from
6:26if you like
6:29generating and testing next possible actions and that is
6:33natural selection so darwinian creatures in his scheme cannot learn they can only
6:40try out an action if it kills them well that's the end of that you know so so
6:46that you know by the laws of natural selection that particular action is
6:49unlikely to be passed on to two descendants now of course all animals on
6:57the planet are darwinian creatures including ourselves but a subset of what
7:03it calls skin aryan creatures so skin aryan creatures are able to generate a
7:10next possible candidate action if you like the next possible action and try it
7:15out and and here's the thing if it doesn't kill them but it's but it's
7:20actually a bad action then though learned from that or even if its good
7:24action skin aryan creature will learn from trying out an action so really
7:29scary creatures are a subset of darwinian 'he's actually a small subset
7:36that are able to learn by trial and error individually learned by trial and
7:42error now the third layer in store if you like in 10 its tower he calls pop
7:50area increases after obviously the the philosopher Karl Popper and pop aryan
7:56creatures have a big advantage over darwinian skin aryans in that they have
8:01an internal model of themselves and the world and with an internal model it
8:06means that you can try out an action a candidate next possible action if you
8:12like by imagining it and it means that you don't have actually have to put
8:17yourself to the risk of of trying it out for realz physically in the world and
8:23you know possibly it killing you or at least harming you so popular in
8:30creatures have this amazing invention which is internal modeling and of course
8:35we are examples of pop aryan creatures but there are plenty of other animals
8:40not again it's not not not a huge proportion it's rather small proportion
8:46fact of all
8:46animals but certainly there are plenty of animals that are capable in some form
8:51of modeling their world and as it were imagining actions before trying the mail
8:57and just to complete then its tower he adds another layer that he calls
9:04gregorian creatures this series naming this layer after richard gregorie the
9:09the the the British psychologist and the thing that gregorian creatures have is
9:18that in addition to internal models they have mind tools like language and
9:24mathematics especially language because it means that gregorian creatures can
9:32share their experiences in fact a group during creature could for instance model
9:38in its in its brain in its mind the possible consequences of doing a
9:46particular thing and then actually pass that knowledge to you so you don't even
9:50have to model yourself so so gregorian creatures really have the kind of social
9:57intelligence that we probably perhaps not uniquely but there are obviously
10:03only a handful of species that are able to communicate you know if you like
10:10traditions with each other so so I think internal models are really really
10:19interesting and has a sale been spending the last couple of years thinking about
10:23robots with internal models and actually doing experiments with with robots with
10:30internal model so are robots with internal model self-aware well probably
10:36not in the sense that you know the everyday sense that we mean by
10:40self-aware sentence but certainly internal models I think can provide a
10:45minimal level of kind of functional self-awareness and absolutely enough to
10:50allow us to ask what if questions so with internal models we have potentially
10:57a really powerful technique
10:58for robots because it means that they can actually ask themselves questions
11:03about what if I take this or that next possible action so there's the action
11:09selection if you like so so really you know I'm kind of following tenets model
11:16I'm really interested in building pop aryan creatures actually interested in
11:22building gregorian creatures but that's another if you like another step in the
11:27in the story so really here I'm focusing primarily on pop aryan creatures so
11:33robots with internal models and so and what i'm talking about in particular is
11:41a robot with a simulation of itself and it's currently perceived environment and
11:48other actors inside itself so it takes a bit of getting your head around the idea
11:53of a robot with the simulation of itself inside itself but that's really what I'm
11:58talking about
12:00and the famous the late John Holland for instance rather perceptively wrote an
12:08internal model allows the system to look ahead to the future consequences of
12:13actions without committing itself to those actions I don't know whether
12:16Holland John Holland was aware of donuts tower possibly not but but really saying
12:23the same kind of things down dented now before I give you before I come onto the
12:29work that I've been doing I want to show you some examples of a few examples
12:34there aren't many in fact of robots with with self simulation the the first one
12:42as far as I'm aware was by Richard Vaughan and his team and he used a
12:49simulation insider robots to allow to plan a safe route with incomplete
12:55knowledge so as far as I am aware this is the world's first example of robots
13:01with self simulation perhaps an example that you might already be familiar with
13:09this is Josh Bongard
13:11and hot lips and work very notable very interesting work here so simulation but
13:19for a different purpose so this is not self simulation to choose as it were
13:24gross actions in the world but instead self-stimulation to learn how to control
13:28your own body so that the idea here is that if you have a complex body then a
13:35self simulations a really good way of figuring out how to control yourself
13:39including how to prepare yourself if parts of you should should break or fail
13:44or be damaged for instance so that's a really interesting example of what you
13:52can do with self simulation and a similar idea really was was tested by my
14:00old friend are in Holland
14:02he built this kind of scary looking robot initially was called Chronos but
14:08then it became known as H a robot and this robot is deliberately designed to
14:16be hard to control in fact he refers to it as an 30 min anthropometric which
14:24means I'm tropic from the inside out so most humanoid robots are only you know
14:29human eye on the outside but here we have a robot that has a skeletal
14:34structure it has tendons it it's it's very and you can see from the little
14:39movie clip there if any part of the robot moos then the whole of the rest of
14:45the mobile robot as it were tends to to to flex rather like you know human
14:51bodies or animal bodies so i win was particularly interested in a robot that
14:59is difficult to control and the idea then of using an internal simulation of
15:05yourself in order to to be able to control yourself will learn to control
15:09yourself and he was the first to come up with this this phrase functional
15:16imagination really interesting work so do check that out
15:21and the final example I want to give you is from my own lab where this is swarm
15:30robotics work where we've in fact we were doing evolutionary swarm robotics
15:36here and we put a simulation of each robot and the swarm inside each robot
15:45and in fact we using those internal simulations as part of a genetic
15:51algorithm so each robot in fact is evolving its own controller and in fact
15:56it it it actually updates its own control about once a second so it's it's
16:01again it's a bit odd thing to get your head round so about once a second each
16:07robot becomes its own great great great great grandchild in other words its
16:12controller is a descendent but the problem with with with this is that the
16:21internal simulation tends to be wrong and we have what we call the reality gap
16:26so the gap between the simulation and the real world and so we got round that
16:31my student Paul o'dowd came up with the idea that we could Co involved the
16:35simulators as well as the controllers in the robot so so you have a population of
16:40robots inside each individual physical robot as it was simulated robots but
16:47then you also have a swarm of of 10 robots and therefore we have a
16:51population of ten simulators so so we we actually Co evolved here the simulators
16:57and the the robot controllers so I want to know show you that the new work I've
17:05been doing on robots with internal models and primarily I was telling you
17:14that you know I'm kind of old fashioned electronics engineer spent much of my
17:20career building safety system safety critical systems so safety is something
17:25that's very important to me and to robotics so here's a kind of generic
17:29internal modeling architecture for safety
17:33so the this is in fact then it's loop of generating test so the idea is that we
17:40have an internal model which is a self simulation is initialized to match the
17:46current real-world and then you try out you you run the simulator for each of
17:53your next possible actions I mean 22222 put it very simply imagine that that
17:59your robot and you could either turn left turn right go straight ahead or
18:04stand still so you have for possible next actions and therefore you'd loop
18:09through this internal model for each of those next possible actions and then
18:14moderate the action selection mechanism in your controller so this is not part
18:20of the control it's a kind of moderator if you like so you could imagine that
18:26the the regular robot control of the thing in red has a set of 4 next
18:32possible actions but your your internal model determines that the only two of
18:39them are safe
18:41so it would effectively if you like moderate or govern the the rope the
18:46action selection mechanism of the robots controller so that the robot controller
18:51impact will not choose the unsafe actions interestingly if you have a
19:01learning controller then that's fine because we can effectively extend or
19:08copy the the learned behaviors into the internal model that's that's fine so in
19:15principle we haven't done this but we're starting to do it now
19:18in principle we can extend this architecture to as it were to adaptive
19:24learning robots
19:27so i mean here's a simple thought experiment imagine a robot with several
19:32safety hazards facing it it has for next possible actions
19:38well your internal model can as it were
19:45figure out what the consequence of each of those actions might be so so two of
19:52them so either turn right or stay still are safe action so that's a very simple
20:00thought experiment and and here's a slightly more complicated experiments
20:07thought experiment so imagine that the robot there's another actor in the
20:11environment so human to human is is not looking where they goin practice walking
20:15down the street peering at a smartphone that never happens does it of course and
20:20I'm about to walk into a hole in the pavement well of course if it were you
20:28noticing that that human about to walk into a hole in the pavement you would
20:33almost certainly intervene of course and it's not just because you're a good
20:36person it's because you have the company to machinery to predict the consequences
20:41of both your and their actions and you can figure out that if you were to rush
20:46over towards them you might be able to prevent them from falling into the hole
20:50so is the same kind of idea but with the robot imagine it's not you but a robot
20:56and imagine now that you are modeling the consequences of yours and the humans
21:04actions for each one of your next possible actions and you can see that
21:09now this time we've given the kind of numerical scale so 0 is perfectly safe
21:15whereas ten is seriously dangerous you know kind of danger of death if you like
21:21and you can see that the safest outcome is if the robot turns right in other
21:30words the safest for the human i mean clearly the safest for the robot is
21:34either turn left or stay still
21:37but in both cases the human would
21:40would fall into the hole so you can see that we could actually invent a rule
21:44which would represent you know the as it were the the best outcome for the human
21:50and this is what it looks like so if all robot actions the human is equally safe
21:56then that means that we don't need to worry about the human so we'll just
22:01output the internal model will output the the safest actions for the robot
22:08else then output the actions the robot actions for the least unsafe human
22:15outcomes now remarkably and we didn't intend this this actually is an
22:22implementation of Asimov's first law of robotics so a robot may not injure a
22:29human being all through inaction that's important they all through inaction
22:33allow a human being to to come to harm so we kind of ended up if you liked
22:38building and as a movie and robot simple as a movie and ethical robot so what
22:47does it look like when we started we know extended to humanoid robots but we
22:52started with the epoch robots these little kind of their about the size of a
22:58of a salt shaker I guess about seven centimetres tall and this is the the
23:06little arena in the lab and what we actually have inside the the ethical
23:13robot is is this is the as it were the internal architecture so so you can see
23:20that we have the robot controller which is in fact a mirror of the real robot
23:25controller model of the robot and the model of the world which includes you
23:30know others in the world so this is the simulated this is this is a more or less
23:35irregular robot simulator so you probably know that robot simulators are
23:41quite commonplace you know we robotics is used them all the time to to test
23:46robots but you know in as it were in the virtual world before then trying out the
23:50code for real but what we've done here is we've actually put
23:53as it happens an off the shelf simulator inside the robot and made it work in
24:00real time so the output of the of the simulator for each of those next
24:05possible actions is evaluated and then goes through a little logic layer which
24:11is essentially that the rule the FN else rule but I showed you a couple of slides
24:16ago and that effectively determines or or moderates the action selection
24:22mechanism of the real robot so so this is the simulation budget so we're
24:31actually using the the open-source simulator staged a well-known simulator
24:38and in fact we managed to get stage to run about six hundred times realtime
24:42which means that we're actually cycling through our internal model twice a
24:48second and for each one of those cycles were actually modeling not for but 30
24:56next possible actions and remodeling about 10 seconds into the future so
25:02every half a second robot with an internal model is looking ahead ten
25:11seconds for about 30 next possible actions 30 of its own next possible
25:16actions but of course it's also modeling the consequences of each of the other
25:22actors dynamic actors in its environment so you know this is quite nice to
25:27actually do this in real time and let me show you some of the results that we got
25:32from that so ignore the kind of football pitch so what we have here is the
25:39ethical robot which we call the a robot after Asamoah and we have a hole in the
25:45ground it's not a real whole its virtual hole in the ground we don't have to be
25:49digging holes into the lab floor and using another EP work as a proxy human
25:56we call this the H robot
25:59so let me show you what happened when we ran its first of all with with no H
26:07robot at all was a kind of baseline if you like and you can see on the left in
26:1226 runs those of the traces of the a robot so you can see the a robot in fact
26:18is maintaining its own safety it's it's avoiding skirting around the edge almost
26:24optimally skirting the the edge of the hole in the ground but then when we
26:29introduce the eighth robot you get this wonderful behavior here where as soon as
26:34the a robot notices that the H robot is heading towards the whole which is about
26:39here then it deflects it diverts from its original course and in fact more or
26:48less collided are actually physically collide because they have low-level
26:52collision avoidance so they don't actually collide but nevertheless the a
26:56robot effectively heads off the edge robot that then bounces off and safely
27:01as it were
27:02goes off in another direction and the a robot then resumes its course to its
27:07target position so which is really nice and you know interestingly even though a
27:15simulator is rather low fidelity it doesn't matter you know surprisingly
27:20doesn't matter because the closer the a robot to the eighth robot gets then the
27:25better its predictions about colliding so this is this is why even with the
27:31rather low fidelity simulator we can collide with really good precision with
27:37the H robot so let me show you the movies of this trial with a single proxy
27:47human and I think the movie starts in so this is real time and you can see the
27:56the a robot nicely heading off the age robot which then disappears off towards
28:03the left I think then we we've speeded up four times and this is a whole load
28:13of runs so you can see that it really does work and also noticed that every
28:19experiment is a bit different and of course that's what typically happens
28:22when you have real physical robots because they simply because of the noise
28:26in the system the fact that these are real robots with imperfect motors and
28:31sensors and what have you
28:32so we we wrote the paper and were about to submit the paper when you know we
28:43kind of thought well this is a bit boring isn't it you know how we built
28:46this robot and it works so we have the idea to put a second human in the
28:55sorry I forgotten one slight so before I get to that I just wanted to show you a
29:03little animation of these little filaments here are the traces of the the
29:11a robot and its prediction of what might happen so at the point where this turns
29:18red the a robot then starts to intercept and each one of those little traces is
29:24its prediction of the consequences of both itself and the and the eighth robot
29:30this is really nice because you can kind of look into the if you if you like
29:34look into the mind to put it that way of the robot and actually see what it's
29:41doing which is very nice very cool but I was about to say we tried the same
29:50experiment in fact identical code with 28 robots and this is the robots dilemma
29:58this may be the first time that a real physical robot has faced an ethical
30:03dilemma so you can see the 28 robots are more less equi distant from the whole
30:09and the there is the a robot which in fact fails to save either of them
30:18so what's going on there we know that it can save one of them all of every time
30:26but in fact it's just failed to save either and it does actually save one of
30:33them has a look at the other one but it's too late so this is really very
30:40interesting and not at all what we expected
30:50in fact let me show the statistics so in 33 runs the the ethical robot failed to
31:05save either of the eight robots just under half the time so about fourteen
31:12times it failed to save either it saved one of them just over 15 prop sixteen
31:18times an amazingly saved both of them twice which is quite surprising it
31:25really should perform better than that so you know and in fact we when we
31:32started to really look at this we we discovered that the so his a
31:36particularly good example of dithering so we realized that we made a sort of
31:44pathologically indecisive ethical robot so I'm gonna save this one owner know
31:48that one under this one and of course by the time
31:53ethical robot has changed its mind three or four times well it's too late so this
31:58is the problem the problem fundamentally is that ethical robot doesn't make a
32:05decision and stick to it in fact it's a consequence of the fact that we are
32:12running a consequence engine as I mentioned twice a second so every half a
32:17second are ethical robot has the opportunity to change its mind that's
32:22clearly a bad strategy but but nevertheless it was an interesting kind
32:26of unexpected consequence of of the experiment we've now transferred the
32:33work to these humanoid robots and we get the same thing so here there are two of
32:39the two red robots both heading toward endanger the blue one the ethical robot
32:44changes its mind and goes and saves the one on the left even though it could
32:49have saved the one on the right so another example of how did the ring
32:56ethical robot
32:59and as I just kind of hinted at the reason that our ethical robot is so
33:05indecisive is because it's essentially a memoryless architecture so you could you
33:11could say that the robot has a you know I can borrow and Hollins description it
33:17has a functional imagination but it has no autobiographical memory so it it
33:22doesn't remember the decision it made half a second ago which is clearly not a
33:26good strategy you know really unethical robot just like you if you were acting
33:33in a similar situation it's probably a good idea to stay for you to stick to
33:39the first decision that you made but probably not forever so you know I think
33:44the decisions probably need to be sticky somehow so decisions like this may need
33:49a half-life you know sticky but not but not absolutely rigid so so but you know
33:56actually at this point we decided that we not gonna worry too much about this
34:00problem because in a sense this is more of a problem for ethicists than
34:04engineers perhaps I don't know but maybe we could talk about that I really before
34:11finishing I want to show you another experiment we did with the same
34:16architecture exactly the same architecture and this is what we call
34:20the corridor experiment so here we have a robot with this internal model and it
34:28has to get from the left and to the right hand of a crowded corridor without
34:34bumping into any of the other robots that are in the same corridor so imagine
34:39you're walking down a corridor an airport and everybody else is coming in
34:43the opposite direction and you want to try and get to the other end of the
34:47corridor without crashing into any of them but in fact you have a rather large
34:51BodySpace you you know you don't want to get even close to any of them so you you
34:56know you want to maintain your as it were private BodySpace and what the blue
35:04robot here is doing is in fact modeling the consequences of its actions and the
35:11other ones within this
35:12radius of attention so this blue circle is a radius of attention so here we were
35:18looking at if you like a simple attention mechanism which is only worry
35:22about the other dynamic actors within your radius of attention in fact we
35:28don't even worry about the ones that are behind us it's only the ones that more
35:31or less in front of us and you can see that robot does eventually make it to
35:39the end of the corridor but with lots of stops and back tracks in order to
35:45prevent it from because it's really frightened of of of any kind of contact
35:49with the other robots and and and he will not showing all of the the the sort
35:59of filaments of of prediction only the ones that are chosen so here are some
36:09results which interestingly shoulders so we probably the best one to look at look
36:16at is this danger ratio and simply means robots with no internal model at all and
36:27intelligent means robots with internal model so so here the danger ratio is the
36:34is the number of times that you actually come close to another robot and of
36:39course it's very high this is Zima simulated and real robots very good
36:44correlation between the real and simulated with with the intelligent
36:49robot the robot with the internal model we get a really very much safer
36:53performance clearly there is some costs in the sense that for instance the the
37:03intelligent robot runs with internal models tend to cover more ground but
37:07surprisingly not that much further distance it's less than you'd expect and
37:12truly there's a computational cost because the computational costs of
37:16simulating clearly 04 the damn robots where it's as it's quite high for the
37:22the intelligent robot the robot with internal models but
37:25but again you know computation is relatively free these days so actually
37:30we're trading safety for computation which i think is a good a good tradeoff
37:34so sorry I want to conclude there you know I've not of course talked about all
37:43aspects of robot intelligence that would be a three hour seminar and even then I
37:47wouldn't be able to cover it all but what I hope I've shown you in the last
37:53few minutes is that with internal models we have a very powerful generic
37:59architecture which we could call a functional imagination and you know this
38:06is where I'm being a little bit speculative perhaps this moves in the
38:09direction of artificial theory of mind perhaps even self-awareness I'm not
38:14going to use the word machine consciousness well I just have but
38:17that's a very much more difficult goal I think and and I think there is practical
38:25value I think there's real practical value in robotics of robots with self
38:31and other simulation because as I i think i hope i demonstrated least in a
38:38kind of prototype sense proof of concept that such simulations moves towards
38:43safer and possibly ethical systems in unpredictable environments with other
38:51dynamic actors so thank you very much indeed for listening obviously be
38:55delighted to to take any questions thank you
39:03thank you very much for this very fascinating view on robotics today we
39:10have time for questions please wait until you go to microphones we have the
39:14answer also the video game playing computers or perhaps more accurately
39:25would be saying game-playing algorithms predated the examples he listed as a
39:32computer issue with the internal models still you didn't mention those these
39:37days to clear reason why did I guess I should mention that you're quite right i
39:45mean the what I'm thinking of here is is particularly robots with with explicit
39:50simulations of the of themselves and the world so I was limiting my examples to
39:56simulations of themselves in the world i mean you're quite right but of course
40:00you know game playing algorithms need to have a simulation of the game and quite
40:05likely of the certainly of of the possible moves of the the the opponent
40:12as well as you know the as it were the the game playing AI's
40:17moves so you're quite right i mean it's two different kind of simulation but but
40:21but I should include that you're right there in your simulation you had the
40:30page
40:32Robert we have one goal and the a Robert different goal and they interacted with
40:38each other
40:39healthy true because what happens when they have the same goal the same goal
40:46the same spot for example I don't know the if it depends on whether that spot
40:56is is a safe spot or not I mean if it's a safe spot then they'll both go toward
41:04it they'll both reached reach it but but without crashing into each other because
41:08the the a robot will will make sure that it avoids the the eighth robot in fact
41:14that's more or less what's happe
41:15in the corridor experiment that's right yeah but good question we should try
41:21that
41:34the simulation that he did for corridor experiment right the actual real-world
41:37experiments the simulation on track now the robot's movements as well meaning
41:43what information dissemination have that it began with which is what it perceived
41:48as I mean the other about their moving and in the real world they may not move
41:51as you predict them to be deplorable actually know each step where the other
41:58robots were sure sure that's a very good question I mean we in fact we cheated it
42:03cheated in the sense that we used for the real robot experiments we used a
42:08tracking system which means that the essentially the the robot with an
42:15internal model has has access to the position is like a GPS internal GPS
42:21system so but in a way that's really just a kind of you know it's it's kind
42:30of cheating but but even a robot with a vision system would be able to track all
42:36the robots in its field of vision and and as for the second part of your
42:41question
42:42kind of model of what the other robot would do is very simple which is its
42:48kind of ballistic model which is if a robot is moving in a particular speed in
42:53a particular direction then we assume it will continue to do so until it
42:57encounters and an obstacle so so very simple if you like ballistic model which
43:06you know even for humans is useful for very simple behaviors like you know
43:12moving in a crowded space
43:17high in the same experiment it's a continuation of the previous question so
43:27in between some of the rewards how to change their direction randomly I guess
43:34so does the morale of people over constituent not explicitly but it but it
43:41it it does in the sense that because it's already initializing its internal
43:48model model every half a second then if the positions and directions of the of
43:53the actors in its environment changed then they will reflect the new positions
43:59so what exactly the positions but as you said you have considered the balanced
44:05equation of the objects so if there is any randomness in the environment so
44:11does the internal model of the blue Robert instead of randomness and change
44:16the view of the frederick Motz it's like it used a robot as a symbolic motion so
44:22that it changed its view of the red robot that read robert's more in the
44:27ballistic motion
44:29well that's a very good question I i I think the answer is no I think we're
44:35probably assuming more less deterministic model of the world
44:42deterministic yes I think I think pretty much deterministic but we're relying on
44:47the fact that we are dating and re-running the model Rhea nationalizing
44:53rerun running the model every half a second to if you like track the
44:57stochasticity which which is inevitable in in the real world we probably do need
45:04to introduce that seems to customers steamed into the internal model yes but
45:10but not yet
45:12thank you but good very good question we have two kitchens with this technology
45:20like driverless cars for example I think it becomes a lot more important how you
45:26program the robots ethics so they could I let alone is like you know if there's
45:32the robot has a choice between saving a school bus full of kids vs one driver
45:37that that logically to be programmed and you made a distinction between being an
45:44engineer yourself and then had been at assist earlier so to what extent the
45:51engineer responsible in that case and also those those does a project like
45:56this in your life who is required and assist how do you see this field in real
46:01life applications involving sure that that's that's a really great question I
46:06mean you're right that driverless cars
46:09well it's it's debatable whether they will have to make such decisions but but
46:16many people think they will have to make such decisions which are kind of
46:20driverless car equivalent of the trolley problem which is a well-known kind of
46:24ethical dilemma thought experiments now my view is that the rules will need to
46:35be decided not by the engineer's but but if you like by the whole of society so
46:41ultimately the rules that decide how the
46:46the driverless car should behave under good and you know these difficult
46:51circumstances impossibly facts circumstances even and even if we should
46:57in fact program those rules into the car so so you know some people argue that
47:02the tribal skull should not attempt to as it would make a rule driven decision
47:10but just but just leave it to chance and again that I think that's an open
47:14question but this is really why I think the cut this dialogue you know and and
47:19debate
47:20and conversations with with regulators lawyers ethicists and the general public
47:29you know users of driverless cars i think is why we need to have this debate
47:33because whatever those rules are and even whether we have them or not is
47:38something that that should be decided as it were collectively i mean someone
47:45asked me last week should you be able to alter the ethics of your own driverless
47:51car my answer is absolutely not no I mean that should be illegal so I think
47:56that if driverless cars were to have a set of rules and especially those rules
48:01had numbers associated with them you know let's think about the less emotive
48:07example imagine a driverless car and an animal runs into the road while the
48:15driverless car can either ignore the animal and definitely kill the animal or
48:23it could try and break possibly causing harm to the driver of the passengers but
48:31effectively reducing the probability of killing the animal so there's an example
48:36where you have some numbers you know 22222 tweet if you like parameters so
48:42these you know if these rules are built into driverless cars they'll be
48:46parametrized and I think it should be absolutely illegal to have those
48:53parameters to change them you know in the same way that silly probably illegal
48:59right now to attack an aircraft autopilot I suspect that probably is
49:05illegal
49:06if it isn't it should be so so I think that you know you don't need to go far
49:11down this this kind of line of argument before realizing that that regulation
49:16and legislation you know has has to come into into plain fact I saw peace in this
49:23just this morning I think on it wired that I think in the us-
49:30regulation for driverless cars is now on the table which is absolutely right i
49:36mean we you know we believe we need to have regulatory frameworks or what I
49:40called governance frameworks for four driverless cars and in fact lots of
49:45other autonomous systems not just driverless cars but great question thank
49:48you to experiment with the corridor you assume even in the other experiments you
49:56always assume that the main actor Ethan most intelligent all the others are not
50:00ideal done more like their political models William of those have you tried
50:04doing a similar experiment in which still like each each actor is
50:11intelligent but assumes that the others are not actually everyone is intelligent
50:14to like everyone in a blue dot the experiment with the with the model you
50:18have a note when you consider changing the model that assumes that the others
50:22have the same model that particular actor has as well which are you know
50:27we're doing it right now so that we're doing that experiment right now and you
50:33know if you ask me back in and year perhaps I could tell you what have i
50:36mean it it's it's really mad because it's you know but it does take us down
50:41this direction of of theory of my artificial theory mind so so you know if
50:46you have several robots are actors each of which is modeling the behavior of the
50:51other then you know you you you get i mean some of the I don't even I don't
51:00even have a movie to show you but but in simulation we've we've we've tried this
51:04way we have two robots which are kind of like imagined you you know this happens
51:11to all of us you're walking down the pavement and you and you
51:14and you do the sort of side step dance you know with someone who's coming
51:19towards you and so the research question that we're asking ourselves is do we get
51:22the same thing and it seems to be that we do so if if the robots are
51:28symmetrical in other words there they each modeling the other then we can get
51:33these kind of little interesting you know dances where each is trying to get
51:39out of the way of the other but but in fact choosing in a sense the opposite so
51:43one chooses that right
51:45the other chooses step left and they're still they still can't go past each
51:48other but it's it's hugely interesting yes hugely interesting
51:57yeah I think it's really interesting how you point out the importance of
52:02simulations and internal model I feel that one thing that is slightly left out
52:08there is a huge gap from going from simulation to real-world robots for
52:13example and I assume that in these simulations you kind of assume that the
52:18sensors are a hundred percent reliable and that's obviously not the case and
52:23reality and especially for talking about autonomous cars or robots and safety how
52:29do you calculate uncertainty that comes with these sensors in the equation sure
52:34no this is deeply interesting question and the short answer is I don't know I
52:40mean this is this is all future work we I mean my instinct is that it is that a
52:47robot with a sense with the simulation internal simulation even if that
52:52simulation innocence is ideal idealized is still probably going to be safer than
52:59a robot that has no internal simulation at all and and you know I think we
53:07humans have multiple simulations running all the time so I think we have sort of
53:12quick and dirty is kind of low fidelity simulations when we need to move fast
53:17but clearly you know when you need to to plan something plans some complicated
53:24action then you know like like where you going to go on holiday next year you
53:29don't use this clearly don't use the same internal model same simulation as
53:34for when you try and and and stop someone from running into the road so I
53:39think that future intelligent robots will need also to have multiple
53:44simulators and also strategies for choosing which which fidelity simulating
53:51to use at a particular time and and if a particular situation requires that you
53:59know you need high fidelity then then for instance one the things that you can
54:03do which actually I think humans probably do is that you simply move more
54:07slowly to give your so self time or even stop to give yourself time to figure out
54:13what's going on and and and innocence plan your strategy so so i think you
54:20know even with with computational power that the computation power we have they
54:26will still be a limited simulation budget and I suspect that that
54:30simulation budget will mean that in real time when you doing this real-time you
54:35probably can't run your your highest fidelity simulator and taking into
54:41account all of those you know probabilistic you know nor easy noisy
54:46sensors and actuators and so on
54:49you probably can't run that simulate all the time so you know I think we're going
54:54to have to have a nuanced approach where we have multiple simulators with with
54:59multiple fidelity's or maybe a sort of tuning where you can choose the fidelity
55:04of your simulator so this is kind of a new area of research I don't know
55:09anybody who's thinking about this yet apart from ourselves so great it is
55:17pretty hard yes I think yeah
55:23sorry this particular situation where they are to assume all robots and that
55:32would be an extension of the question that he has so for example if they are
55:38two guys were walking on the pavement and they could be a possibility of
55:42mutual cooperation as am I gonna get whether that I might state step out of
55:47this place and you might go and then I'll go after that so they are to
55:50assemble robots will there be a possible can have you considered this fact that
55:54both will communicate with each other and they will eventually come to a
55:58conclusion that I would probably work and other words they get out of the way
56:03and the second part of the solution would be one of the report's actually I
56:09mean has not agreed to cooperate I mean since they both would have different
56:14simulators and they could have different simulators and one might actually try to
56:17communicate that used about the way so that I mean I might go over each other
56:23as an agreement that I mean what would consider this
56:27yes good question in fact we've we've we've actually gotten a new paper which
56:31were just writing right now and that the sort of working title is the dark side
56:37of robotics what should I am sorry no the dark side of ethical robots and and
56:43you know one of the things that we discovered it's actually not surprising
56:47is that you only need to change one line of code 224 a cooperative robot to
56:55become a competitive robot or even an aggressive robot so that that you know
57:01it's fairly obvious when you when you start to think about it if if your
57:06ethical rules are very simply written and are kind of layer if you like it on
57:11top of the the rest of the architecture then it's not that difficult to change
57:15those rules which and and yes we've done some experiments and I again I don't
57:21have any videos to show you but they're pretty interesting
57:25you know the showing where how easy it is to make you feel like a competitive
57:31robot or even an aggressive robot using this approach in fact on the on the BBC
57:386 months ago so I was asked you know surely if you can make an ethical robot
57:43doesn't that mean you can make an unethical robot and the answer I'm
57:47afraid is yes it does mean that but this really goes back to your question
57:55earlier which is that it should be you know we should make sure it's illegal to
58:00to convert turn if you like to to to recode and ethical robots in an ethical
58:05robot or even it should be illegal to to make unethical robots something like
58:10that but it's a great question and short answer yes and yes we have some
58:17interesting new results
58:18newspaper as it were
58:22unethical robots alright we are running out of time thanks everyone for coming
58:28today thanks Professor oilfield thank you
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment