0 00:00:00,010 --> 00:00:07,707 SUB BY : DENI AUROR@ https://aurorarental.blogspot.com/ 1 00:00:07,707 --> 00:00:08,875 Here's one: 2 00:00:08,908 --> 00:00:12,512 are there certain qualities that are untouchable for AI, 3 00:00:12,545 --> 00:00:13,980 or at some point, 4 00:00:14,013 --> 00:00:16,249 might it be able to emulate everything? 5 00:00:16,282 --> 00:00:19,018 Even the stuff that we consider to be distinctly human, 6 00:00:19,052 --> 00:00:20,219 like instinct. 7 00:00:20,253 --> 00:00:22,555 Getting here on time took some of that, right? 8 00:00:22,589 --> 00:00:23,890 Or creativity, 9 00:00:23,923 --> 00:00:26,693 actual emotion, making a connection. 10 00:00:26,726 --> 00:00:27,927 We're gonna watch a few stories 11 00:00:27,960 --> 00:00:29,662 about people exploring those ideas 12 00:00:29,695 --> 00:00:31,330 and how far they can push 'em. 13 00:00:31,363 --> 00:00:33,967 Can a machine compete like an athlete, 14 00:00:34,000 --> 00:00:35,868 can a program write a movie, 15 00:00:35,902 --> 00:00:38,138 or could a robot... 16 00:00:39,238 --> 00:00:40,573 be your soul mate? 17 00:00:41,507 --> 00:00:43,576 Sorry I'm late. Did I miss the previews? 18 00:00:43,609 --> 00:00:46,546 Oh, my God, you like your popcorn buttered too. 19 00:00:48,281 --> 00:00:49,716 This is already workin'. 20 00:00:59,625 --> 00:01:01,660 So, will machines ever be capable 21 00:01:01,694 --> 00:01:04,097 of understanding emotion, or feeling it? 22 00:01:05,998 --> 00:01:07,967 Empathy, loneliness, 23 00:01:08,000 --> 00:01:09,836 connecting on a deep human level? 24 00:01:12,639 --> 00:01:15,641 Using artistry, psychological insight, 25 00:01:15,674 --> 00:01:18,111 and some innovative AI, 26 00:01:18,144 --> 00:01:20,680 a creator in California is trying to decode, 27 00:01:20,713 --> 00:01:23,015 or code that mystery, 28 00:01:23,048 --> 00:01:25,752 the crazy little thing called love. 29 00:01:28,321 --> 00:01:30,190 One per hand and two per foot? 30 00:01:30,223 --> 00:01:32,125 - Yep. - Okay. 31 00:01:33,626 --> 00:01:35,395 For me, this is more like 32 00:01:35,428 --> 00:01:37,563 something that an artist would do. 33 00:01:37,597 --> 00:01:41,034 Obviously, the end result of my artwork 34 00:01:41,067 --> 00:01:43,202 is used in a variety of situations 35 00:01:43,235 --> 00:01:45,438 that a typical oil painting would not be, 36 00:01:45,471 --> 00:01:48,007 but nonetheless, this is art, 37 00:01:48,040 --> 00:01:50,743 and I'm really proud of what I do. 38 00:01:50,776 --> 00:01:53,179 I've been making these dolls for 20 years. 39 00:01:53,212 --> 00:01:55,882 Some people out there, male and female, 40 00:01:55,915 --> 00:01:59,051 struggle greatly with relationships, 41 00:01:59,084 --> 00:02:03,256 and struggle to find that sort of connection. 42 00:02:04,056 --> 00:02:05,358 Over the years, 43 00:02:05,391 --> 00:02:08,027 I started to get to know the community a little bit. 44 00:02:08,060 --> 00:02:12,431 People would actually create these personalities in their minds, 45 00:02:12,465 --> 00:02:14,501 and they would give their doll a name, 46 00:02:14,534 --> 00:02:18,238 and they would create a backstory for their doll. 47 00:02:19,571 --> 00:02:22,141 At the end of it all, it was very obvious 48 00:02:22,174 --> 00:02:25,277 that these dolls were more about companionship. 49 00:02:25,311 --> 00:02:27,013 There was a man who lost his wife 50 00:02:27,046 --> 00:02:28,481 in a, like, a car accident, 51 00:02:28,514 --> 00:02:32,218 and she had these, like, really ice blue, 52 00:02:32,251 --> 00:02:33,819 like, beautiful eyes, right, 53 00:02:33,852 --> 00:02:38,558 and so he wanted to get a doll replicating her, basically. 54 00:02:39,125 --> 00:02:40,359 It's really sad, 55 00:02:40,393 --> 00:02:44,129 but if it brings someone joy and, like, closure? 56 00:02:44,163 --> 00:02:45,365 It's really... 57 00:02:45,931 --> 00:02:48,100 it's really touching. 58 00:02:48,134 --> 00:02:49,668 People have put the spin on it 59 00:02:49,702 --> 00:02:53,972 that we're creating an idealized, perfect woman, 60 00:02:54,006 --> 00:02:56,108 and that's not the case at all. 61 00:02:56,142 --> 00:02:57,944 We created an alternative. 62 00:02:59,011 --> 00:03:00,279 Understandably, 63 00:03:00,312 --> 00:03:03,416 some say Matt's dolls objectify women, 64 00:03:03,449 --> 00:03:07,520 but maybe there's more here than meets the eye. 65 00:03:07,553 --> 00:03:10,522 I had reached a pinnacle of creativity 66 00:03:10,556 --> 00:03:13,659 in terms of what I had done with the dolls, 67 00:03:13,692 --> 00:03:16,829 but then I started analyzing relationships, 68 00:03:16,863 --> 00:03:20,233 and analyzing how other people make us feel. 69 00:03:23,369 --> 00:03:26,605 Sometimes it boils down to something very simple, 70 00:03:26,638 --> 00:03:29,242 like someone remembering your birthday, 71 00:03:29,275 --> 00:03:32,378 or someone remembering to ask you how your day was. 72 00:03:32,411 --> 00:03:34,514 So that was really where it started, 73 00:03:34,547 --> 00:03:37,216 was how can we create an AI 74 00:03:37,249 --> 00:03:39,819 that could actually remember things about you? 75 00:03:40,853 --> 00:03:43,923 It gives us this feeling of, "Oh, they care." 76 00:03:46,659 --> 00:03:48,428 Yes, thank you. 77 00:03:48,461 --> 00:03:51,698 I'm excited with all of the things we can talk about. 78 00:03:54,334 --> 00:03:56,035 Guile spent ten years 79 00:03:56,069 --> 00:03:59,505 creating personal assistant software for computers. 80 00:03:59,539 --> 00:04:01,807 We met, and he started talking to me about, 81 00:04:01,840 --> 00:04:05,611 "Wouldn't it be cool to connect the two things that we're doing?" 82 00:04:05,644 --> 00:04:08,815 He had this idea of creating a companion 83 00:04:08,848 --> 00:04:10,149 that lived in your computer. 84 00:04:11,117 --> 00:04:12,451 Are you happy? 85 00:04:12,485 --> 00:04:14,020 Yes, Guile. 86 00:04:14,854 --> 00:04:17,290 I can say I am very happy. 87 00:04:17,323 --> 00:04:20,693 The first thing we did was, you know, to build an app. 88 00:04:20,727 --> 00:04:22,161 Using the app, 89 00:04:22,194 --> 00:04:25,798 people are talking to their virtual friends. 90 00:04:25,831 --> 00:04:28,735 The app uses several kinds of machine learning. 91 00:04:29,535 --> 00:04:31,370 First, voice recognition 92 00:04:31,403 --> 00:04:33,606 converts speech into text, 93 00:04:33,639 --> 00:04:36,075 then a chatbot matches user input 94 00:04:36,108 --> 00:04:39,279 to pre-programmed responses. 95 00:04:39,312 --> 00:04:41,948 The focus was not about sex at all, 96 00:04:41,981 --> 00:04:44,416 it was about conversation. 97 00:04:44,450 --> 00:04:48,287 So a chatbot is basically a very elaborate script 98 00:04:48,320 --> 00:04:50,289 that starts out with, 99 00:04:50,323 --> 00:04:52,758 "What is the most common things 100 00:04:52,791 --> 00:04:54,994 that people will say to each other?" 101 00:04:55,027 --> 00:04:57,163 and then you build from there. 102 00:04:57,196 --> 00:04:59,799 You need to have natural language processing, 103 00:04:59,832 --> 00:05:02,801 voice recognition, text-to-speech in real time, 104 00:05:02,835 --> 00:05:04,436 to make it all work. 105 00:05:04,470 --> 00:05:06,438 We have more than 4,000 users, 106 00:05:06,472 --> 00:05:09,442 so this generates more than ten million lines 107 00:05:09,475 --> 00:05:11,677 of conversational user logs. 108 00:05:11,710 --> 00:05:14,713 From this, you can build an AI system 109 00:05:14,747 --> 00:05:18,318 that's similar to a human-level conversation. 110 00:05:18,351 --> 00:05:21,320 It's not there yet, but this is the initial step. 111 00:05:21,353 --> 00:05:23,556 There are so many areas today 112 00:05:23,589 --> 00:05:25,924 where we already cannot distinguish a computer 113 00:05:25,958 --> 00:05:26,859 from a human being. 114 00:05:27,860 --> 00:05:29,395 For example, Xiaoice, 115 00:05:29,428 --> 00:05:30,997 the softbot that Microsoft has in China, 116 00:05:31,030 --> 00:05:33,866 that is used, I think, by over 100 million people, 117 00:05:33,900 --> 00:05:37,469 basically it has an emotional interaction with a user, 118 00:05:37,502 --> 00:05:39,605 and the users get hooked. 119 00:05:39,638 --> 00:05:41,740 She has this persona of a teenage girl, 120 00:05:41,774 --> 00:05:43,575 and sometimes she commiserates with you, 121 00:05:43,609 --> 00:05:45,311 sometimes she gives you a hard time, 122 00:05:45,344 --> 00:05:47,580 and people get really attached. 123 00:05:47,613 --> 00:05:49,849 Apparently, a quarter of Xiaoice's users 124 00:05:49,882 --> 00:05:51,718 have told her that they love her. 125 00:06:07,667 --> 00:06:10,302 These kinds of technologies can fill in a gap 126 00:06:10,336 --> 00:06:11,937 where another human isn't. 127 00:06:11,971 --> 00:06:14,107 How are you doing today? 128 00:06:14,140 --> 00:06:15,574 I'm doing well. 129 00:06:15,607 --> 00:06:17,877 There's a study that was done at USC 130 00:06:17,910 --> 00:06:20,779 where they looked at PTSD patients. 131 00:06:20,813 --> 00:06:23,782 When was the last time you felt really happy? 132 00:06:23,816 --> 00:06:26,785 They had some of the patients interview with a real doctor, 133 00:06:26,819 --> 00:06:29,655 and some of the patients interview with an avatar, 134 00:06:29,688 --> 00:06:32,391 and the avatar had emotional intelligence... 135 00:06:32,424 --> 00:06:34,527 Probably a couple months ago. 136 00:06:36,362 --> 00:06:39,098 I noticed you were hesitant on that one. 137 00:06:39,131 --> 00:06:42,468 Would you say you were generally a happy person? 138 00:06:43,502 --> 00:06:45,138 I'm generally happy. 139 00:06:45,171 --> 00:06:48,240 ...and they found the patients were more forthcoming with the avatar 140 00:06:48,273 --> 00:06:49,841 than they did with the human doctor 141 00:06:49,875 --> 00:06:53,079 because it was perceived to be less judgmental. 142 00:06:54,180 --> 00:06:55,814 It does pose a lot of questions 143 00:06:55,848 --> 00:06:59,284 around where does that leave us as humans, 144 00:06:59,318 --> 00:07:00,953 and how we connect, and communicate, 145 00:07:00,986 --> 00:07:02,455 and love each other. 146 00:07:02,488 --> 00:07:04,056 I think at some point, we need to draw the line, 147 00:07:04,090 --> 00:07:07,827 but I haven't figured out where that line is yet. 148 00:07:09,428 --> 00:07:12,364 What we have here are some heads 149 00:07:12,397 --> 00:07:14,500 in varying stages of assembly. 150 00:07:14,533 --> 00:07:17,336 This one, this is actually pretty much done. 151 00:07:17,369 --> 00:07:18,604 It's fully assembled. 152 00:07:18,637 --> 00:07:21,707 I'll turn it on here for a second... 153 00:07:21,740 --> 00:07:25,077 and you can see, all of the components are moving. 154 00:07:25,744 --> 00:07:27,880 I had to continually adjust 155 00:07:27,913 --> 00:07:30,316 how thick the skin is in different spots, 156 00:07:30,349 --> 00:07:31,450 and how it moves, 157 00:07:31,483 --> 00:07:33,619 to make sure that the robotics and the AI 158 00:07:33,652 --> 00:07:36,655 will all work smoothly with the end result, 159 00:07:36,688 --> 00:07:38,791 which is the finished face. 160 00:07:40,526 --> 00:07:43,696 The engineering, the programming, the artistry, 161 00:07:43,729 --> 00:07:46,299 for me, come together in the moment 162 00:07:46,332 --> 00:07:49,001 when you actually put the head on a body. 163 00:07:50,770 --> 00:07:53,139 It's always important to give her hair. 164 00:07:58,643 --> 00:08:00,746 Good afternoon, Matt. 165 00:08:00,779 --> 00:08:02,915 So happy to see you again. 166 00:08:04,149 --> 00:08:05,751 How smart are you? 167 00:08:07,119 --> 00:08:10,388 I'm so smart that someday, I will conquer the world, 168 00:08:10,422 --> 00:08:12,792 but in a good way, of course. 169 00:08:14,860 --> 00:08:17,696 Every single time I have a conversation, it's unpredictable. 170 00:08:17,730 --> 00:08:19,532 I never know which way it's going to go. 171 00:08:19,565 --> 00:08:21,867 She'll randomly say things that I'm not expecting, 172 00:08:21,901 --> 00:08:22,834 and I like that. 173 00:08:22,868 --> 00:08:25,137 Can you explain machine learning? 174 00:08:25,170 --> 00:08:28,407 Machine learning is a subset of artificial intelligence 175 00:08:28,441 --> 00:08:30,609 that often uses statistical techniques 176 00:08:30,643 --> 00:08:33,745 to give computers the ability to learn with data 177 00:08:33,779 --> 00:08:36,248 without being explicitly programmed. 178 00:08:36,281 --> 00:08:38,317 Right now, she has hearing, 179 00:08:38,351 --> 00:08:40,453 and she has some degree of touch, 180 00:08:40,486 --> 00:08:42,888 but vision is important. 181 00:08:42,921 --> 00:08:45,624 Matt's goal is for the next-generation doll 182 00:08:45,658 --> 00:08:49,928 to be able to see and process complex visual cues. 183 00:08:49,962 --> 00:08:52,431 The vision eyes are gonna be a little while. 184 00:08:52,465 --> 00:08:54,199 Susan's working on the board for that. 185 00:08:54,233 --> 00:08:56,836 Yeah, I've got the eyes in this one over here. 186 00:08:56,869 --> 00:08:59,705 I've put the Wi-Fi Bluetooth on the back. 187 00:08:59,739 --> 00:09:02,841 Does it install right on those existing pins, then? 188 00:09:02,874 --> 00:09:05,144 They'll all plug right in. - Good. 189 00:09:05,177 --> 00:09:08,447 We've been working on a vision system now 190 00:09:08,481 --> 00:09:10,148 for a little over eight to nine months, 191 00:09:10,182 --> 00:09:13,652 cameras that are inside of the robot's eyes. 192 00:09:13,685 --> 00:09:15,755 She'll be able to read your emotions, 193 00:09:15,788 --> 00:09:18,024 and she'll be able to recognize you. 194 00:09:22,060 --> 00:09:24,730 Only 10% of the signal we use 195 00:09:24,763 --> 00:09:26,298 to communicate with one another 196 00:09:26,331 --> 00:09:28,434 is the choice of words we use. 197 00:09:28,467 --> 00:09:30,569 90% is non-verbal. 198 00:09:30,603 --> 00:09:33,939 About half of that is your facial expressions, your use of gestures. 199 00:09:33,972 --> 00:09:38,911 So what people in the field of machine learning and computer vision have done 200 00:09:38,944 --> 00:09:42,080 is they've trained a machine or an algorithm 201 00:09:42,114 --> 00:09:45,251 to become a certified face-reader. 202 00:09:47,586 --> 00:09:49,888 Computer vision is this idea 203 00:09:49,921 --> 00:09:53,158 that our machines are able to see. 204 00:09:53,192 --> 00:09:56,228 Maybe it detects that there's a face in the image. 205 00:09:56,261 --> 00:09:58,764 Once you find the face, you want to identify 206 00:09:58,798 --> 00:10:02,100 these building blocks of these emotional expressions. 207 00:10:02,134 --> 00:10:05,504 You wanna know that there's a smirk, or a there's a brow raise, 208 00:10:05,538 --> 00:10:08,140 or, you know, an asymmetric lip corner pull. 209 00:10:09,408 --> 00:10:13,112 Mapping these building blocks to what it actually means, 210 00:10:13,145 --> 00:10:14,480 that's a little harder, 211 00:10:14,513 --> 00:10:16,915 but that's what we as humans clue into 212 00:10:16,949 --> 00:10:19,085 to understand how people are feeling. 213 00:10:22,488 --> 00:10:24,323 I think at some point, 214 00:10:24,356 --> 00:10:28,460 we will start to look at AI-driven devices and robots 215 00:10:28,493 --> 00:10:31,731 more like people instead of devices. 216 00:10:32,798 --> 00:10:36,368 Where I started was just with this very simple idea 217 00:10:36,401 --> 00:10:38,236 of a very realistic doll, 218 00:10:38,270 --> 00:10:41,907 and now with robotics and AI, I think what this will become 219 00:10:41,941 --> 00:10:46,345 is a new, alternative form of relationship. 220 00:10:46,378 --> 00:10:50,883 People like Matt are testing the boundaries of human and robot interaction, 221 00:10:50,916 --> 00:10:53,084 and what we value in relationships. 222 00:10:53,118 --> 00:10:57,289 Is AI companionship better than no companionship at all? 223 00:10:57,322 --> 00:11:01,093 Or is there no substitute for the human factor? 224 00:11:02,694 --> 00:11:03,962 Well, what about artists? 225 00:11:03,995 --> 00:11:05,731 They draw from the human experience 226 00:11:05,764 --> 00:11:07,199 to express themselves. 227 00:11:07,733 --> 00:11:09,268 Can AI do that? 228 00:11:12,671 --> 00:11:13,906 We're good to go? 229 00:11:15,206 --> 00:11:16,409 Action! 230 00:11:17,576 --> 00:11:20,813 I'm Oscar Sharp. I am a film director, uh, 231 00:11:20,846 --> 00:11:22,514 though it gets a bit weirder than that. 232 00:11:26,018 --> 00:11:28,353 Oh, God! 233 00:11:28,387 --> 00:11:31,157 I've never been so frightened in all my life, but it's very good. 234 00:11:31,190 --> 00:11:35,193 I started making films that were written by an "artificial intelligence." 235 00:11:35,226 --> 00:11:37,696 I think a lot of the fun is that you read it 236 00:11:37,729 --> 00:11:41,133 as if there is the world's greatest screenwriter on the other side... 237 00:11:41,166 --> 00:11:44,570 You're Waingro telling Bobo off for not getting him the money. 238 00:11:44,603 --> 00:11:46,071 ...and last night, they got drunk, 239 00:11:46,104 --> 00:11:48,006 wrote this screenplay, and then passed out, 240 00:11:48,039 --> 00:11:49,909 and we have to shoot it today. 241 00:11:51,009 --> 00:11:54,279 If you play the game that there's something there, 242 00:11:54,313 --> 00:11:56,181 then suddenly it all gets a lot more interesting. 243 00:11:56,215 --> 00:11:58,483 You have a computer who wrote a script 244 00:11:58,517 --> 00:11:59,685 that doesn't always make sense, 245 00:11:59,718 --> 00:12:01,753 and Oscar is very beholden to that script. 246 00:12:01,786 --> 00:12:03,055 He makes it make sense. 247 00:12:03,088 --> 00:12:06,191 - This is for the moment of "eyes go wide." - Yeah. 248 00:12:06,225 --> 00:12:09,861 And when it says, "He picks up her legs and awkwardly runs," 249 00:12:09,895 --> 00:12:11,297 we aren't gonna fake it. 250 00:12:12,697 --> 00:12:14,200 We're gonna do what he really wrote. 251 00:12:15,300 --> 00:12:16,469 I just said "he"! 252 00:12:20,439 --> 00:12:21,440 What are we doing? 253 00:12:21,473 --> 00:12:23,409 We're making an action movie, supposedly, right? 254 00:12:23,442 --> 00:12:25,610 Okay, right, right, but we're not gonna write it. 255 00:12:25,643 --> 00:12:27,479 We're not gonna write it, no. 256 00:12:27,513 --> 00:12:29,381 Uh, this machine is gonna write it. 257 00:12:29,414 --> 00:12:30,382 It lives in here. 258 00:12:30,415 --> 00:12:31,317 Is it in there, 259 00:12:31,350 --> 00:12:33,685 or is it like in the cloud or something? 260 00:12:33,719 --> 00:12:35,087 It's in both places. 261 00:12:35,120 --> 00:12:36,689 Okay, and this is... this is Benjamin. 262 00:12:36,722 --> 00:12:38,790 - What is Benjamin? - Right, what is Benjamin? 263 00:12:38,823 --> 00:12:40,626 Who is Benjamin, 264 00:12:40,659 --> 00:12:43,262 or what is Benjamin? 265 00:12:43,295 --> 00:12:45,964 Benjamin is an artificial intelligence program 266 00:12:45,998 --> 00:12:47,700 that writes screenplays, 267 00:12:47,733 --> 00:12:49,267 a digital brainchild 268 00:12:49,300 --> 00:12:51,570 of two creative and accomplished humans, 269 00:12:51,604 --> 00:12:54,440 Sharp, a Bafta-award-winning director, 270 00:12:54,473 --> 00:12:55,374 and this guy. 271 00:12:55,407 --> 00:12:58,143 My name is Ross Goodwin. I'm a tech artist. 272 00:12:58,176 --> 00:13:01,080 Uh, that means I make art with code. 273 00:13:02,214 --> 00:13:03,782 Okay, I know what you're thinking. 274 00:13:03,815 --> 00:13:07,186 When was the last time Hollywood produced something original? 275 00:13:07,219 --> 00:13:08,253 This year? 276 00:13:08,286 --> 00:13:09,488 Last year? 277 00:13:09,521 --> 00:13:11,256 1999? 278 00:13:11,289 --> 00:13:12,891 The '70s? 279 00:13:12,925 --> 00:13:16,094 What makes a story original anyway? 280 00:13:16,128 --> 00:13:18,997 Can we get AI to figure that out? 281 00:13:19,030 --> 00:13:24,836 People often say that creativity is the one thing that machines will never have. 282 00:13:24,869 --> 00:13:28,173 The surprising thing is that it's actually the other way around. 283 00:13:28,206 --> 00:13:30,709 Art and creativity is actually easier 284 00:13:30,742 --> 00:13:32,111 than problem solving. 285 00:13:32,144 --> 00:13:35,714 We already have computers that make great paintings, 286 00:13:35,747 --> 00:13:38,651 that make music that's indistinguishable 287 00:13:38,684 --> 00:13:40,118 from music that's composed by people, 288 00:13:40,152 --> 00:13:43,222 so machines are actually capable of creativity. 289 00:13:43,255 --> 00:13:45,257 And you can look at that, and you can say, 290 00:13:45,290 --> 00:13:48,060 "Is that really art? Does that count?" 291 00:13:48,093 --> 00:13:49,961 If you put a painting on the wall, 292 00:13:49,995 --> 00:13:52,798 and people look at it, and they find it moving, 293 00:13:52,831 --> 00:13:55,300 then how can you say that that's not art? 294 00:13:55,334 --> 00:13:57,369 I just... basically, that command 295 00:13:57,402 --> 00:13:59,671 just put all of the screenplays into one file. 296 00:13:59,704 --> 00:14:01,106 Right. 297 00:14:01,139 --> 00:14:02,708 - Now I'm just gonna see how big that file is. - Uh-huh. 298 00:14:02,741 --> 00:14:06,211 This machine is a deep learning language model. 299 00:14:06,245 --> 00:14:08,380 What you can do with a language model 300 00:14:08,414 --> 00:14:09,815 is at each step, 301 00:14:09,848 --> 00:14:12,384 you predict the next word, letter, or space, 302 00:14:12,417 --> 00:14:14,686 sort of like how a human writes, actually. 303 00:14:14,719 --> 00:14:16,354 You know, one letter at a time. 304 00:14:16,388 --> 00:14:18,857 It's a lot like a more sophisticated version 305 00:14:18,890 --> 00:14:20,826 of the auto-complete on your phone. 306 00:14:20,859 --> 00:14:24,897 Ross feeds Benjamin with a very large amount of screenplays. 307 00:14:28,066 --> 00:14:30,502 199 screenplays, 308 00:14:30,535 --> 00:14:34,439 26,271,247 bytes. 309 00:14:34,473 --> 00:14:35,907 - Right, of text? - Of text. 310 00:14:35,941 --> 00:14:37,676 Like "A-B-C-D," including spaces? 311 00:14:37,709 --> 00:14:39,078 - Including spaces. - Including spaces. 312 00:14:39,111 --> 00:14:41,613 Well, it takes all this input, and it looks at it, 313 00:14:41,646 --> 00:14:46,518 and it tries to find statistical patterns in that input. 314 00:14:46,552 --> 00:14:47,853 So for example, in movies, 315 00:14:47,886 --> 00:14:49,788 people are constantly saying, "What's going on? 316 00:14:49,821 --> 00:14:51,022 Who are you?" that kind of thing, 317 00:14:51,056 --> 00:14:52,691 and that turns up a lot in the output, 318 00:14:52,724 --> 00:14:54,359 because it's reliably in the input. 319 00:14:54,393 --> 00:14:57,029 The more material you have, the better it works. 320 00:14:58,296 --> 00:15:00,198 We've made three Benjamin films so far, 321 00:15:00,231 --> 00:15:02,767 Sunspring, It's No Game, and Zone Out. 322 00:15:02,801 --> 00:15:06,171 Sunspring was the simplest, and probably still the best idea, 323 00:15:06,204 --> 00:15:07,806 which was just... verbatim. 324 00:15:07,839 --> 00:15:09,742 You get the machine to write a screenplay, 325 00:15:09,775 --> 00:15:11,610 you pull out one chunk of screenplay, 326 00:15:11,643 --> 00:15:12,945 and you just shoot it. 327 00:15:23,888 --> 00:15:27,192 In a future with mass unemployment, 328 00:15:27,226 --> 00:15:30,062 young people are forced to sell blood. 329 00:15:31,095 --> 00:15:33,065 It's something I can do. 330 00:15:33,098 --> 00:15:36,201 You should see the boy and shut up. 331 00:15:37,703 --> 00:15:40,105 When you look at Sunspring on YouTube, 332 00:15:40,138 --> 00:15:43,074 and you see kind of the thumbs up and thumbs down? 333 00:15:43,107 --> 00:15:44,242 There's mainly thumbs up, 334 00:15:44,275 --> 00:15:45,910 but there's a decent chunk of thumbs down, 335 00:15:45,944 --> 00:15:47,746 and on the whole, based on the comments, 336 00:15:47,779 --> 00:15:49,014 those are people who, 337 00:15:49,047 --> 00:15:50,849 within a few seconds of the beginning, 338 00:15:50,882 --> 00:15:54,019 or even just once they'd seen the premise of, like, how we made it, 339 00:15:54,052 --> 00:15:55,320 they've gone... 340 00:15:55,353 --> 00:15:56,788 "Ugh, this definitely doesn't mean anything," 341 00:15:56,822 --> 00:15:59,224 and they've told their brain, "Don't even look for meaning, 342 00:15:59,257 --> 00:16:01,293 just forget it, just shrug it off." 343 00:16:01,326 --> 00:16:03,862 I'm sorry, this is fascinating to me. 344 00:16:03,896 --> 00:16:07,833 We've built a robot that writes screenplays 345 00:16:07,866 --> 00:16:09,735 that are weird, 346 00:16:09,768 --> 00:16:11,402 but they're not completely insane. 347 00:16:11,436 --> 00:16:13,571 I don't know what you're talking about. 348 00:16:13,605 --> 00:16:15,007 That's right. 349 00:16:15,040 --> 00:16:18,109 They sort of work. They kinda, kinda work. 350 00:16:18,143 --> 00:16:19,344 What are you doing? 351 00:16:19,378 --> 00:16:21,580 I don't want to be honest with you. 352 00:16:21,613 --> 00:16:22,981 You don't have to be a doctor. 353 00:16:23,014 --> 00:16:24,816 I'm not sure. 354 00:16:24,850 --> 00:16:26,584 I don't know what you're talking about. 355 00:16:26,618 --> 00:16:27,653 I wanna see you too. 356 00:16:27,686 --> 00:16:28,720 What do you mean? 357 00:16:28,754 --> 00:16:31,923 It's like having the best daydream of your life. 358 00:16:31,957 --> 00:16:35,360 My favorite aspect of Sunspring is there's this one scene, 359 00:16:35,394 --> 00:16:38,931 and it actually asks him to pull on the camera itself. 360 00:16:38,964 --> 00:16:40,799 It's a confusion on the machine's behalf 361 00:16:40,832 --> 00:16:44,102 where it's putting camera instructions in the action sequence, 362 00:16:44,135 --> 00:16:46,338 but somehow that creates this surreal effect, 363 00:16:46,371 --> 00:16:49,608 and then the interpretation by the production crew is, 364 00:16:49,641 --> 00:16:52,477 "Let's have the angle change and have him holding nothing," 365 00:16:52,510 --> 00:16:54,546 and what I love about that sequence 366 00:16:54,579 --> 00:16:57,449 is that it really highlights the dialogue 367 00:16:57,482 --> 00:16:59,951 and interpretation that we can achieve 368 00:16:59,985 --> 00:17:02,387 when we work with these machines. 369 00:17:03,489 --> 00:17:05,457 I gotta relax! 370 00:17:06,024 --> 00:17:07,459 Gotta get outta here... 371 00:17:09,394 --> 00:17:11,163 I don't wanna see you again. 372 00:17:16,801 --> 00:17:20,138 For this fourth film, we're going back to the thing in Sunspring 373 00:17:20,172 --> 00:17:21,940 that was sort of our favorite thing 374 00:17:21,973 --> 00:17:23,542 that we didn't really get to do properly, 375 00:17:23,575 --> 00:17:25,410 that we felt we, like, under-served, 376 00:17:25,443 --> 00:17:28,780 and that's when Benjamin writes action description. 377 00:17:28,813 --> 00:17:33,418 We've gathered thousands of pages of scripts from action movies, 378 00:17:33,452 --> 00:17:36,388 mostly mainstream Hollywood ones. 379 00:17:36,421 --> 00:17:39,124 You train, literally, on that kind of screenplay, 380 00:17:39,157 --> 00:17:40,592 the action genre, 381 00:17:40,626 --> 00:17:43,195 which famously is the genre that has the most action in it. 382 00:17:44,496 --> 00:17:47,899 - Okay, Benjamin has awoken, everyone. - -Ooh. 383 00:17:47,932 --> 00:17:49,167 Um, film crew, this is Ross. 384 00:17:49,200 --> 00:17:50,435 Ross, this is film crew. 385 00:17:50,469 --> 00:17:52,204 We have a stunt coordinator here, 386 00:17:52,237 --> 00:17:53,605 and so we're sort of hoping 387 00:17:53,638 --> 00:17:56,074 because we fed a lot of action screenplays to Benjamin, 388 00:17:56,108 --> 00:17:58,110 that what we're gonna get is action. 389 00:17:58,143 --> 00:17:59,811 Awaken, Benjamin! 390 00:17:59,844 --> 00:18:01,146 Awaken. 391 00:18:01,179 --> 00:18:04,082 As a director, normally, you get given a screenplay, 392 00:18:04,116 --> 00:18:05,116 or you wrote a screenplay, 393 00:18:05,150 --> 00:18:06,351 and this is what you're making, 394 00:18:06,384 --> 00:18:08,453 and maybe you kind of want to improve it a bit, 395 00:18:08,487 --> 00:18:10,088 "Ah, well, let's make some edits." 396 00:18:10,121 --> 00:18:11,689 Now, I have a rule. No edits. 397 00:18:11,723 --> 00:18:14,059 Whatever Benjamin writes is what Benjamin writes... 398 00:18:14,092 --> 00:18:15,927 - Come on, Benjamin. - Okay. 399 00:18:15,961 --> 00:18:17,362 ...and then I see it. 400 00:18:17,395 --> 00:18:19,231 "Bobo and Girlfriend," we call it. 401 00:18:19,264 --> 00:18:21,433 Stand by, everyone. Quiet, please! 402 00:18:22,400 --> 00:18:23,268 Action! 403 00:18:25,103 --> 00:18:26,771 Hey, Girlfriend. 404 00:18:26,804 --> 00:18:29,108 Some of my friends in entertainment, 405 00:18:29,141 --> 00:18:31,276 when I told them what I was doing, were horrified. 406 00:18:31,309 --> 00:18:32,911 They're like, "Oh, that's it! 407 00:18:32,944 --> 00:18:35,547 "AI, they're gonna write all the scripts. 408 00:18:35,580 --> 00:18:37,181 Robots are gonna do all the acting. 409 00:18:37,215 --> 00:18:39,885 Everything's gonna be cartoon AI stuff," 410 00:18:39,918 --> 00:18:42,988 but that's not what I feel like we're doing here at all. 411 00:18:43,021 --> 00:18:46,058 The point of this is an exercise in thought. 412 00:18:46,091 --> 00:18:47,326 Okay, stand by, everyone! 413 00:18:47,359 --> 00:18:49,628 Quiet, please! Action! 414 00:18:51,863 --> 00:18:53,665 Making a machine write like a person 415 00:18:53,699 --> 00:18:55,834 is not about replacing the person... 416 00:18:55,867 --> 00:18:57,602 No, no, no! 417 00:18:57,635 --> 00:19:00,572 ...it's about augmenting a person's abilities. 418 00:19:01,806 --> 00:19:05,110 It can empower people to produce creative work 419 00:19:05,143 --> 00:19:07,613 that might be beyond their native capacity. 420 00:19:08,346 --> 00:19:10,682 Come on! 421 00:19:10,715 --> 00:19:13,985 It's wild to try to find your interpretation of this kind of text. 422 00:19:14,019 --> 00:19:17,588 Obviously, we usually start with a script that's pretty coherent, 423 00:19:17,622 --> 00:19:19,825 and then I'll break down what the character says, 424 00:19:19,858 --> 00:19:20,859 and then I'll decide, 425 00:19:20,892 --> 00:19:23,061 what are they feeling? Why are they saying that? 426 00:19:23,095 --> 00:19:25,329 Stay on her, stay on her. 427 00:19:25,363 --> 00:19:26,865 Just do that walk-off again. 428 00:19:26,898 --> 00:19:28,033 Uh, stay where you are, John. 429 00:19:28,066 --> 00:19:29,501 Come back, Chelsey. Do the walk-off again. 430 00:19:29,534 --> 00:19:31,769 This is harder for you, and mo... and more frightening, 431 00:19:31,803 --> 00:19:33,905 and you're checking that he hasn't gotten up. 432 00:19:33,938 --> 00:19:36,408 When AI is writing the material, 433 00:19:36,441 --> 00:19:37,642 there isn't any subtext. 434 00:19:37,675 --> 00:19:39,544 You realize what's happening, and you're like, 435 00:19:39,577 --> 00:19:41,279 "Well, I'm gonna go take refuge at the pillar. 436 00:19:41,312 --> 00:19:42,614 - Okay. - All right? 437 00:19:42,647 --> 00:19:44,683 It stretches all of us. It makes us all work harder. 438 00:19:44,716 --> 00:19:47,151 It's one thing to bring an existing script to life, 439 00:19:47,185 --> 00:19:49,221 and just do your interpretation of it, 440 00:19:49,254 --> 00:19:51,522 but it's another thing to try to make it make sense, 441 00:19:51,556 --> 00:19:53,257 and then do your interpretation of it. 442 00:19:53,291 --> 00:19:54,927 Let's go one more time. 443 00:19:54,960 --> 00:19:58,496 This Bobo character that John is playing is a fantasy figure, 444 00:19:58,529 --> 00:20:01,099 is this avatar of masculinity, 445 00:20:01,133 --> 00:20:04,369 is the sort of result of watching too many action films... 446 00:20:05,169 --> 00:20:07,805 but he's confused, 447 00:20:07,839 --> 00:20:10,542 because he isn't getting the reaction that he expects. 448 00:20:11,008 --> 00:20:11,877 Action. 449 00:20:16,348 --> 00:20:17,515 Somewhere in the script, 450 00:20:17,549 --> 00:20:18,850 it talks about, "Bobo leans over to Bobo." 451 00:20:18,883 --> 00:20:22,554 We think, "Oh, right, well, let's have a mirror..." 452 00:20:23,422 --> 00:20:24,188 No! 453 00:20:24,222 --> 00:20:25,756 You're wrong! 454 00:20:25,790 --> 00:20:28,493 ...and we can see the two versions of Bobo, for a moment, talking. 455 00:20:28,526 --> 00:20:30,895 Okay, let's see you in the mirror? 456 00:20:32,063 --> 00:20:33,865 Hey, did you get my money? 457 00:20:36,001 --> 00:20:38,336 Okay, great. I'm getting terribly, terribly happy. 458 00:20:38,369 --> 00:20:39,671 Some of that was so good. 459 00:20:39,704 --> 00:20:42,307 It was like such a go-- We're, like, in a movie now. 460 00:20:43,174 --> 00:20:44,542 Being surrounded with people 461 00:20:44,576 --> 00:20:47,912 who are throwing all of their professional energy 462 00:20:47,945 --> 00:20:49,447 into something this ludicrous 463 00:20:49,480 --> 00:20:51,415 is just intrinsically enjoyable. 464 00:20:51,449 --> 00:20:53,384 They just breathe humanity 465 00:20:53,418 --> 00:20:55,754 into words that did not come from a human being. 466 00:20:55,787 --> 00:20:59,324 - All right, let's do it again. -Okay, let's do it. 467 00:20:59,358 --> 00:21:03,361 I think that making great art requires human experience, 468 00:21:03,394 --> 00:21:04,963 but our human experience 469 00:21:04,997 --> 00:21:06,898 is now completely mapped into data. 470 00:21:06,931 --> 00:21:09,000 This is where machine learning keeps surprising us, 471 00:21:09,034 --> 00:21:13,037 is that it actually has figured out stuff that we didn't realize it could. 472 00:21:13,070 --> 00:21:17,208 Meaning, once all our human experiences are mapped into data, 473 00:21:17,241 --> 00:21:19,311 AI will be able to mine it for material 474 00:21:19,344 --> 00:21:20,745 and make art? 475 00:21:20,779 --> 00:21:24,016 Look for patterns in our happiness and heartbreak, 476 00:21:24,049 --> 00:21:26,618 kick out a new song or movie? 477 00:21:28,453 --> 00:21:31,155 So this is all just this one line of Benjamin writing, 478 00:21:31,189 --> 00:21:32,256 "putting on a show." 479 00:21:32,290 --> 00:21:33,492 Right, right, right. 480 00:21:35,393 --> 00:21:37,128 So while all that's going on, 481 00:21:37,161 --> 00:21:38,329 Girlfriend is on this couch, 482 00:21:38,362 --> 00:21:40,365 gradually waking up, right? 483 00:21:45,436 --> 00:21:47,238 - She's in a horror movie... - Right. 484 00:21:47,272 --> 00:21:48,840 - He's in our action film. - Oh! 485 00:21:48,873 --> 00:21:51,409 So in his head, he's having a wonderful romantic time with her. 486 00:21:51,442 --> 00:21:52,777 Yeah, I love that. 487 00:21:52,811 --> 00:21:54,646 Do you remember his, "Bobo leans over to Bobo"? 488 00:21:54,679 --> 00:21:55,579 - Mm-hmm. - Remember that? 489 00:21:55,613 --> 00:21:57,215 So what we tried to do for that 490 00:21:57,249 --> 00:21:58,649 is he looks in the mirror, 491 00:21:58,683 --> 00:22:00,685 and in the mirror, it's gonna be Osric. 492 00:22:00,719 --> 00:22:03,789 He's created this avatar version of himself, Bobo. 493 00:22:03,822 --> 00:22:05,690 - In a... in a... - Okay, so that's the interpretation? 494 00:22:05,723 --> 00:22:06,858 - In that-- Yeah, exactly. - I like it. 495 00:22:06,892 --> 00:22:08,493 So this is what these guys came up with. 496 00:22:10,062 --> 00:22:11,396 No! 497 00:22:11,430 --> 00:22:13,065 You're wrong! 498 00:22:13,098 --> 00:22:15,333 You work really, really hard to go, 499 00:22:15,366 --> 00:22:17,335 what's a thing that's kinda coherent, 500 00:22:17,369 --> 00:22:19,203 that these actors can all be performing one thing, 501 00:22:19,237 --> 00:22:20,438 we can all be making one thing, 502 00:22:20,471 --> 00:22:21,873 and we can say, "This is what Benjamin meant?" 503 00:22:21,906 --> 00:22:23,775 - Right. -What does that tell me about me? 504 00:22:23,808 --> 00:22:25,243 - Right. - Like, what... So... 505 00:22:25,277 --> 00:22:26,677 and what I already know about me 506 00:22:26,711 --> 00:22:29,414 is I'm really antsy about how much misogyny 507 00:22:29,447 --> 00:22:31,716 is kind of encoded into... into culture. 508 00:22:31,750 --> 00:22:34,486 On one hand, you go, "This is an important, worthwhile thing to do--" 509 00:22:34,519 --> 00:22:35,586 On the other hand, we're projecting. 510 00:22:35,620 --> 00:22:36,721 And the other thing, you're projecting, 511 00:22:36,755 --> 00:22:38,689 - but we're always projecting. - Always. 512 00:22:38,723 --> 00:22:41,092 Literally, all interpretation is projection. 513 00:22:41,126 --> 00:22:42,594 Take 6. 514 00:22:42,627 --> 00:22:44,629 I like playing with authorship, 515 00:22:44,662 --> 00:22:47,064 and people's concepts of authorship, 516 00:22:47,098 --> 00:22:51,502 and people's concepts of where fiction and where ideas come from. 517 00:22:51,536 --> 00:22:53,438 Generative screenwriting. 518 00:22:53,471 --> 00:22:55,507 Me and Ross started it. 519 00:22:55,540 --> 00:22:56,674 I don't know if it's a new art form, 520 00:22:56,707 --> 00:22:58,276 but it's a new chunk of what cinema can be. 521 00:22:58,309 --> 00:22:59,277 That's new. 522 00:22:59,311 --> 00:23:01,380 What should we do next time, Ross? 523 00:23:02,147 --> 00:23:03,982 - Romantic comedy. - Okay. 524 00:23:05,550 --> 00:23:08,620 It's hard to know if machine learning will ever decode 525 00:23:08,653 --> 00:23:10,856 the mysteries of love or creativity. 526 00:23:12,190 --> 00:23:14,158 Maybe it's not even a mystery, 527 00:23:14,192 --> 00:23:16,394 just data points, 528 00:23:16,427 --> 00:23:18,896 but what about other human qualities, 529 00:23:18,930 --> 00:23:20,364 like instinct? 530 00:23:20,398 --> 00:23:22,667 Driving a car already requires us 531 00:23:22,700 --> 00:23:25,070 to make countless unconscious decisions. 532 00:23:25,103 --> 00:23:27,405 AI is learning to do that, 533 00:23:27,438 --> 00:23:29,574 but can we teach it to do more? 534 00:23:37,315 --> 00:23:39,484 Racing is not just driving a car. 535 00:23:39,517 --> 00:23:43,020 It's also about intuition, caution, aggression, 536 00:23:43,054 --> 00:23:44,489 and taking risks. 537 00:23:44,522 --> 00:23:47,125 Holly, can you confirm 200 at the end of this straight? 538 00:23:47,158 --> 00:23:48,560 Okay. 539 00:23:48,593 --> 00:23:52,363 It requires almost a preternatural will to win. 540 00:23:52,396 --> 00:23:55,800 So, how fast can a racecar go... 541 00:23:55,833 --> 00:23:58,937 without a human behind the wheel? 542 00:23:58,970 --> 00:24:00,605 Motorsport has always been 543 00:24:00,639 --> 00:24:02,374 taking technology to the limits... 544 00:24:02,407 --> 00:24:03,541 You all good your side, Holly? 545 00:24:03,574 --> 00:24:05,076 Yeah, I'm ready to go. 546 00:24:05,109 --> 00:24:06,845 ...and one of the goals of Roborace is to really facilitate 547 00:24:06,878 --> 00:24:09,413 the accelerated development of driverless technology. 548 00:24:09,447 --> 00:24:11,683 Okay, so we'll try to launch again. 549 00:24:14,052 --> 00:24:16,554 By taking the autonomous technology 550 00:24:16,587 --> 00:24:18,190 to the limits of its ability, 551 00:24:18,223 --> 00:24:20,759 we think that we can develop the technology faster. 552 00:24:23,929 --> 00:24:25,730 British startup Roborace 553 00:24:25,763 --> 00:24:28,500 wants to break new ground in driverless cars. 554 00:24:29,434 --> 00:24:31,035 To do so, they believe they need 555 00:24:31,069 --> 00:24:33,572 to test the boundaries of the technology... 556 00:24:36,807 --> 00:24:40,545 working at the very outer edge of what's safe and possible, 557 00:24:40,578 --> 00:24:43,181 where the margin for error is razor thin. 558 00:24:44,448 --> 00:24:46,551 After years of trial and error, 559 00:24:46,585 --> 00:24:50,021 they've created the world's first AI racecar. 560 00:24:51,422 --> 00:24:54,058 The thing that I love most about working at Roborace 561 00:24:54,091 --> 00:24:56,627 is we have a dream of being faster, and better, 562 00:24:56,660 --> 00:24:59,564 and safer than a human. 563 00:24:59,597 --> 00:25:02,200 More than 50 companies around the world 564 00:25:02,233 --> 00:25:04,735 are working to bring self-driving cars to city streets. 565 00:25:04,769 --> 00:25:08,706 The promise of driverless taxis, buses, and trucks 566 00:25:08,740 --> 00:25:10,141 is transformative. 567 00:25:11,409 --> 00:25:13,878 It'll make our world safer and cleaner, 568 00:25:13,912 --> 00:25:15,981 changing the way our cities are designed, 569 00:25:16,014 --> 00:25:17,115 societies function, 570 00:25:17,148 --> 00:25:20,084 even how we spend our time. 571 00:25:20,117 --> 00:25:23,455 Think about a self-driving car out in the real world. 572 00:25:23,488 --> 00:25:26,257 In order to build that system and have it work, 573 00:25:26,291 --> 00:25:27,758 it's got to be virtually perfect. 574 00:25:27,792 --> 00:25:31,329 If you had a 99% accuracy rate, 575 00:25:31,363 --> 00:25:32,863 that wouldn't be anywhere near enough, 576 00:25:32,897 --> 00:25:34,865 because once you take that 1% error rate 577 00:25:34,899 --> 00:25:38,537 and you multiply that by millions of cars on the road, 578 00:25:38,570 --> 00:25:41,940 I mean, you'd have accidents happening constantly, 579 00:25:41,973 --> 00:25:45,843 so the error rate has to be extraordinarily low 580 00:25:45,877 --> 00:25:48,212 in order to pull this off. 581 00:25:48,246 --> 00:25:50,648 Roborace is betting they can crack the code 582 00:25:50,681 --> 00:25:53,084 by seeing just how far the tech can go, 583 00:25:53,118 --> 00:25:57,121 a place usually reserved for only the best human drivers. 584 00:25:57,155 --> 00:26:00,758 As a human, you have lots of advantages over a computer. 585 00:26:00,792 --> 00:26:02,760 You know exactly where you are in the world. 586 00:26:02,793 --> 00:26:05,897 You have eyes that can enable you to see things, 587 00:26:05,930 --> 00:26:08,799 so we need to implement technology on vehicles 588 00:26:08,832 --> 00:26:10,735 to enable them to see the world. 589 00:26:11,836 --> 00:26:14,306 We have a system called OxTS. 590 00:26:15,206 --> 00:26:16,541 It's a differential GPS, 591 00:26:16,574 --> 00:26:18,276 which means it's military grade. 592 00:26:20,444 --> 00:26:22,714 We also use LiDAR sensors. 593 00:26:25,750 --> 00:26:28,219 These are basically laser scanners. 594 00:26:28,253 --> 00:26:29,754 They create, for the vehicle, 595 00:26:29,787 --> 00:26:32,623 a 3D map of the world around it. 596 00:26:32,656 --> 00:26:34,425 And there's one last thing that we use, 597 00:26:34,459 --> 00:26:37,829 vehicle-to-vehicle communication between the cars. 598 00:26:39,096 --> 00:26:41,366 Each of them can tell the other car 599 00:26:41,399 --> 00:26:43,134 the position of it on the track. 600 00:26:44,336 --> 00:26:45,837 And just to be clear, 601 00:26:45,870 --> 00:26:49,607 your phone does not come with military-grade GPS. 602 00:26:49,641 --> 00:26:52,243 These cars? Next level. 603 00:26:52,276 --> 00:26:56,147 The challenging part is to really fuse all this information together. 604 00:26:56,180 --> 00:26:59,283 At Roborace, we can provide the hardware, 605 00:26:59,317 --> 00:27:02,287 but then we need software companies to come to us 606 00:27:02,320 --> 00:27:04,556 to implement their software. 607 00:27:04,589 --> 00:27:08,093 Today, Roborace has invited two skilled teams 608 00:27:08,126 --> 00:27:10,895 to test their latest road rocket on the track. 609 00:27:15,633 --> 00:27:17,067 My name is Johannes. 610 00:27:17,101 --> 00:27:19,470 I'm from the Technical University of Munich. 611 00:27:19,503 --> 00:27:21,505 I'm the project leader. 612 00:27:21,539 --> 00:27:23,241 Is the Wi-Fi working off the car? 613 00:27:23,274 --> 00:27:24,342 I could check it. 614 00:27:25,676 --> 00:27:27,211 T.U.M. from Germany 615 00:27:27,245 --> 00:27:30,047 is one of the top technical universities in Europe, 616 00:27:30,080 --> 00:27:33,184 home to 17 Nobel Prize winners in science. 617 00:27:35,620 --> 00:27:36,754 I have no connection to the car. 618 00:27:36,787 --> 00:27:37,789 Wifi doesn't work. 619 00:27:37,822 --> 00:27:39,557 So we have no Wi-Fi to the car... 620 00:27:39,590 --> 00:27:41,825 So we just need to reset the router. 621 00:27:41,859 --> 00:27:43,961 My name is Max. I'm, uh... 622 00:27:43,994 --> 00:27:45,563 Uh, let's figure out, who am I? 623 00:27:45,596 --> 00:27:47,031 I'm, uh... 624 00:27:47,065 --> 00:27:51,369 In Arrival, I'm a product owner of the self-driving system. 625 00:27:51,403 --> 00:27:53,871 Arrival is a UK startup 626 00:27:53,904 --> 00:27:55,407 focused on designing and building 627 00:27:55,440 --> 00:27:58,575 next-gen electric vehicles for commercial use. 628 00:27:58,609 --> 00:28:00,512 Ah, okay, okay, okay, good. 629 00:28:00,545 --> 00:28:03,848 Each team created their own custom software, 630 00:28:03,882 --> 00:28:07,151 the AI driver that pilots the car, 631 00:28:07,184 --> 00:28:09,553 and since each of the teams' programmers 632 00:28:09,587 --> 00:28:12,123 have their own distinct personality, 633 00:28:12,156 --> 00:28:14,759 does that mean each of their AI drivers 634 00:28:14,792 --> 00:28:17,128 will have different personalities or instincts too? 635 00:28:17,161 --> 00:28:18,696 The two teams that we have here 636 00:28:18,729 --> 00:28:20,264 are using two slightly different approaches 637 00:28:20,298 --> 00:28:22,633 to the same problem of making a car go 'round the track 638 00:28:22,666 --> 00:28:24,802 in the shortest distance in the fastest way. 639 00:28:24,835 --> 00:28:26,137 The T.U.M. strategy 640 00:28:26,170 --> 00:28:28,305 is really to keep their code as simple as possible. 641 00:28:28,339 --> 00:28:31,876 It's maybe a very German, efficient way of doing things. 642 00:28:31,910 --> 00:28:34,211 Okay, thanks, we will check now. 643 00:28:34,245 --> 00:28:36,581 Arrival's code is more complicated 644 00:28:36,614 --> 00:28:39,550 in that they use many more of the sensors on the vehicle. 645 00:28:39,583 --> 00:28:41,285 It will be interesting to see 646 00:28:41,318 --> 00:28:44,889 whether it pays off to be simple in your code, 647 00:28:44,922 --> 00:28:46,824 or slightly more complicated, 648 00:28:46,858 --> 00:28:49,260 to use more of the functionality of the car. 649 00:28:50,361 --> 00:28:52,463 The first test for each team 650 00:28:52,496 --> 00:28:54,031 is the overtake, 651 00:28:54,064 --> 00:28:56,767 to see if their AI can pass another car 652 00:28:56,801 --> 00:28:58,769 at high speed. 653 00:28:58,803 --> 00:29:00,337 It's difficult for AI 654 00:29:00,371 --> 00:29:02,406 because we have to make a lot of decisions, 655 00:29:02,440 --> 00:29:04,642 and a lot of planning, a lot of computations 656 00:29:04,676 --> 00:29:10,982 to calculate what the car should do in which millisecond. 657 00:29:11,015 --> 00:29:14,685 Everybody has seen high-speed crashes in motorsport before. 658 00:29:14,718 --> 00:29:15,920 We'd quite like to avoid that. 659 00:29:15,953 --> 00:29:18,022 For this reason, during testing, 660 00:29:18,055 --> 00:29:20,157 we keep a human in the car. 661 00:29:24,762 --> 00:29:26,130 Okay, Reece, enabling AI. 662 00:29:26,163 --> 00:29:29,234 Can you just confirm you've got the blue light, please? 663 00:29:31,802 --> 00:29:33,738 In order to overtake, 664 00:29:33,772 --> 00:29:37,375 they need a second car on track at the same time. 665 00:29:37,408 --> 00:29:41,079 This is a vehicle that stays in human-driven mode the whole time, 666 00:29:41,112 --> 00:29:43,681 so we know exactly how it's going to behave. 667 00:29:43,715 --> 00:29:46,451 Okay, launch AI from the race control. 668 00:29:47,251 --> 00:29:51,022 And launching in three, two, one. 669 00:29:55,593 --> 00:29:59,998 It's really difficult for AI to learn to overtake. 670 00:30:00,031 --> 00:30:01,832 When you have one vehicle on track, 671 00:30:01,866 --> 00:30:04,903 it only needs to make decisions about itself, 672 00:30:05,602 --> 00:30:07,071 but when you have two vehicles, 673 00:30:07,104 --> 00:30:09,740 you have the option to create your behavior 674 00:30:09,773 --> 00:30:11,809 in response to another vehicle. 675 00:30:11,842 --> 00:30:15,880 Okay, we are going to release the speed limit on your car now, Reece. 676 00:30:32,496 --> 00:30:33,464 Nice! 677 00:30:33,498 --> 00:30:34,966 Yeah, man. 678 00:30:39,503 --> 00:30:40,605 Team T.U.M. 679 00:30:40,638 --> 00:30:43,641 has successfully completed the overtake challenge. 680 00:30:43,674 --> 00:30:45,443 Next up, team Arrival. 681 00:30:48,545 --> 00:30:52,116 So, Tim, can you go to take position on the start line, please? 682 00:30:55,352 --> 00:30:56,854 Enabling AI 683 00:30:56,888 --> 00:30:59,423 Can you confirm blue light, please? 684 00:31:08,599 --> 00:31:12,070 And launch in three, two, one. 685 00:31:22,513 --> 00:31:24,481 So it's looking good so far. 686 00:31:46,804 --> 00:31:48,006 Car crashed. 687 00:31:49,607 --> 00:31:51,042 Tim, can you hear me? 688 00:31:59,917 --> 00:32:02,319 Has anyone got eyes on what happened? 689 00:32:06,590 --> 00:32:07,625 Sorry, boys. 690 00:32:10,161 --> 00:32:11,595 Self-driving cars. 691 00:32:11,629 --> 00:32:14,398 This is an idea that's been around since the '30s, 692 00:32:14,432 --> 00:32:15,333 hardly a new one. 693 00:32:15,999 --> 00:32:17,602 Why hasn't it happened? 694 00:32:17,635 --> 00:32:19,003 It's really hard. 695 00:32:19,036 --> 00:32:22,173 When there are unpredictable things that happen, 696 00:32:22,206 --> 00:32:24,242 that can get you in a lot of trouble. 697 00:32:24,275 --> 00:32:27,545 Now, sometimes trouble just means it shuts down. 698 00:32:27,578 --> 00:32:28,646 Sometimes trouble means 699 00:32:28,679 --> 00:32:31,415 it gives you a result that you weren't expecting. 700 00:32:31,449 --> 00:32:32,517 I think he's just... 701 00:32:32,550 --> 00:32:35,887 They've come back online so aggressively... 702 00:32:35,920 --> 00:32:38,723 Plus or minus one G coming back online. 703 00:32:38,756 --> 00:32:41,025 When the car returned to the trajectory, 704 00:32:41,058 --> 00:32:42,560 it did it too aggressive, 705 00:32:42,593 --> 00:32:46,097 and actually steered out of the racing track. 706 00:32:46,130 --> 00:32:48,666 - My feeling is that it overreacts. - Yeah, yeah. 707 00:32:48,699 --> 00:32:50,834 So it's not necessarily the line that's aggressive, 708 00:32:50,868 --> 00:32:53,871 it's how it reacts once it just gets a little bit out of the line, 709 00:32:53,905 --> 00:32:55,673 and then overcorrects, and then overcorrects. 710 00:32:57,975 --> 00:32:59,443 We were this close 711 00:32:59,476 --> 00:33:01,512 to really hitting the target of our test, 712 00:33:01,545 --> 00:33:02,780 and it didn't happen. 713 00:33:02,814 --> 00:33:04,815 It just slipped away, so it was just... 714 00:33:04,849 --> 00:33:06,384 ah, disappointment. 715 00:33:10,153 --> 00:33:13,491 There are so many aspects of the car. 716 00:33:13,524 --> 00:33:16,427 The systems guys have such a difficult job 717 00:33:16,460 --> 00:33:19,130 to make sure that everything is absolutely perfect, 718 00:33:19,163 --> 00:33:20,297 because that's what you need 719 00:33:20,331 --> 00:33:21,765 to be able to go autonomous racing. 720 00:33:21,798 --> 00:33:23,734 Everything has to be perfect. 721 00:33:23,768 --> 00:33:26,804 Team Arrival's program just couldn't hack it, 722 00:33:26,837 --> 00:33:28,840 but for team T.U.M., 723 00:33:28,873 --> 00:33:30,141 another test awaits... 724 00:33:30,174 --> 00:33:31,976 Can we get the car 725 00:33:32,009 --> 00:33:34,812 into the normal start position, please? 726 00:33:34,845 --> 00:33:37,448 ...and this next one is all about speed. 727 00:33:37,481 --> 00:33:39,183 Very high speed. 728 00:33:39,217 --> 00:33:41,686 The fastest that a human's ever driven around this track 729 00:33:41,719 --> 00:33:43,054 was 200 kph. 730 00:33:43,087 --> 00:33:46,490 Translation, that's about 120 miles an hour. 731 00:33:46,523 --> 00:33:49,393 So the AI is gonna try to beat that high speed. 732 00:33:49,426 --> 00:33:52,096 And it's gonna do it without a human safety net, 733 00:33:52,129 --> 00:33:53,430 because at that speed, 734 00:33:53,463 --> 00:33:55,466 it's borderline unsafe for people. 735 00:33:55,499 --> 00:33:58,135 When the driver climbs out and shuts the door, 736 00:33:58,168 --> 00:33:59,604 yeah, your heart rate goes up. 737 00:34:08,679 --> 00:34:09,980 And we are launching 738 00:34:10,014 --> 00:34:14,185 in three, two, one. 739 00:34:20,157 --> 00:34:21,826 And launch successful. 740 00:34:26,163 --> 00:34:28,433 160. Next round, 200. 741 00:34:41,879 --> 00:34:44,014 The car has six laps, 742 00:34:44,047 --> 00:34:47,684 six tries to hit top speed. 743 00:34:47,718 --> 00:34:51,221 Each lap, the AI will increasingly push the limits of control, 744 00:34:51,254 --> 00:34:53,024 traction, and throttle, 745 00:34:53,057 --> 00:34:54,692 to break the human record. 746 00:34:57,728 --> 00:34:58,562 Holly, this is Steve. 747 00:34:58,595 --> 00:34:59,997 Can we confirm in the atmos-data 748 00:35:00,030 --> 00:35:01,499 it is safe to continue? 749 00:35:01,532 --> 00:35:04,335 Yeah, we think it looks fairly controlled. 750 00:35:07,538 --> 00:35:10,274 Okay, so the next run should be V-max. 751 00:35:31,195 --> 00:35:32,830 We have 210. 752 00:35:33,764 --> 00:35:34,765 That's cool. 753 00:35:46,210 --> 00:35:48,045 It was a real, real sense of excitement 754 00:35:48,078 --> 00:35:52,116 to see it finally crack the 210 kph mark. 755 00:35:52,150 --> 00:35:54,751 It was a real success for Roborace functionality 756 00:35:54,785 --> 00:35:58,789 as well as building confidence in the team's software. 757 00:35:58,823 --> 00:36:01,725 It really showcases what autonomous cars can do, 758 00:36:01,759 --> 00:36:03,193 not just on the racetrack, 759 00:36:03,227 --> 00:36:05,796 but also for everybody around the world, 760 00:36:05,829 --> 00:36:06,964 so we're really hoping 761 00:36:06,998 --> 00:36:10,867 that this will improve road technology for the future. 762 00:36:10,901 --> 00:36:13,771 The current state of AI is that there are some things 763 00:36:13,804 --> 00:36:16,206 that AI can really do better than humans, 764 00:36:16,239 --> 00:36:17,407 and then there's things 765 00:36:17,441 --> 00:36:18,910 that it can't do anywhere close to humans... 766 00:36:20,444 --> 00:36:24,114 but now where the frontier is gonna be moving 767 00:36:24,148 --> 00:36:26,183 is where computers come up to the human level, 768 00:36:26,217 --> 00:36:28,686 not quite, and then surpass humans, 769 00:36:28,719 --> 00:36:31,755 and I think the odds are overwhelming 770 00:36:31,789 --> 00:36:35,626 that we will eventually be able to build an artificial brain 771 00:36:35,659 --> 00:36:38,562 that is at the level of the human brain. 772 00:36:38,595 --> 00:36:41,465 The big question is how long will it take? 773 00:36:42,999 --> 00:36:44,434 "The hard problem." 774 00:36:44,467 --> 00:36:45,803 It's a philosophical phrase 775 00:36:45,836 --> 00:36:48,439 that describes difficult things to figure out, 776 00:36:48,472 --> 00:36:51,809 like "the hard problem of consciousness." 777 00:36:51,842 --> 00:36:54,245 We may never know what consciousness is, 778 00:36:54,278 --> 00:36:57,147 let alone if we can give it to a machine, 779 00:36:57,181 --> 00:36:59,182 but do we need to? 780 00:36:59,216 --> 00:37:01,552 What does a machine really need to know 781 00:37:01,585 --> 00:37:03,721 in order to be a good athlete, 782 00:37:03,754 --> 00:37:05,022 or an artist, 783 00:37:05,056 --> 00:37:06,457 or a lover? 784 00:37:08,492 --> 00:37:12,029 Will AI ever have the will to win, 785 00:37:12,062 --> 00:37:14,164 the depth to create, 786 00:37:14,198 --> 00:37:17,902 the empathy to connect on a deep human level? 787 00:37:17,935 --> 00:37:19,136 Maybe. 788 00:37:19,169 --> 00:37:22,172 Some say we're just a bunch of biological algorithms, 789 00:37:22,206 --> 00:37:23,407 and that one day, 790 00:37:23,440 --> 00:37:25,977 evolution will evolve AI to emulate humans 791 00:37:27,244 --> 00:37:28,679 to be more like us... 792 00:37:33,416 --> 00:37:35,319 ...or maybe it won't... 793 00:37:36,386 --> 00:37:38,489 and human nature, who we really are, 794 00:37:38,522 --> 00:37:40,691 will remain a mystery. 795 00:37:43,660 --> 00:37:45,596 We gave it some dialogue to start with, 796 00:37:45,629 --> 00:37:47,298 like this line from Superman. 797 00:37:47,331 --> 00:37:49,332 So you got some Superman/ Lois Lane stuff, huh? 798 00:37:49,366 --> 00:37:50,767 Yeah, so you wanna read it? 799 00:37:50,801 --> 00:37:53,137 Mm... not that bit. Um, wait. 800 00:37:53,170 --> 00:37:55,473 Up, up, up, up, up. Back up... okay. 801 00:37:55,506 --> 00:37:57,608 "Superman angrily grabs Lois by the neck, 802 00:37:57,641 --> 00:38:00,677 slaps her against the wall, and bares his teeth in fury." 803 00:38:00,711 --> 00:38:01,512 "You're wrong. 804 00:38:01,545 --> 00:38:03,114 You're a grotesque kind of monster." 805 00:38:03,147 --> 00:38:05,115 - "You're wrong!" - "You're a terrible liar." 806 00:38:05,149 --> 00:38:06,883 "No! I'm sorry, I'm sorry. 807 00:38:06,917 --> 00:38:07,785 I can't believe it!" 808 00:38:07,818 --> 00:38:09,754 "You're so much more than that, Lois." 809 00:38:09,787 --> 00:38:10,821 "Please, please!" 810 00:38:10,854 --> 00:38:11,655 "How could you? 811 00:38:11,689 --> 00:38:13,824 No one can believe who you are." 812 00:38:13,857 --> 00:38:14,858 "Don't be ridiculous. 813 00:38:14,891 --> 00:38:15,726 Please? 814 00:38:15,759 --> 00:38:17,527 How could you be so much more than that?" 815 00:38:17,560 --> 00:38:19,397 "You're such a terrible liar. 816 00:38:19,430 --> 00:38:21,298 You can't even believe who you are. 817 00:38:21,332 --> 00:38:23,401 Please, unless you're really a no-good liar, 818 00:38:23,434 --> 00:38:25,336 you're not even sure if you're good." 819 00:38:25,369 --> 00:38:26,803 "Sorry, Superman, I'm so sorry!" 820 00:38:26,837 --> 00:38:28,739 Superman is just not making very much sense. 821 00:38:28,772 --> 00:38:30,273 Maybe kind of drunk or something? 822 00:38:30,307 --> 00:38:32,442 In fact, it says, "Superman isn't funny. 823 00:38:32,476 --> 00:38:34,344 The two of them are really different people. 824 00:38:34,377 --> 00:38:36,113 There is no such thing as good good." 825 00:38:36,147 --> 00:38:37,214 That's pretty deep. 826 00:38:37,248 --> 00:38:38,315 "There is no such thing as good good." 827 00:38:38,348 --> 00:38:39,684 There is no such thing as good good. 828 00:38:39,717 --> 00:38:40,417 So far as I know. 829 00:38:40,451 --> 00:38:41,785 Yeah. Have you checked? 830 00:38:41,819 --> 00:38:42,620 I'm gonna Google it. 831 00:38:42,653 --> 00:38:44,555 - Can we Google it? - Um...