id
int64 1
325
| guest
stringlengths 3
45
⌀ | title
stringlengths 3
75
⌀ | text
stringlengths 1
528
| start
stringlengths 9
11
| end
stringlengths 9
11
|
---|---|---|---|---|---|
1 | Max Tegmark | Life 3.0 | so that it becomes as good as possible at this thing. | 1:07:12.920 | 1:07:15.080 |
1 | Max Tegmark | Life 3.0 | So if you think of a landscape with some valley, | 1:07:15.080 | 1:07:19.680 |
1 | Max Tegmark | Life 3.0 | where each dimension of the landscape | 1:07:20.720 | 1:07:22.120 |
1 | Max Tegmark | Life 3.0 | corresponds to some number you can change, | 1:07:22.120 | 1:07:24.120 |
1 | Max Tegmark | Life 3.0 | you're trying to find the minimum. | 1:07:24.120 | 1:07:25.640 |
1 | Max Tegmark | Life 3.0 | And it's well known that | 1:07:25.640 | 1:07:26.760 |
1 | Max Tegmark | Life 3.0 | if you have a very high dimensional landscape, | 1:07:26.760 | 1:07:29.040 |
1 | Max Tegmark | Life 3.0 | complicated things, it's super hard to find the minimum. | 1:07:29.040 | 1:07:31.840 |
1 | Max Tegmark | Life 3.0 | Quantum mechanics is amazingly good at this. | 1:07:31.840 | 1:07:35.840 |
1 | Max Tegmark | Life 3.0 | Like if I want to know what's the lowest energy state | 1:07:35.840 | 1:07:38.240 |
1 | Max Tegmark | Life 3.0 | this water can possibly have, | 1:07:38.240 | 1:07:39.720 |
1 | Max Tegmark | Life 3.0 | incredibly hard to compute, | 1:07:41.720 | 1:07:42.560 |
1 | Max Tegmark | Life 3.0 | but nature will happily figure this out for you | 1:07:42.560 | 1:07:45.400 |
1 | Max Tegmark | Life 3.0 | if you just cool it down, make it very, very cold. | 1:07:45.400 | 1:07:48.000 |
1 | Max Tegmark | Life 3.0 | If you put a ball somewhere, | 1:07:49.800 | 1:07:50.880 |
1 | Max Tegmark | Life 3.0 | it'll roll down to its minimum. | 1:07:50.880 | 1:07:52.240 |
1 | Max Tegmark | Life 3.0 | And this happens metaphorically | 1:07:52.240 | 1:07:54.280 |
1 | Max Tegmark | Life 3.0 | at the energy landscape too. | 1:07:54.280 | 1:07:56.320 |
1 | Max Tegmark | Life 3.0 | And quantum mechanics even uses some clever tricks, | 1:07:56.320 | 1:07:59.280 |
1 | Max Tegmark | Life 3.0 | which today's machine learning systems don't. | 1:07:59.280 | 1:08:02.520 |
1 | Max Tegmark | Life 3.0 | Like if you're trying to find the minimum | 1:08:02.520 | 1:08:04.160 |
1 | Max Tegmark | Life 3.0 | and you get stuck in the little local minimum here, | 1:08:04.160 | 1:08:06.960 |
1 | Max Tegmark | Life 3.0 | in quantum mechanics you can actually tunnel | 1:08:06.960 | 1:08:08.760 |
1 | Max Tegmark | Life 3.0 | through the barrier and get unstuck again. | 1:08:08.760 | 1:08:11.840 |
1 | Max Tegmark | Life 3.0 | That's really interesting. | 1:08:13.480 | 1:08:14.320 |
1 | Max Tegmark | Life 3.0 | Yeah, so it may be, for example, | 1:08:14.320 | 1:08:16.120 |
1 | Max Tegmark | Life 3.0 | that we'll one day use quantum computers | 1:08:16.120 | 1:08:19.160 |
1 | Max Tegmark | Life 3.0 | that help train neural networks better. | 1:08:19.160 | 1:08:22.840 |
1 | Max Tegmark | Life 3.0 | That's really interesting. | 1:08:22.840 | 1:08:23.680 |
1 | Max Tegmark | Life 3.0 | Okay, so as a component of kind of the learning process, | 1:08:23.680 | 1:08:27.040 |
1 | Max Tegmark | Life 3.0 | for example. | 1:08:27.040 | 1:08:27.880 |
1 | Max Tegmark | Life 3.0 | Yeah. | 1:08:27.880 | 1:08:29.440 |
1 | Max Tegmark | Life 3.0 | Let me ask sort of wrapping up here a little bit, | 1:08:29.440 | 1:08:33.080 |
1 | Max Tegmark | Life 3.0 | let me return to the questions of our human nature | 1:08:33.080 | 1:08:36.880 |
1 | Max Tegmark | Life 3.0 | and love, as I mentioned. | 1:08:36.880 | 1:08:40.000 |
1 | Max Tegmark | Life 3.0 | So do you think, | 1:08:40.000 | 1:08:41.640 |
1 | Max Tegmark | Life 3.0 | you mentioned sort of a helper robot, | 1:08:44.280 | 1:08:46.000 |
1 | Max Tegmark | Life 3.0 | but you could think of also personal robots. | 1:08:46.000 | 1:08:48.640 |
1 | Max Tegmark | Life 3.0 | Do you think the way we human beings fall in love | 1:08:48.640 | 1:08:52.480 |
1 | Max Tegmark | Life 3.0 | and get connected to each other | 1:08:52.480 | 1:08:54.680 |
1 | Max Tegmark | Life 3.0 | is possible to achieve in an AI system | 1:08:54.680 | 1:08:58.040 |
1 | Max Tegmark | Life 3.0 | and human level AI intelligence system? | 1:08:58.040 | 1:09:00.360 |
1 | Max Tegmark | Life 3.0 | Do you think we would ever see that kind of connection? | 1:09:00.360 | 1:09:03.720 |
1 | Max Tegmark | Life 3.0 | Or, you know, in all this discussion | 1:09:03.720 | 1:09:06.160 |
1 | Max Tegmark | Life 3.0 | about solving complex goals, | 1:09:06.160 | 1:09:08.520 |
1 | Max Tegmark | Life 3.0 | is this kind of human social connection, | 1:09:08.520 | 1:09:10.760 |
1 | Max Tegmark | Life 3.0 | do you think that's one of the goals | 1:09:10.760 | 1:09:12.560 |
1 | Max Tegmark | Life 3.0 | on the peaks and valleys with the raising sea levels | 1:09:12.560 | 1:09:16.280 |
1 | Max Tegmark | Life 3.0 | that we'll be able to achieve? | 1:09:16.280 | 1:09:17.360 |
1 | Max Tegmark | Life 3.0 | Or do you think that's something that's ultimately, | 1:09:17.360 | 1:09:20.040 |
1 | Max Tegmark | Life 3.0 | or at least in the short term, | 1:09:20.040 | 1:09:21.760 |
1 | Max Tegmark | Life 3.0 | relative to the other goals is not achievable? | 1:09:21.760 | 1:09:23.640 |
1 | Max Tegmark | Life 3.0 | I think it's all possible. | 1:09:23.640 | 1:09:25.120 |
1 | Max Tegmark | Life 3.0 | And I mean, in recent, | 1:09:25.120 | 1:09:27.600 |
1 | Max Tegmark | Life 3.0 | there's a very wide range of guesses, as you know, | 1:09:27.600 | 1:09:30.840 |
1 | Max Tegmark | Life 3.0 | among AI researchers, when we're going to get AGI. | 1:09:30.840 | 1:09:33.720 |
1 | Max Tegmark | Life 3.0 | Some people, you know, like our friend Rodney Brooks | 1:09:35.120 | 1:09:37.640 |
1 | Max Tegmark | Life 3.0 | says it's going to be hundreds of years at least. | 1:09:37.640 | 1:09:41.040 |
1 | Max Tegmark | Life 3.0 | And then there are many others | 1:09:41.040 | 1:09:42.200 |
1 | Max Tegmark | Life 3.0 | who think it's going to happen much sooner. | 1:09:42.200 | 1:09:44.040 |
1 | Max Tegmark | Life 3.0 | And recent polls, | 1:09:44.040 | 1:09:45.520 |
1 | Max Tegmark | Life 3.0 | maybe half or so of AI researchers | 1:09:46.840 | 1:09:48.640 |
1 | Max Tegmark | Life 3.0 | think we're going to get AGI within decades. | 1:09:48.640 | 1:09:50.920 |
1 | Max Tegmark | Life 3.0 | So if that happens, of course, | 1:09:50.920 | 1:09:52.720 |
1 | Max Tegmark | Life 3.0 | then I think these things are all possible. | 1:09:52.720 | 1:09:55.040 |
1 | Max Tegmark | Life 3.0 | But in terms of whether it will happen, | 1:09:55.040 | 1:09:56.840 |
1 | Max Tegmark | Life 3.0 | I think we shouldn't spend so much time asking | 1:09:56.840 | 1:10:00.600 |
1 | Max Tegmark | Life 3.0 | what do we think will happen in the future? | 1:10:00.600 | 1:10:03.240 |
1 | Max Tegmark | Life 3.0 | As if we are just some sort of pathetic, | 1:10:03.240 | 1:10:05.160 |
1 | Max Tegmark | Life 3.0 | your passive bystanders, you know, | 1:10:05.160 | 1:10:07.040 |
1 | Max Tegmark | Life 3.0 | waiting for the future to happen to us. | 1:10:07.040 | 1:10:09.280 |
1 | Max Tegmark | Life 3.0 | Hey, we're the ones creating this future, right? | 1:10:09.280 | 1:10:11.640 |
1 | Max Tegmark | Life 3.0 | So we should be proactive about it | 1:10:11.640 | 1:10:15.520 |
1 | Max Tegmark | Life 3.0 | and ask ourselves what sort of future | 1:10:15.520 | 1:10:16.920 |
1 | Max Tegmark | Life 3.0 | we would like to have happen. | 1:10:16.920 | 1:10:18.240 |
1 | Max Tegmark | Life 3.0 | We're going to make it like that. | 1:10:18.240 | 1:10:19.920 |
1 | Max Tegmark | Life 3.0 | Well, what I prefer is just some sort of incredibly boring, | 1:10:19.920 | 1:10:22.720 |
1 | Max Tegmark | Life 3.0 | zombie like future where there's all these | 1:10:22.720 | 1:10:24.320 |
1 | Max Tegmark | Life 3.0 | mechanical things happening and there's no passion, | 1:10:24.320 | 1:10:26.040 |
1 | Max Tegmark | Life 3.0 | no emotion, no experience, maybe even. | 1:10:26.040 | 1:10:28.040 |
1 | Max Tegmark | Life 3.0 | No, I would of course, much rather prefer it | 1:10:29.600 | 1:10:32.040 |
1 | Max Tegmark | Life 3.0 | if all the things that we find that we value the most | 1:10:32.040 | 1:10:35.240 |
1 | Max Tegmark | Life 3.0 | about humanity are our subjective experience, | 1:10:36.240 | 1:10:40.680 |
1 | Max Tegmark | Life 3.0 | passion, inspiration, love, you know. | 1:10:40.680 | 1:10:43.000 |
1 | Max Tegmark | Life 3.0 | If we can create a future where those things do happen, | 1:10:43.000 | 1:10:48.000 |
1 | Max Tegmark | Life 3.0 | where those things do exist, you know, | 1:10:48.000 | 1:10:50.840 |
1 | Max Tegmark | Life 3.0 | I think ultimately it's not our universe | 1:10:50.840 | 1:10:54.560 |
1 | Max Tegmark | Life 3.0 | giving meaning to us, it's us giving meaning to our universe. | 1:10:54.560 | 1:10:57.960 |
1 | Max Tegmark | Life 3.0 | And if we build more advanced intelligence, | 1:10:57.960 | 1:11:01.840 |
1 | Max Tegmark | Life 3.0 | let's make sure we build it in such a way | 1:11:01.840 | 1:11:03.680 |
1 | Max Tegmark | Life 3.0 | that meaning is part of it. | 1:11:03.680 | 1:11:08.680 |
1 | Max Tegmark | Life 3.0 | A lot of people that seriously study this problem | 1:11:09.120 | 1:11:11.400 |
1 | Max Tegmark | Life 3.0 | and think of it from different angles | 1:11:11.400 | 1:11:13.600 |
1 | Max Tegmark | Life 3.0 | have trouble in the majority of cases, | 1:11:13.600 | 1:11:16.880 |
1 | Max Tegmark | Life 3.0 | if they think through that happen, | 1:11:16.880 | 1:11:19.160 |
1 | Max Tegmark | Life 3.0 | are the ones that are not beneficial to humanity. | 1:11:19.160 | 1:11:22.520 |
1 | Max Tegmark | Life 3.0 | And so, yeah, so what are your thoughts? | 1:11:22.520 | 1:11:25.560 |
1 | Max Tegmark | Life 3.0 | What's should people, you know, | 1:11:25.560 | 1:11:29.400 |
1 | Max Tegmark | Life 3.0 | I really don't like people to be terrified. | 1:11:29.400 | 1:11:32.040 |
1 | Max Tegmark | Life 3.0 | What's a way for people to think about it | 1:11:33.440 | 1:11:35.040 |