00:00 I think this is something that you've spoken about a little bit.
00:01 It was something we worked really hard on because my job in government is to say, "Hang
00:06 on, there is a potential risk here," not a definite risk, but a potential risk of something
00:11 that could be bad.
00:13 My job is to protect the country, and we can only do that if we develop the capability
00:18 we need in our safety institute and then go in and make sure we can test the models before
00:22 they are released.
00:23 Delighted that that happened today, but what's your view on what we should be doing?
00:28 We've talked about the potential risk.
00:30 We don't know, but what are the types of things governments like ours should be doing to manage
00:34 and mitigate against those risks?
00:37 I deal with regulators throughout the world because of Starlink being communications,
00:42 rockets being aerospace, and cars being vehicle transport.
00:47 I'm very familiar with dealing with regulators, and I actually agree with the vast majority
00:51 of regulations.
00:52 There's a few that I disagree with from time to time, but 0.1% probably, or less than 1%
00:57 of regulations I disagree with.
01:00 There is some concern from people in Silicon Valley who have never dealt with regulators
01:06 before and they think that this is going to just crush innovation and slow them down and
01:10 be annoying.
01:11 It will be annoying, it's true.
01:12 They're not wrong about that.
01:13 I think we've learned over the years that having a referee is a good thing.
01:26 What would your observation be on AI and the impact on labor markets and people's jobs
01:32 and how they should feel about that as they think about this?
01:37 I think we are seeing the most disruptive force in history here.
01:43 We will have for the first time something that is smarter than the smartest human.
01:49 And that, I mean, it's hard to say exactly what that moment is, but there will come a
01:56 point where no job is needed.
01:59 You can have a job if you want to have a job for personal satisfaction, but the AI will
02:05 be able to do everything.
02:06 I'm somebody who believes work gives you meaning.
02:08 I think a lot about that as, I think work is a good thing, it gives people purpose in
02:12 their lives.
02:13 And if you then remove a large chunk of that, what does that mean and where do you get that
02:18 from?
02:19 Where do you get that drive, that motivation, that purpose?
02:21 I mean, you were talking about it.
02:22 You work a lot of hours.
02:23 This balance between free speech and moderation is something we grapple with as politicians.
02:29 You were grappling with your own version of that and you moved away from a manual human
02:36 way of doing it, the moderation, to the community notes.
02:40 And I think it was an interesting change.
02:43 It's not what everyone else has done.
02:45 It'd be good, what was the reasoning behind that and why do you think that is a better
02:49 way to do that?
02:50 Yeah, part of the problem is if you empower people as censors, then there's going to be
02:57 some amount of bias they have.
03:00 And then whoever appoints the censors is effectively in control of information.
03:04 So then the idea behind community notes is, well, how do we have a consensus driven, I
03:10 mean, so it's not really censoring it, but a consensus driven approach to truth.
03:15 [BLANK_AUDIO]
Comments