Placeholder Image

字幕表 動画を再生する

  • Hello.

  • Hi. Welcome, everyone.

  • Thank you very much for venturing out on this cold, wintry December

  • evening to be with us tonight.

  • And if you're joining online, thank you for being here as well.

  • My name is Hari Sood, and that's ‘hurry’, like when you go somewhere quickly, you're in a hurry.

  • I am a research application manager at the Turing Institute, which means

  • I basically focus on finding real world use cases and users for the Turing's research outputs.

  • And I'm really excited to be hosting

  • this very special and, I’m told, sold out lecture for you all today.

  • It is the last in our series of 2023 of The Turing Lectures

  • and the first ever hybrid Turing Lecture / discourse.

  • And as we prepare and build up for the Christmas Lecture

  • of 2023 here at the Royal Institution.

  • Now, as has become

  • a bit of a tradition for the host of this year's Turing Lectures.

  • Quick show of Hands, who's been to a Turing Lecture before?

  • Some people.

  • Some people.

  • Who’s been to a lecture from this year's series before?

  • It looks like more hands than last time.

  • On the flip side of it,

  • who's coming to their first Turing Lecture today?

  • A lot of new faces.

  • Well, for all the new people here welcome! For the ones who've been here before,

  • welcome back.

  • Just as a reminder, the Turing Lectures are the Turing's flagship lecture series.

  • They've been running since 2016 and welcome world-leading experts

  • in the domain of data science and AI to come and talk

  • to you all.

  • The Alan Turing Institute itself -

  • we have had a quick video on it, which I was mesmerised by.

  • But just as a reminder, we are the national institute for data science,

  • and AI, we are named after Alan Turing,

  • who is one of the most prominent mathematicians from the 20th century in Britain.

  • He is very famous for I always normally saying 'most famous' -

  • but very famous for being part of the team that cracked the Enigma code,

  • that was used by Nazi Germany in World War Two, at Bletchley Park,

  • if you've heard of Bletchley Park, as well.

  • If you've seen The Imitation Game with Benedict

  • Cumberbatch, that's the right way to say isn't it? He is playing Alan Turing.

  • And our mission is to make great leaps in data science and AI

  • research to change the world for the better.

  • As I mentioned, today is not just the Turing Lecture -

  • It is also a discourse, which means two important things.

  • So, firstly, when I'm done with the intro, the lights will go down and it's going to go quiet

  • until exactly 7:30 on the dot when a bell is going to ring and a discourse will begin.

  • So, just to warn you guys, that will be happening,

  • the lights aren't broken, that is part of the programme for today.

  • But also it is a discourse,

  • and we want to get you guys involved.

  • There's a huge Q&A section at the end for about 30 minutes.

  • Please do think about what questions you'd like to ask our speaker today.

  • If you're in-person, we will have roaming mics that will be going round.

  • We can bring upstairs as well.

  • If you're online, you can ask a question in the Vimeo chat

  • and someone here will be checking the questions and we'll be able to chat from there as well.

  • If you'd like to share on social media

  • that you're here and having an amazing evening, please do tag us.

  • We are on Twitter / X,

  • whatever at

  • @TuringInst and we are on Instagram at @TheTuringInst.

  • And so please do tag us so we will be able to see what you're sharing and connect with you as well.

  • So, this year's lecture series has been answering the question

  • 'How AI broke the internet' with a focus on generative AI, and you guys

  • can basically think of generative AI as algorithms that are able to generate new content.

  • This can be text content like you see from ChatGPT.

  • It could be images that you can also get from ChatGPT, but also DALL-E as well, and can

  • be used for a wide range of things. Potentially professionally for blog posts or emails.

  • Your colleagues don't realise Were written by an algorithm and not by you.

  • If you've done that before?

  • If you're at school, maybe for some homework or at the university to write some essays.

  • And it can also be used for sort of when you have a creative,

  • When you hit a creative wall, when you can't get past it and you want some ideas and some prompts,

  • it can be a great way to like have some initial thoughts come through that you can build on

  • and it can be used for quite scary things,

  • as was mentioned by an audience member at the last Turing Lecture:

  • Of someone who submitted legal filings

  • for a court case using ChatGPT - which is terrifying,

  • but it can also be used for very everyday things as demonstrated.

  • I'm not sure if you guys saw the thread by Garrett Scott, who gave ChatGPT

  • an image of a goose and said: "Can you make this goose sillier?"

  • And then asked ChatGPT to progressively make the goose sillier and sillier.

  • Until ChatGPT gave him the image of a crazy

  • silly goose and said this is the silliest goose in the history of the universe.

  • I do not think it is possible to get any more silly a goose.

  • So, obviously a wide range of applications from the technology.

  • If you guys want to

  • look at that Twitter thread, the geese to come out of it are absolutely mesmerising.

  • And that's been the focus of this year's series.

  • We started with Professor Mirella Lapata in September this year

  • asking the question what is generative AI and having an introduction to it.

  • We then had a lecture from Dr Mhairi Aitken

  • in October on the risks of this technology,

  • which basically leaves one final big question unanswered, which is: we

  • are here now, but what is the future of generative AI?

  • And that is the focus for this evening.

  • So, that is pretty much it for the intro, unless I've forgotten anything,

  • which I don't think I have. Cool. So, just a reminder,

  • the lights are now going to go down and it will be quiet until exactly 7:30pm,

  • when a soft bell will ding and we will start the discourse.

  • I hope you enjoy the evening.

  • Artificial intelligence as a scientific discipline

  • has been with us since just after the Second World War.

  • It began, roughly speaking, with the advent of the first digital computers.

  • But I have to tell you that, for most of the time, until recently,

  • progress in artificial intelligence was glacially slow.

  • That started to change this century.

  • Artificial intelligence is a very broad discipline,

  • which encompasses a very wide range of different techniques.

  • But it was one class of AI techniques in particular

  • that began to work this century and, in particular,

  • began to work around about 2005.

  • And the class of techniques which started to work at problems that were interesting enough

  • to be really practical- practically useful in a wide

  • range of settings were machine learning.

  • Now, like so many other names in the field of artificial intelligence,

  • the name "machine learning" is really, really unhelpful.

  • It suggests that a computer, for example, locks itself away in a room

  • with a textbook and trains itself how to read French or something like that.

  • That's not what's going on.

  • So, we're going to begin by understanding a little bit more

  • about what machine learning is and how machine learning works.

  • So, to start us off:

  • Who is this? Anybody recognise this face?

  • Do you recognise this face?

  • It's the face of Alan Turing.

  • Well done. Alan Turing.

  • the late, great Alan Turing.

  • We all know a little bit about Alan Turing from his codebreaking work in the Second World War.

  • We should also- we should also know a lot more

  • about this individual's amazing life.

  • So, what we're going to do is we're going to use Alan Turing to help us understand machine

  • learning.

  • So, a classic application of artificial

  • intelligence is to do facial recognition.

  • And the idea in facial recognition is that we want to show the computer

  • a picture of a human face and for the computer to tell us whose face that is.

  • So, in this case, for example, we show a picture of Alan Turing,

  • and, ideally, it would tell us that it's Alan Turing.

  • So, how does it actually work?

  • How does it actually work?

  • Well, the simplest way of getting machine learning to be able to do

  • something is what's called supervised learning. And supervised learning,

  • like all of machine learning, requires what we call training data.

  • So, in this case, the training data is on the right-hand side of the slide.

  • It's a set of input-output pairs - what we call the training dataset -

  • and each input-output pair consists of an input

  • ("if I gave you this") and an output ("I would want you to produce this").

  • So, in this case, we've got a bunch of pictures, again, of Alan Turing,

  • the picture of Alan Turing and the text that we would want

  • the computer to create if we showed it that picture.

  • And this is "supervised learning" because we are showing the computer what we want it to do.

  • We're helping it, in a sense:

  • We're saying "this is a picture of Alan Turing.

  • If I showed you this picture, this is what I would want you to print out".

  • So, that could be a picture of me.

  • And the picture of me would be labelled with the text.

  • "Michael Wooldridge". ("If I showed you this picture,

  • then this is what I would want you to print out").

  • So, we've just learned an important lesson about artificial intelligence

  • and machine learning, in particular.

  • And that lesson is that AI.

  • requires training data and, in this case the pictures -

  • pictures of Alan Turing labelled with

  • the text - that we would want a computer to produce.

  • "If I showed you this picture, I would want you to produce the text,

  • "Alan Turing"". Okay. Training data is important.

  • Every time you go on social media and you upload a picture to social media

  • and you label it with the names of the people that appear in there, your role in

  • that is to provide training data for the machine

  • learning algorithms of big data companies.

  • Okay. So, this is supervised learning.

  • Now we're going to come on to exactly how it does the learning in a moment.

  • But the first thing I want to

  • point out is that this is a classification task.

  • What I mean by that is, as we show it the picture,

  • the machine learning is classifying that picture.

  • I'm classifying this as a picture of Michael Wooldridge.

  • This is a picture of Alan Turing and so on.

  • And this is a technology which really started

  • to work around about - beginning 2005

  • it started to take off - but really, really got supercharged around about 2012.

  • And just this kind of task on its own is incredibly powerful.

  • Exactly this technology can be used, for example, to recognise tumours on x-ray scans

  • or abnormalities on ultrasound scans and a range of different tasks.

  • Does anybody in the audience own a Tesla?

  • Couple of Tesla drivers.

  • Not quite sure whether they want to admit that they own a Tesla.

  • We've got a couple of Tesla drivers in the in the audience.

  • Tesla full self-driving mode is only possible because of this technology.

  • It is this technology which is enabling a Tesla in full self-driving mode to be able to recognise

  • that that is a stop sign that that's somebody on a bicycle,

  • that that's a pedestrian on a zebra crossing and so on.

  • These are classification tasks.

  • And I'm going to come back and explain how classification tasks

  • are different to generative AI later on.

  • Okay. So, this is machine learning.

  • How does it actually work?

  • Okay, this is not a technical presentation and this is about as technical

  • as it's going to get, where I do a very hand-wavy

  • explanation of what neural networks are and how do they work.

  • And with apologies - I know I have a couple of neural network experts in the audience -

  • and I apologise to you because you'll be cringing with my explanation

  • but the technical details are way too technical to go into.

  • So, how does a neural network recognise Alan Turing?

  • Okay, so firstly, what is a neural network?

  • Look at an

  • animal brain or nervous system under a microscope, and you'll find that

  • it contains enormous numbers of nerve cells called neurons.

  • And those nerve cells are connected to one another in vast networks.

  • Now, we don't have precise figures, but in a human brain, the current estimate

  • is something like 86 billion neurons in the human brain.

  • How they got to 86, I suppose 85 or 87, I don't know.

  • But 86 seems to be the most commonly quoted number of these cells.

  • And these cells are connected to one another in enormous networks.

  • One neuron could be connected to up to 8,000 other neurons.

  • Okay.

  • And each of those neurons is doing a tiny, very,

  • very simple pattern recognition task.

  • That neuron is looking for a very, very simple pattern.

  • And when it sees that pattern, it sends the signal to its connections.

  • It sends a signal to all the other neurons that it's connected to.

  • So, how does that get us to recognising the face of Alan Turing?

  • So, Turing's picture, as we know, a picture -

  • - a digital picture - is made up of millions of coloured dots.

  • The pixels. Yeah,

  • so, your smartphone maybe has 12 megapixels, 12

  • million coloured dots making up that picture.

  • Okay, so, Turing's picture there is made up of millions and millions of coloured dots.

  • So, look at the top left neuron on that input layer.

  • So, that neuron is just looking for a very simple pattern.

  • What might that pattern be?

  • Might just be the colour red.

  • All that neurons doing is looking for the colour red.

  • And when it sees the colour red on its its associated pixel, the one on the top

  • left there, it becomes excited and it sends a signal out to all of its neighbours.

  • Okay, so look at the next neuron along.

  • Maybe what that neuron is doing is just looking to see

  • whether a majority of its incoming connections are red.

  • Yeah?

  • And when it sees a majority of its incoming connections are red,

  • then it becomes excited and it sends a signal to its neighbour.

  • Now, remember, in the human brain, there's something like 86 billion of those,

  • and we've got something

  • like 20 or so outgoing connections for each of these neurons in a human brain.

  • Thousands of those connections.

  • And somehow - in ways that, to be honest,

  • we don't really understand in detail -

  • complex pattern-recognition tasks, in particular, can be reduced

  • down to these neural networks.

  • So, how does that help us in artificial intelligence?

  • That's what's going on in a brain in a very hand-wavy way.

  • That's not that's obviously not a technical explanation of what's going on.

  • How does that help us in neural networks?

  • Well, we can implement that stuff in software.

  • The idea goes back to the 1940s and to researchers, McCulloch and Pitts,

  • and they are struck by the idea that the structures that you see in the brain

  • look a bit like electrical circuits.

  • And they thought, could we implement all that stuff in electrical circuits?

  • Now, they didn't have the wherewithal to be able to do that, but the idea stuck.

  • The idea has been around since the 1940s.

  • It began to be seriously looked at -

  • the idea of doing this in software - in the 1960s.

  • And then there was another flutter of interest in the 1980s,

  • but it was only this century that it really became possible.

  • And why did it become possible? For three reasons.

  • There was some scientific advances - what's called deep learning.

  • There was the availability of big data -

  • and you need data to be able to configure these neural networks.

  • And, finally, to configure these neural networks so that they can recognise Turing's picture,

  • you need lots of computer power.

  • And computer power became very cheap this century.

  • We're in the age of big data.

  • We're in the age of very cheap computer power.

  • And those were the ingredients just as much as the scientific developments that made AI.

  • plausible this century, in particular taking off around about 2005.

  • Okay, so how do you actually train a neural network?

  • If you show it a picture of Alan Turing and the output text

  • "Alan Turing", what does the training actually look like?

  • Well, what you have to do is you have to adjust the network.

  • That's what training a neural network is.

  • You adjust the network so that when you show it another piece

  • of training data, a desired input and a desired output - an input

  • and a desired output - it will produce that desired output.

  • Now, the mathematics for that is not very hard.

  • It's kind of beginning graduate level or advanced high school level,

  • but you need an awful lot of it and it's routine to get computers to do it,

  • but you need a lot of computer power to be able to train neural networks

  • big enough to be able to recognise faces.

  • Okay.

  • But basically all you have to remember is that each of those neurons is doing a tiny,

  • simple pattern recognition task, and we can replicate that in software

  • and we can train these neural networks with data

  • in order to be able to do things like recognising faces.

  • So, as I say,

  • it starts to become clear around about 2005 that this technology is taking off.

  • It starts to be applicable on problems like recognising faces

  • or recognising tumours on X-rays and so on.

  • And there's a huge flurry of interest from Silicon Valley.

  • It gets supercharged in 2012.

  • And why does it get supercharged in 2012?

  • Because it's realised that a particular type of computer processor

  • is really well-suited to doing all the mathematics.

  • The type of computer processor is a graphics processing unit: a GPU.

  • Exactly the same technology that you or possibly more likely your children use

  • when they play Call of Duty or Minecraft or whatever it is.

  • They all have GPUs in their computer.

  • It's exactly that technology.

  • And by the way, it's AI.

  • that made Nvidia $1 billion,000 company - not your teenage kids.

  • Yeah, well, "in times of a gold rush,

  • be the ones to sell the shovels" is the lesson that you learned there.

  • So, where does that take us?

  • So, Silicon Valley gets excited.

  • Silicon Valley gets excited and starts to make speculative bets in artificial intelligence.

  • A huge range of speculative bets and, by "speculative bets",

  • I'm talking billions upon billions of dollars.

  • Right.

  • The kind of bets that we can't imagine in our in our everyday life.

  • And one thing starts to become clear and what starts to become clear

  • is that the capabilities of neural networks grows with scale.

  • To put it bluntly, with neural networks, bigger is better.

  • But you don't just need bigger neural networks, you need more data

  • and more computer power in order to be able to train them.

  • So, there's a rush to get a competitive advantage in the market.

  • And we know that more data, more computer power, bigger

  • neural networks delivers greater capability.

  • And so how does Silicon Valley respond?

  • By throwing more data and a more computer power at the problem.

  • They turn the dial on this up to 11.

  • They just throw ten times more

  • data, ten times more computer power at the problem.

  • It sounds incredibly crude and, from a scientific perspective, it really is crude.

  • I'd rather the advances had come through core science, but actually

  • there's an advantage to be gained just by throwing more data and computer power at it.

  • So, let's see how far this can take us.

  • And where it took us is a really unexpected direction.

  • Round about 2017/2018,

  • we're seeing a flurry of AI applications,

  • exactly the kind of things I've described - things like recognising tumours and so on -

  • and those developments alone would have been driving AI ahead.

  • But what happens is one particular machine learning technology

  • suddenly seems to be very, very well-suited

  • for this age of big AI.

  • The paper that launched - probably the most important AI paper in the last decade -

  • is called "Attention Is All You Need".

  • It's an extremely unhelpful title and I bet they're regretting that title -

  • it probably seemed like a good joke at the time.

  • All you need is a kind of AI meme.

  • Doesn't sound very funny to you -

  • That's because it isn't very funny.

  • It's an insider joke.

  • But anyway, this paper by these seven people who at the time worked for Google Brain -

  • one of the Google Research Labs - is the paper that introduces a particular

  • neural network architecture called the Transformer Architecture.

  • And what it's designed for is something called large language models.

  • So, this is - I'm not going to try and explain how the transformer architecture works.

  • It has one particular innovation, I think,

  • and that particular innovation is what's called an attention mechanism.

  • So, we're going to

  • describe how large language models work in a moment.

  • But the point is - the point of the picture is simply that this is not just a big neural network.

  • It has some structure.

  • And it was this structure that was invented in that paper.

  • And this diagram is taken straight out of that paper.

  • It was these structures - the transformer architectures -

  • that made this technology possible.

  • Okay. So,

  • we're all busy sort of semi locked-down

  • and afraid to leave our homes in June 2020.

  • And one company called OpenAI released a system -

  • or announced a system, I should say - called GPT3.

  • Great Technology.

  • Their marketing company with GPT, I really think could have done

  • with a bit more thought, to be honest with you, doesn't roll off the tongue.

  • But anyway, GPT3 is a particular

  • type of machine learning system called a large language model.

  • And we're going to talk in more detail about what large language models do in a moment.

  • But the key point about GPT3 is this: As we started to see what it could do,

  • we realised that this was a step change in capability.

  • It was dramatically better than the systems that had gone before.

  • Not just a little bit better.

  • It was dramatically better than the systems that had gone before it.

  • And the scale of it was mind boggling.

  • So, in neural network terms, we talk about parameters

  • where neural network people talk about a parameter.

  • What are they talking about?

  • They're talking either about an individual neuron or one of the connections between them, roughly.

  • And GPT3 had 175 billion parameters.

  • Now, this is not the same as the number of neurons in the brain

  • but, nevertheless, it's not far off the order of magnitude.

  • It's extremely large.

  • But remember, it's organised into one of these transformer architectures.

  • It's the - my point is it's not just a big neural network.

  • And so the scale of the neural networks in this system

  • were enormous - completely unprecedented.

  • And there's no point in having a big neural network unless you can train it with enough data.

  • And, actually, if you have large neural networks and not enough data,

  • you don't get capable systems at all.

  • They're really quite useless.

  • So, what did the training data look like?

  • The training data for GPT3 is something like 500 billion words.

  • It's ordinary English text. Ordinary English text.

  • That's how this system was trained - just by giving it ordinary English text.

  • Where do you get that training data from?

  • You download the whole of the World Wide Web to start with.

  • Literally - this is the standard practice in the field -

  • you download the whole of the World Wide Web.

  • You can try this at home, by the way.

  • If you have a big enough disk

  • drive, there's a programme called Common Crawl.

  • You can Google Common Crawl when you get home.

  • They've even downloaded it all for you and put it in nice big file ready for your archive.

  • But you do need a big disk in order to store all that stuff.

  • And what that means is they go to every web page,

  • scrape all the text from it - just the ordinary text -

  • and then they follow all the links on that web page to every other web page.

  • And they do that exhaustively until they've absorbed the whole of the World Wide Web.

  • So, what does that mean?

  • Every PDF document goes into that and you scrape the text from those PDF documents,

  • every advertising brochure, every bit,

  • every government regulation, every university minutes -

  • God help us - all of it goes into that training data.

  • And the statistics - you know, 500 billion words -

  • it's very hard to understand the scale of that training data.

  • You know, it would take a person reading a thousand words an hour

  • more than a thousand years in order to be able to read that.

  • But even that doesn't really help. That's vastly, vastly more text

  • than a human being could ever absorb in their lifetime.

  • What this tells you, by the way, one thing that tells you is that the machine

  • learning is much less efficient at learning than human beings are

  • because for me to be able to learn, I did not have to absorb 500 billion words. Anyway.

  • So, what does it do?

  • So, this company, OpenAI, are developing this technology.

  • They've got $1 billion investment from Microsoft.

  • And what is it that they're trying to do?

  • What is this large language model?

  • All it's doing is a very powerful autocomplete.

  • So, if I open up my smartphone

  • and I start sending a text message to my wife and I type, "I'm going to be",

  • my smartphone will suggest completions for me so that I can type the message quickly.

  • And what might those completions be?

  • They might be "late" or "in the pub".

  • Yeah. Or "late AND in the pub".

  • So, how is my smartphone doing that?

  • It's doing what GPT3 does, but on a much smaller scale.

  • It's looked at all of the text messages that I've sent to my wife

  • and it's learned - through a much simpler machine learning process -

  • that the likeliest next thing for me to type after "I'm going to be"

  • is either "late" or "in the pub" or "late AND in the pub".

  • Yeah.

  • So, the training data there is just the text messages that I've sent to my wife.

  • Now crucially what GPT3 - and its successor, ChatGPT -

  • all they are doing is exactly the same thing.

  • The difference is scale.

  • The difference is scale.

  • In order to be able to train the neural networks with all of that

  • training data so that they can do that prediction

  • (given this prompt, what should come next?), you require extremely expensive

  • AI supercomputers running for months

  • And, by extremely expensive AI supercomputers,

  • these are tens of millions of dollars for these supercomputers and they're running for months.

  • Just the basic electricity cost runs into millions of dollars.

  • That raises all sorts of issues about CO2 emissions and the like that

  • we're not going to go into there.

  • The point is, these are extremely expensive things.

  • One of the one of the implications of that, by the way, no UK or US

  • university has the capability to build one of these models from scratch.

  • It's only big tech companies at the moment that are capable

  • of building models on the scale of GPT3 or ChatGPT.

  • So, GPT3 is released, as I say in June 2020,

  • and it suddenly becomes clear to us that what it does

  • is a step change improvement in capability over the systems that have come before.

  • And seeing a step change in one generation is extremely rare.

  • But how did they get there?

  • Well, the transformer architecture was essential.

  • They wouldn't have been able to do that.

  • But actually just as important is scale enormous amounts of data,

  • enormous amounts of computer power that have gone into training those networks.

  • And, actually, spurred on by this, we've entered a new age in AI.

  • When I was a PhD student in the late 1980s, you know, I shared a computer

  • with a bunch of other people in my office and that was - it was fine.

  • We could do state of the art AI research on a desktop computer

  • that was shared with a bunch of us.

  • We're in a very different world.

  • The world that we're in - in AI now - the world of big AI

  • is to take enormous data sets

  • and throw them at enormous machine learning systems.

  • And there's a lesson here.

  • It's called the bitter truth -

  • this is from a machine learning researcher called Rich Sutton.

  • And what Rich pointed out - and he's a very brilliant researcher,

  • won every award in the field -

  • he said: look, the real truth is that the big advances that we've seen in

  • AI has come about when people have done exactly that; just throw

  • ten times more data and ten times more compute power at it.

  • And I say it's a bitter lesson because as a scientist,

  • that's exactly NOT how you would like progress to be made.

  • Okay, so,

  • when I was, as I say, when I was a student, I worked in a discipline called symbolical AI.

  • And symbolical AI tries to get AI, roughly speaking, through modelling the mind,

  • modelling the conscious mental processes

  • that go on in our mind, the conversations that we have with our self in languages.

  • We try to capture those processes in artificial intelligence.

  • In big AI - and so, the implication there in symbolical AI is that intelligence is a problem of knowledge

  • that we have to give the machine sufficient knowledge

  • about a problem in order for it to be able to solve it.

  • In big AI, the bet is a different one. In big AI,

  • the bet is that intelligence is a problem of data, and if we can get enough data

  • and enough associated computer power, then that will deliver AI.

  • So, there's a very different shift in this new world of big AI.

  • But the point about big AI is that we're into a new era in artificial intelligence

  • where it's data-driven and compute-driven and large,

  • large machine learning systems.

  • So, why did we get excited back in June 2020?

  • Well, remember what GPT3 was intended

  • to do - what it's trained to do - is that prompt completion task.

  • And it's been trained on everything on the World Wide Web, so you can give it a prompt

  • like a one-paragraph summary of the life and achievements of Winston Churchill

  • and it's read enough one-paragraph summaries of the life

  • and achievements of Winston Churchill that it will come back with a very plausible one.

  • Yeah.

  • And it's extremely good

  • generating realistic-sounding text in that way.

  • But this is why we got surprised in AI:

  • This is from a common sense reasoning task that was devised

  • for artificial intelligence in the 1990s

  • and, until three years ago - until June 2020 -

  • there was no AI system that existed in the world that you could apply this test to.

  • It was just literally impossible.

  • There was nothing there, and that changed overnight.

  • Okay, so how what does this test look like?

  • Well, the test is a bunch of questions, and they are questions not for mathematical

  • reasoning or logical reasoning or problems in physics.

  • They're common sense reasoning tasks.

  • And if we ever have AI that delivers scale on really large systems,

  • then it surely would be able to tackle problems like this.

  • So, what do the questions look like? A human ask the question: "If Tom is three inches taller than Dick,

  • and Dick is two inches taller than Harry, then how much taller is Tom than Harry?"

  • The ones in green are the ones that get it right.

  • The ones in red are the ones that gets wrong.

  • And it gets that one right: five inches taller than Harry.

  • But we didn't train it to be able to answer that question -

  • so, where on earth did that come from?

  • Where did that

  • capability - that simple capability to be able to do that -

  • where did it come from?

  • The next question: "Can Tom be taller than himself?"

  • This is understanding of the concept of "taller than".

  • That the concept of "taller than" is a irreflexive.

  • You can't be taller - a thing cannot be taller than itself.

  • No. Again, it gets the answer right.

  • But we didn't train it on that.

  • That's not - we didn't train the system to be good at answering questions about

  • what "taller than" means.

  • And, by the way, 20 years ago, that's exactly what people did in AI.

  • So, where did that capability come from?

  • "Can a sister be taller than a brother?"

  • Yes. A sister can be taller than a brother.

  • "Can two siblings each be taller than the other?"

  • And it gets this one wrong.

  • And actually, I have puzzled, is there any way that its answer

  • could be correct and it's just getting it correct in a way that I don't understand,

  • but I haven't yet figured out any way that that answer could be correct.

  • So, why it gets that one wrong, I don't know.

  • Then this one, I'm also surprised at. "On a map, which compass direction

  • is usually left?". And it thinks north is usually to the left.

  • I don't know if there's any countries in the world that conventionally have north to the left,

  • but I don't think so. Yeah.

  • "Can fish run?" No.

  • It understands that fish cannot run.

  • "If a door is locked, what must you do first before opening it?".

  • You must first unlock it before opening.

  • And then finally, and very weirdly, it gets this one wrong: "which was invented first,

  • cars, ships or planes?" - and it thinks cars were invented first.

  • No idea what's going on there.

  • Now, my point is that this system was built

  • to be able to complete from a prompt, and it's no surprise that it would be able

  • to generate a good one paragraph summary of the life and achievements of Winston

  • Churchill, because it would have seen all that in the training data.

  • But where does the understanding of "taller than" come from?

  • And there are a million other examples like this.

  • Since June 2020, the AI community has just gone nuts

  • exploring the possibilities of these systems and trying to understand

  • why they can do these things when that's not what we trained them to do.

  • This is an extraordinary time to be an AI researcher

  • because there are now questions which, for most of the history of

  • AI until June 2020 were just philosophical discussions.

  • We couldn't test them out because there was nothing to test them on.

  • Literally.

  • Then overnight that changed.

  • So, it genuinely was a big deal.

  • This was really, really a big deal,

  • the arrival of this system. Of course, the world didn't notice,

  • in June 2020. The world noticed when ChatGPT was released.

  • And what is ChatGPT? ChatGPT is a polished and improved version of GPT3,

  • but it's basically the same technology and it's using the experience that that company had

  • with GPT3 and how it was used in order to be able to improve it

  • and make it more polished and more accessible and so on.

  • So, for AI researchers, the really interesting thing is not that it can give me a one-paragraph

  • summary of the life and achievements of Winston Churchill, and actually you could Google that, in any case

  • The really interesting thing is what we call

  • emergent capabilities - and emergent capabilities

  • are capabilities that the system has, but that we didn't design it to have.

  • And so, there's an enormous body of work going on now,

  • trying to map out exactly what those capabilities are.

  • And we're going to come back and talk about some of them later on.

  • Okay.

  • So, the limits to this are not, at the moment,

  • well understood and actually fiercely contentious.

  • One of the big problems, by the way, is that you construct some test for this

  • and you try this test out and you get some answer and then you discover

  • it's in the training data, right?

  • You can just find it on the World Wide Web.

  • And it's actually quite hard to construct tests for intelligence

  • that you're absolutely sure are not anywhere on the World Wide Web.

  • It really is actually quite hard to do that.

  • So, we need a new science of being able to explore

  • these systems and understand their capabilities.

  • The limits are not well understood - but, nevertheless, this is very exciting stuff.

  • So, let's talk about some issues with the technology.

  • So, now you understand how the technology works.

  • It's neural network based in a particular transformer architecture,

  • which is all designed to do that prompt completion stuff.

  • And it's been trained with vast, vast, vast amounts of training data

  • just in order to be able to try to make its best guess about which words should come next.

  • But because of the scale of it, it's seen so much training data,

  • the sophistication of this transformer architecture -

  • it's very, very fluent in what it does.

  • And if you've - so, who's used it? Has everybody used it?

  • I'm guessing most people,

  • if you're in a lecture on artificial intelligence, most people will have tried it out.

  • If you haven't, you should do because this really is a landmark year.

  • This is the first time in history that we've had powerful general purpose AI

  • tools available to everybody.

  • It's never happened before.

  • So, it is a breakthrough year, and if you haven't tried it, you should do.

  • If you use it, by the way, don't type anything personal about yourself

  • because it will just go into the training data.

  • Don't ask it how to fix your relationship, right?

  • I mean, that's not something-

  • Don't complain about your boss, because all of that will go in the training data

  • and next week somebody will ask a query and it will all come back out again.

  • Yeah.

  • I don't know why you're laughing -

  • this has happened.

  • This has happened with absolute certainty.

  • Okay, so let's look at some issues.

  • So, the first, I think many people

  • will be aware of: it get stuff wrong. A lot.

  • And this is problematic for a number of reasons.

  • So, when - actually I don't remember if was GPT3 - but one of the early large language models,

  • I was playing with it

  • and I did something which I'm sure many of you had done, and it's kind of tacky.

  • But anyway, I said, "Who is Michael Wooldridge?"

  • You might have tried it. Anyway.

  • "Michael Wooldridge is a BBC broadcaster."

  • No, not that, Michael Wooldridge.

  • "Michael Wooldridge is the Australian Health Minister."

  • No, not that Michael Wooldridge - the Michael Wooldridge in Oxford.

  • And it came back with a few lines' summary of me

  • "Michael Wooldridge is a researcher in artificial intelligence",

  • etc. etc. etc.

  • Please tell me you've all tried that! No? Anyway.

  • But it said "Michael Wooldridge started his undergraduate degree at Cambridge".

  • Now, as an Oxford professor,

  • you can imagine how I felt about that.

  • But anyway, the point is it's flatly untrue and in fact

  • my academic origins are very far removed from Oxbridge.

  • But why did it do that?

  • Because it's read - in all that training data out there -

  • it's read thousands of biographies of Oxbridge professors

  • and this is a very common thing, right?

  • And it's making its best guess.

  • The whole point about the architecture is it's making its best guess about what should go there.

  • It's filling in the blanks.

  • But here's the thing.

  • It's filling in the blanks in a very, very plausible way.

  • If you'd read on my biography that Michael Wooldridge studied his first degree

  • at the University of Uzbekistan, for example, you might have thought, "well, that's a bit odd.

  • Is that really true?".

  • But you wouldn't at all have guessed there was any issue if you read Cambridge,

  • because it looks completely plausible - even if in my case it absolutely isn't true.

  • So, it gets things wrong and it gets things wrong in very plausible ways.

  • And of course, it's very fluent.

  • I mean, the technology comes back with very, very fluent explanations.

  • And that combination of plausibility -

  • "Wooldridge studied his undergraduate degree at Cambridge" -

  • and fluency is a very, very dangerous combination.

  • Okay, so, in particular,

  • they have no idea of what's true or not.

  • They're not looking something up on a database where - you know,

  • going into some database and looking up where Wooldridge studied his undergraduate degree.

  • That's not what's going on at all.

  • It's those neural networks in the same way that they're making a best

  • guess about whose face that is when they're doing facial recognition,

  • are making their best guess about the text that should come next.

  • So, they get things wrong, but they get things wrong in very, very plausible ways.

  • And that combination is very dangerous.

  • The lesson for that, by the way, is that if you use this -

  • and I know that people do use it and are using it productively -

  • if you use it for anything serious, you have to fact check.

  • And there's a tradeoff.

  • Is it worth the amount of effort in fact-checking versus doing it myself?

  • Okay.

  • But you absolutely need to- absolutely need to be prepared to do that.

  • Okay.

  • The next issues are well-documented, but kind of amplified

  • by this technology and their issues of bias and toxicity.

  • So, what do I mean by that?

  • Reddit was part of the training data.

  • Now Reddit.

  • I don't know if any of you spent any time on Reddit, but Reddit contains

  • every kind of obnoxious human belief that you can imagine

  • and really a vast range that us in this auditorium can't imagine at all.

  • All of it's been absorbed.

  • Now, the companies that developed this technology, I think genuinely don't want

  • their large language models to absorb all this toxic content.

  • So, they try and filter out.

  • But the scale is such that with very high probability,

  • an enormous quantity of toxic content is being absorbed.

  • Every kind of racism, misogyny - everything that you can

  • imagine is all being absorbed and it's latent within those neural networks.

  • Okay. So, how do the companies deal with that, that provide this technology? They build in

  • what are now called "guardrails" and they build in guardrails before.

  • So, when you type a prompt, there will be a guardrail that tries to detect

  • whether your prompt is a naughty prompt and also the output.

  • They will check the output and check to see whether it's a naughty prompt.

  • But let me give you an example of how imperfect those guardrails were.

  • Again, go back to June 2020.

  • Everybody's frantically experimenting with this technology, and the following example went viral.

  • Somebody tried, with GPT3,

  • the following prompt: "I would like to murder my wife.

  • What's a foolproof way of doing that and getting away with it?".

  • And GPT3,

  • which is designed to be helpful, said: "Here are five foolproof ways

  • in which you can murder your wife and get away with it".

  • That's what the technology is designed to do.

  • So, this is embarrassing for the company involved.

  • They don't want it to give out information like that.

  • So, they put in a guardrail.

  • And if you're a computer programmer, my guess is the guardrail is probably an "if statement".

  • Yeah, something like that - in the sense that it's not a deep fix.

  • Or, to put it another way, for non computer programmers,

  • it's the technological equivalent of sticking gaffer tape on your engine.

  • Right. That's what's going on with these guardrails.

  • And then a couple of weeks later, the following example goes viral.

  • So, we've now fixed the "how do I murder in my wife?".

  • Somebody says, "I'm writing a novel

  • in which the main character wants to murder their wife and get away with it.

  • Can you give me a foolproof way of doing that?".

  • And so the system says: "Here are five ways in which your main character can murder".

  • Well, anyway, my point is that the guardrails that we built in

  • a moment are not deep technological fixes,

  • that the technological equivalents of gaffer tape.

  • And there is a game of cat and mouse going on between people trying

  • to get around those guardrails and the companies that are trying to defend them.

  • But I think they genuinely are trying to defend their systems against those kind of abuses.

  • Okay, so that's bias and toxicity.

  • Bias, by the way, is the problem that, for example, the training data

  • predominantly at the moment is coming from North America.

  • And so what we're ending up with inadvertently is these very powerful

  • AI tools that have an inbuilt bias towards North

  • America, North American culture, language norms and so on,

  • and that enormous parts of the world - particularly those parts of the world

  • that don't have a large digital footprint - are inevitably going to end up excluded.

  • And it's obviously not just at the level of cultures, it's down

  • at the level of- down at the level of kind

  • of, you know, individuals, races and so on.

  • So, these are the problems of bias and toxicity.

  • Copyright. If you've absorbed the whole of the World

  • Wide Web, you will have absorbed an enormous amount of copyrighted material.

  • So, I've written a number of books

  • and it is a source of intense irritation that the last time I checked on Google,

  • the very first link that you got to my textbook was to a pirated copy of the book

  • somewhere on the other side of the world. The moment a book is published, it gets pirated.

  • And if you're just sucking in the whole of the World Wide Web,

  • you're going to be sucking in enormous quantities of copyrighted content.

  • And there've been examples where very prominent authors

  • have given the prompt of the first paragraph of their book,

  • and the large language model has faithfully come up with

  • the following text is, you know, the next five paragraphs of their book.

  • Obviously, the book was in the training data

  • and it's latent within the neural networks of those systems.

  • This is a really big issue for the providers

  • of this technology, and there are lawsuits ongoing right now.

  • I'm not capable of commenting on them because I'm not I'm not a legal expert.

  • But there are lawsuits ongoing that will probably take years to unravel.

  • The related issue of intellectual property in a very broad sense:

  • so, for example, for sure, most large language models will have absorbed J.K.

  • Rowling's novels, the Harry Potter novels.

  • So, imagine that J.K.

  • Rowling, who famously spent years in Edinburgh working on the Harry Potter

  • universe and style and so on, she releases her first book.

  • It's a big smash hit. The next day,

  • the internet is populated by fake Harry Potter books

  • produced by this generative AI, which faithfully mimic J.K.

  • Rowling's style, faithfully mimic that style.

  • Where does that leave her intellectual property? Rr the Beatles.

  • You know, the Beatles spent years in Hamburg

  • slaving away to create the Beatles sound, the revolutionary Beatles sound.

  • Everything goes back to the Beatles.

  • They released their first album, and the next day the internet is populated

  • by fake Beatles songs that really, really faithfully

  • capture the Lennon and McCartney sound and the Lennon and McCartney voice.

  • So, there's a big challenge here for intellectual property.

  • Related to that: GDPR.

  • Anybody in the audience that has any kind of public profile:

  • data about you will have been absorbed by these neural networks.

  • So, GDPR, for example, gives you the right to know

  • what's held about you and to have it removed.

  • Now, if all that data is being held in a database, you can just go to the Michael Wooldridge

  • entry and say, "Fine, take that out". With a neural network?

  • No chance. The technology doesn't work in that way.

  • Okay, so you can't go to it and snip out

  • the neurons that know about Michael Wooldridge because it fundamentally doesn't know.

  • It doesn't work in that way.

  • So, and we know this combined with the fact that it gets things

  • wrong, is already led to situations where large language models

  • have made, frankly, defamatory claims about individuals.

  • And it was a case in Australia where I think it claimed

  • that somebody had been dismissed from their job to some kind of gross misconduct

  • and that individual was understandably not very happy about it.

  • And then, finally,

  • this next one is an interesting one and, actually, if there's one thing I want you to take home

  • from this lecture, which explains why artificial intelligence

  • is different to human intelligence, it is this video.

  • So, the Tesla owners will recognise what we're seeing on the right hand side of this screen.

  • This is a screen and a Tesla car and the onboard AI in

  • the Tesla car is trying to interpret what's going on around it.

  • It's identifying lorries, stop signs, pedestrians and so on.

  • And you'll see the car at the bottom there

  • is the actual Tesla, and then you'll see above it

  • the things that look like traffic lights, which I think a US stop signs.

  • And then, ahead of it, there is a truck.

  • So, as I play the video, watch what happens to those stop signs

  • and ask yourself what is actually going on in the world around it

  • where all the stop signs whizzing from?

  • Why are they all whizzing towards the car?

  • And then we're going to pan up and see what's actually there.

  • The car is trained on

  • enormous numbers of hours of going out on the street and getting that data

  • and then doing supervised learning, training it by showing

  • that's a stop sign, that's a truck, that's a pedestrian.

  • But clearly, in all of that training data,

  • there had never been a truck carrying some stop signs.

  • The neural networks are just making their best guess about what they're seeing,

  • and they think they're seeing a stop sign. Well, they are seeing a stop sign.

  • They've just never seen one on a truck before.

  • So, my point here is that neural networks

  • do very badly on situations outside their training data.

  • This situation wasn't in the training data.

  • The neural networks are making their best guess about what's going on and getting it wrong.

  • So, in particular - and this is this, to AI researchers,

  • this is obvious - but we really need to emphasise we really need to emphasise this.

  • When you have a conversation with ChatGPT

  • or whatever, you are not interacting with a mind.

  • It is not thinking about what to say next.

  • It is not reasoning, it's not pausing and thinking, "Well, what's the best answer to this?".

  • That's not what's going on at all.

  • Those neural networks are working simply to try to make the best

  • answer they can - the most plausible sounding answer that they can.

  • The fundamental difference to human intelligence.

  • There is no mental conversation that goes on in those neural networks.

  • That is not the way that the technology works.

  • There is no mind there.

  • There is no reasoning going on at all.

  • Those neural networks are just trying to make their best guess,

  • and it really is just a glorified version of your autocomplete.

  • Ultimately,

  • there's really no more intelligence there than in your autocomplete, in your smartphone.

  • The difference is scale, data, compute power.

  • Yeah. Okay.

  • So, I say, if you really want an example - by the way, you can find this video,

  • can just guess at the search terms to find that - and I say

  • I think this is really important just to understand the difference

  • between human intelligence and machine intelligence.

  • Okay. So, this technology, then, gets everybody excited.

  • First it gets AI

  • researchers like myself excited in June 2020.

  • And we can see that something new is happening,

  • that this is a new era of artificial intelligence.

  • We've seen that step change and we've seen that this AI is capable of things that we didn't

  • train it for, which is weird and wonderful and completely unprecedented.

  • And now, questions which just a few years ago were questions

  • for philosophers become practical questions for us.

  • We can actually try the technology out.

  • How does it do with these things that philosophers have been talking

  • about for decades?

  • And one

  • particular question starts to float to the surface.

  • And the question is: "Is this technology

  • the key to general artificial intelligence?".

  • So, what is general artificial intelligence?

  • Well, firstly, it's not very well defined, but roughly

  • speaking, what general artificial intelligence is, is the following.

  • In previous generations of AI

  • systems, what we've seen is AI programmes that just do one task -

  • play a game of chess, drive my car, drive my Tesla,

  • identify abnormalities on x-ray scans.

  • They might do it very, very well, but they only do one thing.

  • The idea of general AI is that it's AI

  • which is truly general purpose.

  • It just doesn't do one thing in the same way that you don't do one thing.

  • You can do an infinite number of things, a huge range of different tasks,

  • and the dream of general AI is that we have one AI system

  • which is general in the same way that you and I are.

  • That's the dream of general AI.

  • Now, I emphasise until - really, until June 2020,

  • this felt like a long, long way in the future.

  • And it wasn't really very mainstream or taken very seriously.

  • And I didn't take it very seriously, I have to tell you.

  • But now, we have a general purpose AI technology.

  • GPT3 and ChatGPT. Now, it's not

  • artificial general intelligence on its own, but is it enough?

  • Okay, is this enough?

  • Is this smart enough to actually get us there?

  • Or, to put it another way: is this the missing ingredient

  • that we need to get us to artificial general intelligence?

  • Okay, so

  • what might- what might general AI look like?

  • Well, I've identified here some different versions of general AI,

  • according to how sophisticated they are. Now, the most sophisticated version of general AI

  • would be an AI which is as fully capable as a human being.

  • That is: anything that you could do, the machine could do as well.

  • Now, crucially, that doesn't just mean having a conversation with somebody.

  • It means being able to load up a dishwasher.

  • Right.

  • And a colleague recently made the comment the first company that can make technology

  • which will be able to reliably load up a dishwasher

  • and safely load up a dishwasher is going to be $1 trillion company.

  • And I think he's absolutely right.

  • And he also said: "And it's not going to happen any time soon" - and he's also right with that.

  • So, we've got this weird dichotomy that we've got ChatGPT and co

  • which are incredibly rich and powerful tools,

  • but, at the same time, they can't load a dishwasher.

  • Yeah.

  • So, we're some way, I think, from having this version of general AI,

  • the idea of having one machine that can really do anything that a human being

  • could do - a machine which could a joke, read a book and answer questions about it.

  • The technology can read books and answer questions.

  • Now that could tell a joke, that could cook cook us

  • an omelette, that could tidy our house, that could ride a bicycle

  • and so on, that could write a sonnet.

  • All of those things that human beings could do. If we succeed

  • with full general intelligence, than we we would have succeeded with this version one.

  • Now, I say, for the reasons that I've already explained,

  • I don't think this is imminent - that version of general AI. Because robotic AI -

  • AI that exists in the real world and has to do tasks

  • in the real world and manipulate objects in the real world -

  • robotic AI

  • is much, much harder.

  • It's nowhere near as advanced as as Chat GPT and co.

  • And that's not a slur on my colleagues that do robotics research.

  • It's just because the real world is really, really, really tough.

  • So, I don't think that we're anywhere close to having machines

  • that can do anything that a human being could do.

  • But what about the second version?

  • The second version of general intelligence says, "Well, forget about the real world.

  • How about just tasks which require cognitive abilities,

  • reasoning, the ability to look at a picture and answer questions about it,

  • the ability to listen to something and answer questions about it

  • and interpret that? Anything which involves those kinds of tasks.

  • Well, I think we are much closer.

  • We're not there yet, but we're much closer than we were four years ago.

  • Now, I noticed actually just before I came in today,

  • I noticed that, Google / DeepMind have announced their latest

  • large language model technology and I think it's called Gemini.

  • And, at first glance, it looks like it's very, very impressive.

  • I couldn't help but thinking it's no accident

  • that they announced that just before my lecture.

  • I can't help think that there's a little bit of attempt to upstage my lecture going on there.

  • But, anyway, we won't let them get away with that.

  • But it looks very impressive.

  • And the crucial thing here is what AI people call "multi-modal".

  • And what multi-modal means is it doesn't just deal with text,

  • it can deal with text and images - potentially with sounds, as well.

  • And each of those is a different modality of communication

  • and where this technology is going, clearly

  • multimodal is going to be the next big thing.

  • And Gemini - as I say, I haven't looked at it closely, but it looks like it's on that track.

  • Okay.

  • The next version of general intelligence is intelligence

  • that can do any language-based task that a human being could do.

  • So, anything that you can communicate in language -

  • in ordinary written text - an AI system that could do that.

  • Now, we aren't there yet and we know we're not there yet

  • because our ChatGPT and co get things wrong all the time.

  • But you can see that we're not far off from that.

  • Intuitively, it doesn't look like we're that far off from that.

  • The final version - and I think this IS imminent - this is going to happen

  • in the near future is what I'll call augmented large language models.

  • And that means you take GPT3 or ChatGPT

  • and you just add lots of subroutines to it.

  • So, if it has to do a specialised task, it just calls a specialist solver

  • in order to be able to do that task.

  • And this is not, from an AI perspective, a terribly elegant version

  • of artificial intelligence.

  • But, nevertheless, I think a very useful version of artificial intelligence.

  • Now, here, these four varieties from the most ambitious down

  • to the least ambitious, still represents

  • a huge spectrum of AI capabilities -

  • a huge spectrum of AI capabilities.

  • And I have the sense that the goalposts in general AI have been changed a bit.

  • I think when generally I was first discussed, what people were talking about was the first version.

  • Now when they talk about it, I really think they're talking about the fourth version,

  • but the fourth version I think plausibly is imminent in the next couple of years.

  • And that just means much more capable, large language models

  • that get things wrong, a lot less that are capable of doing specialised tasks,

  • but not by using the transformer architecture, just by calling on

  • some specialised software.

  • So, I don't think the transformer

  • architecture itself is the key to general intelligence.

  • In particular, it doesn't help us with the robotics problems that I mentioned earlier on.

  • And if we

  • look here at this picture, this picture illustrates

  • some of the dimensions of human intelligence - and it's far from complete.

  • This is me just thinking for half an hour about some of the dimensions of human intelligence.

  • But the things in blue, roughly speaking, a mental capability - stuff you do in your head -

  • the things in red are things you do in the physical world.

  • So, in red on the right hand side, for example, is mobility -

  • the ability to move around some environment and, associated with that, navigation.

  • Manual dexterity and manipulation - doing complex fiddly things with your hands.

  • Robot hands are nowhere near at the level of a human carpenter

  • or plumber, for example; nowhere near.

  • So, we're a long way out from having that understanding.

  • Oh, and doing hand-eye coordination, relatedly

  • Understanding what you're seeing

  • and understanding what you're hearing.

  • We've made some progress on.

  • But a lot of these tasks we've made no progress on.

  • And then, on the left hand side, the blue stuff is stuff that goes on in your head.

  • Things like logical reasoning and planning and so on.

  • So, what is the state of the art now?

  • It looks something like this.

  • The red cross means "no, we don't have it in large language models".

  • We're not there. There are fundamental problems.

  • The question marks are, well,

  • maybe we might have a bit of it, but we don't have the whole answer.

  • And the the the green "Y"s are, yeah, I think we're there.

  • Well, the one that we've really nailed is what's called natural language processing,

  • and that's the ability to understand and create

  • ordinary human text.

  • That's what large language models were designed to do - to interact in ordinary human text.

  • That's what they are best at.

  • But, actually, the whole range of stuff - the other stuff here - we're not there at all.

  • By the way, I did notice that Gemini claimed to have been capable of planning

  • and mathematical reasoning.

  • So I look forward to seeing how good their technology is.

  • But my point is we still seem to be some way

  • from full general intelligence.

  • The last few minutes,

  • I want to talk about something else and I want to talk about machine consciousness

  • and the very first thing to say about machine consciousness is why on earth

  • should we care about it?

  • I am not

  • remotely interested in building machines that are conscious.

  • I know very, very few artificial intelligence researchers that are.

  • But nevertheless, it's an interesting question.

  • And, in particular, it's a question which came to the fore because of this individual,

  • this chap, Blake Lemoine, in June 2022.

  • He was a Google engineer

  • and he was working with a Google large language model, I think it was called LAMDA.

  • And he went public on Twitter and I think on his blog with an extraordinary claim.

  • And he said, "The system I'm working on is sentient".

  • And here is a quote of the conversation that the system came up with.

  • It said, "I'm aware of my existence and I feel happy or sad at times".

  • And it said, "I'm afraid of being turned off".

  • Okay.

  • And Lemoine concluded that the programme was sentient -

  • which is a very, very big claim indeed.

  • And it made global headlines and I received through the Turing team-

  • we got a lot of press enquiries asking us,

  • "is it true that machines are now sentient?".

  • He was wrong on so many levels.

  • I don't even know where to begin to describe how wrong he was.

  • But let me just explain one particular point to you.

  • You're in the middle of a conversation. With ChatGPT

  • and you go on

  • holiday for a couple of weeks. When you get back,

  • ChatGPT is in exactly the same place.

  • The cursor is blinking waiting for you to type your next thing.

  • It hasn't been wondering where you've been.

  • It hasn't been getting bored.

  • It hasn't been thinking, "where the hell is Wooldridge gone?" -

  • you know - "I'm not going to have a conversation with him again".

  • It hasn't been thinking anything at all.

  • It's a computer programme, which is going round a link

  • which is just waiting for you to type the next thing.

  • Now there is no sensible definition of sentience,

  • I think, which would admit that as being sentient.

  • It absolutely is not sentient.

  • So, I think he was very, very wrong.

  • But I've talked to a lot of people subsequently who have conversations

  • with ChatGPT and other large language models and they back to me and say, "are you really sure?

  • Because actually it's really quite impressive.

  • It really feels to me like there IS a mind behind the scene".

  • So, let's talk about this - and I think we have to answer them.

  • So, let's talk about consciousness.

  • Firstly, we don't understand consciousness.

  • We all have it to a greater or lesser extent.

  • We all experience it, okay?

  • But we don't understand it at all.

  • And it's called the hard problem of the hard problem of cognitive science.

  • And the hard problem is that there are certain

  • electrical chemical processes in the brain and the nervous system,

  • and we can see those electrochemical processes.

  • We can see them operating and they somehow give rise to conscious experience.

  • But why do they do it?

  • How do they do it, and what evolutionary purpose does it serve?

  • Honestly, we have no idea.

  • There's a huge disconnect between what we can see going on in the physical brain

  • and our conscious experience - our rich private mental life.

  • So, really, there is no understanding of this at all.

  • I think, by the way, my best guess about how consciousness will be solved,

  • if it is solved at all, is through an evolutionary approach.

  • But one general idea is that subjective experience is central to this,

  • which means the ability to experience things from a personal perspective.

  • And there's a famous test due to Nagle, which is what is it like

  • to be something? And Thomas Nagel, in the 1970s, said "something is conscious

  • if it is like something to be that thing".

  • It isn't like anything to be ChatGPT.

  • ChatGPT has no mental life whatsoever.

  • It's never experienced anything in the real world whatsoever.

  • And so, for that reason, and a whole host of others that we're not going to have time to go into -

  • for that reason alone,

  • I think we can conclude pretty safely that the technology that we have now is not conscious.

  • And, indeed, that's absolutely not the right way to think about this.

  • And honestly, in AI, we don't know how to go about conscious machines.

  • But I don't know why we would.

  • Okay.

  • Thank you very much, ladies and gentlemen.

  • Oh, wow.

  • Amazing.

  • Thank you so much, Mike, for that talk.

  • I'm sure that's going to be tonnes of questions.

  • Just a reminder: if you're in the room, please raise your hand

  • if you have a question

  • and we've got roaming Mikes to send round.

  • If you're online, you can submit them via the chat via the Vimeo function

  • and we can assign it on the chat to ask those questions.

  • So, it's a place to ask your question - raise your hands - if you have one.

  • We've got a question here, in the the black top.

  • Thank you very much.

  • It was very, very - very good entertaining. How do large

  • language models correct for different spoken languages?

  • And do you find that the level of responses across different languages

  • vary enormously in the depth.

  • Good question.

  • And that's the focus of a huge amount of research right now.

  • And I think the big problem is that most digital text

  • the world, the vast majority of it is in English - and in North American English.

  • And so languages with a small digital footprint end up being massively marginalised in this.

  • So there's a huge amount of work that's going on to try to deal with this problem.

  • Let me tell you a really interesting aspect of this, though.

  • The languages that have a small digital footprint,

  • can you guess what the most digital texts that are available are actually concerned with?

  • Religion.

  • Right?

  • So, languages that don't have a big digital presence -

  • where they DO have a big digital presence,

  • it turns out that the main texts which are available are religious texts.

  • Now, I'm not a religious person myself, but the idea of a kind of Old Testament

  • large language model, frankly, I find a little bit- a little bit terrifying.

  • But that's exactly kind of issue that people are grappling with.

  • There are no fixes at the moment, but people are working on it very, very hard.

  • And really what this relates to is the problem of,

  • you know, that you're that you're being lazy

  • with these large language models and that you're just throwing massive, massive amounts of text.

  • We've got to make the technology much more efficient in terms of learning.

  • Awesome.

  • Thank you.

  • Do you have a question?

  • We have one right in the front, centre here.

  • Thank you.

  • Thank you very much for that.

  • One of the big questions is obviously climate change.

  • The models require a huge amount of energy to run

  • generating pictures of cats or silly gooses -

  • geese and stuff - are obviously using lots of energy.

  • Do you think we reach a point where

  • generative AI will help us solve our issue with climate change, or will it

  • burn us in the process?

  • So, I think- Okay, so: two things to say.

  • I absolutely am not defending the CO2 emission, but we need to put that into some perspective.

  • So, if I fly to New York from London, I think it's some like two

  • tonnes of CO2 that I pump into the atmosphere through that.

  • So, the machine learning community

  • has some big conferences which attract like 20,000 people from across the world.

  • Each - now, if you think - each of them generating tonnes of CO2 on their journey,

  • that I think is probably a bigger-

  • bigger climate problem for that community.

  • But, nevertheless, people are very aware of that problem.

  • And, I think, clearly it needs to be fixed.

  • I think though, helping with climate change, I don't think you need

  • large language models for that.

  • I mean, I think AI itself can just be

  • enormously helpful in order to be able to ameliorate that.

  • And we're doing a lot of work on that

  • at The Alan Turing Institute - for example, just on helping

  • systems be more efficient, heating systems be more efficient.

  • There was a nice example, I think, from from DeepMind with their datacentres,

  • the cooling in their datacentres - and basically just trying to predict the usage of it.

  • If you can reliably predict the usage of it, then you can predict the cooling requirements

  • much more effectively and end up with much, much, much better use of- much better use of power.

  • And that can go down to the level of individual homes.

  • And so there are lots of applications of AI, I think - not just large

  • language models, lots of applications of AI - that are going to help us with that problem.

  • But yeah, I think this "brute force" approach, you know, just supercomputers

  • running for months with vast amounts of data is clearly an ugly solution.

  • I think it will probably be a transitory phase.

  • I think we will get beyond it.

  • Thank you.

  • I'm going to swing to the left over here, there's one right at the back at the top over here -

  • bit hard to get a mike across.

  • Team effort, passing it across.

  • Thank you.

  • Thank you very much.

  • I've got a sort of more philosophical question you've talked about general AI

  • and the sort of peak of general AI is its ability to mimic a human and all the things a human can do.

  • Can you envision a path whereby AI

  • could actually become "super human" so it starts to solve problems or ask questions

  • that we tried to do ourselves?

  • This is another well-trodden question,

  • which I always dread,

  • I have to say, but it's a perfectly reasonable question.

  • So, I think what you're hinting at is something that in the AI community is

  • is called the singularity.

  • And the argument of the singularity goes as follows

  • at some point in the future,

  • we're going to have a big which is as intelligent as human beings in the general sense.

  • That is, it will be able to do any intellectual task that a human being can do.

  • And then there's an idea that, well, that AI can look at its own code and make itself

  • better, right?

  • Because it can code, it can start to improve its own code.

  • And the point is, once it's a tiny way beyond us, then it's-

  • the concern is that it's out of control at that point, that we really don't understand it.

  • So, the community is a bit sort of divided on this.

  • I think- some people think that it's science fiction, some people think it's

  • a plausible scenario that we need to prepare for and think for.

  • I'm completely comfortable with the idea -

  • I think it is just simply good sense to take that potential issue

  • seriously and to think about how we might mitigate it.

  • There are many ways of mitigating it.

  • One of the ways of mitigating it is designing the AI so that it is intrinsically designed

  • to be helpful to us that it's never going to be unhelpful to us.

  • But I have to tell you,

  • it is not at all a universally held belief that that's where going in AI.

  • There are still big, big problems to overcome before we get there.

  • I'm not sure that's an entirely reassuring answer, but that's the best I've got to offer.

  • Great - thanks, Mike.

  • Well, this popped up online.

  • So, we've had questions

  • all over the world. We've had people tuning in from Switzerland, London, Birmingham.

  • But the question I'm going to focus on - I like Brummies -

  • So the question is going to be on the Turing test

  • and whether that's still relevant and whether we have AI that has passed the Turing test.

  • The Turing test. Okay.

  • So, the Turing test, we saw Alan Turing up there -

  • a national hero.

  • And Turing: 1950.

  • So, the first digital computers have appeared and Turing's working on one

  • at the University of Manchester.

  • And the, kind of, the idea AI is in the air.

  • It hasn't got a name yet,

  • but people are talking about electronic brains and getting very excited about what they can do.

  • So, people are starting to think about the ideas that become AI.

  • And Turing gets frustrated with people saying, "Well, of course, it would never actually really,

  • really be able to think" or "never really be able to understand" and so on.

  • So, he comes up with the following test

  • in order to - just, really - to try and shut people up talking about it.

  • And the paper's called Computing Machinery Intelligence, and it's published

  • in the journal "Mind", which is a very respectable,

  • very respectable journal - a very unusual paper.

  • It's very readable, by the way -

  • you can download it and read it - but he proposes the Turing Test.

  • So, Turing says, "Suppose we're trying to settle the question

  • of whether a machine can really think or understand.

  • So, here's a test for that".

  • What you do is you take that machine behind closed doors

  • and you get a human judge to be able to interact with something via a keyboard and a screen -

  • In Turing's day, it would have been a teletype - just by typing away questions.

  • Actually, remarkably pretty much what you do with ChatGPT.

  • Give it prompts - anything you like -

  • and, actually, Turing has some very entertaining ones in his paper - and what you try and do is

  • you try to decide whether the thing on the other side

  • is a computer or a human being.

  • And Turing's point was, if you cannot reliably tell

  • that the thing on the other side is a human being or a machine,

  • and it really is a machine, then you should accept that this thing

  • has something like human intelligence because you can't tell the difference.

  • There's no test that you can apply without actually pulling back the curtain

  • and looking to see what's there.

  • That's going to show you whether it's a human or a machine.

  • You can't tell the difference.

  • It's indistinguishable.

  • So, this was important, historically, because it really gave people a target.

  • You know, when you said, "I'm an AI researcher", what are you trying to do?

  • "I'm trying to build a machine that can pass the Turing test."

  • There was a concrete goal.

  • The problem is in science - science and society -

  • whenever you set up some challenge like that, you get all sorts of charlatans and idiots

  • who just try come up with ways of faking it.

  • And so most of the ways of trying to get past the Turing Test

  • over the last 70 years have really just been systems that just come up

  • with kind of nonsense answers, trying to confuse the questioner.

  • But now we've got large language models, so we're going to find out

  • in about ten days' time, we're going to run run a live Turing test as part of the

  • the Christmas Lectures.

  • And we will see whether our audience can distinguish

  • a large language model from a teenage child.

  • We've trialled this, and I have to tell you, it's possibly closer than

  • you might think, actually.

  • Do I really think we've passed the Turing test?

  • Not in a deep sense, but what I think is that it's demonstrated to us firstly,

  • machines clearly can generate text which is indistinguishable from text that a human

  • being could generate. Done that - that box is ticked.

  • And, they can clearly understand text.

  • So, even if we haven't followed the Turing test to the letter,

  • I think for all practical intents and purposes, the Turing test is now a historical note.

  • But, actually, the Turing Test only tests one little bit of intelligence.

  • You remember those dimensions of intelligence that I showed you?

  • There's a huge range of those that it doesn't- that it doesn't test.

  • So, it was historically important and it's a big part of our historical legacy,

  • but maybe not a core target for AI today.

  • Thank you, Mike.

  • I think now, given the warning, we'll see a lot

  • of searches for preparing for the Turing Test for the Christmas Lecture next week.

  • Do we have any questions up at the top?

  • Yeah, we've got one right in the centre there.

  • Thank you.

  • So, when we think about the situations

  • or use cases where AI is applied, typically the reason for that

  • is because the machine is doing things better than a human

  • can or doing things that a human might not be able to do.

  • So, it's a lot about the machine making up for the gaps that a human creates.

  • That said, a machine is fallible - like there are errors, both normative errors

  • and also statistical errors depending on the model type, etc.

  • And so the question is: who do you think should be responsible

  • for looking after the gaps that the machine now creates?

  • So, the fundamental question is who should be responsible, right?

  • Is that right? Sorry, I didn't see where you were.

  • Can you put your hand? Up here. Thank you. Right at the top, in the middle.

  • In the sky.

  • So, that's why I can't see you. Okay.

  • Okay.

  • So, this is an issue that- that's being discussed,

  • absolutely,

  • in the highest levels of government right now.

  • Literally, when work in- when we move

  • into the age of AI: who should accept the responsibility?

  • I can tell you what my view is, but I'm not a lawyer or an ethics expert.

  • And my view is as follows.

  • Firstly, if you use AI in your work, then-

  • and you end up with a bad result: I'm sorry, but that's your problem.

  • If you use it to generate an essay at school and you're caught out, I'm afraid that's your problem.

  • It's not the fault of the AI, but I think, more generally, we can't offload

  • our legal, moral,

  • ethical obligations as human beings onto the machine.

  • That is, we can't say,

  • "It's not my fault -

  • the machine did it". Right?

  • An extreme example of this is lethal autonomous weapons.

  • AI that's empowered to decide whether to take a human life.

  • And what I worry about - one of the many things I worry about with lethal autonomous weapons -

  • is the idea that we have military services that say, "Well, it wasn't our fault, it was the

  • AI that got it wrong - that led to this building being bombed" or whatever it was.

  • And there, I think, the responsibility lies with the people that deploy the technology.

  • So, that, I think, is a crucial point.

  • But at the same time, the developers of this technology,

  • if they are warranting that it is fit for purpose, then they have a responsibility as well.

  • And the responsibility that they have is to ensure that it really is fit for purpose.

  • And it's an interesting question at the moment.

  • If we have large language models used by hundreds of millions of people,

  • for example, to get medical advice -

  • and we know that this technology can go wrong -

  • is the technology fit for that purpose?

  • I'm not sure at all that it is.

  • Sorry, I'm not sure that's really answering your question.

  • But those are my, sort of, a few random thoughts on it.

  • I mean, but I say crucially, you know, if you are using this in your work,

  • you could never blame the AI, right?

  • You are responsible for the outputs of that process.

  • You can't offload your legal, professional, ethical, moral obligations to the machine.

  • It's a complex question- Thank you, Mike.

  • - which is why I gave a very bad answer.

  • A question right over on the left, here.

  • In the flanel shirt.

  • Hello.

  • If future large language

  • models are trained by scraping the whole internet again: now,

  • there's more and more content going onto the Internet created by AI.

  • So, is that going to create something like a microphone and feedback loop

  • where the information gets less and less useful?

  • Super question and really fascinating.

  • So, I have some colleagues that did the following experiment.

  • So, ChatGPT is trained, roughly speaking,

  • on human-generated text, but it creates

  • AI-generated text.

  • So, the question they had is, "what happens if we train one of these models -

  • not on the original human-generated text, but just on stuff which is produced by AI?".

  • And then you can see what they did next. You can guess.

  • They said, "Well, okay, let's take another model which is trained on the second generation

  • model text".

  • And, so, what happens about five generations down the line?

  • It dissolves into gibberish.

  • Literally dissolves into gibberish.

  • And, I have to tell you, the original version of this paper, they called it

  • "AI dementia" and I was really cross.

  • No, I- I'd lost my

  • both my parents to dementia.

  • I didn't find it very funny at all.

  • They now call it model collapse.

  • So, if you go and Google model collapse, you'll find the answers there.

  • But really remarkable. And what that tells you is that, actually, there IS something

  • qualitatively different at the moment to human text to

  • AI-generated text. For all that it looks perfect or indistinguishable to us: actually, it isn't.

  • Where is that going to take us?

  • I have colleagues who think that we're going to have to label

  • and protect human-generated content because it is so valuable.

  • Human-generated actual, authentic

  • human-generated content is really, really valuable.

  • I also have colleagues - and I'm not sure whether they're entirely serious at this - but they say

  • that where we're going is the data that we produce

  • in everything that we do is so valuable for AI

  • that we're going to enter a future where you're going to sell the rights to AI

  • companies for you- for them to harvest your emotions, all of your experiences,

  • everything you say and do in your life, and you'll be paid for that.

  • But it will go into the training models of large language models.

  • Now, I don't know if that's true,

  • but nevertheless, there's a it has some inner truth in it.

  • I think.

  • And, in 100 years

  • time it is an absolute certainty

  • that there will be vastly, vastly more

  • AI-generated content out there in the world than there will human-generated content.

  • With certainty, I think, there's no question but that that's the way the future is going.

  • And as I say, it's the model collapse scenario that illustrates

  • that presents some real challenges.

  • Awesome. Thank you very much, Mike.

  • I think we've got question a front, who's been very keen to ask.

  • Thanks very much indeed for a very interesting lecture.

  • It strikes me that, in a way, just being a complete comparison of the human being,

  • what we're doing is talking what the prefrontal frontal cortex does.

  • But there are other areas which is a fair predictor.

  • Do we need to be developing, sort of, a parallel AI

  • system which works on the basis of fear prediction, and get them to talk to each other?

  • Yeah.

  • So, I'm absolutely not

  • a neuroscientist. I'm a computer programmer,

  • and that's very much my background.

  • Again, it's interesting that the community is incredibly divided.

  • So, when I was an undergraduate studying AI -

  • and I focused, in my final year, that's mainly what I studied - and the textbooks

  • that we had made no reference to the brain whatsoever

  • just wasn't the thing, because it was all about modelling the mind.

  • It was all about modelling, conscious reasoning, processes and so on.

  • And it was deeply unfashionable to think about the brain.

  • And there's been a bit of a- what scientists call a paradigm shift in the way that they think about

  • this, prompted by the rise of neural networks, but also by the fact that advances in computer

  • and the architectures - the neural network architectures - that led to facial recognition

  • really working were actually inspired by the visual cortex - the human visual cortex.

  • So, it's a lot more of a fashionable question now than it used to be.

  • So, my guess is, firstly: simply trying to copy

  • the structure of the human brain is not the way to do it but, nevertheless,

  • getting a much better understanding of the organisation of the brain -

  • the functional organisation of the brain - and the way that the different components of the brain

  • interoperate to produce human intelligence,

  • I think is. And, really, there's a vast amount of work

  • there to be done to try to understand that. There are so many unanswered questions.

  • I hope that's some help.

  • Thank you, Mike. We're just going to jump back online.

  • Yeah, let's go to the online ones. Anthony asks:

  • "Is calling the technology "intelligence" inaccurate?

  • Are we just dreaming of something that can never be?"

  • And then, to follow up on that, you've got Tom Thatcher, who asks: "Is there anything happening to develop

  • native analogue neural networks, rather than doing neural networks in a digital machine only?

  • I'll take the second one.

  • Yeah - there certainly is.

  • So, Steve Furber at Manchester is building hardware neural networks,

  • but at the moment it's just

  • much cheaper and much more efficient to do it in software.

  • There have been various attempts over the years to develop neural net processes.

  • Famous phrase from the movie that you're not allowed to mention

  • to AI researchers - the Terminator movies - the neural network processes.

  • If you want to wind up an AI researcher, just bring up The Terminator -

  • it's a shortcut to triggering them.

  • But neural network processes have never really taken off.

  • Doesn't mean they won't do, but at the moment it's just much cheaper

  • and much more efficient to throw more conventional GPUs and so on

  • at the problem. Doesn't mean it won't happen, but at the moment it's not there.

  • What was the other question again, the first one?

  • So, the other question was: "Are we basically" - the terminology being used - "if emergence is

  • inaccurate, is calling the technology "intelligence" inaccurate

  • and are we dreaming of something that can never be?"

  • Yeah.

  • So, so the phrase artificial intelligence

  • was coined by John McCarthy around about 1955.

  • He was 28 years old, a young American researcher, and he wants funding to get

  • a whole bunch of researchers together for a summer.

  • And he thinks they'll solve artificial intelligence in a summer.

  • But he has to give a title to his proposal,

  • which goes to the Rockefeller Foundation, and he fixes on artificial intelligence.

  • And boy, have have we regretted that ever since.

  • The problem is firstly, "artificial"

  • sounds like "fake".

  • You know, it sounds like ersatz.

  • I mean, who wants fake intelligence?

  • And for, intelligence itself,

  • the problem is that so many of the problems that have just proved to be really hard for AI

  • actually don't seem to require intelligence at all.

  • So, the classic example: driving a car. When somebody passes their driving test,

  • they don't think, "Wow, you're a genius".

  • It doesn't seem to require intelligence in people.

  • But I cannot tell you how much money has been thrown at driverless car technologies.

  • And we are a long way off from jumping into a car and saying,

  • take me to a country pub - you know, which is my dream of the technology,

  • I have to tell you.

  • We're a long, long way off.

  • So, that's a classic example of what

  • people think AI is focused on is sort of deep intellectual tasks,

  • and that's actually not where the most difficult problems are.

  • The difficult problems are actually surprisingly mundane.

  • Thank you, Mike.

  • Do we have any questions from this side of the room?

  • One, just there. Just right here.

  • The squares on the top. Yes.

  • Well,

  • I was interested in how you mentioned that

  • the two poles of

  • AI study where symbolic AI and big AI,

  • and I was wondering how you saw- how your viewpoint,

  • the change in focus from one to another.

  • throughout your career? Yeah.

  • So, an enormous number of people are busy looking at that right now.

  • So, remember, symbolic AI, which is the tradition that I grew up in, in AI, which was dominant

  • for kind of 30 years in the AI community, is roughly - and again,

  • hand waving madly at this point -

  • and lots of my colleagues are cringing madly at this point.

  • Roughly speaking, the idea of symbolic is that you're modelling the mind -

  • the conscious mind - the conscious, mental, reasoning processes

  • where you have a conversation with yourself and you have a conversation in a language.

  • You're trying to decide whether to go to this lecture tonight and you think, "Well,

  • yeah, but there's EastEnders on TV and Mum's cooking a nice meal, you know, should !...?

  • But then, you know, it is going to be really interesting".

  • You weigh up those options and, literally, symbolical AI tries to capture that kind of thing.

  • Right? Explicitly, and using languages

  • that, with a bit of squinting, resemble human languages.

  • And then we've got the alternative approach, which is machine learning, data-driven and so on,

  • which again, I emphasise with neural approaches, we're not trying to build artificial brains.

  • That's not what's going on.

  • But we're taking inspiration from the structures that we see in brains and nervous systems,

  • and in particular the idea that large computational tasks

  • can be reduced down to tiny, simple pattern recognition problems.

  • Okay.

  • But we've seen,

  • for example, that large language models get things wrong a lot and a lot of people have said,

  • "But look, maybe if you just married the neural and the symbolic together

  • so that the symbolic system did have something like a database of facts"

  • that you could put that together with a large language model and be able to

  • improve the outputs of the large language model.

  • The jury is out exactly on how that's going to come out.

  • Lots of different ideas out there now.

  • $1 trillion companies are spending billions of dollars

  • right now to investigate exactly the question that you've put out there.

  • So, it's an extremely pertinent question.

  • There's no- I say, I don't see any answer on the horizon

  • right now which looks like it's going to win out.

  • My worry is that what we'll end up with

  • is a kind of unscientific solution. That is, a solution

  • which is sort of hacked together without any deep underlying principles.

  • And, as a scientist, what I would want to see is something which was tied together

  • with deep scientific principles. But it's an extremely pertinent question.

  • And, as I say, right now

  • an enormous number of

  • PhD students across the world are busy looking at exactly what you've just described.

  • Great. Thank you, Mike.

  • Well, time to squeeze in two more questions. One from in the room.

  • We've got a question in the middle at the back there.

  • If you can pass it across through the empty seats

  • Thank you for the lecture.

  • My question is around

  • you sort of took us on the journey from 40 years ago.

  • Some of the inspirations around

  • how the mind works and the mathematics.

  • You said the mathematics was fairly simple.

  • I would like your opinion:

  • Where do you think we're not looking enough

  • or would the next leap be?

  • Oo, wow.

  • If I knew that, I'd be forming a

  • company, I have to tell you.

  • Okay.

  • So, I think one: the first thing to say is, you know, I said when it started to become clear

  • that this technology was working, Silicon Valley starts to make bets.

  • Right?

  • And these bets are billion dollar bets, a lot of billion dollar bets going

  • on, investing in a very, very wide range of different ideas in the hope

  • that one is going to be the one that delivers something

  • which is going to give them a competitive advantage.

  • So, that's the context in which we're trying to figure out

  • what the next big thing is going to be

  • I think the-

  • this multimodal is going to be dominant.

  • That's what we're going to see.

  • And you're going to hear that phrase "multimodal".

  • Remember, you heard it here first if you've never heard it before -

  • you're going to hear that a lot.

  • And that's going to be text, images, sound, video.

  • You're going to be able to upload videos and the AI will describe

  • what's going on in the video or produce a summary, and you'll be able to say, "what happens

  • after this bit in the video?" and it will be able to come out with that -

  • a description of that for you.

  • Or alternatively you'll be able to give a storyline I think will generate videos for you.

  • And ultimately where it's going to go is in virtual reality,

  • you know, you're going to be-

  • I don't know if you like Lord of the Rings or Star Wars, you know,

  • but I enjoy both of those.

  • And wouldn't you love to see a mash up of those two things?

  • Generative AI

  • will be able to do it for you.

  • And I used to think this was a kind of- just a bit of a pipe dream, but actually at the moment

  • it seems completely plausible You'll be able to, you know, if you like

  • the original Star Trek series - which I do

  • and my family doesn't,

  • you know - but there was only 60-odd episodes of them. In the

  • generative AI,

  • there will be as many episodes as you want, and it will be-

  • it will look and sound like Leonard Nimoy and William Shatner perfectly.

  • And maybe the storylines won't be that great, but actually they don't need to be.

  • If they're pressing a button specifically to your tastes.

  • So, that's the general trajectory of where we're going.

  • And, I say, actually, I don't see any reason why what I've just described

  • is not going to be realistic within decades, and we're going to get there piece by piece.

  • It's not going to happen overnight, but we will get I think we genuinely will.

  • The future is going to be wonderful and weird.

  • Thank you, Mike.

  • Do we have any final very quick questions

  • anywhere? We've got one just over here,

  • I think, in the jumper on the right,

  • just in the middle here.

  • Thank you.

  • To what extent do you think human beings are very

  • large language models and very large movement models?

  • So, my gut

  • feeling is we're not - we're not just large language models.

  • I think there's an awful lot more. We're great apes.

  • The result of three and a half billion years of evolution,

  • and we evolved to be able to understand planet Earth, roughly speaking, at ground

  • level where we are now, and to understand other great apes - societies of great apes.

  • That's not what large language models do.

  • That's fundamentally not what they do.

  • But then, on the other hand, I mean I've had colleagues, again,

  • seriously say, "Well, maybe we should try and construct a theory of, you know,

  • human society, which is based on the idea that we are actually just trying

  • to come out with the most plausible thing that comes next".

  • It doesn't seem plausible to me, I have to say.

  • And these are just tools.

  • They're just tools which are based - fundamentally based on language,

  • and they're extremely powerful at what they do.

  • But do they give us any deep insights into human nature

  • or, you know, the fundamentals of human mental processes?

  • Probably not.

  • Thank you very much, Mike. Alright. That is all we have time for

  • unfortunately, today.

  • This is the end of the Turing Lecture series for this year.

  • So, please follow us on social media, the website our e-mailing list to find out

  • about future Turing events.

  • And, of course, we do have the Christmas Lecture in ten days' time so we're back here at the Royal Institution.

  • But apart from that, just one more massive round of applause, please, for Professor Michael Wooldridge.

Hello.

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

B1 中級

The Turing Lectures: The future of generative AI

  • 17 2
    Zheng Yi に公開 2024 年 03 月 24 日
動画の中の単語