#2: Is the Path to Empowerment Paved by AI? - Eric Xing

Welcome back to the DeepView Conversations. Today for our

second episode, I traveled all the way from

New Jersey to the United Arab Emirates, specifically

Abu Dhabi, which is the home of the Mohamed bin Zayed University of

Artificial Intelligence. This is the world's first and only

university that is focused exclusively on AI research. And

right now I'm speaking with Dr. Eric Jing, who is the

president of MBZ UAI. Eric, thanks so

Well, thank you for traveling all the way here. I appreciate your diligence

and I love to chat

So I want to start with you, right? And as we've kind

of talked about before, the cameras started rolling, right? Your background

is in research, and you've been a professor, and you still kind

of wear that research hat. How do you go from a professor

of computer science to the president of MBZUAI

Well, to be honest, I never thought I'd become a

president of a university at any time. I'm

still right now affiliated with Carnegie Mellon. I've been working

as a professor in Carnegie Mellon since 2004. I

advise students, I teach classes, I

write grant proposals, and write

papers. In fact, I'm still doing all this right now, except

for classroom teaching. So research is very, very strong, very,

very important for me. That's why I haven't

thought about becoming a university administrator. During my career, when

I became senior, I did receive numerous invitations

and requests to be department chairs, deans,

even in one case another president, in

different institutions in the United States or in

UK and other places in the world. And I

didn't leave Carnegie Mellon. So

when I got an invitation for interviewing

here, I thought it's maybe

a learning experience for me to see the

place and also to maybe offer my my

insights or maybe experiences or opinions to

the leadership in here. During the interview, my interaction

with the host here is mainly based

on my identity as a researcher, as

a developer, and as a scientist. So I give them

basically the viewpoint you

know, from that angle, you know, how a university should be

built, should be operated, and what's

the best function and reputation

for university. And I also know that I'm

not the only one they interviewed, you know, they interviewed quite a few people. All

of them are much more senior and and

recognized that me. So I wasn't taking it too seriously

to be a real job interview. It's really an exchange, an

intellectual exchange. So I went back home. But then After

some time, to my surprise, they made an offer to

me. I wasn't ready at

all to take this offer because at

that time I was still very busy working with my students

and working on a few exciting projects. That was

at the beginning of the COVID. In fact, the

trip to UAE for the interview was the last trip I made

before travel shut down. in the States and

around the world. As we entered the

COVID season, things changed a

lot, even back home in the States. You

are locking down at home. You don't really get to

see your students so often. You teach through a camera, which

all were very new to all of us. And you started

getting yourself isolated from the rest of the world. Which

is good and bad. I didn't like that experience. I feel research

is important for science to be done collaboratively

with colleagues, with students. So I missed that part.

But on the other hand, it gave me the time to think. and

to really ask myself again where

I'm heading. So with this offer, which is

definitely non-trivial, it is supposed to

be bringing me to a place that I've never had previous

interaction and experience with, and also it's a job I've not

done before. So I consulted

some of my colleagues, mentors, and

friends. Professor Sir Michael Bradley, the former

acting president of the university, he

made me a very, very interesting case. He said, look, this

is more than just a university administrator. It

is about creating something from

scratch with a clean slate. And also, it

is in the region where no

such university ever existed. And the

country and the region needs such an institution to

really be the beacon of a new culture.

In a sense, I have the opportunity to to

influence and to shape a new type of culture, to

really help, you know, a

whole people, a whole country, you know, to

seek new directions, you know, or create new

results and also to redefine you

know, some of the decision process and the operation and

even culture, you know, in the region. So in a sense, there is

a chance for you to impact a country to

reshape the culture. So that's quite exciting to

me. I'm always surrounded by adventures in trying new things, and

this is definitely very new. Then I asked another friend

and mentor, Professor Michael Jordan, who has

been my advisor back in UC

Berkeley when I was doing my PhD. But then we kept the

friendship all the way. And he

also strongly suggested

to me to consider this opportunity more seriously. He

said, Hui, you are a professor

and a researcher with a good track

record of being productive and being creative on the

scientific part. But what could be

the next big stage? And that's what he asked,

right? Yeah, he understands that I want to

continue to do science, but bigger science also requires bigger

resources. And also, from a pure

career development standpoint, getting such

an offer at a relatively young age really actually

means that you have the

opportunity to experiment and also to try

different new ideas. In a sense, you have a chance to fail and

do it again, do it differently and so forth. So

he said this is a once in a life opportunity because

not every country are trying to build a university of

this and also not any time such

thing can happen. This is maybe just a once-in-a-life opportunity.

So these are the two strongest positive inputs

I can remember. There are other inputs as well that

are as positive. But of course, I also received a lot of warnings

and negative feedbacks. It is very

risky. It could ruin my career and reputation if I screw the job. or

I may run into a very different culture that I

just cannot understand and work with,

all these pragmatic issues and other things. So

I took a step back and I

took advantage of being in the height of the COVID with

very few people to talk to if I don't

actively seek out and this summer, in that thinking.

And in the end, I think I

will be living in my house probably

for the next God knows how many months and

years. I wouldn't be able to achieve too many

exciting things anyway. And this

particular opportunity is indeed once in a life, right?

And from a technical standpoint, I

started to see AI as a rising

sun on the horizon. the established

university infrastructure is already having

difficulties in coping with the new needs,

the new culture, and

the new evaluation and merit

metrics in this new discipline. But it is

very hard to change an established organization. So

even from a self-professional growth

and development point of view, it

is an opportunity to take your fate into your own hands and

also maybe create an environment that can help many of

our colleagues, students, and of

course the country and the people in

this region to a brand new set

up of such an institution. So I see that as a,

I start to appreciate that's becoming a huge opportunity that I can really

make a difference because I am an active researcher. and

it is not so common for an active

researcher and teacher to jump

into a university professor role. Usually, the

typical presidents you see are already somewhat

removed from the daily scientific life and

teaching life. Therefore, it

can be difficult for them to appreciate, you know, from

a close distance, the difficulties, the

needs, and all that, you know, in programming

and in organizing the university, right? So

all this converged to, you know, a

decision that, you know, I'm going to give it a try. And

then, you know, it took me a

year to make that decision. I actually felt

really appreciative that the university was

able to wait for me for that year. In fact, also the

Board of Trustees, they had the patience and also the trust

on me, you know, to onboard myself. I

remember it took me some effort, took

everybody a lot of effort, in fact, to even make the trip to here because all

the flights were unpredictable. And even getting

a COVID test certificate is also unpredictable.

But you need that to board a plane, which is a 24-hour fresh kind of certificate. a

lot of logistic challenges. So eventually I made

it to Abu Dhabi one day before the first

class of the students opened. I remember I got

myself prepared in a temporary

residence that I basically borrowed a

room from a very kind host, and

I delivered the opening speech of

that class, again, from a Zoom

meeting. Very

simple, not as fancy as today, the recording devices, which is coming

from my computer camera. So that's basically how the whole thing started.

At that time, the university was really, really very young and weak.

It has barely, I think, less

than a dozen faculty and only one

class of students, which was

very good. But at least I didn't have

any control or

knowledge about that because I was not admitted under

my kind of direction. So we literally walk

into the unknown and that basically

If you run a startup, you're going to want to pay attention to this. Startups, especially

AI startups, need good cloud providers. Lucky for you,

Google Cloud exists, and they offer an experience, the Google for Startups

Cloud Program that's tailored specifically to you. The program

offers eligible startups up to $200,000 in Google Cloud

credits for up to two years. For AI startups, that number is $350,000. Plus,

you get the mentorship and experience of Google Cloud's top-notch team. What

are you waiting for? Start building today with Google Cloud at cloud.google.com slash

startup. That's C-L-O-U-D dot G-O-O-G-L-E dot

Yeah, it's quite a journey. Now, about AI specifically,

right? And I know you've had an interesting journey with that. You started studying

molecular biology, and you kind of jumped into machine learning. You've

said of AI that it could usher in an age of empowerment. The

printing press ushered in the Age of Enlightenment, and

you've kind of equated the two as somewhat similar in how society-changing

AI could be. What does that Age of Empowerment look like? And

from where we're sitting now, what's the gap to get there? And

I'm glad that you asked this question because, you know, many people would ask

how to achieve artificial superintelligence and so on.

And then what is that, right? So, yeah, I think I

like to see any innovation and breakthrough from

a utility standpoint, what

function it delivers. and how such functionality

change people's life, change society, and

maybe even change civilization. So

it is in that context I call AI

to be equivalent to the

breakthrough leading to the age of enlightenment,

which is the printing press. There are many, many other inventions. The

invention of the steam engine, for example, discovery of

electricity, all these are huge. But I

think AI and

also printing press, all the invention of letters

and writing are at a different level compared

to all these technical innovations. Because if

you look at, if you put AI

and the printing press and maybe the

invention of paper and writing next to each other, they have something in

common. It is actually the backbone of human

civilization in their experience with

information. It's about how we

process and store and use

information in a

massively different scope of

magnitude. have

the technology of writing, use letters and characters

and papers. then

the information can be recorded rather than through

verbal communications. I

remember, is that Plato?

I think Plato is a student of Socrates. Socrates

didn't actually believe that knowledge, one should write

any book. He believed in in-person experience of

teaching. But it is

Plato who convinced him, or maybe who took

on the job of taking down all Socrates'

teaching and ideas into writing. And now we

appreciate how important that is, because that knowledge can

pass down to generations as it is. So in

a sense, it is a embodiment of

the antiquity time, information technology. But

at that time, you need to copy the book. And the

people who own the book need to be

literate and also need to have that kind of status, have

the opportunity to access to that. So knowledge belongs to a very few

people. Printing press is

a major advancement unlocking the knowledge from a few

people to everyone because it allows every

people to have their own copy of the book which began

from the Bible and they can form their own interpretations and

understanding and then after that people start to

think basically based on that from those knowledges and

then they write more books right so we have really

even the university evolving from a theoretical school to you

know study every topics Now, why

AI is directly related

to that train of evolution? If

you look at large language model, for example, it is trained not

on one book, It is not trained on

one sector of knowledge. It is literally trained on

everything ever written into language through

civilization and also in any media. In

a sense, you can imagine it is about turning all the

libraries in the world into one

device and then make it available next to your

fingertip. You can basically now easily get knowledge and

learn from it using this interface, which

makes the knowledge broadcast

and storage and accessibility to

a different level. Then AI also not only retrieves

knowledge for you, it actually started to be

able to use the knowledge to solve problems, as we can see. They can

at least answer many of your questions. And they can also

start producing, for example, code

and softwares. And nowadays, people

are trying to develop a lot

of applications in different sectors based on that, because

knowledge can be actionated to solve problems. And

in that sense, I call it an empowerment after

enlightenment. It is not about just giving you the

space to think. It is now giving you the tools to

make you a more powerful enabler

or actor in whatever occupation that you do.

And then, in a sense, leave you out space to

do maybe more challenging, intellectually

challenging things. And so that empowerment comes

together with even a next level of enlightenment

where, you know, everybody can do

very often time, you know, a very advanced level of

work without necessarily spending 20 years of

training, you know, in school. And then people should

not start rethinking, you know, what I'm going to learn now, you know,

in the in the next five years and what kind of problems I

should tackle. So we can refocus on problems

and maybe needs rather than just

focusing on repeating, executing

The difference to me between the large language models that

we're kind of dealing with today, right, and something like the printing press is,

you know, I think of all these technological innovations as levels of

automation, right? The printing press, as you said, got, instead

of having to handwrite books and pass them out, right, we could, we

have copies of stuff. it created more accessibility. The

jump to language models and

the kind of generative AI we're seeing now, to me, what you're starting

to do is you're starting to take the human out of the loop to a degree. Do

you see some sort of risk in that environment where maybe

for some people it might unlock them to explore greater problems? Is

there a concern that for other people it just acts as a crutch for human

intelligence, human creativity, because you have everything accessible,

I don't deny that there is a risk. In fact, any new technology you

know, comes together with both

opportunities and risk. Just like, you know, even

we talk about something that we are very familiar with,

you know, medicine, for example, you know, there are always a side effect that

we need to control. That's why FDA is formed. Genetic engineering, you

know, we use them now to dramatically raise the productivity of

crops. And we can engineer, you

know, into the crops, you know, or the

cattle's or the livestock's some nutritional deficiencies

and so forth. But you can also use

that technology to create very harmful viruses

or other forms of toxicities. So

it's really about how to mitigate and

how to regulate those risks. I think AI

is not an exception. But on the other hand,

also not a unique

threat that we didn't deal with before. In

the past, people figured out that there is a clear distinction between

the science, the technology, and the product, right?

And where you set the brick and

where you regulate is well

understood, that you probably should focus more

on regulating the users, on regulating the distributions, but

on the intellectual part, which is the science, and

maybe even the technology, which is the production part, you use

a different type of management. Typically,

for the science part, you don't even manage, or you'd manage very little, because

the creativity, once you shut it down, the cost

is too big to bear. In fact, many of the assumptions

about the current rush into AI

regulation discussion and even panic is

also based on an unclear understanding

about really the power of AI and how powerful

and how potentially dangerous it can be. It

is not well understood, especially by the general public.

Sometimes, even from a tactical standpoint, even

if there is a risk, how material that

risk is, and it actually had a big impact

on the risk and benefit calculations. So

I want to say a few words about really where we are in AI in

terms of its danger and its risk. Yeah, AI

is more than automation for sure, because AI can

automatically learn skills from training

materials. So it is a one-step beyond

just a programmed automation.

It does have the ability to acquire knowledge and to

become more and more powerful once

you plug them into some data sources and

also run an algorithm. But on the other hand, AI

also has a lot of limitations. It

is not a living creature that has self-identity,

self-drive, agency, free

will, and so forth. In fact, it is unclear to me at least where

all those will come from based on the current

architecture and our current mathematics of doing things. I

think down the road, new architectures beyond transformers, new

form of data beyond text and image will be feeding into

the next generation systems to create more

functions. But these functions actually are still quite

a few steps away from the type of higher

level intelligence or human-like intelligence I

just mentioned, like an agency, like identity, like a will, and

so forth. For example, how to even do more

complex reasoning, multiple steps reasoning. You want to

plan a shopping trip, you know, a travel, and

purchase a flight ticket, you know,

a plan on hotels. That requires multiple steps of

reasoning. I don't think right now the current

version of GPTs is able to do well on that job. Now,

talking maybe a slightly more challenging task. How to run

a store? How to run a company? How

to win a campaign, a battle?

These are reasonings which are typically a

hundred steps. conditioning on real-world

environments, and also conditioning on

embodied experiences from sensory and from anything. And

these are very difficult. In fact, you can probably already find

out that you cannot ask, for example, a GPT or AI

assistant to teach you how to swim, to teach you how to play the

music. Because why even different

music mentors, why

we have master classes? You know, those musicians, when they teach you,

it's not about reading from a book and say, hey, you apply

two grams of force onto a key and hold for

this many milliseconds and to produce a sound. It's

a very different experience of music teaching. And

I don't think right now our AI system is

even designed to do that. And certainly they cannot do that. So

I think the current AI system is still

on the steep learning curve to gain more capability, whether

it has the risk that people

are now, you know, fear

about. I think it's premature to say

that. Now people say disinformation, deepfake,

and all that, these are risks

not coming from AI technology. It is coming from

the users of the technology. It's just like a

gun. You definitely can cause a lot of harm by using

the gun. But is that the mistake of the gun?

People are still making better guns these days, and

for whatever specific purposes. But how

to regulate its usage is the key. So I

would like to remind policymakers and the

general public to really be

a little bit more concise and

careful about where the focus is

in such AI risk and AI regulation conversations.

The AI regulation conversation is very broad, right?

There are some instances that are trying to be very focused, such

as with deepfakes and job loss related to

deepfakes and the creative industries. There are others that

are trying to mitigate and prepare for these hypothetical risks

that we don't really know if they'll happen at all. And

And I want to add one more thing. In fact, even if you want to regulate

and counter that, you actually need AI to do it. Because like

the deepfake images and

the videos, the misinformation, it's just

too much for people or for a

manual devices or approach to discover

and to contain all of them. Unless you shut down the

computer, every computer in the world, they will be generated. And

As you just said, the problem with this technology and

regulating this technology is it requires

a global effort. If one country regulates really

hard, that's not going to stop China or other countries from

not regulating at all. And I think there's some concern about that.

the way you kind of framed the regulation question was to focus

Exactly. Yeah. That's why I think, for example, you

said about adversarial scenarios at certain countries

like China, will be doing what they need to do anyway,

and you cannot stop them. I think the best thing

that one can do is to advance

your own technology, in fact, to make AI even better.

Even we are under a material risk, we have

the means to take care of that, to counter that.

And do you think it would be an achievable thing? There

has been talk about the idea of international government, international

governance of AI that kind of supersedes any

individual efforts. You know, in the U.S. it's even more fragmented because

you have states doing things and then the federal government is trying to figure stuff

out. the EU is moving, and that'll impact where

companies deploy. But the idea of getting

every developer around the world from universities to

corporations to agree to a set of guidelines that maybe

Should it be feasible? I'm actually optimistic. It's absolutely

achievable. But on the other hand, I have to also

point out that if you look at the age of Enlightenment after

the printing press, it took roughly 100 years. before

all the warring states to eventually align

on some common principles and also to learn

how to live with this technology. In

West Europe, different religious sectors

fought to death just because a

different interpretation of a Bible principle. you

know, after all the people learn about it and have their own interpretations. But

eventually, you know, they learn how to settle things. And

also they learn how to live with a new society that

all can read and have many books and so forth. And

I can also name some counter examples where printing press

was actually banned. And after some hundred years,

those regions become much, much less developed. than

the West European countries. Why I mention this?

because I want to emphasize the importance of experimenting

and also the time. We are just barely five

years or maybe even three years into the age of

large-language models, right? So the AI era

has just started, and it is also evolving very fast.

We actually don't even see, you know, where it will platoon and

how a steady state would look like. And that's why I

think at this point, rush

into a premature set

of very stiff laws or regulations isn't

going to be very helpful because it may cause

missed opportunities and also maybe

an awkward misfit of all these

policies. with

respect to the real world, how they evolve. So I would say that

first, from a scientific standpoint,

we need to collect enough data points and

make careful observations and analysis. about

the opportunity and the risk, and

also the trajectory of the technological development. Thinking

ahead of the curve and be able to predict where

that technology is evolving, where I should put a

gate or a checkpoint down the road, rather

than passively you know, chase, you

know, a running kind of objects

and interfere with some of the, you know,

very organic and natural and creative and productive kind

of activities of scientific

advancements. The trick is just trying to figure out where it's going to go so

you can get ahead of it. Yeah, and it will take some time. But of

course, I'm optimistic that it wouldn't take 100 years. It wouldn't be as terrible

and as costly as the religious wars

Now, you talked a minute ago about we don't want to miss the

opportunity, right? And we've talked

about how impactful this technology can be. I've

been obsessed with this idea of trying to narrow down on

what the true promise of AI is, because like you said, I think a lot

of people misunderstand what the technology is itself. Part

of that is the fault of the kind of hype of corporations that

we've been seeing and social media is not helping. Part

of that is the fact that for a lot of subject

areas within AI, we don't have unified definitions, or people say

one thing when they're talking about something else. But

trying to narrow in on the idea of what is the promise of this technology and

how it can help people, because I think in

a lot of cases, people are afraid. People

are afraid about losing their jobs. People are afraid about existential

risk scenarios, which get kind of thrust

upon them from some of the companies developing this

technology. And it's set up as something that seems

to have so many, it's set up as something that's

dual use, but it seems to have so many negative risks that

I think for a lot of people, there is a question of why

develop this at all, if the risks can range across

this realm of severity. So to you,

Yeah, that has a close connection to really the

level of functionality that the AI will

be achieving step by step. And at

every functionality, you can accordingly kind

of define the cost and also

the opportunity. For example, Henry

Ford used to say that every American need to own a car. And

I think it's very bad news to the train

drivers and maybe to the carriage drivers. But

look, if every American needs to own a car, the number of

workers to produce those cars will be humongous. In fact, it

ends up creating a job. Now imagine that in our household,

we want to have a robotic maid to just take

care of all the household

businesses. Wouldn't that be an amazing product

that everybody wouldn't object to? I don't think right now people object to

washing machines and so on. So that's a

huge market. And the key technology to

get there is AI. You need to basically have a

robot that is able to plan, able to act,

be able to understand, communicate with you, and also

get things done. So we are actually, honestly, many

years away even from achieving that. But in achieving

that, I would imagine that everybody wants to be in

that business and join the production line and

make a lot of money out of it. Yeah, you will be losing, you know, of course,

you know, the clean ladies may be losing their job, but they will get a

better job elsewhere, because you will be joining the

production line in different steps of producing that amazing machine.

And also, you know, we talk about AI being

replacing some even white collar fancy jobs, like

a accountant, like a legal

worker. because we see HHBT are

passing the bar examinations, passing the medical examinations. Wouldn't

it be nice that it freed up these very smart people to

think about new practices in

those domains? So I think Sam

Altman, from his business standpoint, defined a

few tiers of AI. I

don't want to comment on whether those steps are too ambitious or

maybe less realistic. But

I think the way of thinking is very clear. It is aiming

for higher and higher functionality. And what

I can say from a technological standpoint, It

also will be dependent on

multiple generations of data, architecture,

and algorithmic evolutions. We're not there.

In fact, we don't have the tool right now to achieve even his second and third goals.

And within these second, third, and fifth goals, in fact, I can define

multiple concrete reasoning and

also acting tasks that requires

a lot of innovations. So at this university, in

fact, we were involved in some of this work, such as developing the

post-LLM models, including the word models,

that can even do a trivial thing of

simulating the word. If you cannot simulate the word, how

can you even teach a guy to pour a glass of water? That's actually not

written in the book. There are many other things that require

you to be able to simulate real-world experiences, and also generating

the next word based on your action on the real world, and

so on and so forth. And only in that way, you

can start to do thought experiments. You can do a new form of reasoning by

simulation, which could overcome some of

the limitations of the current language models. Then once you

have the word model, you can go one step further into

an agent model, where he has the word model as the brain,

but also he needs to have other mental devices, such

as perception, and

goal, and belief, and all these sort

of things. Each of them requires another

model architectures. So I can go on and speak maybe

for an hour about the technical steps that is required, even

to achieve what we probably can see as a very normal

and tedious level of intelligence. Then

I want to also say a few words about the other opportunities that is beyond

our daily life. which is the AI for science. Doing

scientific research is always a function of a technology environment.

Science requires more than brain, it requires the tool. Microscope,

telescope actually play the bigger role in advancing biology and

physics. AI is actually one such tool. Right now

we have, for example, in the space of medicine and

biology, there's a huge amount of genomic data, proteomic

data, and all sorts of orbit data that

is not meant even for people to analyze because they are too big and

too complex. Before the modern AI

of large language models and foundation models, we

can do very little in understanding and utilizing the

data. And now there is a new

movement, in fact, in computational biology of building foundation

models to distill in an unprecedented

way, you know, new form of knowledges and also predictability and

the simulation or capability of biological

systems. Imagine that. What if you can now

use an AI model to make proposals in

a very, very short amount of time of hundreds, if not thousands, of

drug design that can be applied to cure a

particular disease. And then you don't have to even apply the

drug to a real biological system

like a cell or a model organism. You can actually

also do the same in a word model of

the biology, which simulates all biological

phenomenons and the responses up to perturbation. You

can pretty much do this whole loop of innovation and

design and experimentation on a computer.

And that actually is now becoming a very, very exciting

goal for many of my colleagues, including students and

the professors in this university. Likewise, you can

draw the same experiences or expand the same experience in physics, in

agriculture, in many other sciences. So I think AI will

become, again, the modern day new microscope, which

tells you how to, helps you how to understand the complex data,

just like the classical microscope did. which is also about helping

you to understand information that you cannot see too small or

So you were talking about limitations of the current architecture. And

something else you mentioned that I want to tie into this as well is

that you refer to AI as our

new age microscope in the ways it'll

advance science. and not

as a species, right? And I think a lot of the source

of where the fear is coming from and where the hype is coming from, and I

think those two are very interrelated, is this language around

AI. And some researchers have taken issue with the

term AI at all because of this. as

the creation of a new species, as a species to supplant the

human species, as the kind of next thing.

And we are creating this being thing, replicable,

whatever that is. And it's vague if you try to pin it down, but the

way you're talking about it, with the limitations and the

focus on it as a specific tool, It's

very different. And so, you know, one, I

want to start with what those limitations are. You know,

we talk about hallucination, we talk about biases and

these kinds of inherent issues to the architecture. What's also

the importance of the language and how

Whether AI is a species and so on, these kinds of

discussions rarely happen among scientists

and engineers. I call them the consequence

of people living in different dimensions. Imagine

that there is a tribe somewhere in the Amazon jungle that

has never seen an aircraft or

maybe a car. and then all of a sudden they see

a person landing from

an aircraft. It is not it's

going to be surprising that they take that as a new species. In

fact, if we look at the ancient mosaics

and the pictures and sculptures in Egypt

and in other ancient civilizations, you actually see

a lot of such depictions of a god-like

figure or magical species, which,

I don't know, may actually represent a machinery. That's very

possible. So there is a key

distinction between technical and a

scientific discussion from a maybe

romantic and humanity discussions. For

example, this very word of intelligence. Intelligence is

not a very scientific design. I regret actually our

discipline is called artificial intelligence. because the other

fields are called physics and biology. It's a noun. Intelligence

is a noun, but it is derived from an adjunct. It's,

you are intelligent, and so on. It's a very subjective thing. For

example, imagine that I have in my hand a

calculator, but I didn't tell you. In fact,

especially for people who didn't know the existence of calculator, he would be amazed how

fast and how well I can do not just the trivial

calculations, but also square roots and the integrations and everything. I

can be called very intelligent, but it is a calculator. Once

you are in the same dimension, you see how it is made and

how it is supposed to function, you would be less impressed

and nervous. For example, large language

model. We built it. In fact, we are building it here as

well in MBCI. We are among the few universities, if not the only one in the world,

who actually pre-train large language model of our own. It

is a model. It is a language model doing

one thing, which is the next world prediction. So

in that function, many of the utilities somehow

can be encapsulated and dispatched. For

example, answering all your questions in

a certain scope. Next World

Prediction is like a linguistic simulator of

simulating how a person would

react to a prompt or a question through language.

It is literally a definable and a

measurable function. But human beings may

respond in a certain way. In fact, even between different human

beings, we do different. Some people recite some text. They

need to have a script to read from it. Some other

people improvise this. Some people are half-half. So

the approach of achieving that to a

receiver is actually less relevant. They draw on

their experiences. Now you sit on the end of the language

model, you see all your difficult questions, your difficult math

being answered. It doesn't prove that the machine actually understands

it, you know, or really, really digests it, or do it

the way a human would do. They can do it in a very different

way, and that way, by now, happens to be next-world prediction.

So that's why among scientists, among engineers, we

tend to be less amazed by this technology. And

we knew it is going to, for example, I'm not surprised at all that in one

day, in February soon, that a large-language model could

get a gold medal in the International Math Olympiad. Because

that type of reasoning, that type of problem-solving is

doable through a natural prediction model. As long as

you have a big enough brain to remember zillions

of words. you know, a prefix,

the sentence, and then the same amount next to that. But

I also know that there are some questions that they cannot answer, even through a

very, very simple interface. I've tried that. I reserved that

question list. If you are interested, I can send it to you. But I tested even

the latest strawberry. They still cannot answer. But

to a general public, they need education, I think. Right now, the

AI literacy is very, very different from

how much we know about algebra. Some people are very poor at algebra, but

they know what that is. They wouldn't be surprised to

see a calculator. But on the other hand, AI

literacy is much worse right now. When people see something amazing,

they literally get scared. I think there is a need to enhance

the education of AI. It's not about just math and programming. There

are other dimensions regarding AI utility, AI

measurements. and business and products that

people should know. Then they can have a much, much,

hopefully, authentic and realistic view about AI, even

without knowing the mathematics and programming. You don't need to be a mathematician or

That's the fascinating thing to me about AI is that it

is possibly one of the most multidisciplinary technologies that

I can think of. You have the technical aspects,

the computer science and the math, but there's also a lot of philosophy, human

psychology, neuroscience, linguistics, the

list kind of goes on. in terms of what you're talking about and the

core of what language models are and what they do,

which is a point that, you know, I appreciate you for bringing up because it's not

talked about often enough. And this kind

of race to intelligence and what intelligence is and what human intelligence is.

And the fascinating thing to me is the kind of stark differences between

language and intelligence and how these two

are not necessarily related. And

in interacting with these chatbots, right, I feel

like we, as a species, communicate our intelligence, our

thought, through language. This is our medium. This

is what sets us apart. And when we see something talking

to us in this medium of how we

convey our intelligence, it's easy for us to assume that it

That extrapolation is very natural for humans to make.

But even so, I guess it is not difficult

for us to realize that even if we don't speak,

we still function and we can still communicate. You have gestures,

you have many sensories that

can sense visual, audio,

temperature, all sorts of things. And not

all of this actually were described in language. In fact, language can

not be used to describe everything. There are many, many reasoning that

is existing outside of language. So

I would really draw a difference

between the functionality of communicating and

functioning linguistically versus

being intelligent. Because intelligence is, again, a very,

very subjective experience. It's a user experience. It's not a measurable

I think the fascinating thing about what we're seeing with

generative AI is that- There is one more thing, which

is oftentimes we call the soul. you

know, the spirit of a creature. I

have to say that I am very limited in my knowledge in understanding that

part. I don't know what that exactly is. Can we assume

that just by becoming knowledgeable enough, I

automatically become soulful? I'm not sure.

I think there's a difference between a machine and a biological creature.

And that difference is actually at least not well understood at this point.

I think the older generation philosophers had time to

think about it. They actually debated

on what is the status of human

reasoning and human morality and

will in the definition

of human versus machineries. I

think that discussion and that debate and thinking

should continue or restart because now we

have a new technology, which is AI. At that time,

they had this discussion because of the

latest invention, which is the calculus, the the

orbiting planets and so forth, which was fascinating. It may give you the

feeling that, wow, I'm like a god. I can predict the position

of the planets. I can predict the ellipse and so forth. And

then people start to think, oh, what is knowledge? What

is reasoning? What is conscious? So forth. I

think now it's time to rethink about that. Because with AI, which we

actually can make in a mechanical way using a

very, very simple training kind

of strategy, say, next world prediction, how

can we redefine what is ration and what

is mind and so forth? But I would

caution people from

making the jump of directly connecting

a very, very capable and seemingly

smart artificial system to

a human-like creature which has the

ego and also maybe

even the malicious intent to come after us. I think

there are a lot of steps, even philosophical ones and cognitive

ones, between them. And drawing an automatic extrapolation

is not very convincing, at least to a

The point you brought up about the soul, I find that

fascinating. And that's the kind of these differences

between knowledge and having been trained on every human work

ever written. the different facets that

make up what human intelligence is. And it goes beyond

knowing things. The stuff that makes humans human

I had one interesting debate with a colleague. For example, we

train generative AI, such as the Zara, for producing movies. Now,

let's imagine that the system watch a lot of Animal Kingdom

movies, like animals fight each other, eat

each other, and so forth, so that they can produce similar movies. Along

the way, are they going to be learning the will of

killing and of predator and prey

type of mindset? Well, nobody actually

knows, at least at this point. But it's a fascinating question to

Intention is a big thing here that's missing.

But there's also a lot that's not known about language models.

The real risk I do see in a potentially

out of control AI system is not about

that they kill people and so forth. It's about their

environmental footprint. Imagine that if

you do have an AI system that is told to

infinitely iterate on some very,

very difficult objectives that actually can be measured, but

they can go very, very big. What if the machine decides

to keep training themselves and burn a

heck of energy and so forth, and causing really environmental

disasters? To me, that's a real risk we have to be careful

about. There has to be checkpoints and the bricks, so

that at least a person should be allowed to turn it off. If

you want to build a system that is never having a switch, probably

no good. But these are something you

can already bear in your mind when you do the design. And

we could call it also a regulation in the science space. But

I don't want to overgeneralize that. It's not like

everywhere. It's only a very, very

definable and measurable risk that is indeed very,

very grave. But on the other hand, I think

Right. And a lot of these risks come from putting

these systems in a place where they can make decisions without human oversight. And

I guess the immediate and obvious remedy is don't do

that. Maintain human oversight. You won't have those problems.

Yeah, yeah. That's true. So I

think it's an ongoing discussion. But again, there is still

a difference between the scientific kind

of brain work, or maybe even

the technological experimental work from

the production and deployments and

the usage. I would imagine

that it's more effective and productive to focus on

the latter. Also, one more thing. I

think AI, compared to other sciences,

has the advantage that it is still one step

away from the physical world. Imagine that you are a biologist, you

do virology and genetic

engineering. Well, the products of

your experimentation is a physical being. If

you somehow smash the glass, or you somehow cause

a lab leak, it actually does enter the

environment. And they will stay there, or they could spread,

just like what we see in COVID. So these are very mature

risks. And therefore, it requires rather

stringent regulations and control. AI maybe

conceptually have that kind of same risk, but

it is fortunately still in a digital

and a virtual environment. In the worst case, you just

throw off your iPhones and shut down your machines. You go back

to the cage, the caves. We will still

survive, right? And without, you know, for

example, computer. Of course, I know it's very difficult these days, but I

mean, it's a different level of risk. That

existential risk is different from a biological disaster,

a chemical skill, a nuclear kind of explosion,

Part of the regulation question, and all

these topics are so passionately debated in the public space.

is the idea of open source versus a lot of what

we see from the companies, which is super, super closed source.

We're not even going to tell you what we trained on. That has to come

out through reports from investigative journalists. And

it trickles out and is fought over in court because of copyright concerns.

But the debate hinges on this idea

that science needs to be open. Science

needs to be explainable and replicable. And if

it's not open, it's not really science. And it's the distinction between product

and science. But there's also concerns about

what open source could mean for security risks. And you

were talking about chemical and biological situations

and how AI open source can maybe help bad

actors in those realms. So as

it relates to that and making sure that nothing bad happens, what

Well, I don't

want to reiterate the philosophy in science

that there is a fundamental value of share

and open and so forth. Definitely, I'm for that. More

than that, I actually don't think secrecy is

a recipe for security. I

think there's a fundamental issue. People thought by keeping things secret, you

become safer. I doubt it. I

don't think history gave us too many examples of that. Because

eventually, you will lose the secrecy. I think having

more people to understand the technology, to be able

to investigate and experiment creates

a much larger pool of people

and also a large pool of knowledge to contain

the risk if that risk happens. That's kind of, again, maybe

a more philosophical and abstract argument. But

I can give you a few examples. Right

now, we don't actually know exactly how

GPTs are trained. Do

you think that makes the technology safer or less safe? Well,

I don't know. I feel less comfortable

if I'm using something that I have

no knowledge from. Yes, I'm not going to break my car and

do it myself or understand how

it is made. But at least I know that there

are more than one manufacturer who knows how to do it. If

my car breaks down, like I have to drive a Mercedes, I don't have to go back

to Mercedes. I can go to any car shop to have it fixed. That

makes me feel safe. So under the same principle, I think important

technologies, I'm not even arguing a commercial reason. I think many of

the arguments was because of a commercial interest and other

geopolitical interest. But from a security standpoint, open

source has the advantage of

sharing the knowledge to more

people, and also keeping the

technology transparent for scrutiny, and therefore, for

a better opportunity to regulate it. So I think that

part is very important, even from a pure security

standpoint. Now, of course, there are even more

arguments about the economical and the scientific kind of

value from it. Economy is obvious. You don't want

a lucrative and important technology to be

monopolized by one of your companies. That's why in

the US, you have the antitrust laws and the anti-monopoly laws,

right? So if your company, Rockefeller, once

they own all the petroleum production, the

government orders it to break it. There is always an interest

of healthy competition among entities for

key technology and key products to actually better the

quality of technology itself and also to make it affordable and

accessible to people. So that's a commercial argument I want to make. Open

source gives startups, even

universities, a share of their impact

and also their ability to

innovate in this market. That's scientifically even

more important. I think no science actually gets

developed very well if it's kept in secret. So

for the sake of advancing the science of artificial

intelligence, especially in the United States, definitely the

U.S. is right now the current leader. I

think by having the open

source community very robustly

grounded in the United States, actually end up helping the

United States to be the leader for

the past many decades in IT technology. Look at, for

example, the development of

operating systems like Linux, the search engines.

and many of the novel IT innovations

is because there is this openness of

technology to be played and experimented and

advanced, not just by a designated player, but

by the entire academic and

industrial community at large. I think that capability and

freedom is nowhere to see, in fact, elsewhere

in the world, not even in Europe. In

fact, that's part of the reason I become excited to take

on this job in here, because I do like to

see such a mindset and such a value,

a principle in education, in technology, to

impact regions which are yet to be impacted. So

here is a young country that

wants to have a voice and have a stake in the global technological

advancements. And they have the

resource and also the ambition and also the enthusiasm of

being a leader. And absolutely, I think by

you know, moving the,

maybe I'm lacking the right word, by learning or

maybe by advocating and advancing, amplifying in that

kind of openness, that kind of transparency, you

know, from the very beginning of the foundation of this university and

of the talent we are training, I believe is

at the end of the day helping, you know, to grow this technology. And

that's also why in our university, yeah, we indeed take a very, very

open, but of course, compliant and

regulated approach to develop our AI.

You may have heard about, of course, everybody heard about Lama from Meta,

right? And it's a huge benefit

and support to

the community. And they benefit a lot, a lot of researchers working

on that and maybe inspired

a lot of new advancements. In here, we

have a project called LLM360, which actually went

even multiple steps further. We not only open-sourced the

model weights, we open-sourced the

checkpoints and also the code that

you use to produce the model. And our goal is

to let researchers to be able to reproduce the result,

good and bad, because otherwise you can only use

it and augment it. But you

still have this dependency of waiting for the next version, you

know, for you to rebuild everything from scratch. And

in our case, We want to make the community even bigger

with this higher degree of transparency. And I

think, does that give us any risk? Well,

you can never rule out the possibility of a bad player somewhere

in the world to take that and turn it into somewhat

a malicious tool. I

don't think, you know, that's going

to be stopped if we just close source it. People will

figure out from else places. But on the other hand, by

opening the technology in such a way,

we end up having more people understanding how

to augment and how to manage

and contain the risk from this machine and how to make

the machine better. So again, it's a debatable topic, but

I just want to share with you my view from a developer and

a researcher who makes these kind of technologies, I think

And we talked about this a little bit, but the thing that I think is important

in the realm of AI literacy and helping ordinary

people who don't spend their time reading

research papers or writing research papers about AI, What

are the biggest misconceptions that you see people

make about artificial intelligence in general

Well, all the things that we just discussed, like taking

AI as some sort of a creature that

could spin out of control from human hands, is definitely

one misperception. And

also exaggerating the power

of AI in real human

life, and also in scientific exploration, In

my opinion, it's also a misperception because AI

is very powerful, but

it is powerful in certain dimensions. For example, it is very powerful now

in playing the game of chess and Go. But it

doesn't mean that it plays other things well, like planning

your trips. So it takes time to get there. Another

misperception is that AI is

only about large language models and the GPTs. In

fact, AI is a very, very broad technology that

is more than actually we

expected to be coexisting with

us already. Our search engine is also an AI tool. And

I don't think people want to ban such engine. And

also, now everybody is driving

a car with a GPS or

with your iPhone next to you to

show you the road. I can tell you that I'm old enough that I live

in an age where I use a paper map to drive. It's painful. I don't want

to go back that age. People don't realize that that function is also

from AI. So AI is already everywhere. So

using the GPT as the

main or the only representation of AI is

itself a misperception. And therefore, the discussion around

AI sometimes is

very, very confused and ambiguous. Maybe we

should say that, how about we discuss large language models? That's actually

a lot more precise. and

better scoped. But when you say AI, literally

you are talking about how about we take a Google

map from your iPhone. That's a part of AI. And

who knows what will come along that line, the next product. I think

self-driving is one of the major things That

is coming very close to what

it's already realized. It's just like right now the policy is

lagging behind for its deployment. In

fact, I think a policy is more needed over

there in terms of deployment. Because if you look at the statistics, the

accidents committed by an autonomous driving car actually is

lower than people. And the

reason why it is not widely adopted is a policy and

infrastructure issue, not necessarily a technological issue.

And so I think there are a lot of examples in

such a space where especially decision makers

and policy makers will be benefiting from

having scientists and developers next to

them to be part of the discussion. I have to say

that I've been to events in various organizations

in the government bodies where a scientist

like myself could amount to

5% of the participants, or even less. So

in that environment, it is amplifying, in

We talk about people kind of trying to educate themselves better. And

like you were saying, it's really difficult. A

lot of this is ambiguous. A lot of the language changes depending on who you

talk to. What should people

be doing about AI right now? How

should they be kind of going about thinking

about it? I think there's a

degree that some people are afraid that it's too complex, too

out of reach to have an opinion on it or to think

about it too much. What would you tell them they

Good question. I

have to say that I don't want

to push the blame to the public just because they don't do

enough. I think the general public, even

including decision makers, policy makers, I

genuinely believe that they wanted to know more, but they don't know where

to learn and how to learn. So our educational system

should be the starting point that our current curriculum regarding

AI or computer science probably it's time to revisit

them. In fact, for the past, I

would say, several decades, if you look at the curriculum,

there hasn't been a fundamental change. It's literally just about how

much I teach along the same line. Which program

language? Okay, I can start it from Fortran in the 50s,

and then C, and then Java, and now Python, and so forth. But

it's still about teaching a different language. mathematics, well

there are new material coming in, let's retire the older ones

and bring in new ones. But yeah, you need to know linear algebra,

you need to know calculus, and so forth. So it's still the

very, very conventional way of thinking, and the statistics, for example,

are still using a year to teach regressions and so forth. With

now AI and with functions like

GPT or search engine and so forth, in

fact, we already are in the crisis where some people are using

GPTs to write scientific reviews of the submitted

publications. And people are getting very frustrated,

but you cannot stop it. It's happening, right? So then your attitude towards

even published papers may need to be augmented. Are

they really the product of the the researchers or

it's a mix and so forth. So I think this awareness of

the changing technology should be first taught to the public,

you know, in some curriculum. And that material does not require you

to know calculus and know linear algebra and know all this fancy

mathematical stuff. And in fact, it could be taught even

in elementary school and high school. So I think bringing

the AI into the background of

all teaching materials makes them not a specialty,

but a very, very horizontal, fundamental

science is probably where we want to begin from. Start early

and start very, very simple but broad to

create awareness about what we

already are using AI to do, and how

well they can do, and what else we set

forth for them to do. That kind of high-level knowledge

can be already taught to the general public

without any emphasis on technical specialty.

Then you can talk about the connections of

AI with all disciplines, not just AI

for science or AI making a product. For

example, artists are already using generative

AI to create visual arts,

music, and many other things. How

do we appreciate and interpret these

kinds of outcomes? and how can we redefine copyrights

and maybe authorship

and all that kind of things. I think this kind

of issues are already happening, but

very few people really put time

to think about it, or not even maybe knowing it. So

I think the AI literacy is really about awareness

of such an entity, such a

capability. We all know, for example, the capability of linear

algebra, or not linear algebra, just algebra. That's why

we are not afraid of a calculator, because we actually know algebra, even

though we don't know how to do algebra, but we know what is algebra. Right

now, we don't have such a literacy in AI. And

that's where, from the educational standpoint, I

would like to see universities and even

going further earlier, high schools, elementary schools, start

thinking about that. My son is in high school right now and

I sometimes look into

his curriculum and try to understand how

much their teachers and principals actually know

about all this. I think it's fun to keep an eye on it. Then

given that's happening, I have more to say to

the general public. Hey, go read

such material. Go watch a few informative

videos on YouTube and other places. But on

the other hand, go ignore certain kind of information that

you get from all different sources. There is definitely a

difference of good and bad information. And there

is also a judgment that you have to grow to be

able to tell what is really

valuable and authentic versus

what may be something suspicious. I think, again,

I'm not a fan of just

cleaning up the things to make it like a

hospital and a clean room. In fact, if

you look at evolution, both in

terms of a biological entity or even evolution

of our brain, You evolve and

you become healthier because you have adversaries. You

have all these viruses, you have all these pathogens, you

have actually all these misinformations from

your world. And you co-evolve with it. You

end up developing your own immune to be stronger. And

keeping you away from all this isn't going to make you healthier. In

fact, it makes you weaker. I don't think the politicians

and the general public appreciate that. They want to be as

safe as possible, away from all risk and all

kind of pathogens, psychological pathogens, you

know, or biological pathogens. I think it's going to weaken

our civilization. And then when you have an adversary, we

talk about China, we talk about some other countries, what if they don't do that? You

will eventually run into an adversary that is healthier than

you, and that will be even worse. So I would really

advocate a culture of openness and

also of brevity to embrace this

reality and then try to take on it. Thanks so much,

Creators and Guests

Eric Xing
Guest
Eric Xing
President @ MBZUAI. Researcher, educator, entrepreneur, and administrator in computer science, artificial intelligence, and healthcare.
#2: Is the Path to Empowerment Paved by AI? - Eric Xing