One year has passed since generative AI started captivating the public imagination and making headlines across the globe. What are the key reflections from the year before and what are some of the emerging capabilities that will shape the future?
Peter Hallinan, Leader, Responsible AI, Amazon Web Services
Sara Hooker, Head, Research, Cohere
Pilar Manchón, Senior Director, Engineering, Google
Andrew Ng, Founder, DeepLearning.AI
Deepa Seetharaman, AI Reporter, The Wall Street Journal (moderator)
This is the full audio from a session at the World Economic Forum’s AI Governance Summit 2023, on 15 November, 2023.
Watch the session here: https://www.weforum.org/events/ai-governance-summit-2023/sessions/state-of-gen-ai-views-from-the-frontier
AI Governance Alliance: https://initiatives.weforum.org/ai-governance-alliance/home
Check out all our podcasts on wef.ch/podcasts:
Join the World Economic Forum Podcast Club on Facebook.
Podcast transcript
This transcript has been generated using speech recognition software and may contain errors. Please check its accuracy against the audio.
Deepa Seetharaman, AI Reporter, The Wall Street Journal: Hello. Welcome to the State of Generative AI: Views from the Frontier. We have so much to get through, so I'm just going to do a quick intro for all of our speakers and then jump in. This is Andrew Ng, Founder and CEO of DeepLearning.AI. We have Pilar Manchón, Senior Director of Research and Strategy at Google AI. We have Sara Hooker, Director at Cohere For AI and Peter Hallinan, Leader of Responsible AI at AWS. I thank you guys for coming.
Andrew, I just want to start with you and I'm just going to jump right in. We've had such an intense year of AI development and I'm curious if you can talk a little bit about anything that's surprised you about some of the recent advancements technologically? And has anything surprised you about the way we are talking about governing and living with these systems?
Andrew Ng, Founder, DeepLearning.AI: Yes. AI is a general-purpose technology like electricity, and I think we're in the process of identifying a lot of use cases. If I were to ask you what is electricity good for, it's always hard to answer that question because it's useful for so many things and AI is like that too.
So a lot of work remains with new generative AI, but also other things like supervised learning, labelling things that work well years ago to figure out the applications.
But I would say here's my biggest surprise. I was not expecting that in 2023, I'll be spending this much time to try to convince governments not to outlaw open-source software or to pass laws that effectively make open-source software impractical for many organizations to keep contributing to, because that's a fundamental building block. It is very democratizing,
I think earlier this morning, Jeremy Jurgens talked about how we don't want to leave nations behind. Guess what? If some of the lobbyists’ attempts to shutter open-source succeed, there'll be a lot of losers and a tiny number of winners. Almost everyone in this room, almost all nations, will be a loser if those lobbying efforts succeed. And that deeply concerns me.
Deepa Seetharaman: I'm curious if any of the other panellists, if you agree with Andrew's perspective on open-source being outlawed or if there are other models to think about?
Sara Hooker, Head, Research, Cohere: Yes, I just don't think it's binary. I think it's really interesting. I lead a research team and we build both these models as well as the next generation models. But as Andrew said, we come from a community of AI researchers where open-source is just very core to how we've developed our field and to progress within the field.
But frankly, I also wrestle with the fact that we're not in a conference centre anymore and our work is used by millions of people around the world.
I actually think- So, our lab open-sources and publishes, we're actually open to seeing a large multilingual model next year. So I say this is someone who's actively wrestling with these questions because I actually think maybe the mistake is we treat it as open source or not open source. And perhaps more interesting as a technical conversation is how do we have responsible release? What does it look like to still give researchers and independent organizations and all these organizations that are necessary for progress, access to all of this, while also making sure that we're developing the technology for model traceability and for data traceability. Because right now I think the main concern driving these questions of nuance around open source is the reality that these models are being used in ways that are powerful for good, but also can be used in ways that are unintended by the researchers who build them.
Deepa Seetharaman: Can you just describe what you mean when you say traceability?
Sara Hooker: Maybe I'll give an example, because when I mentioned we're actively grappling with that, we're releasing a multilingual model because most of that models right now are English-first. They serve English.
Andrew is completely correct by saying who is left behind. Well, we're releasing AYA because it serves many more languages, but when we release this model, it will be weights. And that means that when we drop, release weights, anyone can just copy it. It's a file and you can just- essentially we lose our ability to track where it's being used.
I think there's interesting urgency to technical questions around this. Can we have model signatures in the sense can we trace where models are used? But it is an extremely challenging technical problem. And so, I also don't want to minimize the amount of work that would be needed to have serious model traceability.
Deepa Seetharaman: Pilar, how does Google think about this? What is the idea that open source represents sort of, I don't know, a challenge to safety? I mean, how are you thinking through that particular question yourself?
Pilar Manchón, Senior Director, Engineering, Google: Well we release a lot of open software? Right? And we release a lot of models and we release a lot of tools. So, we are active contributors to the community and we completely support it.
But you know, in terms of completely open sourcing some of the models, then you have to take into consideration the benefits and the downsides, the trade-offs. And in this case, like Sara said, it's not a binary decision. It's more about when you release, what you release, with what level of traceability, control, transparency and responsibility.
So, I think that we have to find the right kind of balance, and that's what we are trying to do with Google as well with an open architecture, open infrastructure that enables you to use not only Google's models, but also any model, any open source model or any model that you have access to so that people can choose the level of risk that they want to undertake, how they want to work, how much testing is done and the level of transparency of each of those models.
So, I think that the answer is a little more complex, but we're dealing with not only the complexity of all the technologies that we're still researching about, but a complexity of releasing them in a safe way and allowing for research and also making sure that other countries don't fall behind, other communities don’t fall behind, and we democratize it. But we have to be careful.
Deepa Seetharaman: There's a lot of interesting ideas there, but I just wanted to key in on that word, safe. I mean, AI safety — we've been talking a lot about that as a concept, but it's not particularly well defined. I think most of us don't quite know specifically what that means. I'm curious: What is safe in these scenarios? What is the appropriate level of safety?
Andrew Ng: So, I feel like AI does have risks. And I think that if you look at different applications of AI for example, media, social media, if you build a medical device, if you build a self-driving car, all of those things could cause very significant harm and I think deserve to be regulated.
And I think the problem with a lot of the current breakthroughs proposals is rather than regulating the application layer, they tend to regulate at the technology layer.
So, for example, the White House executive order from a few weeks ago was starting to propose reporting requirements and maybe other burdens in the future for, basically, if you build a big AI system, then you have, you know, starting to have burdensome regulatory requirements. And I think some proposals in Europe also have a similar flavour.
And the problem with regulating the technology layer is we know that if you want more of something, don't add friction to it. To me, the heart of the question is do we think the world is better off with more intelligence, human intelligence, or artificial intelligence. Yes, intelligence can be used for nefarious purposes, but I think that as the world became more educated, more intelligent, we became better off, even though, you know, there were some nefarious uses of intelligence.
So, the problem with the current regime proposals is it adds friction to the creation of intelligence. Whereas in contrast, I think we need good AI laws to say if you serve a billion users, which, by the way, means you have the capacity for burdensome compliance, then let's get transparency and regulation and safety and auditing. But if the same laws place a similar burden on a small start-up or big company, or a very small research team at a big company, then this is about letting companies climb up and then pull the ladder up behind themselves so that no one else can follow them. And that's unfortunately where a lot of proposals are headed.
Pilar Manchón: I think that that is super important and obviously regulating the applications and it remains in the context of even what was said in the previous panel is super important not the technology itself, because it could be used for anything, so that's 100%.
But I think that is very important that we think in terms of the users that are able to use the technology but don't understand it deep enough to know what the collateral impact of what they're doing is.
So, it's not only - safety doesn't only mean using it in a safe way or creating it for people who intend to do bad things with it. But also for unintended collateral effects for people who do not understand what they're doing well enough to know, to know better.
Deepa Seetharaman: Peter, I'd love to hear from you.
Peter Hallinan: Yes, I'd just like to add maybe couple together the open source and the safety issues, if you will.
So just speaking for AWS, right? We're a big proponent of open source software. We support the PyTorch framework, we’re supporting llama, and part of that is simply to offer options. We don't have the perspective that there is going to be one model to rule them all. There are in fact going to be a variety of base models, a variety of specialized models. But, you know, there's a lot to learn about these models still.
And when you have open source models available, people can do research, they can explore things they can learn, and that improves safety across the board.
So, I think these are these are highly coupled issues. And yes, you know, one has to strike a balance. There are issues with knowledge can be used for good or for bad, but it's better to have a smaller set of sort of known unknowns and a smaller set of unknown unknowns than a larger set, I think. And I think that open source work contributes to reducing both of those.
Deepa Seetharaman: I'm curious, Sara, if you can talk a little bit about the broader discussion we're having globally around AI safety and the risks, especially existential risks versus near-term risks, there seems to be - that also is a binary conversation. And I'm curious if you could talk about whether that framing is helpful, how that's shaped the way we all think about AI.
Sara Hooker: Yes, maybe you ask because perhaps I'm a bit grumpy about this. So, I mean, I think firstly, the notion of safety, right? We talk about safety. We often talk about a lot of desirable concepts like this. We talked about interpretability for ages as if it's a finish line and one day we're like, it's safe, it's interpretable, rather than a spectrum.
I think there's a more nuanced divide, which in some ways has created a lot of value-driven divides that have kind of polarized a research community as people who build this technology about where do we place the emphasis in terms of risk.
Like these models — it's very rare as a researcher that you build something that is used overnight, that research and direct world impact collide and it only happens a few times in history. So, I think what researchers are grappling with is this technology is being used right now, but also the pace of acceleration is felt by, I think, everyone in a nuanced way.
However, how that's translated is in this divide about whether we focus on longer term risk, which may be harder to measure progress against, because, you know, when you think about something long term that's existential, essentially what you're saying is something devastating that we may not be able to articulate now, but as a future threat. Or do you focus on the reality of these models being deployed every day?
You mentioned, you know, how do users know what they don't know? How do you calibrate hallucinations or misinformation, which is something which I think we're going to talk about in much more urgent tones, but we're not talking about yet enough. I think in many ways this is for me one of the risks that is most present and we have to articulate. And that's why I think we can't treat open source as a binary we have to acknowledge open source is really critical, but it amplifies risk.
What should we do? And I think that's a much more interesting conversation to have because then we can funnel the resources we need to really equip ourselves for what's coming next year, which is that elections are going to be held all over the world and we don't have traceability tools and we don't have good ways to implement.
So how do we navigate this divide? I think what I always try and state is the tools for both existential risks and present risks require better auditing tools at scale. We have large models with millions of data points that are trained on, but also being used in very different ways.
Whether you care about risk which is perhaps more long term, like bio risk or cyber threats, or if you care about things that are very fundamental and present today, like hallucinations, we still need the ability to audit.
And that is very difficult, because take red teaming: If I asked Andrew, what does red teaming mean to you? And then I ask Pilar, what does it mean to you? You may give me totally different answers, like how long should it go for? Should it be your friends in a Slack thread? Should it be a dedicated group that does in an ongoing way for a production AI system? We have to have these crucial, precise conversations even about the reality of how we tackle any risk and create the tools we have possible.
And so that's why I think it's okay if some people in this room feel very strongly about bio risk, I'm not going to try to dissuade you, although I can at lunch - that maybe we have to really care about what's happening right now. But I do think what's important is we have a more precise conversation about the limitations of our current tools, even for present day risks, let alone, the longer-term risks.
Deepa Seetharaman: You said a lot of really interesting things, but the one I wanted to get back to is the idea that open source amplifies risk. And Andrew, I was just curious, you know, if that's the case, what is the problem with additional regulation and barriers, if open source technologies amplify risk, if they're more vulnerable to problems?
Andrew Ng: Actually, I think what Peter said was the opposite. Not that open source amplifies risk, but the transparency actually helps to reduce the risk.
Peter Hallinan: It’s both, right? I mean, you get people doing diverse things. I mean, you're not going to have guarantees of watermarked output from open source text-to-image synthesizers, for sure. The ecosystem is more complicated. But on the other hand, you gain a lot of understanding.
I think the focus on safety, there's so much temptation to focus just on a foundation model. We're basically in a process of experimenting and co-engineering new human workflows with new technologies. It's very hard to put each of these AIs into the single box, right? Some of them are quite simple. Some of them are quite complicated.
I think one has to sort of approach this on a use case by case basis, where use cases are defined extremely narrowly, so narrowly, that they'll give anybody in marketing conniptions, right.
But, you know, face recognition, for example, is not a use case. There's many different applications of face recognition technology. But you have to think very carefully. Am I trying to do virtual proctoring? Am I trying to look up a found child in a database of missing children? Am I trying to index an actor within a video dataset? All of these are different use cases. They get tuned differently.
Gen AI has dangled in front of us this beautiful model that can do so many different things. And yet, as we deploy it, we need to go back to the basics of narrow use cases. What in this particular situation makes sense?
Your question earlier about what is safe enough, right? You give me a model that does anything. I can't answer your question. But if you give me an application domain, a specific narrow use case, I can answer the question.
And more importantly, we're deploying - I mean, lots of people, lots of enterprises, lots of individuals are trying these technologies out - you have to kind of scope the challenge, the deployment challenge, the building challenge, to who's actually doing it.
If you make it a broad use case, people get stuck. But if it's a narrow use case, then you can have a development team which is not world-class philosophers and ethicists. You can have reasonable people make reasonable decisions about how to do this safely.
So I think you're sort of narrowing in, thinking carefully about risk, which, by the way, is a social decision-making process. It's not a turn the crank and this is the risk kind of thing.
And then really understanding that there is a shared responsibility model.
I know that it's been understood in security, for example, AWS has a shared security model where AWS takes care a part and the customer takes care a part.
But in ML (Machine Learning), it's endemic with the technology. ML is really about statistics. We're rolling out statistics. Okay? And once you put privacy in play, the deployer has visibility on their data. The builder does not. The deployer must understand how to test.
Testing is not easy. Okay? That requires that you introspect, that you think about what's acceptable in your particular use case. It's a time-consuming process. It takes a lot of social discourse and discussion, just as risk assessment does. But that's key to this. Anyway, I'll pause there. I get very excited about this stuff.
Andrew Ng: I think Peter is right. So, the thing about AI, and Peter mentioned the term foundation models. so large companies are training these large, and increasingly start ups, are training the base AI models from, say, reading a trillion words on the internet and that’s a core technology component.
Many of you will have used ChatGPT or Bard or other tools like that as a consumer too. There’s one segment I think is underappreciated, which is these tools are a fantastic building block for others to write software on top of them, not just to use as a consumer tool.
So maybe one quick example. You know, in previous lives I built applications for, say, email routeing. Customer sends an email, what department do I route this to? And with traditional AI techniques, it might have taken me and very good AI teams like six months to build something like that. Thanks to this generation of tools, there are now hundreds of thousands of people they can build in maybe a week what used to take me six months. And so this dramatic lowering of the barrier to entry means that there is starting to be and there will be a lot more AI applications out there in the world.
And this comes back to the point of AI being a general purpose technology. Is not just ChatGPT or Bard. It's being used in corporations for email routeing it’s being used to help with legal documents, with nascent approaches to help with health care. I've been working with the former CEO of Tinder, Renate Nyborg, on AI applied to relationship mentoring. But there are going to be far more applications than any one of us can probably imagine at this point.
And the problem with the regulations on open source is if you slow down the work on the technology, on the foundations model, you're saying let's slow down AI for all of these wonderful applications, most of which we have not even imagined yet, as opposed to if you were to say, “oh, if you want to use AI to build the medical device. Well, I know there are risks to that, you have to prove your medical device is safe before we put in the human body.” Or if you want to use AI for underwriting. Well, we know we don't want underwriting to be biased or unfair. So, I know what the risks are. Let's really regulate that.
And I think that's why I really agree with what Peter is saying, that regulating at the technology foundation model layer there is just saying we want less AI - that would damage most of the world. But if we regulate the application layer, then we can be realistic about the risk without kind of slowing the world down.
Deepa Seetharaman: What is relationship mentoring?
Andrew Ng: Oh, so my team at AI Fund we decided to apply AI to relationship coaching.
And you might wonder, like. Andrew, I'm an AI guy. What do I know about romantic relationships? And in fact, if you don't believe me, you can actually ask my wife. My wife would confirm that I know nothing about romance, but I wrote an AI together with Renate Nyborg, former CEO of Tinder, and then my team wound up building with, collaborating with her to build her products that she announced a few weeks ago called Meeno that is a romantic relationship meant to help individuals think through relationships.
I think the US Surgeon General has declared loneliness an epidemic in the United States - it is actually worse for you to be that lonely than to, say, smoking 15 cigarettes a day, I think. And so, you know, Renate, with a little bit of help from us, is trying to use AI to help with, I think, a really, really important global problem.
Pilar Manchón: I might add something because I think what you said was super interesting and I am in agreement with you probably 99%.
The other 1%. Something that I think we all know is that the legal system, the regulation and the collateral impact of everything that we are doing, always comes far behind what we're doing.
And we all kind of acknowledge that AI is not only accelerating, accelerating in itself, but accelerating everything. So, it's hard to find a domain, a science and industry anywhere where AI is not having some kind of an impact.
And if you start thinking about reducing and analyzing each of those use cases and the millions of other use cases that we haven’t even thought about that we could never do or use for, there is no regulation, there is no law, there is no precedents.
We people that are here, we struggle to keep up date with the latest of the latest. And if you include there the morality of the ethical or the values of what if we apply this to that. Try to think about a judge, trying to think about whether something is legal or ethical or whether there is collateral damage.
If we run so fast that the rest of society cannot come with us safely, then we're going to create a whole generation of casualties, of the collateral, unintended impact of this renaissance and revolution that is, on the one hand wonderful, on the other hand, unprecedented in size and speed.
So, we do need to take that into consideration. I’m as excited as you are about the renaissance of all of these wonderful things that we can do with AI. But at the same time we have to think about who else is there who is not in AI that has to follow and has to suffer the consequences of what won't necessarily go all the way right.
Andrew Ng: I empathize with what you're saying. You make a good point and I just worry that the difficulty of doing it right is being used as an excuse to do it wrong.
Deepa Seetharaman: I'm curious how you guys - how anyone on the panel that wants to address this: What's to come? You know, we have all talked about agents. Like AI agents, the idea that you might have a system that interacts with other systems that do things that are potentially helpful for you. For me, it would be reading all the emails from my kid's pre-school. That would be very helpful to have an AI agent that does that. But what are your thoughts on the feasibility of those kinds of systems, like how quickly can they come or what are the technological challenges that might stand in their way?
Peter Hallinan: I'll just say these systems are here now. If there is something, big coming this year. Well, it's already been announced. The ability to hook LLMs up to stuff and start doing things is just very attractive.
Now, I hope - take this as a clear directive - do not do things like try and steer a power grid or anything that's sort of a risky connection with these.
But these will start on the consumer side. Just as OpenAI has released recently, I mean, there's going to be lots of opportunities to hook these models up to things and have little apps and notice that in a chat that, oh, you're asking for the value of this thing and then it spins up a little script, you know, write it and do the calculation on the fly. All of that kind of stuff is here. Now how good it is. Okay. It will get better over time.
I raise it because it complicates this business of shared responsibility and testing. The notion of privacy is critical throughout. What happens when someone has signed up to use sort of an orchestrated agent system and they want their data to be theirs, as they should. And yet the system is spinning up like little programmes to execute various calculations that are needed.
And it's derive the structure of the programme from the context of the chat. Like how is that actually tested and verified? It begins to - we know how to do that. But you know, what we're beginning to do is integrate a lot of pieces together and it just takes care and thought and, you know, sort of step by step.
I mean, I don't think you can turn it off, but it's here. And I think it's partially exciting and partially test. Please test.
Sara Hooker: I don't know. I would say it's here, but it's pretty clunky. Maybe I'll describe the technical problem. I mean, you're trying to use large models with the infrastructure of the internet and the infrastructure of the internet was built in very fractured ways. The whole notion of API design is because people choose different ways of doing it in different places.
So what you'll notice now is that what's compelling about this idea of agents is in some ways we leverage external knowledge all the time, right? Our ability to connect with other humans has been amplified by having the internet or having a phone, which is probably very close to you wherever you're sitting now, you probably have your phone somewhere close. That's something which is an auxiliary tool of information.
The reality, though, is that you have to make all this work with the internet and it's going to be fractured. So, you'll notice people are starting by very particular use cases.
I think in the short term, this is going to be the reality because it will be hard to pivot and create more general agents.
I agree completely with Peter, the idea of safety. What does it mean to be in the loop? You know, how do you - if an agent conducts a transaction for you, what's the accountability if it goes wrong?
And that's a very basic sample, but there's much more perhaps problematic ones. So we have to think about what does intervention points look like? I will say that's a medium-term problem, but we need to start working on it now.
This wide idea of what's exciting, I think for me what's really interesting is things like multilinguals. So, making these models more robust in different parts of the world. Multimodal, so how - the original vision of AI was let's impart to machine skills reserved for humans. But the way it was implemented throughout computer science history has ben these disparate fields, you would have audio and computer vision and language. What's exciting about this moment is that we have the compute power perhaps to crudely do multimodal. Right now our main station seems to be throwing a lot of compute at it, but it's the first step in having a more nuanced approach.
I would also say for me, adaptive computation is one of the most interesting ideas that is really important because if you think about it, we're addicted to this formula of bigger is better. Why do we do that? Because we essentially throw a huge model at every single data point the same amount of times. And that's not how humans approach our environment. We typically apply more compute capacity to things which are more difficult. We squint if we don't understand something, but we largely ignore things that are easy.
This idea of how can we have adaptive compute is for me one of the most fundamental questions of how can we avoid this ladder to the moon, where we we’re trying to just really use this crude tool of parameter counts to try and approach more and more intelligent systems.
Deepa Seetharaman: What’s adaptive compute?
Sara Hooker: It actually, I think, is a few things, some that are actually already in production. You can think of a mixture of experts as adaptive compute, but it's not - a mixture of experts right now, frankly, as kind of a efficiency solution. It's just to reduce the total number of flops, but it's not truly modular or specialized.
If you squint hard, you'll say it's specialized, you'll say every expert is a different thing. But the reality for people who've worked on it is that we don't have good ways of enforcing specialization.
But the ideal thing is you have different models which are specialized in different things, but it's also things like early exit, like why do we have to show every example to the model the same amount of times? It's also things like what is the critical subset of data points to train on?
A lot of our work as a lab has been showing maybe you don't have to train on all of the internet. Maybe in fact it really matters what you pay attention to.
But for me, this is one of the most interesting because it moves us away from this paradigm of uniformity, where you're treating all data points the same, but you're also applying all weights to all data points. And it's a very interesting direction.
I don't know. Andrew, it looks like you want to say something. Go for it.
Andrew Ng: It makes sense. There's an emerging paradigm called data-centric AI, where the idea is instead of trying to get as much data as possible, instead of just focusing on big data, focus on good data so that you can focus your attention on what's actually most useful to expend your computation on.
I just had some other things - Sara mentioned multimodal. Just to make some predictions about upcoming trends, I think we've all seen the text processing revolution. I think the vision image processing revolution is maybe a couple of years behind the text processing revolution and images and text do come together with multimodal.
But what I'm seeing is that computers are starting to really perceive or see the world much better than ever before. And rather than image generation, I've seen the breakthrough in image analysis.
So, this will have consequences, for example, with, maybe self-driving cars where they will perceive the environment much more accurately than before.
So if you have, you know, a business with a lot of images. I think there could be consequences for this.
And then I think- what else? I think agents, we’ve chatted a lot about agents already, but this is one of the Wild West areas of AI research right now, frankly. So, I think of- the term agents is not well-defined. People use it in different ways.
But this concept that right now you can prompt or you can tell a Large Language Model like ChatGPT or Bard what to do and it does it for you. That's there now. But this idea that we can say: “Dear AI, help me do market research for the top competitors of this firm” and it will decide by itself the steps to do that. First to do a web search for the competitors, then visit each of the competitor’s websites. Then generate some research, go and do all those steps. So this idea, you can have a computer, figure out a multi-step plan, and then carry the multi-step plan. That's kind of at the heart of the agents concept.
And right now, what I'm seeing is I've seen fantastic demos. They look amazing. But, you know, most of us, we just can't get them to work,for the most practical commercial things just yet, despite the amazing demos. But this is coming. A lot of us are working on and paying attention to it, and I think when that becomes more widespread, it would be an exciting breakthrough.
Deepa Seetharaman: How long until we have agents that can book flights for us?
Sara Hooker: Go for it, Andrew.
Andrew Ng: I think for verticalized applications, it might be quite easy.
In fact, even now versions of ChatGPT can decide to browse the web, decide when to visit another web page, whether to scroll down the web page. And I think that and even now, one of the biggest application sectors of Large Language Models has been customer operations or customer service representatives. And so, if you go to a website and chat to a customer service representative, these bots are integrated to take action, such as it has to decide at some point is it going to issue a refund or not, or call a database query to answer your question about when was your order shipped and when is it going to arrive? So these AI models, they can start to take some actions by querying databases or sometimes even something, as you know, maybe risky as issuing a refund. You don't want to get that wrong. That is already starting to get there.
Deepa Seetharaman: I just want to remind the audience that we are taking questions, which I'll take in a couple of minutes. I have a lot of questions, but I'll just limit myself to one. I'm curious how you see the field continuing to develop over the next 5 to 10 years. We’ve talked about agents and that being both here and also, you know, some years away. But what are the other applications, other things that people are working on and trying to try to push us towards?
Andrew Ng: I have a suggestion for many of you from different businesses, which is: whenever there's a new wave of technology, the media and societal interest tends to be at the technology of the tooling layer, because it's fun to talk about it’s cutting edge.
But it turns out that the only way for the tooling layer, for the technology layer, to be successful, like the clouds and the open AI, the API service and so on, the only way for that to be successful is if the applications built on top of them are even more successful so that they can generate enough revenue to pay for all these tools that we read about in the media.
For whatever reason, in earlier ways of technology innovation as well, a lot of the attention is on the technology layer rather than the application layer. But for this whole ecosystem to be successful, almost by definition, the applications have to generate even more revenue.
And I think that's where a lot of the richest opportunities lie. To look at your business, figure out what are the specific use cases in your business, and then to go do that.
Actually some of my friends have done and my teams do this too, is work with businesses to try to analyze, if you have 10,000 or 100,000 employees, what are all these people actually doing? And to go through a systematic process of taking jobs which comprise many different tasks, to take jobs, break them down into tasks and figure out which of the tasks that are amenable to AI augmentation or automation. And I find that when we go through that brainstorming exercise, pretty much every time we find tons of opportunities that you end up being exciting for businesses to pursue.
Pilar Manchón: Personally, what I would like to see rather than where we're going is more more work on human value alignment.
And it's really easy to understand what that means. And we all have a general concept of, you know, we all have a certain set of values.
But the reality is that when you come down to it, your values, my values, the values in the West, the values in the East - so that's not about human right alignment. It is alignment with a certain set of values that you can be transparent about, that you can provide control over in that you can hold yourself or the model accountable for.
So it's not only to get the models themselves to be aligned with a certain set of values, but have enough control, transparency, accountability and flexibility so that we can all have versions of that model, applications of those models that align with the values we want and we can feel safe about.
And we don't have to agree on all those values. There is a core set of values that most of us, I guess I hope, agree on. But there are certain things that will differ, and I think that is extremely important that that happens sooner rather than later, so that the democratization of the usage of these technologies can go further and can go beyond what we think our values are, into all kinds of communities, geographies and domains.
And the second thing that I'm also super excited about is the application of all this AI to the different fields of science, because we have already seen examples of how AI can help change overnight, you know, challenges that have been in different fields for decades or centuries. And all of a sudden, you know, something that took five years to do and that takes you 5 minutes to do.
And as we open all those tools in and we let people just go crazy and do all kinds of experimentations with it, we're going to see an unprecedented number of disruptions and breakthroughs and new ways of seeing the world that is going to change who we are as a society. So, I think that's where we're going..
Deepa Seetharaman: I want to take a couple of questions. Does anyone in the crowd have a question? Yes, please.
Audience member 1: Thank you so much for a great panel I’m Landry Signé, senior fellow at the Brookings Institution and Executive Director at the Thunderbird School of Global Management in DC.
So there's a couple of dimensions that I would like you to elaborate on. So, with Gen AI, we have the pacing challenge, the incredible speed of development and also the coordination challenge, the multiplicity of actor and of usages which are also being made.
So we are here discussing AI governance. How do you think that various stakeholders could work together to address that pacing and coordination challenges, knowing that the public sector, the ability of the public sector to evolve with speed or with velocity, is pretty much different from the one of the private sector, let alone the civil society and a diversity of stakeholders, and what participation means, because we are also speaking about the imperative of including services like civil society. But what level of participation would also be construed as meaningful? Thank you.
Andrew Ng: It's tough. I think education is going to be key. [inaudible] is teaching Generative AI for everyone on Coursera, but I think helping everyone having a basic understanding will be important so that all the stakeholders participate. But I think Peter is going to say something.
Peter Hallinan: Well, I mean, it's a core question, and yet it's almost a question that's impossible to answer.
I think we do what we can. We engage in venues like this to discuss.
I think anybody in the field, whether you're a deployer, a builder, just a user, should be engaged with government as government considers regulation.
I think you should get out and try it. There's all sorts of organizations which exist today to facilitate conversations. You know, I'm speaking for AWS and Amazon, we fund lots of research for third parties. There's just so many different levers that you need to pull to engage people in these discussions.
And I don't know that there is any one lever and I don't know - there's so many different speeds at which different organizations, whether they're civil sector or private sector, government, move, you know, so how to steer it all? I don't know. The best you can do is contribute and engage.
Pilar Manchón: I'd like to add something as well. The way I think about this is like when you go to the beach and there is a lifeguard or there is no lifeguard, and if there is no lifeguard because regulation has not made it that far, then you're swimming on that beach of your own risk. So education is important to understand the risks that you are undertaking as a developer, as a user, as an organization, etcetera.
And regulation can at least be transparent in terms of is there a safeguard of some sort here or not? Are you swimming at your own risk in that particular area?
Peter Hallinan: Yes, well, you can also bring your family to the beach to watch out for you. I think that's how I would think of it.
Deepa Seetharaman: We have time for one more.
Audience member 2: Thanks a lot for a nice discussion. Daniel Dobos, Swisscom research director.
Andrew you mentioned nicely basically a comparison to electricity. And looking at the history of electricity, people have discussed a lot about all the risk that will bring us. What people will use it for, what people will misuse it for. Same with connectivity. What people will do if they now have information at any given moment.
So let me try to bring you a little bit in the future of, I don't know, five years, ten years, 20 years, and ask my question, will we sit here, I don't know, in five, ten, 20 years and discuss what the biggest risk is that we made a critical infrastructure. And the biggest risk is that we don't have it available anymore and our services cannot work without it anymore.
Andrew Ng: Oh, I see relatively little risk of us deploying AI, and then for some reason AI becoming unavailable unless some really horrible regulation shuts it down.
I feel like AI has risks and I think a lot of things that that Sara, Peter, Pilar described- one of the challenges of AI, it is different than previous technologies and I think something that Sara alluded to is that there's different boundary conditions to earlier technologies. So, we don't really know as well exactly what is going to work and what is not going to work, which is why the way we manage it and govern it is different.
But I can tell you that I work with a lot of product teams that are doing just fine in terms of testing extensively, deploying responsibly, having human in the loop until we're confident it is safe.
So, I think that a lot of fears for AI- is not that AI is harmless and will never do harm. But I think that a lot of fear is overblown.
Deepa Seetharaman: Anyone else agree with that? A lot of the fears are overblown.
Sara Hooker: I tend to agree in the sense that I always think the best way forward with risk is to allocate resources to the risks that we see every day.
I probably disagree with Andrew a little bit in the sense that I do think there are enormous risks that even happen with our models deployed right now and that we need to allocate.
But I do agree in a sense that we need more scrutiny for domain-sensitive areas. We need to allocate core fundamental research.
You know, one of the most promising things I've seen recently is that every country wants to start an AI safety institute, which I think is actually not a bad thing. I think it will funnel needed research and strengthen within government technical talent, which has been notoriously difficult for governments to attract in the West. And I think it's really important that you have technical people informing what are the realities of how these models succeed and are not brittle.
What I will say and where we agree is that, for me, there's been a lot of anxiety around long-term existential risk, which for me feels like in some ways a way that sometimes displaces conversations about the reality of how these models are deployed. And I think that I always ask, well, how do we measure progress along these axes of extension? And we don't have a measure of progress because there are many possible risks, and it's hard to quantify appropriately what is the actual probability or likelihood of an existing entity.
Andrew Ng: So, I actually spoke with quite a few people about the existential risk, and candidly, I don't get it. Many of them are very vague.
Many of them are vague and fluffy statements and I can't disprove that AI could wipe us out no more than I can disprove that radio waves emitted from Earth won't attract aliens to come and wipe us out. But it is so fluffy, I don't know what to do about that because I can't disprove a negative.
And I agree with Sara. This is a distraction from, you know, frankly, is there disinformation or misinformation on media or social media? Those are some short-term things where we could pass transparency and safety types of regulations and take actions that this other thing is a huge distraction from.
Oh, by the way, when I speak of US government officials, many of them kind of roll their eyes. Whereas interestingly, Europe is taking extinction risk more seriously. So, there is a divergence.
And one of the things I see is there is a faction in the US government that tragically, because of real or perceived adversaries potentially having access to open source, there is a faction in the US governments that would welcome a slowing down with of open source because of potential adversaries, bureaucracies, having access to it.
In contrast the European move to slow down open source, it seems, I don't understand really, frankly, I think that if we were to slow down open source Europe would be one of the places that is shut out because of the concentration in the US right now.
I feel like a lot of the, I think the theory, of slowing down open source in the US is flawed, I don't think it's a good idea. And then I think it is even more obviously not a good idea for Europe, because Europe would be one of the places that is shut out if some of these laws come to pass.
Deepa Seetharaman: I know I'm pushing it, but I am going to take one more question. Is anyone? Yes, please.
Audience member 3: Thank you very much.
Andrew, you recently tweeted about your son creating a mess with a chocolate cookie that he found in the pantry. But in that tweet, you brought out what I think is one of the most important points, which is it just might be easier to align AI with human values than aligning humans with human values.
And I think that is the biggest risk as far as, you know, coming from a country like India, that's one of the biggest risks that we see. Because even as we speak of AGI (Artificial General Intelligence) etcetera, you know, behind every smart algorithm, there is a smarter human being still. Any thoughts on how do you fix this problem?
Andrew Ng: Yes, this is a great point. So, what happened, one or two weeks ago my son got into the pantry, stole chocolate, made a mess. I was kind of slightly annoyed as a parent. And I tweeted out, at that moment I definitely felt like I had better tools to align AI with human values that I had tools to align my two-year-old son with human values.
And more seriously, I feel like the tools for aligning AI with human values, they are better than most people think. They're not perfect. But if you use ChatGPT or Bard and try to get it to give you detailed instructions for committing harm or committing a criminal act, it's actually really difficult to get AI to do that because it turns out that if we teach an AI, we want it to be honest, helpful and harmless. It really tries to do that.
And we can set the numbers in the AI to kind of very directly have it do that.
Whereas, how do you convince someone not to invade Ukraine? I don't know how to do that.
So, I actually sincerely find that we have better tools, more powerful, than the public probably appreciates for just telling the AI to do what we want. And then while it will fail to do so in some cases, it tends to get a lot of publicity, AI is probably already safer than most people think, which is not to say we should not also have maybe every country have a view on AI safety and keep on investing significantly in it.
Deepa Seetharaman: Thank you all.
Michele Mosca and Donna Dodson
December 20, 2024