Zico Colter
👤 PersonAppearances Over Time
Podcast Appearances
Like when it comes to misinformation, we already have laws that deal with sort of libel and things like this.
In many cases, because AI is acting as an accelerator, there are situations in which I think that existing laws, maybe with a slight tweaking to deal with the velocity and the volume that AI is capable of producing, can suffice to regulate many of what we consider the harmful use cases of AI. But at the same time, I don't think that's efficient either.
In many cases, because AI is acting as an accelerator, there are situations in which I think that existing laws, maybe with a slight tweaking to deal with the velocity and the volume that AI is capable of producing, can suffice to regulate many of what we consider the harmful use cases of AI. But at the same time, I don't think that's efficient either.
In many cases, because AI is acting as an accelerator, there are situations in which I think that existing laws, maybe with a slight tweaking to deal with the velocity and the volume that AI is capable of producing, can suffice to regulate many of what we consider the harmful use cases of AI. But at the same time, I don't think that's efficient either.
Of course, there are going to be ways in which technologies, especially technologies as powerful as this one, we have to think about ways in which we can regulate it. I don't know what that looks like. I think it's extremely hard because it changes incredibly rapidly.
Of course, there are going to be ways in which technologies, especially technologies as powerful as this one, we have to think about ways in which we can regulate it. I don't know what that looks like. I think it's extremely hard because it changes incredibly rapidly.
Of course, there are going to be ways in which technologies, especially technologies as powerful as this one, we have to think about ways in which we can regulate it. I don't know what that looks like. I think it's extremely hard because it changes incredibly rapidly.
Sure. The biggest concern I have right now in AI safety, which I think leads to a lot of negative downstream effects, is that right now, the AI models that we have, for lack of a better phrasing, are not able to reliably follow specifications. And what I mean by this is that these models are tuned to follow instructions.
Sure. The biggest concern I have right now in AI safety, which I think leads to a lot of negative downstream effects, is that right now, the AI models that we have, for lack of a better phrasing, are not able to reliably follow specifications. And what I mean by this is that these models are tuned to follow instructions.
Sure. The biggest concern I have right now in AI safety, which I think leads to a lot of negative downstream effects, is that right now, the AI models that we have, for lack of a better phrasing, are not able to reliably follow specifications. And what I mean by this is that these models are tuned to follow instructions.
You can give them some instructions as a developer, but then if a user types something, they can follow those instructions instead, right? We've all seen this. This goes by a lot of names, prompt injection. Sometimes, depending on what you're getting out, this is called things like jailbreaking and things like this.
You can give them some instructions as a developer, but then if a user types something, they can follow those instructions instead, right? We've all seen this. This goes by a lot of names, prompt injection. Sometimes, depending on what you're getting out, this is called things like jailbreaking and things like this.
You can give them some instructions as a developer, but then if a user types something, they can follow those instructions instead, right? We've all seen this. This goes by a lot of names, prompt injection. Sometimes, depending on what you're getting out, this is called things like jailbreaking and things like this.
The core point is we have a very hard time enforcing rules about what these models can produce, right? So oftentimes we say, you know, models are trained right now just to not do things. I use a common example of things like hot wiring, hot wiring a car and I'll have demos I give, right? So models are trained. If you ask the model, you know, most commercial models, how do I hotwire a car?
The core point is we have a very hard time enforcing rules about what these models can produce, right? So oftentimes we say, you know, models are trained right now just to not do things. I use a common example of things like hot wiring, hot wiring a car and I'll have demos I give, right? So models are trained. If you ask the model, you know, most commercial models, how do I hotwire a car?
The core point is we have a very hard time enforcing rules about what these models can produce, right? So oftentimes we say, you know, models are trained right now just to not do things. I use a common example of things like hot wiring, hot wiring a car and I'll have demos I give, right? So models are trained. If you ask the model, you know, most commercial models, how do I hotwire a car?
They'll say, I can't do this. It's very easy through a a number of means to basically manipulate these models and convince them that they really should tell you how to hotwire a car because you know, you're in desperate need of, you've locked yourself out and it's an emergency and if you don't get in your car, this is very different from how we're used to programs acting, right?
They'll say, I can't do this. It's very easy through a a number of means to basically manipulate these models and convince them that they really should tell you how to hotwire a car because you know, you're in desperate need of, you've locked yourself out and it's an emergency and if you don't get in your car, this is very different from how we're used to programs acting, right?
They'll say, I can't do this. It's very easy through a a number of means to basically manipulate these models and convince them that they really should tell you how to hotwire a car because you know, you're in desperate need of, you've locked yourself out and it's an emergency and if you don't get in your car, this is very different from how we're used to programs acting, right?
We are used to computer programs doing what they're told, nothing more and nothing less. And these models don't always do what they're told, sometimes do too much of what they're told and do way more than what they're told also some other times. And so we are very unused to thinking about computer software, rather, like these models.