Zico Colter
👤 PersonAppearances Over Time
Podcast Appearances
We are used to computer programs doing what they're told, nothing more and nothing less. And these models don't always do what they're told, sometimes do too much of what they're told and do way more than what they're told also some other times. And so we are very unused to thinking about computer software, rather, like these models.
We are used to computer programs doing what they're told, nothing more and nothing less. And these models don't always do what they're told, sometimes do too much of what they're told and do way more than what they're told also some other times. And so we are very unused to thinking about computer software, rather, like these models.
And what that means is, and to be honest, I don't really care if models tell me how to hotwire a car. I just don't. It doesn't matter, right? There's instructions on the internet on how to hotwire a car. They're not really revealing anything that sensitive. However...
And what that means is, and to be honest, I don't really care if models tell me how to hotwire a car. I just don't. It doesn't matter, right? There's instructions on the internet on how to hotwire a car. They're not really revealing anything that sensitive. However...
And what that means is, and to be honest, I don't really care if models tell me how to hotwire a car. I just don't. It doesn't matter, right? There's instructions on the internet on how to hotwire a car. They're not really revealing anything that sensitive. However...
As we start to integrate these models into larger systems, as we start to have agents that go out and do things, that parse the internet and go out and do things, if all of a sudden they're running their model, parsing untrusted third-party data, that data can essentially gain control of those models. To a certain extent.
As we start to integrate these models into larger systems, as we start to have agents that go out and do things, that parse the internet and go out and do things, if all of a sudden they're running their model, parsing untrusted third-party data, that data can essentially gain control of those models. To a certain extent.
As we start to integrate these models into larger systems, as we start to have agents that go out and do things, that parse the internet and go out and do things, if all of a sudden they're running their model, parsing untrusted third-party data, that data can essentially gain control of those models. To a certain extent.
And this is from a sort of cybersecurity standpoint, not the normal cybersecurity, but sort of from a concept of cybersecurity. This is sort of like these models have a buffer overflow in all of them that we know about. And most importantly, that we don't know how to patch and fix. We don't know how to fix this yet with models. To be clear, I think we can make a lot of progress.
And this is from a sort of cybersecurity standpoint, not the normal cybersecurity, but sort of from a concept of cybersecurity. This is sort of like these models have a buffer overflow in all of them that we know about. And most importantly, that we don't know how to patch and fix. We don't know how to fix this yet with models. To be clear, I think we can make a lot of progress.
And this is from a sort of cybersecurity standpoint, not the normal cybersecurity, but sort of from a concept of cybersecurity. This is sort of like these models have a buffer overflow in all of them that we know about. And most importantly, that we don't know how to patch and fix. We don't know how to fix this yet with models. To be clear, I think we can make a lot of progress.
We are making progress. But this is a real concern about models right now. And the negative effects in a domain like a chatbot are maybe not that concerning. But as you start having much more complex LLM systems, this starts becoming much more concerning.
We are making progress. But this is a real concern about models right now. And the negative effects in a domain like a chatbot are maybe not that concerning. But as you start having much more complex LLM systems, this starts becoming much more concerning.
We are making progress. But this is a real concern about models right now. And the negative effects in a domain like a chatbot are maybe not that concerning. But as you start having much more complex LLM systems, this starts becoming much more concerning.
What I will also say is that, and this is maybe the reason why I placed this concern first, is that I think this fact is something we need to figure out, or kind of all the other downstream concerns that we have about these models get much, much worse. So let me just take an example. Oftentimes, I'm touching a lot of points here I know too, but I think I'll wrap it up soon.
What I will also say is that, and this is maybe the reason why I placed this concern first, is that I think this fact is something we need to figure out, or kind of all the other downstream concerns that we have about these models get much, much worse. So let me just take an example. Oftentimes, I'm touching a lot of points here I know too, but I think I'll wrap it up soon.
What I will also say is that, and this is maybe the reason why I placed this concern first, is that I think this fact is something we need to figure out, or kind of all the other downstream concerns that we have about these models get much, much worse. So let me just take an example. Oftentimes, I'm touching a lot of points here I know too, but I think I'll wrap it up soon.
Oftentimes, people talk about risks like bio risks or cyber attack risks and stuff like this. I'm actually, to your point, I'm very concerned about cyber risks in particular. I think this is essentially already solved in many cases by these models. They can already solve and analyze code to find vulnerabilities. This is extremely concerning.
Oftentimes, people talk about risks like bio risks or cyber attack risks and stuff like this. I'm actually, to your point, I'm very concerned about cyber risks in particular. I think this is essentially already solved in many cases by these models. They can already solve and analyze code to find vulnerabilities. This is extremely concerning.
Oftentimes, people talk about risks like bio risks or cyber attack risks and stuff like this. I'm actually, to your point, I'm very concerned about cyber risks in particular. I think this is essentially already solved in many cases by these models. They can already solve and analyze code to find vulnerabilities. This is extremely concerning.