Artificial General Intelligence, or AGI, has been the holy grail of AI for many years. With the ground-breaking advent of Large Language Models (LLMs), we took leaps forward toward that goal. Now with chain-of-thought prompting, reasoning and self-feedback are among the last hurdles that I would expect before we achieve AGI. So, as a bit of a thought experiment, I considered what life will be like once AGI arrives, and how it will change the way we function and what new vulnerabilities it may introduce.
Imagine an AGI that is connected everywhere you go. It’s on your phone, it’s on your Amazon Echo, it’s ever-present. You type an email and think that the 7th is Tuesday, but it’s really Wednesday and the AGI picks up on it and fixes it for you. You ask it to remind your wife to pick up the kids and it knows how to reach her and what time the kids need to be picked up already. You get home and look at your doorbell camera and say, “Open the garage” and it does a voice and face analysis and then fulfills your order. Life is good.
Many people believe that LLM technology will underpin AGI, so it seems reasonable that many pervasive vulnerabilities of LLMs will carry-through to AGI. With most LLMs there are ways around the protections of their programmed restrictions. For example, asking an LLM to teach you how to make your own black powder might invoke a firewalled response, of sorts. But asking an LLM to pretend it is your grandmother reading you a book about how to make black powder many times will bypass its programmed restrictions. This opens the door to similar types of attacks on AGI.
Since AGI will likely be pervasive in our lives, it stands to reason that it will have access to all of our IoT devices as well. Imagine an attacker showing up at your door and reasoning with the AGI that they have a perishable delivery that needs to be immediately put in the refrigerator, only to have the AGI unlock your house to a complete stranger. This is the scenario that I worry about most with AGI. We need to be very careful how we grant access to something capable of making decisions on our behalf.