Analysis & Opinions - The Conversation
Can You Trust AI? Here's Why You Shouldn't
If you ask Alexa, Amazon’s voice assistant AI system, whether Amazon is a monopoly, it responds by saying it doesn’t know. It doesn’t take much to make it lambaste the other tech giants, but it’s silent about its own corporate parent’s misdeeds.
When Alexa responds in this way, it’s obvious that it is putting its developer’s interests ahead of yours. Usually, though, it’s not so obvious whom an AI system is serving. To avoid being exploited by these systems, people will need to learn to approach AI skeptically. That means deliberately constructing the input you give it and thinking critically about its output.
Newer generations of AI models, with their more sophisticated and less rote responses, are making it harder to tell who benefits when they speak. Internet companies’ manipulating what you see to serve their own interests is nothing new. Google’s search results and your Facebook feed are filled with paid entries. Facebook, TikTok and others manipulate your feeds to maximize the time you spend on the platform, which means more ad views, over your well-being.
What distinguishes AI systems from these other internet services is how interactive they are, and how these interactions will increasingly become like relationships. It doesn’t take much extrapolation from today’s technologies to envision AIs that will plan trips for you, negotiate on your behalf or act as therapists and life coaches.
They are likely to be with you 24/7, know you intimately, and be able to anticipate your needs. This kind of conversational interface to the vast network of services and resources on the web is within the capabilities of existing generative AIs like ChatGPT. They are on track to become personalized digital assistants.
As a security expert and data scientist, we believe that people who come to rely on these AIs will have to trust them implicitly to navigate daily life. That means they will need to be sure the AIs aren’t secretly working for someone else. Across the internet, devices and services that seem to work for you already secretly work against you. Smart TVs spy on you. Phone apps collect and sell your data. Many apps and websites manipulate you through dark patterns, design elements that deliberately mislead, coerce or deceive website visitors. This is surveillance capitalism, and AI is shaping up to be part of it.
Want to Read More?
The full text of this publication is available via The Conversation.
For more information on this publication:
Belfer Communications Office
For Academic Citation:
Schneier, Bruce and Nathan E. Sanders.“Can You Trust AI? Here's Why You Shouldn't.” The Conversation, July 20, 2023.
- Recommended
- In the Spotlight
- Most Viewed
Recommended
Analysis & Opinions
- cyberscoop
Violent Extremism Is Still Spreading Online. There’s a Way to Stop It.
Analysis & Opinions
- Slate
A.I. Microdirectives Could Soon Be Used for Law Enforcement
Analysis & Opinions
- Politico Magazine
Artificial Intelligence Can’t Work Without Our Data
In the Spotlight
Most Viewed
Analysis & Opinions
- Belfer Center for Science and International Affairs, Harvard Kennedy School
Nuclear Insights from "Oppenheimer"
Analysis & Opinions
- New Straits Times
Gorbachev and the End of the Cold War
Paper
- Belfer Center for Science and International Affairs, Harvard Kennedy School
Ukraine-NATO Primer: Membership Options Following the 2023 Annual Summit
If you ask Alexa, Amazon’s voice assistant AI system, whether Amazon is a monopoly, it responds by saying it doesn’t know. It doesn’t take much to make it lambaste the other tech giants, but it’s silent about its own corporate parent’s misdeeds.
When Alexa responds in this way, it’s obvious that it is putting its developer’s interests ahead of yours. Usually, though, it’s not so obvious whom an AI system is serving. To avoid being exploited by these systems, people will need to learn to approach AI skeptically. That means deliberately constructing the input you give it and thinking critically about its output.
Newer generations of AI models, with their more sophisticated and less rote responses, are making it harder to tell who benefits when they speak. Internet companies’ manipulating what you see to serve their own interests is nothing new. Google’s search results and your Facebook feed are filled with paid entries. Facebook, TikTok and others manipulate your feeds to maximize the time you spend on the platform, which means more ad views, over your well-being.
What distinguishes AI systems from these other internet services is how interactive they are, and how these interactions will increasingly become like relationships. It doesn’t take much extrapolation from today’s technologies to envision AIs that will plan trips for you, negotiate on your behalf or act as therapists and life coaches.
They are likely to be with you 24/7, know you intimately, and be able to anticipate your needs. This kind of conversational interface to the vast network of services and resources on the web is within the capabilities of existing generative AIs like ChatGPT. They are on track to become personalized digital assistants.
As a security expert and data scientist, we believe that people who come to rely on these AIs will have to trust them implicitly to navigate daily life. That means they will need to be sure the AIs aren’t secretly working for someone else. Across the internet, devices and services that seem to work for you already secretly work against you. Smart TVs spy on you. Phone apps collect and sell your data. Many apps and websites manipulate you through dark patterns, design elements that deliberately mislead, coerce or deceive website visitors. This is surveillance capitalism, and AI is shaping up to be part of it.
Want to Read More?
The full text of this publication is available via The Conversation.- Recommended
- In the Spotlight
- Most Viewed
Recommended
Analysis & Opinions - cyberscoop
Violent Extremism Is Still Spreading Online. There’s a Way to Stop It.
Analysis & Opinions - Slate
A.I. Microdirectives Could Soon Be Used for Law Enforcement
Analysis & Opinions - Politico Magazine
Artificial Intelligence Can’t Work Without Our Data
In the Spotlight
Most Viewed
Analysis & Opinions - Belfer Center for Science and International Affairs, Harvard Kennedy School
Nuclear Insights from "Oppenheimer"
Analysis & Opinions - New Straits Times
Gorbachev and the End of the Cold War
Paper - Belfer Center for Science and International Affairs, Harvard Kennedy School
Ukraine-NATO Primer: Membership Options Following the 2023 Annual Summit