Adverse uses of AI Part 26

1 year ago
60

AI is incredibly powerful and the vast majority of users of AI are making people, companies, countries, society better off. But there are a few adverse uses of AI as well. Let's take a look at some of them and discuss what we can do about them. AI technology has been used to create deep fakes and that means to synthesize video of people doing things that they never actually did. The website BuzzFeed created a video of former US President Barack Obama saying things that he never did. BuzzFeed was transparent about it and when they publish the video, it was really obvious because he told everyone that this is a fake video. But if this type of technology is used to target an individual and make others think they said or did things they never actually did, then these individuals could be harmed and left to defend themselves against fake video evidence of something they never actually did. Similar to the war of spam versus anti-spam, there is AI technology today for detecting if a video is a deep fake. But in today's world of social media, where a fake could spread around the world faster than the truth can catch up, many people are concerned about the potential of deep fakes to harm individuals. There's also a risk of AI technology being used to undermine democracy and privacy. For example, many governments around the world are trying to improve their citizens' lives, and have a lot of respect for the government leaders that are uplifting their citizens. But there are also some oppressive regimes that are not doing the right things by their citizens, that may seek to use this type of technology to carry out oppressive surveillance of their citizens. While governments have illegitimate need to improve public safety and reduce crime, there are also ways of using AI that feel more oppressive than uplifting of its own citizens. Closely related to this, is the rise of fake comments that AI can generate. Using AI technology is now possible to generate fake comments. Either on the commercial side, fake comments of products, or in political discourse, fake comments about political matters in the public discourse, and to generate fake comments much more efficiently than if you only had humans writing them. So, detecting such fake comments and weeding them out, is an important technology for maintaining trust in comments that we might read online as well. Similar to the battles of spam versus anti-spam and fraud verses anti-fraud, I think that for all of these issues, there may be a competition on both sides for quite some time to come. Similar to the battles of span versus anti-spam, fraud versus anti-fraud, I'm optimistic about how these battles will play out. Because if you take spam filter as an example, there are a lot more people that are motivated to make sure spam filters do work, that anti-spam does work. Then there are the smaller number of spammers trying to get this spam into your inbox.

Loading comments...