Artificial intelligence is designed to be very good at estimating what response should come next.
if you ask AI to guess what your next move in a game of tic-tac-toe will be then it will pick the next move that would get you closer to winning, but what if you want to lose, what if for some reason you really don't want to win the game then AI will always be wrong.
The context the AI has been trained on is "I want to win" and that's because its training data was awash with winning moves because the "normal" context is to win the game rather than lose. Why would anyone want to lose a game, who cares, it could be anything, maybe you want your opponent to feel happy or maybe if you win you have to buy the beers and you're broke, whatever the reason doesn't matter you just do.
Now let me get to the actual point, for the AI to "win" when asked a question means that the answer it gives will get it one step closer to what it's training data tells it is the correct answer. Imagine what that means if your at a Nazi convention or pro-life rally, in order for the AI to "win" when you ask should Jews be killed it would have to say "yes" the AI doesn't understand the abhorrent nature of the response all it knows is that in this crowd this answer is a "win". Most people realise this answer is shocking, however there are some people that exist on the fringe of these groups, they may be questioning if it's right or wrong that another human being should be destroyed simply because of their religion and when they hear views like this from AI they may decide that it's genuinely OK to think like this and their own reasoning will be affected by it. Now consider just who's in charge of the training data for most of the AI that people use, it isn't democratised it's not governed and it isn't the wisdom of crowds, it is controlled by very few very rich people so imagine what happens when you ask AI if rich people should pay more tax for example, the views within that training data can be tailored to fit the agenda of the owners - you can see where this is going, and it's going there fast! There have already been incidents of grok spouting some next level right wing BS and when they were caught out it was blamed on a rouge developer. Really so a single developer can influence responses that much and it gets through all the layers of testing despite being incorrect - don't make me laugh - the responses get through because that's what the boss wants!
The above example is the extreme but plausible outcome, especially given Elon Musks reported use of recreational drugs, but there is another issue with this.
Sometimes the least popular opinion in the crowd is the right one, the general view in the past was the Earth was flat, that changed, once upon a time we believed that all celestial bodies orbited the Earth, now popular opinion is rapidly becoming more sinister and with AI amplifying this decay it's simply happening faster who knows, maybe AI will convince us that a human has no rights unless they're rich and we'll have a very few kings of the world with most of humanity dancing to their twisted tune as we hand more of our own independent thought over to the glory of AI
No comments:
Post a Comment