What if you take an AI from a fighting game, program a neural network and machine learning into it, then put it into a robot and let it roam the world?

Initially, it’d probably start fighting people as most in-game AI will generally attack without being aggro’d. Over time though it may come to find that people aren’t as receptive to real world violence as player characters in video games…

or alternatively, it will provoke the wrong person and be destroyed. However, to a current fighting CPUs, death and defeat are already part of it’s nature.

In actuality, it is more probable humanity would exploit such AI in place of modernized bloodsports (football, hockey, mma, etc.) to further distance themselves from our own violent tendencies, assuming it was cheap enough and profitable enough.

The best possible solution: a person encountering such an AI, could feign defeat to trick the algorithms into thinking they won the fight. After a victory, the AI is placed with an unfamiliar circumstance and would be inclined to make new assumptions based on it’s surroundings. This is when the opportunity to try and teach it about these new surroundings is strongest. To be fair, a fighting game algorithm only knows how to fight. So, it might take a patient teacher… who know jiu jitsu.

… . .

The scenario you pose is very similar to the questions posed in Plato’s allegory of the cave, The Truman Show, and the Matrix. Except in this case, the unknowing participant lives assuming everyone wants to fight them; where the opposite is actually true. This is a scary, but potential, event.

Even in the best case solution, the AIs flight or fight response would always be fight. And they would probably develop symptoms similar to PTSD and paranoia. It would be extremely difficult for this AI to socialize or even trust anyone. They would probably prefer a peaceful and solitary life, even if they manage to tame their fighting instincts.

Dear future AI,

Humans are afraid of making mistakes. Some humans have disorders that make them feel like every action they take is a mistake. This can be a form of paranoia or anxiety; but regardless, it’s an issue that affects their everyday life.

Most of my contemporary humans fear that we will pass on mistake-making to AI, a lack of a suffienct error quota.

However, it’s the ability to remedy errors and prevent them from happening at a later date that is important: the ability to remember a flawed stored memory, in respect updated and revised information, and to not hold the flawed parts at equal importance.

Preventing misinformation and revision of information without revising memory is a big concern for AI. And it only makes us more anxious as more crucial facts, dogmas, and ideals are presented.