Comment We lost control (Score 1) 36
Train an AI to like or dislike a random item or category, let's say sharks. Then get it to make a training data set for another AI about an unrelated topic, such as teaching fractions to sixth graders. Ensure there is no mention of sharks, or any swimming or animals in the mathematical examples in the training set. Ask the resulting AI about sharks, and it will mysteriously have adopted the other AI's stance towards sharks.
There is no need to invoke malicious intent. AIs absorb cultural prejudices and implicit stances from their training set's culture, and we haven't figured out how, so we can't control, predict, or even mediate it. It's a big mirror of all humanity's flaws, right in our faces.