330
OpenAI's offices were sent thousands of paper clips in an elaborate prank to warn about an AI apocalypse
(www.businessinsider.com)
This is a most excellent place for technology news and articles.
It's difficult to draw a clear line between a simple neural network and a human brain when it comes to "intelligence". The rouge, paperclip-making "AI" seems be far closer to an intelligence, while flying autos or text prediction seems closer to mere hand-written code.
I think part of the wisdom in the warning is that any kind of "intelligence" (read: NOT specifically artificial general intelligence) is capable of running away with unforseen scenarios.
Hell, even normal ol' algorithms can have some pretty nasty edge cases that noone spots until it's running in production... Sure it's uncommon, but it's not exactly rare. (just look up the list of zero-day exploits over the years)