There is, in some quarters, concern about high–level machine
intelligence and superintelligent AI coming up in a few decades,
bringing with it significant risks for humanity. In other quarters,
these issues are ignored or considered science fiction. We wanted to
clarify what the distribution of opinions actually is, what probability
the best experts currently assign to high–level machine intelligence
coming up within a particular time–frame, which risks they see with that
development, and how fast they see these developing. We thus designed a
brief questionnaire and distributed it to four groups of experts in
2012/2013. The median estimate of respondents was for a one in two
chance that high-level machine intelligence will be developed around
2040-2050, rising to a nine in ten chance by 2075. Experts expect that
systems will move on to superintelligence in less than 30 years
thereafter. They estimate the chance is about one in three that this
development turns out to be ‘bad’ or ‘extremely bad’ for humanity.
-
A funny thing happened earlier this week. SpaceX founder Elon Musk went on Joe Rogan's podcast, drank a little whisky and...
-
Today AI is successfully used to recognize faces and traffic signs, diagnose diseases and advance remedies, predict rainfall and con...
Highlight
Những điều thú vị khi dùng Trí tuệ nhân tạo của Viettel
Những người dùng Internet tại Việt Nam thường lấy “chị Google” ra để… giải trí. Khi “chị” đọc văn bản hay chỉ đường cho người tham gia gi...
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment