{"id":5731,"date":"2024-11-03T01:31:50","date_gmt":"2024-11-03T01:31:50","guid":{"rendered":"https:\/\/hectorkott.com\/?p=5731"},"modified":"2024-11-03T01:31:50","modified_gmt":"2024-11-03T01:31:50","slug":"what-is-ai-superintelligence-could-it-destroy-humanity-and-is-it-really-almost-here","status":"publish","type":"post","link":"https:\/\/hectorkott.com\/?p=5731","title":{"rendered":"What Is AI Superintelligence? Could It Destroy Humanity? And Is It Really Almost Here?"},"content":{"rendered":"<p><\/p>\n<div>\n<p>In 2014, the British philosopher Nick Bostrom published a book about the future of artificial intelligence with the ominous title <em><a href=\"https:\/\/en.wikipedia.org\/wiki\/Superintelligence:_Paths,_Dangers,_Strategies\">Superintelligence: Paths, Dangers, Strategies<\/a><\/em>. It proved highly influential in promoting the idea that advanced AI systems\u2014\u201csuperintelligences\u201d more capable than humans\u2014might one day take over the world and destroy humanity.<\/p>\n<p>A decade later, OpenAI boss Sam Altman says superintelligence may only be \u201c<a href=\"https:\/\/ia.samaltman.com\">a few thousand days<\/a>\u201d away. A year ago, Altman\u2019s OpenAI cofounder Ilya Sutskever set up a team within the company to focus on \u201c<a href=\"https:\/\/openai.com\/index\/introducing-superalignment\/\">safe superintelligence<\/a>,\u201d but he and his team have now raised a billion dollars to create <a href=\"https:\/\/www.reuters.com\/technology\/artificial-intelligence\/ilya-sutskever-how-ai-will-change-his-new-startup-safe-superintelligence-2024-09-05\/\">a startup of their own<\/a> to pursue this goal.<\/p>\n<p>What exactly are they talking about? Broadly speaking, superintelligence is <a href=\"https:\/\/arxiv.org\/abs\/0712.3329\">anything more intelligent than humans<\/a>. But unpacking what that might mean in practice can get a bit tricky.<\/p>\n<h2>Different Kinds of AI<\/h2>\n<p>In my view, the most useful way to think about <a href=\"https:\/\/proceedings.mlr.press\/v235\/morris24b.html\">different levels and kinds of intelligence in AI<\/a> was developed by US computer scientist Meredith Ringel Morris and her colleagues at Google.<\/p>\n<p>Their framework lists six levels of AI performance: no AI, emerging, competent, expert, virtuoso, and superhuman. It also makes an important distinction between narrow systems, which can carry out a small range of tasks, and more general systems.<\/p>\n<p>A narrow, no-AI system is something like a calculator. It carries out various mathematical tasks according to a set of explicitly programmed rules.<\/p>\n<p>There are already plenty of very successful narrow AI systems. Morris gives the <a href=\"https:\/\/theconversation.com\/twenty-years-on-from-deep-blue-vs-kasparov-how-a-chess-match-started-the-big-data-revolution-76882\">Deep Blue chess program<\/a> that famously defeated world champion Garry Kasparov way back in 1997 as an example of a virtuoso-level narrow AI system.<\/p>\n<figure class=\"wp-caption\"><img loading=\"lazy\" decoding=\"async\" sizes=\"auto, (max-width: 1510px) 100vw, 1510px\" src=\"https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM.png\" alt=\"\" width=\"1510\" srcset=\"https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM.png 1510w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-274x300.png 274w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-821x900.png 821w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-768x842.png 768w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-1401x1536.png 1401w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-696x763.png 696w, https:\/\/singularityhub.com\/uploads\/2024\/11\/Screenshot-2024-11-01-at-5.28.55-PM-1068x1171.png 1068w\" class=\"wp-image-159380\" height=\"1656\"><figcaption class=\"wp-caption-text\">Table: The Conversation * Source: Adapted from <a href=\"https:\/\/proceedings.mlr.press\/v235\/morris24b.html\">Morris et al.<\/a> * Created with <a href=\"https:\/\/www.datawrapper.de\/_\/psufz\">Datawrapper<\/a><\/figcaption><\/figure>\n<p>Some narrow systems even have superhuman capabilities. One example is <a href=\"https:\/\/www.nature.com\/articles\/s41586-021-03819-2\">AlphaFold<\/a>, which uses machine learning to predict the structure of protein molecules, and whose creators <a href=\"https:\/\/theconversation.com\/machine-learning-cracked-the-protein-folding-problem-and-won-the-2024-nobel-prize-in-chemistry-240937\">won the Nobel Prize in Chemistry<\/a> this year.What about general systems? This is software that can tackle a much wider range of tasks, including things like learning new skills.<\/p>\n<p>A general no-AI system might be something like <a href=\"https:\/\/www.mturk.com\">Amazon\u2019s Mechanical Turk<\/a>: It can do a wide range of things, but it does them by asking real people.<\/p>\n<p>Overall, general AI systems are far less advanced than their narrow cousins. According to Morris, the state-of-the-art language models behind chatbots such as ChatGPT are general AI\u2014but they are so far at the \u201cemerging\u201d level (meaning they are \u201cequal to or somewhat better than an unskilled human\u201d), and yet to reach \u201ccompetent\u201d (as good as 50 percent of skilled adults).<\/p>\n<p>So by this reckoning, we are still some distance from general superintelligence.<\/p>\n<h2>How Intelligent Is AI Right Now?<\/h2>\n<p>As Morris points out, precisely determining where any given system sits would depend on having reliable tests or benchmarks.<\/p>\n<p>Depending on our benchmarks, an image-generating system such as <a href=\"https:\/\/openai.com\/index\/dall-e-3\/\">DALL-E<\/a> might be at virtuoso level (because it can produce images 99 percent of humans could not draw or paint), or it might be emerging (because it produces errors no human would, such as mutant hands and impossible objects).<\/p>\n<p>There is significant debate even about the capabilities of current systems. One notable 2023 paper <a href=\"https:\/\/arxiv.org\/abs\/2303.12712\">argued<\/a> GPT-4 showed \u201csparks of artificial general intelligence.\u201d<\/p>\n<p>OpenAI says its latest language model, <a href=\"https:\/\/openai.com\/index\/learning-to-reason-with-llms\/\">o1<\/a>, can \u201cperform complex reasoning\u201d and \u201crivals the performance of human experts\u201d on many benchmarks.<\/p>\n<p>However, <a href=\"https:\/\/arxiv.org\/abs\/2410.05229\">a recent paper from Apple researchers<\/a> found o1 and many other language models have significant trouble solving genuine mathematical reasoning problems. Their experiments show the outputs of these models seem to resemble sophisticated pattern-matching rather than true advanced reasoning. This indicates superintelligence is not as imminent as many have suggested.<\/p>\n<h2>Will AI Keep Getting Smarter?<\/h2>\n<p>Some people think the rapid pace of AI progress over the past few years will continue or even accelerate. Tech companies are investing <a href=\"https:\/\/www.bloomberg.com\/professional\/insights\/technology\/big-tech-2025-capex-may-hit-200-billion-as-gen-ai-demand-booms\/\">hundreds of billions of dollars<\/a> in AI hardware and capabilities, so this doesn\u2019t seem impossible.<\/p>\n<p>If this happens, we may indeed see general superintelligence within the \u201cfew thousand days\u201d proposed by Sam Altman (that\u2019s a decade or so in less sci-fi terms). Sutskever and his team mentioned a similar timeframe in their <a href=\"https:\/\/openai.com\/index\/introducing-superalignment\/\">superalignment article<\/a>.<\/p>\n<p>Many recent successes in AI have come from the application of a technique called \u201cdeep learning,\u201d which, in simplistic terms, finds associative patterns in gigantic collections of data. Indeed, <a href=\"https:\/\/www.nobelprize.org\/prizes\/physics\/2024\/press-release\/\">this year\u2019s Nobel Prize in Physics<\/a> has been awarded to John Hopfield and also the \u201c<a href=\"https:\/\/www.forbes.com\/sites\/samshead\/2019\/03\/27\/the-3-godfathers-of-ai-have-won-the-prestigious-1m-turing-prize\/\">Godfather of AI<\/a>\u201d Geoffrey Hinton, for their invention of the Hopfield network and Boltzmann machine, which are the foundation of many powerful deep learning models used today.<\/p>\n<p>General systems such as ChatGPT have relied on data generated by humans, much of it in the form of text from books and websites. Improvements in their capabilities have largely come from increasing the scale of the systems and the amount of data on which they are trained.<\/p>\n<p>However, there <a href=\"https:\/\/openreview.net\/pdf?id=ViZcgDQjyG\">may not be enough human-generated data<\/a> to take this process much further (although efforts to use data more efficiently, generate synthetic data, and improve transfer of skills between different domains may bring improvements). Even if there were enough data, some researchers say language models such as ChatGPT are <a href=\"https:\/\/www.ft.com\/content\/23fab126-f1d3-4add-a457-207a25730ad9\">fundamentally incapable<\/a> of reaching what Morris would call general competence.<\/p>\n<p>One recent paper has suggested an essential feature of superintelligence would be <a href=\"https:\/\/openreview.net\/pdf?id=Bc4vZ2CX7E\">open-endedness<\/a>, at least from a human perspective. It would need to be able to continuously generate outputs that a human observer would regard as novel and be able to learn from.<\/p>\n<p>Existing foundation models are not trained in an open-ended way, and existing open-ended systems are quite narrow. This paper also highlights how either novelty or learnability alone is not enough. A new type of open-ended foundation model is needed to achieve superintelligence.<\/p>\n<h2>What Are the Risks?<\/h2>\n<p>So what does all this mean for the risks of AI? In the short term, at least, we don\u2019t need to worry about superintelligent AI taking over the world.<\/p>\n<p>But that\u2019s not to say AI doesn\u2019t present risks. Again, Morris and co have thought this through: As AI systems gain great capability, they may also gain greater autonomy. Different levels of capability and autonomy present different risks.<\/p>\n<p>For example, when AI systems have little autonomy and people use them as a kind of consultant\u2014when we ask ChatGPT to summarize documents, say, or let the YouTube algorithm shape our viewing habits\u2014we might face a risk of over-trusting or <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/uploads\/prod\/2022\/06\/Aether-Overreliance-on-AI-Review-Final-6.21.22.pdf\">over-relying on them<\/a>.<\/p>\n<p>In the meantime, Morris points out other risks to watch out for as AI systems become more capable, ranging from people forming parasocial relationships with AI systems to mass job displacement and society-wide ennui.<\/p>\n<h2>What\u2019s Next?<\/h2>\n<p>Let\u2019s suppose we do one day have superintelligent, fully autonomous AI agents. Will we then face the risk they could concentrate power or act against human interests?<\/p>\n<p>Not necessarily. Autonomy and control <a href=\"https:\/\/arxiv.org\/abs\/2002.04087v1\">can go hand in hand<\/a>. A system can be highly automated, yet provide a high level of human control.<\/p>\n<p>Like many in the AI research community, I believe <em>safe superintelligence<\/em> is feasible. However, building it will be a complex and multidisciplinary task, and researchers will have to tread unbeaten paths to get there.<\/p>\n<p><em>This article is republished from <a href=\"https:\/\/theconversation.com\">The Conversation<\/a> under a Creative Commons license. Read the <a href=\"https:\/\/theconversation.com\/what-is-ai-superintelligence-could-it-destroy-humanity-and-is-it-really-almost-here-240682\">original article<\/a>.<\/em><\/p>\n<\/div>\n<p>\nFuente: https:\/\/ift.tt\/u9NvLer<br \/>\nPublicado: November 1, 2024 at 07:34PM<\/p>\n","protected":false},"excerpt":{"rendered":"<p>In 2014, the British philosopher Nick Bostrom published a book about the future of artificial intelligence with the ominous title Superintelligence: Paths, Dangers, Strategies. It proved highly influential in promoting the idea that advanced AI systems\u2014\u201csuperintelligences\u201d more capable than humans\u2014might one day take over the world and destroy humanity. A decade later, OpenAI boss Sam&#8230;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_bbp_topic_count":0,"_bbp_reply_count":0,"_bbp_total_topic_count":0,"_bbp_total_reply_count":0,"_bbp_voice_count":0,"_bbp_anonymous_reply_count":0,"_bbp_topic_count_hidden":0,"_bbp_reply_count_hidden":0,"_bbp_forum_subforum_count":0,"footnotes":""},"categories":[100],"tags":[75,76,77],"class_list":["post-5731","post","type-post","status-publish","format-standard","hentry","category-todos-los-feeds","tag-news","tag-noticias","tag-viral"],"_links":{"self":[{"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/posts\/5731","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/hectorkott.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=5731"}],"version-history":[{"count":1,"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/posts\/5731\/revisions"}],"predecessor-version":[{"id":5732,"href":"https:\/\/hectorkott.com\/index.php?rest_route=\/wp\/v2\/posts\/5731\/revisions\/5732"}],"wp:attachment":[{"href":"https:\/\/hectorkott.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=5731"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hectorkott.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=5731"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hectorkott.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=5731"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}