{"id":70167,"date":"2020-07-09T17:40:29","date_gmt":"2020-07-09T17:40:29","guid":{"rendered":"https:\/\/80000hours.org\/?post_type=podcast&#038;p=70167"},"modified":"2024-04-14T01:09:14","modified_gmt":"2024-04-14T01:09:14","slug":"ben-garfinkel-classic-ai-risk-arguments","status":"publish","type":"podcast","link":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/","title":{"rendered":"Ben Garfinkel on scrutinising classic AI risk&nbsp;arguments"},"author":405,"featured_media":70169,"template":"","categories":[1182,1181,1133,368,1195,470,1183,330,1217],"class_list":["post-70167","podcast","type-podcast","status-publish","has-post-thumbnail","hentry","category-technical-ai-safety-research","category-artificial-intelligence","category-effective-altruism","category-existential-risk","category-future-generations-longtermism","category-future-of-humanity-institute","category-long-term-ai-policy","category-moral-philosophy","category-reasoning-well"],"acf":{"acf_audio_url":"https:\/\/media.transistor.fm\/059e0ce2\/6a3734b1.mp3","acf_highlights_audio_url":"","acf_audio_player_thumbnail_image_url":"https:\/\/images.transistor.fm\/file\/transistor\/images\/episode\/1324815\/thumb_1683544641-artwork.jpg"},"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v23.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Ben Garfinkel on scrutinising classic AI risk arguments - 80,000 Hours<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"How sure are we about this AI stuff?\" \/>\n<meta property=\"og:description\" content=\"Ben Garfinkel, Research Fellow at Oxford&#039;s Future of Humanity Institute, thinks classic AI risk arguments haven&#039;t been subject to sufficient scrutiny.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/\" \/>\n<meta property=\"og:site_name\" content=\"80,000 Hours\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/80000Hours\" \/>\n<meta property=\"article:modified_time\" content=\"2024-04-14T01:09:14+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/brain-box_2.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1024\" \/>\n\t<meta property=\"og:image:height\" content=\"537\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:title\" content=\"How sure are we about this AI stuff?\" \/>\n<meta name=\"twitter:description\" content=\"Ben Garfinkel, Research Fellow at Oxford&#039;s Future of Humanity Institute, thinks classic AI risk arguments haven&#039;t been subject to sufficient scrutiny.\" \/>\n<meta name=\"twitter:image\" content=\"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/brain-box_2.jpg\" \/>\n<meta name=\"twitter:site\" content=\"@80000hours\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/\",\"url\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/\",\"name\":\"Ben Garfinkel on scrutinising classic AI risk arguments - 80,000 Hours\",\"isPartOf\":{\"@id\":\"https:\/\/80000hours.org\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg\",\"datePublished\":\"2020-07-09T17:40:29+00:00\",\"dateModified\":\"2024-04-14T01:09:14+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage\",\"url\":\"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg\",\"contentUrl\":\"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg\",\"width\":2000,\"height\":1333,\"caption\":\"Output from Google's [DeepDream](https:\/\/en.wikipedia.org\/wiki\/DeepDream) AI.\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/80000hours.org\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Podcast\",\"item\":\"https:\/\/80000hours.org\/podcast\/episodes\/\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"Ben Garfinkel on scrutinising classic AI risk&nbsp;arguments\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/80000hours.org\/#website\",\"url\":\"https:\/\/80000hours.org\/\",\"name\":\"80,000 Hours\",\"description\":\"\",\"publisher\":{\"@id\":\"https:\/\/80000hours.org\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/80000hours.org\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/80000hours.org\/#organization\",\"name\":\"80,000 Hours\",\"url\":\"https:\/\/80000hours.org\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/80000hours.org\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/80000hours.org\/wp-content\/uploads\/2018\/07\/og-logo_0.png\",\"contentUrl\":\"https:\/\/80000hours.org\/wp-content\/uploads\/2018\/07\/og-logo_0.png\",\"width\":1500,\"height\":785,\"caption\":\"80,000 Hours\"},\"image\":{\"@id\":\"https:\/\/80000hours.org\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/80000Hours\",\"https:\/\/x.com\/80000hours\",\"https:\/\/www.youtube.com\/user\/eightythousandhours\"]}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Ben Garfinkel on scrutinising classic AI risk arguments - 80,000 Hours","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/","og_locale":"en_US","og_type":"article","og_title":"How sure are we about this AI stuff?","og_description":"Ben Garfinkel, Research Fellow at Oxford's Future of Humanity Institute, thinks classic AI risk arguments haven't been subject to sufficient scrutiny.","og_url":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/","og_site_name":"80,000 Hours","article_publisher":"https:\/\/www.facebook.com\/80000Hours","article_modified_time":"2024-04-14T01:09:14+00:00","og_image":[{"width":1024,"height":537,"url":"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/brain-box_2.jpg","type":"image\/jpeg"}],"twitter_card":"summary_large_image","twitter_title":"How sure are we about this AI stuff?","twitter_description":"Ben Garfinkel, Research Fellow at Oxford's Future of Humanity Institute, thinks classic AI risk arguments haven't been subject to sufficient scrutiny.","twitter_image":"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/brain-box_2.jpg","twitter_site":"@80000hours","schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/","url":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/","name":"Ben Garfinkel on scrutinising classic AI risk arguments - 80,000 Hours","isPartOf":{"@id":"https:\/\/80000hours.org\/#website"},"primaryImageOfPage":{"@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage"},"image":{"@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage"},"thumbnailUrl":"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg","datePublished":"2020-07-09T17:40:29+00:00","dateModified":"2024-04-14T01:09:14+00:00","breadcrumb":{"@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#primaryimage","url":"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg","contentUrl":"https:\/\/80000hours.org\/wp-content\/uploads\/2020\/07\/Google_AI_art_1.jpg","width":2000,"height":1333,"caption":"Output from Google's [DeepDream](https:\/\/en.wikipedia.org\/wiki\/DeepDream) AI."},{"@type":"BreadcrumbList","@id":"https:\/\/80000hours.org\/podcast\/episodes\/ben-garfinkel-classic-ai-risk-arguments\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/80000hours.org\/"},{"@type":"ListItem","position":2,"name":"Podcast","item":"https:\/\/80000hours.org\/podcast\/episodes\/"},{"@type":"ListItem","position":3,"name":"Ben Garfinkel on scrutinising classic AI risk&nbsp;arguments"}]},{"@type":"WebSite","@id":"https:\/\/80000hours.org\/#website","url":"https:\/\/80000hours.org\/","name":"80,000 Hours","description":"","publisher":{"@id":"https:\/\/80000hours.org\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/80000hours.org\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/80000hours.org\/#organization","name":"80,000 Hours","url":"https:\/\/80000hours.org\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/80000hours.org\/#\/schema\/logo\/image\/","url":"https:\/\/80000hours.org\/wp-content\/uploads\/2018\/07\/og-logo_0.png","contentUrl":"https:\/\/80000hours.org\/wp-content\/uploads\/2018\/07\/og-logo_0.png","width":1500,"height":785,"caption":"80,000 Hours"},"image":{"@id":"https:\/\/80000hours.org\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/80000Hours","https:\/\/x.com\/80000hours","https:\/\/www.youtube.com\/user\/eightythousandhours"]}]}},"_links":{"self":[{"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/podcast\/70167"}],"collection":[{"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/podcast"}],"about":[{"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/types\/podcast"}],"author":[{"embeddable":true,"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/users\/405"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/media\/70169"}],"wp:attachment":[{"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/media?parent=70167"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/80000hours.org\/wp-json\/wp\/v2\/categories?post=70167"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}