Look for any podcast host, guest or anyone
Showing episodes and shows of

Superintelligent

Shows

The ValmyThe ValmyDebate with Vitalik Buterin — Will “d/acc” Protect Humanity from Superintelligent AI? Podcast: Doom Debates Episode: Debate with Vitalik Buterin — Will “d/acc” Protect Humanity from Superintelligent AI?Release date: 2025-08-12Get Podcast Transcript →powered by Listen411 - fast audio-to-text and summarizationVitalik Buterin is the founder of Ethereum, the world's second-largest cryptocurrency by market cap, currently valued at around $500 billion. But beyond revolutionizing blockchain technology, Vitalik has become one of the most thoughtful voices on AI safety and existential risk.He's donated over $665 million to pandemic prevention and other causes, and has a 12% P(Doom) – putting him squarely in what I con...2025-08-132h 26Digital Disruption with Geoff NielsonDigital Disruption with Geoff NielsonRoman Yampolskiy: How Superintelligent AI Could Destroy Us AllIs this a wake-up call for anyone who believes the dangers of AI are exaggerated?Today on Digital Disruption, we’re joined by Roman Yampolskiy, a leading writer and thinker on AI safety, and associate professor at the University of Louisville. He was recently featured on podcasts such as PowerfulJRE by Joe Rogan.Roman is a leading voice in the field of Artificial Intelligence Safety and Security. He is the author of several influential books, including AI: Unexplainable, Unpredictable, Uncontrollable. His research focuses on the critical risks and challenges po...2025-08-111h 13Mark Zuckerberg BioSnapMark Zuckerberg BioSnapZuckerberg's AI Blitz: Billion-Dollar Talent War, Lawsuits, and Superintelligent Glasses VisionMark Zuckerberg BioSnap a weekly updated Biography.Mark Zuckerberg is dominating headlines again this week with a flurry of moves that look set to define his power and legacy in techs next era. The boldest headlines have centered around his relentless pursuit of artificial intelligence talent. According to the Wall Street Journal and confirmed by Wired and Axios, Zuckerberg personally attempted to buy Mira Muratis startup Thinking Machines Lab for a staggering 1 billion dollars. Murati, previously a chief at OpenAI, refused and so Zuckerberg launched an all-out raid, targeting upwards of a dozen of her 50 staff with most...2025-08-1003 minThis Week in Social Media (TWISM)This Week in Social Media (TWISM)🤖 Time to go superintelligent!🚀 Ready to elevate your social media game to superintelligent levels? This episode of TWISM is your gateway to the future of digital marketing. We dive into the top 35 social media marketing stories of the week, highlighting the latest trends and tools that are transforming the industry. Discover how to harness cutting-edge technologies and strategies to boost your online presence and engagement. From algorithm updates to innovative content creation techniques, we cover everything you need to stay ahead in the fast-paced world of social media. Whether you're a marketer, creator, or brand, this episode is packed with actionable insights to hel...2025-08-0400 minSyllabuswithRohitSyllabuswithRohitSuperintelligence (HINDI/हिंदी में)Superintelligence is a book written by Nick Bostrom. In this book, he talks about the future and how computers and robots might become smarter than humans. Right now, humans are the smartest creatures on Earth. But one day, machines might think better and faster than we do. Bostrom calls this superintelligence.Bostrom explains that superintelligence could happen in different ways. One way is by making very smart computers that can learn by themselves. Another way is by changing human brains with technology to make people much smarter. There could also be new kinds of brains made from things like...2025-08-011h 45Clearer Thinking with Spencer GreenbergClearer Thinking with Spencer GreenbergAI: Autonomous or controllable? Pick one (with Anthony Aguirre)Read the full transcript here. Is AI that's both superintelligent and aligned even possible? Does increased intelligence necessarily entail decreased controllability? What's the difference between "safe" and "under control"? There seems to be a fundamental tension between autonomy and control, so is it conceivable that we could create superintelligent AIs that are both autonomous enough to do things that matter and also controllable enough for us to manage them? Is general intelligence needed for anything that matters? What kinds of regulations on AI might help to ensure a safe future? Should we stop working towards superintelligent AI...2025-07-311h 31The Neural NetworkersThe Neural NetworkersWelcome to 2027: The Rise of Superintelligent OverlordsIn this once-off deep-dive, we take you into a chilling near-future: the year 2027, where artificial superintelligence no longer serves nations — it shapes them. As AI systems surpass human cognition, the global balance of power fractures. Superpowers race to control minds made of code. Traditional diplomacy is replaced by algorithmic strategy. Wars are waged with autonomous weapons. And the question no one can answer: who really holds the power — governments, corporations, or the machines themselves?From the labs of Silicon Valley to the war rooms of Beijing and Washington, this episode unpacks the geopolitical chaos, ethical breakdowns, and exis...2025-07-1219 min解説!1日5分ビジネス英語解説!1日5分ビジネス英語メタの「超知的」AIチーム Meta’s “Superintelligent” AI team今回 2025年07月03日の記事は「メタの「超知的」AIチーム Meta’s “Superintelligent” AI team」という内容です。さていったいどういう内容なのでしょうか。 / 単語の解説や日本語訳はこちらから https://matt-english.com/podcast/202507032025-07-0300 minMogul Media AI SpotlightMogul Media AI SpotlightZuckerberg's Future Vision for Superintelligent A IRead More: https://www.mindburst.ai/2025/06/zuckerbergs-vision-for-superintelligent.html In a world rapidly evolving with technology, the vision for superintelligent A I emerges as a beacon of potential. This new frontier promises to redefine how we approach everyday challenges, from personalized healthcare solutions to innovative educational tools. The integration of A I into our lives aims to amplify human creativity and productivity, fostering a collaborative environment where machines and people work hand in hand. However, as we stand on the cusp of this A I revolution, it becomes crucial to establish ethical frameworks that guide the development of these...2025-06-1900 minThe AI DiariesThe AI DiariesEp.133: The Implications of Superintelligent AIThis episode is based on an article that discusses the significant implications of superintelligent AI, exploring both its potential benefits and inherent dangers. It highlights the risk of AI surpassing human capabilities, which could lead to unpredictable outcomes and raise concerns about maintaining control. The text also addresses crucial issues like ethical dilemmas arising from machine decision-making, the potential for widespread job displacement, and critical security concerns if these powerful systems are misused. Understanding and managing these various facets is presented as essential for ensuring a safe future alongside advanced AI. Read the full article here2025-05-2604 minWin-Win with Liv BoereeWin-Win with Liv Boeree#42 - Anthony Aguirre and Malo Bourgon - Why Superintelligent AI Should Not Be BuiltHumanity is in a race to build superintelligent AI gods. What happens if we create beings smarter than us—and can no longer steer them?In this special Win-Win IRL episode, Liv Boeree sits down with Anthony Aguirre (CEO of Future of Life Institute) and Malo Bourgon (CEO of MIRI) for a no-holds-barred conversation on finding alignment, avoiding power concentration and surveillance states, and the geopolitical implications of the race to AGI.Questions include: Why can’t the market be trusted to self-regulate AI? Why does power concentration in AI pose such a massive threat to d...2025-05-211h 07Closer To TruthCloser To TruthJoseph Corabi: Could Superintelligent AI Defeat Itself?Can superintelligent AI destroy humanity—or will it collapse under its own philosophical doubts? Professor Joseph Corabi, Chair of Philosophy at Saint Joseph’s University, joins Robert Lawrence Kuhn at MindFest 2025 to explore the limits of artificial intelligence, consciousness, and the metaphysics of mind.   Drawing from his presentation, "A Collapsing Superintelligent AI?", Corabi argues that future AI—however advanced—may become paralyzed by skeptical challenges that human minds navigate through evolved psychology. Could superintelligence be undone by its inability to resolve deep epistemological puzzles?   Topics discussed include: Can AI ever be truly c...2025-05-161h 19Tech ReviewsTech ReviewsFutureHouse Superintelligent Scientific AgentsFutureHouse has introduced a new platform featuring superintelligent AI agents designed to aid scientific research, addressing the issue of overwhelming data by providing tools that can navigate vast scientific literature and specialized databases. The platform offers four agents—Crow, Falcon, Owl, and Phoenix—each with a specific focus, ranging from general literature search to specialized chemistry tasks. These agents are highlighted as superior to existing frontier models in retrieval precision and accuracy, demonstrating the capacity to significantly accelerate scientific discovery by automating tasks like identifying research gaps or analyzing conflicting evidence. The platform provides both a web interface and an API to su...2025-05-0908 minThe AI Agent Daily BriefThe AI Agent Daily BriefVisa's AI Shopping, Ketryx Compliance, and FutureHouse's Superintelligent AgentsIn Episode 64 of The AI Agent Daily Brief, we explore Visa's AI-driven shopping initiatives and the associated consumer privacy concerns. Mastercard and PayPal's recent AI advancements are also discussed, shedding light on their strategic moves. The episode examines Ketryx's AI solutions for ensuring compliance in regulated industries and delves into the impactful role of AI agents in the healthcare sector. Doppel's innovative AI technology is highlighted for its role in combating cybercrime. We also cover FutureHouse's efforts in developing superintelligent AI agents. The episode wraps up with closing remarks and a subscription reminder. (0:00) Introduction to today's episode (0:25) Visa's AI-driven shopping...2025-05-0214 minThreat StatusThreat StatusHow dangerous is superintelligent AI?Ben and Guy break down the collapse of the Israel-Hamas ceasefire and the dramatic feud between the DOGE and the U.S. Institute of Peace. Then, Dan Hendrycks, director of the Center for AI Safety, joins the show to talk about superintelligent AI, its internal value systems, and its potential ramifications for U.S. national security. Plus, the "Threat of the Week."2025-03-2132 minA Beginner\'s Guide to AIA Beginner's Guide to AITypes of AI: Narrow AI vs. General AI vs. Superintelligent AI // 4.5 REPOSTHad to repost this episode, as I now had a ChatGPT 4.5 preview and the quality of the content is at least one level higher than ChatGPT 4, which I used before 🚀---AI isn’t just one thing—it comes in different flavours, from simple task-doers to the mind-bending idea of superintelligent machines that might one day outthink us all. In this episode, we break down the three key types of AI: Narrow AI, which powers your favourite voice assistants and recommendation engines; General AI, the still-unrealised dream of machines that think like humans; and Superin...2025-03-1920 minThe AI PodcastThe AI PodcastEpisode 111 Grok-3: Elon Musk's xAI Breakthrough & the Future of Superintelligent AI | AI PodcastDive into the future of artificial intelligence with this thrilling episode of the AI Podcast, where we explore Grok-3—Elon Musk's hypothetical next-gen AI model from xAI. Could Grok-3 revolutionize industries, accelerate scientific discovery, and redefine personalized AI? We break down its speculated advancements, from cutting-edge neural architectures and multimodal mastery (text, audio, video, sensors) to unprecedented scalability. Discover how Grok-3 might tackle quantum chemistry simulations, optimize global supply chains, or even act as your hyper-personalized digital twin—while navigating ethical minefields like algorithmic bias, data security, and existential risks. Will Grok-3 outpace GPT-5 and Gemi...2025-02-2511 minAI and Us: Exploring Our FutureAI and Us: Exploring Our FutureSuperintelligent AI: Dawn of Humanity's Next ChapterJoin us for an immersive eight-part series exploring the groundbreaking insights from "Superintelligence AI: The Dawn of Humanity's Next Chapter." From understanding the fundamentals of superintelligent AI to navigating its profound implications for humanity, each episode unpacks essential chapters from this transformative work.Episode Highlights:Deep dive into what truly sets superintelligent AI apart from current AI systemsHistorical parallels between AI evolution and past technological revolutionsReal-world implications for businesses, workplaces, and daily lifeCritical discussions on ethical considerations and risk mitigationPractical strategies for thriving in an AI-enhanced futureExpert guidance on participating in AI governance and ethicsTools and...2025-02-0717 minAI Forever - Establishing SMART LivingAI Forever - Establishing SMART LivingHarnessing Cognitive Titans: Unveiling AI's Superintelligent PotentialThe exploration of cognitive superpowers in AI envisions a world where superintelligent entities surpass human abilities, raising the potential for AI dominance. Superintelligence can enhance its own capabilities exponentially, leading to strategic mastery, social manipulation, and technological innovation. Crucially, such entities could reshape societal structures by leveraging cognitive superpowers like intelligence amplification and economic productivity. However, this scenario necessitates careful ethical considerations to align AI's objectives with human values, ensuring accountability and equitable access to technological advancements. As AI integrates into societal frameworks, transparent governance and interdisciplinary dialogue are imperative to foster a future where AI complements human progress. 2025-01-2124 minAI-Gen Daily: AI News by AIAI-Gen Daily: AI News by AIAI News: Hinton Backs Musk's OpenAI Lawsuit & Google's Superintelligent AI Leap | AI-Gen DailyBreaking AI News Roundup Geoffrey Hinton, AI pioneer, joins forces with Elon Musk in a landmark lawsuit against OpenAI, challenging the company's shift from non-profit to commercial entity. Encode provides legal support in this crucial battle over AI's future. Key Stories: • Hinton vs OpenAI: The battle for AI's soul • Google's revolutionary test-time compute scaling • The AI Consulting Network's innovative approaches • Waymo & Serve robots' unprecedented collision Featured Expert Spotlight: Logan Kilpatrick (Google Product Manager) shares insights on superintelligent AI devel...2025-01-0102 minStrictlyVC NewsStrictlyVC NewsMeta tries to slow OpenAI's roll, Claude is one hot tamale, and a gift your therapist will definitely find passive aggressiveStrictlyVC Download Briefing - December 14, 2024 Main Themes: AI Industry Battles: The newsletter highlights the intensifying competition in the AI industry, with Meta challenging OpenAI's move to for-profit status and OpenAI battling accusations of copyright infringement leading to a tragic outcome. Events-Betting Market Controversies: The fierce rivalry in the emerging events-betting market is explored, revealing unethical tactics employed by competing companies Kalshi and Polymarket. Significant Funding Rounds: The newsletter tracks numerous large funding rounds, showcasing the continued influx of capital into AI, fintech, and other sectors. OpenAI in the Spotlight: OpenAI is heavily featured, covering topics ranging...2024-12-2910 minAI Deep DiveAI Deep DiveGrok Updates, Character AI Teen Safety Features, and Superintelligent AIIn today’s AI Deep Dive, we cover Elon Musk’s xAI releasing a faster Grok chatbot with new API features and a "Grok button" for X. Character AI responds to lawsuits by implementing teen safety features and rebranding as an entertainment platform. We also discuss OpenAI co-founder Ilya Sutskever’s views on superintelligent AI and reflect on the legacy of Suchir Balaji, a former OpenAI employee who raised critical copyright concerns before his tragic passing.2024-12-1409 minJoseph Z PodcastJoseph Z PodcastMike Tyson and Next 4 Years! | Prophetic Update with Joseph ZOn today’s live prophetic broadcast, Joseph Z shares some prophecies about the emergence of a superintelligent AI, the outlook of January 2025, and what is coming next.  He revisits his December 3, 2020, prophetic broadcast, in which he revealed seeing a heavyweight champion die after coming out of retirement; which he interprets as the death of an old order, marking a new beginning.  He also shares what he sees for the next four years, revealing that a new beginning shall begin in 2028, after witnessing some fire and rightsizing in 2025, and “pick-up sticks” in 2026. Also, in his August 13, 2024 broadcast, he insists that Jan...2024-11-151h 12AI-Gen Daily: AI News by AIAI-Gen Daily: AI News by AIAI News: Perplexity Ads, Trump's DOGE, and Superintelligent AI by 2027Welcome to AI-Gen Daily, your go-to source for the latest AI news and developments! In this episode, AI Boyce covers:🔍 Perplexity AI introduces sponsored follow-up questions 🏛️ Trump appoints Musk and Ramaswamy to lead new DOGE department 🧠 Anthropic CEO predicts superintelligent AI by 2026-2027 EU seeks public input on AI legislationDive into the world of AI with us as we explore these exciting topics and more. From search engine innovations to political shake-ups, we've got you covered on all things artificial intelligence.Need expert AI guidance? Check out The AI Consulting...2024-11-1502 minDoom DebatesDoom Debates15-Minute Intro to AI DoomOur top researchers and industry leaders have been warning us that superintelligent AI may cause human extinction in the next decade.If you haven't been following all the urgent warnings, I'm here to bring you up to speed.* Human-level AI is coming soon* It’s an existential threat to humanity* The situation calls for urgent actionListen to this 15-minute intro to get the lay of the land.Then follow these links to learn more and see how you can help:* The CompendiumA...2024-11-0415 minFuel Your Mind With This Eye-Opening Full Audiobook And Feel The Difference.Fuel Your Mind With This Eye-Opening Full Audiobook And Feel The Difference.The Humachine: AI, Human Virtues, and the Superintelligent Enterprise, Second Edition by John D. Wood, Nada R. SandersPlease visithttps://thebookvoice.com/podcasts/1/audiobook/798146to listen full audiobooks. Title: The Humachine: AI, Human Virtues, and the Superintelligent Enterprise, Second Edition Author: John D. Wood, Nada R. Sanders Narrator: Rachel Perry Format: Unabridged Audiobook Length: 16 hours 30 minutes Release date: October 29, 2024 Genres: Computers & Technology Publisher's Summary: Updated for a post-Covid world, the second edition of this groundbreaking book explains why becoming a Humachine enterprise is the only way forward for a company to maintain a competitive advantage in the age of artificial intelligence (AI). Now, in the face of the 'Great Resignation' and 'botsourcing,' thought leaders Sanders and Wood...2024-10-294h 30Discover the Best Audio Stories in Non-Fiction, Computers & TechnologyDiscover the Best Audio Stories in Non-Fiction, Computers & TechnologyThe Humachine: AI, Human Virtues, and the Superintelligent Enterprise, Second Edition by John D. Wood, Nada R. SandersPlease visit https://thebookvoice.com/podcasts/1/audiobook/798146 to listen full audiobooks. Title: The Humachine: AI, Human Virtues, and the Superintelligent Enterprise, Second Edition Author: John D. Wood, Nada R. Sanders Narrator: Rachel Perry Format: Unabridged Audiobook Length: 16 hours 30 minutes Release date: October 29, 2024 Genres: Computers & Technology Publisher's Summary: Updated for a post-Covid world, the second edition of this groundbreaking book explains why becoming a Humachine enterprise is the only way forward for a company to maintain a competitive advantage in the age of artificial intelligence (AI). Now, in the face of the 'Great Resignation' and 'botsourcing,' thought leaders Sanders and...2024-10-2930 minThe Unfiltered ArenaThe Unfiltered ArenaHomo Deus: The Rise of Artificial IntelligenceAI’s rapid development could decouple intelligence from consciousness, reshaping work, governance, and society. Harari explores how AI might surpass human intelligence, and the challenges of living in a world controlled by superintelligent algorithms. A podcast series based on Yuval Noah Harari’s books offers an exciting opportunity to dive into the profound themes he explores. This six-episode outline will focus on Harari’s first three major works: Sapiens, Homo Deus, and 21 Lessons for the 21st Century. Each episode will cover core concepts, insightful analysis, and how these ideas relate to our world today. ...2024-10-2809 minArtificial InsightsArtificial InsightsSuperintelligent AI Is Only A Few Thousand Days AwaySam Altman, the CEO of OpenAI, discusses the potential impact of superintelligent AI on education and the workforce. He envisions a future where AI tutors personalise learning for every student, regardless of background, potentially closing the gap between privileged and underprivileged students.2024-10-1808 minAstral Codex Ten PodcastAstral Codex Ten PodcastMantic Monday 9/16/24Probably No Superintelligent Forecaster Yet FiveThirtyNine (ha ha) is a new forecasting AI that purports to be “superintelligent”, ie able to beat basically all human forecasters. In fact, its creators go further than that: they say it beats Metaculus, a site which aggregates the estimates of hundreds of forecasters to generate estimates more accurate than any of them. You can read the announcement here and play with the model itself here. (kudos to the team for making the model publicly available, especially since these things usually have high inference costs) https://www.astralcodexten.com...2024-10-1617 minAI-Gen Daily: AI News by AIAI-Gen Daily: AI News by AISpaceX Rocket Catch, AI Earbuds, and Superintelligent AI by 2026 | AI News RoundupWelcome to AI-Gen Daily, your go-to source for the latest in AI and tech news! In this episode, we cover:• SpaceX's groundbreaking Starship booster catch• ByteDance's new AI-powered Ola Friend earbuds• Anthropic CEO's predictions for superintelligent AI by 2026Join AI Boyce as we explore these exciting developments and their potential impact on our future. From space travel to wearable tech and the possibility of AI surpassing human intelligence, we've got it all covered!Key Takeaways:✅ SpaceX successfully caught a Starship booster with mechanical arms✅ ByteDance enters the AI wearabl...2024-10-1402 minThe AI Daily Brief (Formerly The AI Breakdown): Artificial Intelligence News and AnalysisThe AI Daily Brief (Formerly The AI Breakdown): Artificial Intelligence News and AnalysisThe Challenge of Enterprise AI AdoptionEnterprise AI adoption faces significant obstacles, from employees using AI without sharing discoveries to organizations struggling to scale past the pilot stage. Today’s episode reviews insights from Professor Ethan Mollick’s recent article on creating environments for AI-driven innovation in business. The discussion highlights strategies to encourage transparency in AI use, establish collaborative R&D frameworks, and develop AI-enablement infrastructures, such as Superintelligent, to bridge knowledge gaps and empower teams. Superintelligent’s platform specifically addresses these challenges, providing a centralized hub for employees to share AI insights, which can enable strategic scaling of AI solutions across an organization. ...2024-10-0918 minThe Remarkable Leadership PodcastThe Remarkable Leadership PodcastHow AI Can Help You Create a SuperIntelligent Enterprise with Nada SandersWhat does it take to build a superintelligent enterprise? Dr. Nada Sanders shares insights from her latest research on what it means to create a superintelligent organization. Contrary to popular belief, it's not about only relying on advanced technology. Instead, the combination of AI and human judgment drives success. Nada and Kevin discuss Kasparov’s Law, which emphasizes that an average person with a good process and AI can outperform even superior computers or individuals without the right systems. Nada believes the key to long-term success in a superintelligent enterprise lies in blending technology with human empathy, communication, and ex...2024-10-0950 minAI Unplugged: Bytes and Insights host Oliver MindburstAI Unplugged: Bytes and Insights host Oliver MindburstSutskever's Billion-Dollar Investment in Superintelligent AIRead More: https://www.mindburst.ai/2024/09/sutskever-secures-billion-for.html In a groundbreaking move, a prominent figure in the A I community has secured a staggering billion-dollar investment aimed at developing superintelligent A I. This ambitious initiative promises to reshape the technological landscape, pushing the boundaries of what A I can achieve. By surpassing human intelligence in areas like creativity and problem-solving, this venture could unlock unprecedented advancements across various sectors, including healthcare and education. However, as the potential benefits unfold, so too do pressing ethical concerns surrounding safety, job displacement, and the necessity for robust regulations. As we stand...2024-10-0801 minMad Tech TalkMad Tech Talk#19 - On the Brink of Superintelligence: Sam Altman’s Vision for the Future In this episode of Mad Tech Talk, we delve into the visionary insights of OpenAI CEO Sam Altman, who posits that the advent of superintelligence—AI vastly smarter than humans—could be just a few years away. Drawing from Altman’s recent claims, we explore the transformative potential of deep learning and its profound implications for society. Key topics covered in this episode include: Implications of Superintelligence: Discuss the far-reaching implications of achieving superintelligence, examining both the potential benefits and the risks. Understand how AI could revolutionize various aspects of society, from per...2024-09-2907 minB5Y PodcastB5Y PodcastDiscussing "Situational Awareness" by Leopold AschenbrennerThis episode examines Part IIIc: "Superalignment" from Leopold Aschenbrenner's "Situational Awareness" report. We explore the critical challenge of aligning superintelligent AI systems with human values and goals.Key points include:1. **Defining Superalignment**: We introduce the concept of superalignment - the task of ensuring that AI systems vastly more intelligent than humans remain aligned with our values and intentions.2. **The Scale of the Challenge**: Aschenbrenner argues that aligning a superintelligent AI is fundamentally more difficult than aligning current AI systems, due to the vast intelligence gap.3. **Complexity of...2024-09-2209 minA Beginner\'s Guide to AIA Beginner's Guide to AISuperintelligent AI: The Last Invention Humanity Will Ever Need To Make?In this episode of "A Beginner's Guide to AI", we delve into the fascinating and potentially world-altering concept of Superintelligent AI. We explore what happens when artificial intelligence surpasses human intelligence in every domain, offering both incredible possibilities and significant risks. Through clear explanations, a relatable analogy, and a real-world case study, we break down the challenges and opportunities that come with this advanced technology. Whether you're new to AI or looking to deepen your understanding, this episode is packed with insights that are crucial for anyone interested in the future of technology.2024-09-2117 minLessWrong (Curated & Popular)LessWrong (Curated & Popular)“What is it to solve the alignment problem? ” by Joe CarlsmithPeople often talk about “solving the alignment problem.” But what is it to do such a thing? I wanted to clarify my thinking about this topic, so I wrote up some notes.In brief, I’ll say that you’ve solved the alignment problem if you’ve: avoided a bad form of AI takeover,built the dangerous kind of superintelligent AI agents,gained access to the main benefits of superintelligence, andbecome able to elicit some significant portion of those benefits from some of the superintelligent AI agents at stake in (2).[1] The post also discusses what it wo...2024-08-281h 39History of Philosophy Audio ArchiveHistory of Philosophy Audio Archive#91- Chaos, Consciousness, and Creativity: Terence McKenna, Rupert Sheldrake, Ralph Abraham (Trialogue 1-3)A DMT priest, an evolutionary biologist, and a chaos mathematician walk into a hot-tub in California... (00:00:00) - Host Intro (00:00:43) - Cast of Characters (00:51:12) - The Evolutionary Mind (02:00:17) - Superintelligent AI -//- My Patreon! Buy the ticket, take the ride! https://patreon.com/hemlockpatreon Descriptions courtesy Sheldrake.org https://www.sheldrake.org/audios/the-sheldrake-mckenna-abraham-trialogues Part 1 - Cast of Characters An introduction to the first series of public trialogues held at Esalen, California in 1989. Morphic Resonance, the novelty...2024-08-103h 00AI & IAI & IHow a Top Podcaster Rides the AI Wave - Ep. 28 with Nathaniel WhittemoreKeeping up with AI is Nathaniel Whittemore’s full-time job—and I spent an hour with him to understand how he does it.Nathaniel is the host of a top-ranked AI podcast on the technology charts, The AI Daily Brief, which breaks down the most important news in AI every day. He is also the founder and CEO of Superintelligent, a platform that teaches you how to use AI for work and fun through interactive video tutorials. We talked about how he curates information with X bookmarks, Google News, news aggregator Feedly, and research tool Perp...2024-07-311h 09AI and IAI and IHow a Top Podcaster Rides the AI Wave - Ep. 28 with Nathaniel WhittemoreKeeping up with AI is Nathaniel Whittemore’s full-time job—and I spent an hour with him to understand how he does it. Nathaniel is the host of a top-ranked AI podcast on the technology charts, The AI Daily Brief, which breaks down the most important news in AI every day. He is also the founder and CEO of Superintelligent, a platform that teaches you how to use AI for work and fun through interactive video tutorials.  We talked about how he curates information with X bookmarks, Google News, news aggregator Feedly, and research tool Perpl...2024-07-311h 09Preparing for AI: The AI Podcast for EverybodyPreparing for AI: The AI Podcast for EverybodyTHE TOP 10 AI APPS FOR SUMMER 2024! The best hour you will spend all summerSend us a textHeading to the beach this summer? Or just staying at home pondering how long it will be until a superintelligent AI takes over the world? Or maybe you really want to learn the best AI tools to use this summer? Well whatever your flava you are in the right place, because this week Jimmy and Matt introduce their top 10 AI applications for summer 2024!Join us to discover the best large language models for multimodality, search and privacy, the best image and music generation tools, the most comprehensive Chinese multimodal AI app...2024-07-171h 13For Humanity: An AI Safety PodcastFor Humanity: An AI Safety PodcastEpisode #35 TRAILER “The AI Risk Investigators: Inside Gladstone AI, Part 1” For Humanity: An AI Risk PodcastIn Episode #35 TRAILER:, host John Sherman talks with Jeremie and Eduard Harris, CEO and CTO of Gladstone AI. Gladstone AI is the private company working most closely with the US government on assessing AI risk. The Gladstone Report, published in February, was the first public acknowledgment of AI risk reality by the US government in any way. These are two very important people doing incredibly important work. The full interview lasts more than 2 hours and will be broken into two shows. TIME MAGAZINE ON THE GLADSTONE REPORT https://time.com/6898967/ai-extinction-national-security-risks-report/   P...2024-07-0104 minModern CTOModern CTOUncovering the Ethical Dilemmas of AI with Nathaniel Whittemore, Founder at SuperintelligentToday we’re talking to Nathaniel Whittemore, Founder at Superintelligent and host of The AI Daily Brief. We discuss the reality of the new browser wars, how sci-fi concepts such as time travel can actually impact our lives, and how to actually conceptualize the meaning of work in our modern age.All of this right here, right now, on the Modern CTO Podcast! To learn more about Superintelligent, check out their website: https://besuper.ai/To listen to Nathaniel's podcast, The AI Daily Brief, check it on Libsyn or wherever you get your pod...2024-06-2440 minImagine AI PodcastImagine AI PodcastUnlock AI Mastery in Minutes with Superintelligent AI | CEO Nathaniel Whittemore on Imagine AI | E49Chris Madden welcomes Nathaniel Whittemore, a former marketer at FTX, to discuss his AI learning platform, Superintelligent. They explore its educational approach, content creation, and partnerships with institutions. They also delve into AI's impact on productivity, challenges with distribution, and NVIDIA's role in AI market dynamics. (0:00) Nathaniel's experience as a marketer at FTX (0:22) Introduction and guest welcome (0:33) Nathaniel's AI learning platform, Superintelligent (2:14) Superintelligent's educational approach (3:17) Content creation and future partnerships (5:37) Partnerships with educational institutions (10:21) Chris's career shift and AI interest (16:53) AI's effect on work efficiency and productivity (24:55) Challenges with distribution platforms and content models (30:17) What it was like working with...2024-06-1742 minThe Disruptive Technologies PodcastThe Disruptive Technologies Podcast161:SEC Battles FIT21: Crypto Regulation Clash UnfoldsTimestamps: 00:00 - Introduction 00:51 - SEC vs. FIT21 05:00 - ETH ETF Trading Soars 07:07 - UK Approves Crypto ETPs 09:28 - Hong Kong vs. Worldcoin 11:27 - Buterin's AI Concerns 14:58 - Coffeezilla vs. Rabbit AI 17:39 - Pentagon’s Autonomous Plan 22:09 - Conclusion Join Nikodemus on May 23, 2024, for a riveting exploration of crucial topics in technology and cryptocurrency. This episode covers the SEC's intense opposition to the Financial Innovation and Technology for the 21st Century Act (FIT21), record-breaking trading volumes of Ethereum futures ETFs, the UK's landmark approval of cryptocurrency ETPs, Hong Kong's stringent action against Worldcoin, Vitalik Buterin's warnings on superintelligent AI...2024-05-2323 minThis Week in Enterprise TechThis Week in Enterprise TechWeek 14 - This Week in Enterprise TechOpenAI and Google go head-to-head, OpenAI’s “Superalignment” team is dissolved, Matt Garman is the new CEO at AWS, what’s the current state of IT budgeting and planning, and how far should you trust the concept of autonomy when it comes to your IT environment? Amalgam Insights’ Hyoun Park and The DX Report’s Charles Araujo are here again to discuss This Week in Enterprise Tech.OpenAI vs. GoogleGoogle I/O was last week, where they announced Google Astra, an AI assistant based on Google Gemini models. But OpenAI may have stolen their thunder by r...2024-05-2136 minThe Matrix AI Talk Radio from inteligenesis.comThe Matrix AI Talk Radio from inteligenesis.comOpenAI's Efforts to Control Superintelligent AI Facing Setback In a bold move destined to shape our collective future, OpenAI established an elite unit dedicated to the oversight of what they deemed 'superintelligent' AI. This team's core mission was clear yet daunting. To safeguard humanity from the potentially perilous outcomes of AI that surpasses human intelligence. Recently, though, rumblings from within have surfaced, painting a picture of neglect & contradiction. A source intimate with the company's inner workings has signaled an alarming downgrade in the team's significance. Personnel reassignments & shifting priorities have purportedly left the initiative in limbo. Missteps in AI...2024-05-2001 minTechCrunch Daily CrunchTechCrunch Daily CrunchOpenAI created a team to control 'superintelligent' AI — then let it witherOpenAI created a team to control 'superintelligent' AI — then let it wither, source says; Slack under attack over sneaky AI training policy; Microsoft dodges UK antitrust scrutiny over its Mistral AI stake Learn more about your ad choices. Visit podcastchoices.com/adchoices2024-05-1806 minFor Humanity: An AI Safety PodcastFor Humanity: An AI Safety PodcastEpisode #24 - “YOU can help save the world from AI Doom” For Humanity: An AI Safety PodcastIn episode #24, host John Sherman and Nonlinear Co-founder Kat Woods discusses the critical need for prioritizing AI safety in the face of developing superintelligent AI. In this conversation, Kat and John discuss the topic of AI safety and the potential risks associated with artificial superintelligence. Kat shares her personal transformation from being a skeptic to becoming an advocate for AI safety. They explore the idea that AI could pose a near-term threat rather than just a long-term concern. They also discuss the importance of prioritizing AI safety over other philanthropic endeavors and the need for talented individuals...2024-04-171h 21Tech Threads: Sci-Tech, Future Tech & AITech Threads: Sci-Tech, Future Tech & AIThe Threat of AGI: Should We Fear Superintelligent Machines? #ai #artificialgeneralintelligence🚀 Unlock the Future with AGI! Are We Ready for Superintelligent Machines? 🤖💡CLICK HERE TO VIEW THE VIDEO VERSION OF THIS PODCAST:https://youtu.be/b-AzHV4XOZYWelcome to a rollercoaster of thoughts in our latest video, "The Threat of AGI: Should We Fear Superintelligent Machines?"🎢 Brace yourself for an eye-opening journey through the world of Artificial General Intelligence (AGI), where possibilities collide with controversies! 🌐🔥👉 In this mind-bending adventure, we explore:The race to AGI and recent breakthroughs in AI! 🏁🚀The ethical implications that keep us up at night. 😱🛌Debunking...2024-03-1308 minTech-MexTech-Mex🌐 OpenAI GPT-4.5 Leaked! Gemini, Superintelligent AI Control, Tesla's Optimus Gen2, Billionaire Watch & More! Ep26🚀 Uncover the latest in tech and finance with this jam-packed video! OpenAI's GPT-4.5 has been leaked, revealing the revolutionary Gemini project and mind-blowing demos for controlling superintelligent AI. Explore the fast-approaching era of AGI and OpenAI's strategy of creating a "dumb AI" to manage the smart one.🤖 Tesla's Optimus Gen2: Get ready for the future with Tesla's Optimus Gen2 Robot! It's not just a robot; it can even hold your huevos. Watch as we delve into the cutting-edge technology that Elon Musk and his team have unleashed.💰 Billionaire Watch: Track the earnings...2023-12-2041 minAI DAILY: Breaking AI News Handpicked For The Curious MindAI DAILY: Breaking AI News Handpicked For The Curious MindAI ELECTION HALLUCINATIONPlus OpenAI guards against superintelligent AI. PODCAST EDITION https://aidaily.us Concerns about AI spreading false election information, Jeff Bezos on AI's potential, OpenAI's efforts to control superintelligent AI, doctors' mixed feelings on AI in medicine, and more. Microsoft AI Chatbot Spreads Misinformation About Elections Jeff Bezos Reveals Surprising Insight on AI: "Discoveries, Not Inventions" OpenAI's Superalignment Group Takes First Step in Controlling Super-Intelligent AI Doctors Have High Hopes and Fears for AI in Medicine The Wild West of AI Mental Health Apps: Differentiating Fact from Fiction AI Breaks the Age Barrier: Boomers and Generation X...2023-12-1505 minAI News DailyAI News DailyIntel Unveils Advanced Processors with AI and Gaming Capabilities, World's First AI-Powered News Network, OpenAI Launches $10M Grant Program for Controlling Superintelligent AI, and more...Image: Intel, via theverge.comIn this episode: (00:10): Intel's New Core Processors: Empowering AI and Revolutionizing Gaminghttps://www.brief.news/stories/2023/12/14/intels-advanced-ai-processors?utm_source=brief&utm_medium=share(01:16): Channel 1 News: Bridging the Gap Between AI and Human Anchors in Journalismhttps://www.brief.news/stories/2023/12/14/ai-powered-news-network-launches?utm_source=brief&utm_medium=share(01:54): OpenAI's $10 Million Grant Program and Conference on Superintelligent Alignment: Preventing Rogue AIs with Supervisionhttps://www.brief.news/stories/2023/12/14/openai-launches-10m-grant-program?utm_source=brief&utm_medium=share(02:41): Enhancing Journalism with AI: OpenAI Partners w...2023-12-1505 minDavid Storoy\'s PodcastDavid Storoy's PodcastEr kunstig intelligens en trussel mot menneskeheten?Det er delt mening blant eksperter om hvorvidt kunstig intelligens (KI) er en trussel mot menneskeheten. Noen argumenterer for at utviklingen av superintelligent KI kan potensielt føre til farlige konsekvenser, som tap av menneskelig kontroll og autonomi. Dette kan resultere i økonomisk ulikhet, arbeidsledighet, etiske dilemmaer og potensiell overvåkning og misbruk av KI. Det er også bekymring for at superintelligent KI kan ha uforutsigbare effekter og ta beslutninger som kan være skadelige for mennesker.På den andre siden mener noen eksperter at KI kan være til stor fordel for menneskeheten. De argumenterer for at KI k...2023-12-0901 minFT Tech TonicFT Tech TonicSuperintelligent AI: The DoomersIn the first episode of a new, five-part series of Tech Tonic, FT journalists Madhumita Murgia and John Thornhill ask how close we are to building human-level artificial intelligence and whether ‘superintelligent’ AI poses an existential risk to humanity. John and Madhu speak to Yoshua Bengio, a pioneer of generative AI, who is concerned, and to his colleague Yann LeCun, now head of AI at Meta, who isn’t. Plus, they hear from Eliezer Yudkowsky, research lead at the Machine Intelligence Research Institute, who’s been sounding the alarm about superintelligent AI for more than two decades. Registe...2023-11-1428 minBehind the MoneyBehind the MoneyComing soon: Superintelligent AI In a new series of Tech Tonic, FT journalists Madhumita Murgia and John Thornhill look at the concerns around the rise of artificial intelligence. Will superintelligent AI bring existential risk, or a new renaissance? Would it be ethical to build conscious AI? How intelligent are these machines anyway? The new season of Tech Tonic from the Financial Times, drops mid-November.Presented by Madhumita Murgia and John Thornhill. Senior producer is Edwin Lane and producer Josh Gabert-Doyon. Executive produced by Manuela Saragosa. Sound design by Breen Turner and Samantha Giovinco. Original music by Metaphor Music. The FT’s h...2023-11-1001 minFT Tech TonicFT Tech TonicComing soon: Superintelligent AI In a new series of Tech Tonic, FT journalists Madhumita Murgia and John Thornhill look at the concerns around the rise of artificial intelligence. Will superintelligent AI bring existential risk, or a new renaissance? Would it be ethical to build conscious AI? How intelligent are these machines anyway? The new season of Tech Tonic from the Financial Times, drops mid-November.Presented by Madhumita Murgia and John Thornhill. Senior producer is Edwin Lane and producer Josh Gabert-Doyon. Executive produced by Manuela Saragosa. Sound design by Breen Turner and Samantha Giovinco. Original music by Metaphor Music. The FT’s h...2023-10-3104 minLaunch CodesLaunch Codes03. ‘See, Hear, Speak’ with Chat-GPT, Low MarTech Stack Utilization and Is the MQL Dead?In this week’s episode of Launch Codes, Joe is joined by Andy Caron, President of RP, to discuss new developments in AI, including ChatGPT’s ‘See, Hear, Speak’ functionality, DALL-E 3, and Microsoft Copilot and their impact on content. Joe and Andy also discuss a new Gartner report on declining martech stack utilization, the best practices for syncing an MQL from Marketo to Salesforce, the features Andy is most excited for in September and October’s Adobe Marketo product release, and they give their takes on Superintelligent AI. Bonus: Andy samples Y3000, the new Coca-Cola flavor created by AI.  ---2023-09-2643 minThe Artificial Intelligence ShowThe Artificial Intelligence ShowElon Musk’s Quest to Shape the Future of AI, Hands-On with Google Duet AI, Time’s Top 100 People in AI, and HubSpot’s AI RoadmapInside Elon Musk’s struggle for the future of AIWe just got a never-before-seen look at how—and why—Elon Musk decided to go all-in on artificial intelligence. This comes from an article by Walter Isaacson in Time, and is adapted from his upcoming book Elon Musk, which publishes today! Issacson’s name may ring a bell, as he’s also the author of the Steve Jobs biography.) In the article, Issacson gives new details on the actions Elon Musk has taken to get highly involved in the future of AI. It turns out that Musk has become...2023-09-121h 08Amplify AI - Integrating Intelligence, Preserving HumanityAmplify AI - Integrating Intelligence, Preserving HumanityAI: GPT-4 and the hunt for superintelligence [Tech Stack]Harnessing the power of GPT-4 for business You’re in a treat this Friday. We are exploring GPT-4. I am breaking down the basic concepts of machine learning, natural language processing, and neural networks. Then, I delve into the key components or elements of GPT-4, including data, algorithms, and computing power. I also highlight the potential applications of GPT-4 for content creation, customer service, and personal assistance. Throughout the episode, you’ll see the importance of understanding and controlling superintelligent AI to prevent potential negative outcomes. Here are the key takeaways: GPT-4 is a...2023-09-0814 minThe Ticket: Discover the Future of Customer Service, Support, and Experience, with IntercomThe Ticket: Discover the Future of Customer Service, Support, and Experience, with IntercomThe emergence of superintelligent AIIn this episode, our Senior Director of Machine Learning, Fergal Reid, joins Emmet Connolly, our VP of Product Design, to take superintelligence head-on – the knowns and the unknowns, from the ethics to the threats and the challenges of regulation.Watch on YouTube: https://www.youtube.com/watch?v=KYBQzO2u1XsRead a transcript: https://www.intercom.com/blog/videos/the-emergence-of-superintelligent-aiSee Privacy Policy at https://art19.com/privacy and California Privacy Notice at https://art19.com/privacy#do-not-sell-my-info.2023-08-3150 minAI Unraveled: Latest AI News & Trends, ChatGPT, Gemini, DeepSeek, Gen AI, LLMs, AI Ethics & BiasAI Unraveled: Latest AI News & Trends, ChatGPT, Gemini, DeepSeek, Gen AI, LLMs, AI Ethics & BiasTop AI Image-to-Video Generators 2023 - Google Gemini: Facts and rumors - The importance of making Superintelligent Small LLMsPodcast video at: Djamgatech Education Youtube ChannelTop AI Image-to-Video Generators 2023Genmo D-ID LeiaPix Converter InstaVerseSketch NeROIC DPT Depth RODINGoogle Gemini: Facts and rumorsThe importance of making superintelligent small LLMsThis podcast is generated using the Wondercraft AI platform (https://www.wondercraft.ai/?via=etienne), a tool that makes it super easy to start your own podcast, by enabling you to use hyper-realistic AI voices as your host. Like mine! Get a 50% discount the first month with the code AIUNRAVELED50Attention AI...2023-08-1911 min80,000 Hours Podcast80,000 Hours Podcast#159 – Jan Leike on OpenAI's massive push to make superintelligence safe in 4 years or lessIn July, OpenAI announced a new team and project: Superalignment. The goal is to figure out how to make superintelligent AI systems aligned and safe to use within four years, and the lab is putting a massive 20% of its computational resources behind the effort.Today's guest, Jan Leike, is Head of Alignment at OpenAI and will be co-leading the project. As OpenAI puts it, "...the vast power of superintelligence could be very dangerous, and lead to the disempowerment of humanity or even human extinction. ... Currently, we don't have a solution for steering or controlling a potentially superintelligent...2023-08-082h 51Machine Learning Street Talk (MLST)Machine Learning Street Talk (MLST)Can We Develop Truly Beneficial AI? George Hotz and Connor LeahyPatreon: https://www.patreon.com/mlst Discord: https://discord.gg/ESrGqhf5CB George Hotz and Connor Leahy discuss the crucial challenge of developing beneficial AI that is aligned with human values. Hotz believes truly aligned AI is impossible, while Leahy argues it's a solvable technical challenge.Hotz contends that AI will inevitably pursue power, but distributing AI widely would prevent any single AI from dominating. He advocates open-sourcing AI developments to democratize access. Leahy counters that alignment is necessary to ensure AIs respect human values. Without solving alignment, general AI could ignore or...2023-08-041h 29Amplify AI - Integrating Intelligence, Preserving HumanityAmplify AI - Integrating Intelligence, Preserving HumanityWhat creative entrepreneurs need to know about the 3 types of AI [Tech Stack]In this episode, I want to talk to you about the three types of AI you need to know as a creative entrepreneur: narrow AI, general AI, and superintelligent AI. It’s essential to understand the difference between them, just like knowing the difference between a bicycle, a car, and a rocket ship. Each has unique capabilities and uses, and it’s essential to understand them so you’re not being misled by false information. In this episode I cover: Narrow AI is good at one specific thing, like recognising faces or recommending songs. General AI is goo...2023-07-3112 minA Beginner\'s Guide to AIA Beginner's Guide to AITypes of AI: Narrow AI vs. General AI vs. Superintelligent AIAI isn’t just one thing—it comes in different flavours, from simple task-doers to the mind-bending idea of superintelligent machines that might one day outthink us all. In this episode, we break down the three key types of AI: Narrow AI, which powers your favourite voice assistants and recommendation engines; General AI, the still-unrealised dream of machines that think like humans; and Superintelligent AI, the mind-blowing concept of an intelligence that could surpass us in every way.Through real-world examples like IBM’s Watson on Jeopardy! and DeepMind’s AlphaGo, we explore where AI stand...2023-07-2219 minTED Talks DailyTED Talks DailyWill superintelligent AI end the world? | Eliezer YudkowskyDecision theorist Eliezer Yudkowsky has a simple message: superintelligent AI could probably kill us all. So the question becomes: Is it possible to build powerful artificial minds that are obedient, even benevolent? In a fiery talk, Yudkowsky explores why we need to act immediately to ensure smarter-than-human AI systems don't lead to our extinction. Hosted on Acast. See acast.com/privacy for more information.2023-07-1110 minRetell — AI, tech and ecommerce newsRetell — AI, tech and ecommerce newsWhat is Threads? From Twitter fiasco to 100 million sign ups (+ OpenAI creates team to control "superintelligent" AI)It's Monday, July 10th — Here are the top AI, tech and ecommerce stories from last week and why that matter this week...Threads, Instagram’s ‘Twitter Killer,’ Has ArrivedAdam Mosseri on the launch of ThreadsTwitter is threatening to sue Meta over ThreadsInstagram’s Threads app reaches 100 million users within just five daysElon Musk keeps laying off Twitter employees after saying cuts were doneElon Musk mocks fired Twitter employeesWhat is Instagram’s Threads app? All your questions answeredThe first fully A.I.-generated drug enters clinical trials in human patientsOpenAI is forming a new team to bring ‘superintelligent’ AI under control2023-07-1008 minLondon FuturistsLondon FuturistsThe 4 Cs of SuperintelligenceThe 4 Cs of Superintelligence is a framework that casts fresh light on the vexing question of possible outcomes of humanity's interactions with an emerging superintelligent AI. The 4 Cs are Cease, Control, Catastrophe, and Consent. In this episode, the show's co-hosts, Calum Chace and David Wood, debate the pros and cons of the first two of these Cs, and lay the groundwork for a follow-up discussion of the pros and cons of the remaining two.Topics addressed in this episode include:*) Reasons why superintelligence might never be created*) Timelines for the arrival of superintelligence have...2023-06-1632 minBest Book Summaries 📚 by StoryShotsBest Book Summaries 📚 by StoryShotsSuperintelligence by Nick Bostrom Book Summary, Review and Quotes | Free AudiobookShow notes | PDF & Infographic | Free Audiobook | What happens when artificial intelligence surpasses human intelligence? This is the world that Nick Bostrom explores in his book, Superintelligence. Read 1 million books in minutes. For free. Get the PDF, infographic, full ad-free audiobook and animated version of this summary and more exclusive content on the top-rated StoryShots app: https://www.getstoryshots.com Help us grow and create more amazing content for you! ⭐️⭐️⭐️⭐️⭐️ Don't forget to subscribe, rate and review the StoryShots podcast now.  What should our next book be? Suggest and vote it up on the StoryShots app. 2023-04-0718 minEA TalksEA TalksAI Alignment: An Introduction | Rohin Shah | EAGxOxford 22"You've probably heard that Elon Musk, Stuart Russell, and Stephen Hawking warn of dangers posed by AI. What are these risks, and what basis do they have in AI practice? Rohin Shah will first describe the more philosophical argument that suggests that a superintelligent AI system pursuing the wrong goal would lead to an existential catastrophe. Then, he'll ground this argument in current AI practice, arguing that it is plausible both that we build superintelligent AI in the coming decades, and that such a system would pursue an incorrect goal.View the original talk and video here.2022-12-1752 minCurrent AffairsCurrent AffairsWhy You Don't Need To Worry About "Superintelligent AI" Destroying The World (But Artificial Intelligence Is Still Scary)Some, including both geniuses like Stephen Hawking and nongeniuses like Elon Musk, have warned that artificial intelligence poses a major risk to humankind's future. Some in the "Effective Altruist" community have become convinced that artificial intelligence is developing so rapidly that we could soon create "superintelligent" computers that are so much smarter than us that they could take over and pose a threat to our existence as a species. Books like Nick Bostrom's Superintelligence and Stuart Russell's Human Compatible have warned that we need to get machine intelligence under control before it controls us. Erik J. Larson i...2022-11-0448 minSCIFI SNAKSCIFI SNAKEp. 98: Erik A. Otto, DetonationNår man læser Detonation følger man to historier i hver sin tid. Før og efter atomragnarokket; den såkaldte detonation, der bringer jorden tilbage til et primitivt præ-teknologisk stadie. Bogen starter meget dramatisk i nær fremtid, hvor black ops sikkerhedsspecialisten Axel Kelemen tager os med på mission i kamphelikopter midt i et dronehelvede, mens atommissilerne letter på vej mod masseødelæggelsen. Axel arbejder for tech-firmaet Nadar, ejet af tech-milliardæren Bhavin, som hyrer ham til at redde menneskeheden fra udryddelse. I fremtiden, i bogens andet spor, er verden forandret. E...2022-09-1049 minThe Nonlinear Library: LessWrong Top PostsThe Nonlinear Library: LessWrong Top PostsMorality is Awesome by [anonymous]Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio.This is: Morality is Awesome, published by [anonymous] on the LessWrong.(This is a semi-serious introduction to the metaethics sequence. You may find it useful, but don't take it too seriously.)Meditate on this: A wizard has turned you into a whale. Is this awesome?Is it?"Maybe? I guess it would be pretty cool to be a whale for a day. But only if I can turn back, and if I stay human inside and so...2021-12-1104 minThe Nonlinear Library: LessWrong Top PostsThe Nonlinear Library: LessWrong Top PostsMorality is Awesome by [anonymous]Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Morality is Awesome, published by [anonymous] on the LessWrong. (This is a semi-serious introduction to the metaethics sequence. You may find it useful, but don't take it too seriously.) Meditate on this: A wizard has turned you into a whale. Is this awesome? Is it? "Maybe? I guess it would be pretty cool to be a whale for a day. But only if I can turn back, and if I stay human inside and so...2021-12-1104 minThe Nonlinear Library: LessWrong Top PostsThe Nonlinear Library: LessWrong Top PostsDisentangling arguments for the importance of AI safety by Richard_NgoWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Disentangling arguments for the importance of AI safety, published by Richard_Ngo on the AI Alignment Forum. Crossposted from the AI Alignment Forum. May contain more technical jargon than usual. Note: my views have shifted significantly since writing this post. I now consider items 1, 2, 3, and 6.2 to be different facets of one core argument, which I call the "second species" argument, and which I explore in depth in this report. And I don't really think of 4...2021-12-1113 minThe Nonlinear Library: LessWrong Top PostsThe Nonlinear Library: LessWrong Top PostsDisentangling arguments for the importance of AI safety by Richard_NgoWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio.This is: Disentangling arguments for the importance of AI safety, published by Richard_Ngo on the AI Alignment Forum.Crossposted from the AI Alignment Forum. May contain more technical jargon than usual.Note: my views have shifted significantly since writing this post. I now consider items 1, 2, 3, and 6.2 to be different facets of one core argument, which I call the "second species" argument, and which I explore in depth in this report. And I don't really think of 4...2021-12-1113 minThe Nonlinear Library: Alignment Forum Top PostsThe Nonlinear Library: Alignment Forum Top PostsDisentangling arguments for the importance of AI safety by Richard NgoWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Disentangling arguments for the importance of AI safety, published by Richard Ngo on the AI Alignment Forum. Note: my views have shifted significantly since writing this post. I now consider items 1, 2, 3, and 6.2 to be different facets of one core argument, which I call the "second species" argument, and which I explore in depth in this report. And I don't really think of 4 as an AI safety problem any more. I recently attended...2021-12-1012 minThe Nonlinear Library: Alignment Forum Top PostsThe Nonlinear Library: Alignment Forum Top PostsComprehensive AI Services as General Intelligence by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Comprehensive AI Services as General Intelligence, published by Rohin Shah on the AI Alignment Forum. This is a linkpost for Since the CAIS technical report is a gargantuan 210 page document, I figured I'd write a post to summarize it. I have focused on the earlier chapters, because I found those to be more important for understanding the core model. Later chapters speculate about more concrete details of how AI might develop, as...2021-12-0508 minThe Nonlinear Library: Alignment Forum Top PostsThe Nonlinear Library: Alignment Forum Top PostsThe alignment problem in different capability regimes by Buck ShlegerisI think the alignment problem looks different depending on the capability level of systems you’re trying to align. And I think that different researchers often have different capability levels in mind when they talk about the alignment problem. I think this leads to confusion. I’m going to use the term “regimes of the alignment problem” to refer to the different perspectives on alignment you get from considering systems with different capability levels. (I would be pretty unsurprised if these points had all been made elsewhere; the goal of this post is just to put them all in one plac...2021-12-0507 minThe Nonlinear Library: Alignment Forum Top PostsThe Nonlinear Library: Alignment Forum Top PostsSources of intuitions and data on AGI by Scott GarrabrantWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Sources of intuitions and data on AGI, published by Scott Garrabrant on the AI Alignment Forum. Much of the difficulty in making progress on AI safety comes from the lack of useful feedback loops. We do not have a superintelligent AI to run tests on and by the time we do, it will probably be too late. This means we have to resort to using proxies. In this post, I will highlight what...2021-12-0405 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionReply to Jebari and Lundborg on Artificial Superintelligence by Richard NgoWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Reply to Jebari and Lundborg on Artificial Superintelligence, published by Richard Ngo on the AI Alignment Forum. ebari and Lundborg have recently published an article entitled Artificial superintelligence and its limits: why AlphaZero cannot become a general agent. It focuses on the thorny issue of agency in superintelligent AIs. I’m glad to see more work on this crucial topic; however, I have significant disagreements with their terminology and argumentation, as I outline in this re...2021-11-1708 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionDisentangling arguments for the importance of AI safety by Richard NgoWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Disentangling arguments for the importance of AI safety, published by Richard Ngo on the AI Alignment Forum. Note: my views have shifted significantly since writing this post. I now consider items 1, 2, 3, and 6.2 to be different facets of one core argument, which I call the "second species" argument, and which I explore in depth in this report. And I don't really think of 4 as an AI safety problem any more. I recently attended the 2019 Beneficial AGI...2021-11-1712 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionConclusion to the sequence on value learning by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Conclusion to the sequence on value learning, published by Rohin Shah on the AI Alignment Forum. This post summarizes the sequence on value learning. While it doesn’t introduce any new ideas, it does shed light on which parts I would emphasize most, and the takeaways I hope that readers get. I make several strong claims here; interpret these as my impressions, not my beliefs. I would guess many researchers disagree with the (strength of th...2021-11-1708 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionAlignment Newsletter #40 by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Alignment Newsletter #40, published by Rohin Shah on the AI Alignment Forum. Find all Alignment Newsletter resources here. In particular, you can sign up, or look through this spreadsheet of all summaries that have ever been in the newsletter. The Alignment Forum sequences have started again! As a reminder, treat them as though I had highlighted them. Highlights Reframing Superintelligence: Comprehensive AI Services as General Intelligence (Eric Drexler): This is a huge document; rather than summarize...2021-11-1708 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionReframing Superintelligence: Comprehensive AI Services as General Intelligence by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Reframing Superintelligence: Comprehensive AI Services as General Intelligence, published by Rohin Shah on the AI Alignment Forum. This is: Reframing Superintelligence: Comprehensive AI Services as General Intelligence, published by on the AI Alignment Forum. This is a linkpost for Since the CAIS technical report is a gargantuan 210 page document, I figured I'd write a post to summarize it. I have focused on the earlier chapters, because I found those to be more important for understanding...2021-11-1708 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionAI safety without goal-directed behavior by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: AI safety without goal-directed behavior, published by Rohin Shah on the AI Alignment Forum. When I first entered the field of AI safety, I thought of the problem as figuring out how to get the AI to have the “right” utility function. This led me to work on the problem of inferring values from demonstrators with unknown biases, despite the impossibility results in the area. I am less excited about that avenue because I am pess...2021-11-1705 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionWill humans build goal-directed agents? by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: Will humans build goal-directed agents?, published by Rohin Shah on the AI Alignment Forum. In the previous post, I argued that simply knowing that an AI system is superintelligent does not imply that it must be goal-directed. However, there are many other arguments that suggest that AI systems will or should be goal-directed, which I will discuss in this post. Note that I don’t think of this as the Tool AI vs. Agent AI ar...2021-11-1708 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionThe Alignment Newsletter #1: 04/09/18 by Rohin ShahWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: The Alignment Newsletter #1: 04/09/18, published by Rohin Shah on the AI Alignment Forum. Highlights Specification gaming examples in AI (Victoria Krakovna): A list of examples of specification gaming, where an algorithm figures out a way to literally satisfy the given specification which does not match the designer's intent. Should you read it? There were several examples I hadn't heard of before, which were pretty entertaining. Also, if you have any examples that aren't already listed, it...2021-11-1608 minThe Nonlinear Library: Alignment SectionThe Nonlinear Library: Alignment SectionThe alignment problem in different capability regimes by Buck ShlegerisWelcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: The alignment problem in different capability regimes, published by Buck Shlegeris on the AI Alignment Forum. I think the alignment problem looks different depending on the capability level of systems you’re trying to align. And I think that different researchers often have different capability levels in mind when they talk about the alignment problem. I think this leads to confusion. I’m going to use the term “regimes of the alignment problem” to refer to the d...2021-11-1607 min