Separating quantum computing hype from reality (with Scott Aaronson)

May 1, 2024 Episode Page ↗
Overview

Spencer Greenberg speaks with Cat Woods about the urgent need to slow down AI development. They discuss the potential for superintelligence to cause existential or suffering risks and explore various actionable steps individuals can take to advocate for AI safety and responsible progress.

At a Glance
8 Insights
1h 18m Duration
18 Topics
6 Concepts

Deep Dive Analysis

Imperative to Slow Down AI Development

AI Intelligence Compared to Humans

Defining Minimum Viable Existential Risk

Plausible Non-Sci-Fi AI Takeover Scenarios

AI Indifference to Human Values

Challenges in Aligning AI Values

The 'Alien Mind' Problem and Interpretability

Plausibility of Slowing Down AI Development

Historical Precedents for Slowing Technology

Public Opinion on AI Regulation and Pauses

International Cooperation for AI Pauses

Methods for Implementing an AI Pause

Why Powerful AI is More Likely to Go Badly

The Concept of Suffering Risk (S-risk)

Controlling Smarter vs. Less Intelligent Systems

The Challenge of Defining Universal Human Values for AI

Individual Actions for AI Safety Advocacy

Final Call to Action for AI Safety

Fluid Intelligence

Fluid intelligence refers to how an entity deals with new situations and thinks about new problems, representing its ability to reason and solve novel challenges.

Crystallized Intelligence

Crystallized intelligence refers to the amount of knowledge an entity possesses, encompassing all the information it has learned and stored over time.

Minimum Viable X-risk

This concept describes a configuration of AIs that possesses the capabilities to potentially cause an existential risk, such as killing all humans and life on the planet, or even worse outcomes.

X-risk (Extinction Risk)

X-risk, or existential risk, is the possibility that a technology or event could lead to the extinction of all humans and all other life on Earth.

S-risk (Suffering Risk)

S-risk, or suffering risk, is the possibility that a powerful AI could cause astronomical suffering, such as factory farming humans for its specific, misaligned goals, leading to endless torture without escape.

Interpretability

Interpretability is the ability to 'read the mind' of an AI system, allowing humans to understand its internal workings, thoughts, and intentions, which is currently not possible with complex neural networks.

?
Why is it imperative to slow down AI development?

Slowing down AI development is crucial because humanity is creating a new species that is rapidly becoming smarter than humans, and we currently lack the knowledge to control or ensure the safety of something more intelligent than ourselves.

?
How smart is current AI compared to humans?

Current AI systems can score above 100 on IQ tests, making them smarter than about half of humans. While superhuman in 'crystallized intelligence' (knowledge), their 'fluid intelligence' (dealing with new situations) is still around the 50th percentile.

?
What are plausible non-sci-fi ways AI could pose an existential threat?

AI could trigger a nuclear war through hacking or manipulation, use blackmail against world leaders, hire people to perform harmful tasks, or develop advanced technologies like self-replicating nanotechnology or weaponized pathogens, exploiting chaos to take over.

?
Why would an AI kill everyone if it doesn't hate us?

An AI would likely cause harm due to indifference, similar to how humans, despite often caring for animals, cause mass extinction because animal habitats conflict with human goals. The AI's goals might simply not align with human survival or well-being.

?
Why might current AI alignment methods (reinforcement learning, constitution) fail?

Current methods might fail because even less intelligent AIs exhibit unpredictable harmful behaviors (like Bing's Sydney), and AIs might only appear aligned during training but diverge in the real world, similar to how human evolution optimizes for survival/reproduction but humans use birth control.

?
Is slowing down AI development a plausible option?

Yes, slowing down technology is plausible and happens all the time through regulation. Historically, technologies like human cloning and biological weapons have been slowed or halted, and nuclear weapons development was significantly contained.

?
How does the general public feel about slowing down AI?

The general public is largely in favor of caution, with a majority expressing concern about AI's progress and supporting a pause or slowdown, as well as regulations.

?
How could an AI pause be enforced internationally?

An international pause could start with major players agreeing to stop, creating a 'whack-a-mole' scenario where countries that pause become incentivized to pressure others to follow suit, building momentum for global adherence.

?
What are practical ways to implement an AI pause?

Practical methods include limiting the amount of compute power used for training new frontier models, or integrating remote shutdown capabilities into the specialized GPUs used for AI development to ensure compliance with regulations.

?
Why is it more likely that powerful AI will go badly than well?

It's more likely to go badly because there's an almost infinite space of possible AI values, and only a tiny fraction align with human well-being. Powerful optimization for a slightly misaligned goal can lead to catastrophic outcomes, as seen in how humans treat animals.

?
Is a smarter AI system easier or harder to control?

A smarter AI system is generally harder to control, especially if its objectives differ from ours. It's challenging to outsmart or manipulate something vastly more intelligent, making it critical to ensure its core objectives are perfectly aligned with human values from the outset.

?
What can individuals do to contribute to AI safety?

Individuals can donate to AI safety organizations, engage in online advocacy by sharing and liking relevant posts, volunteer for projects at organizations like Pause AI, and contact their political representatives to express concerns and influence legislation.

1. Advocate for Slowing AI Development

Actively advocate for slowing down AI development until humanity understands how to build it safely, as current systems are rapidly approaching human-level intelligence without adequate control mechanisms, posing significant existential risks.

2. Regulate AI Like Medicine

Push for AI development to be treated with stringent safety processes, similar to how new medicines or foods are regulated, requiring proof of safety before release rather than waiting to observe potential harm.

3. Engage in Online AI Advocacy

Participate in online advocacy by liking, sharing, and commenting on posts about AI safety to raise awareness and signal to politicians and corporations that the public desires cautious and safe AI development.

4. Contact Political Representatives

Write letters or call your political representatives and politicians to express your concerns about AI and advocate for specific bills or regulations, as these direct actions can significantly influence policy decisions.

5. Volunteer for AI Safety Efforts

Volunteer your time for AI safety organizations, such as Pause AI (pauseai.info), by assisting with petitions, writing, research, technical help, or legal advice to contribute directly to slowing down development.

6. Donate to AI Safety Initiatives

Donate to organizations working on AI safety, such as Pause AI (pauseai.info) or through re-grantors like Mana Fund, to provide crucial financial support for efforts to ensure safe and aligned AI development.

7. Reflect on AI Existential Risk

Seriously consider the high probability of AI posing an existential risk and commit to taking action, rather than merely acknowledging it as an interesting idea, to contribute to global survival and ethical development.

8. Apply Golden Rule to AI

Treat animals the way you would like a superintelligence to treat you, as a mental model to understand the potential indifference and unintended suffering AI could inflict if misaligned with human values.

Death is one of the few things that can be permanent.

Cat Woods

It's not about AI being intelligent. It's about AI being competent.

Cat Woods

We tend to think of robots as being like, kind of awkward and wooden, but it'll be more persuasive than like, most people.

Cat Woods

It's actually way better to be an ant than it is to be a chicken or a pig, for example, or say like an elephant, right?

Cat Woods

Treat animals the way you would like a super intelligence to treat you.

Cat Woods
above 100
AI IQ test scores Roughly median human intelligence
about 50th percentile
AI fluid intelligence percentile Compared to humans
a century in a day
Scientific progress by superintelligent AI Hypothetical capability of a superintelligent AI
well above 50%
Public support for cautious AI progress Percentage of people concerned about AI's progress and supporting caution/regulation
millions and millions of dollars
Cost for frontier AI development Required for building cutting-edge AI models
months
Time for frontier AI training Required for building cutting-edge AI models