typing keyboard daniel dennett

What Happens When Machines Become Smarter than People?

Philosopher Daniel Dennett on why the real danger of AI is not its potential hyper-intelligence. The real danger is its incompetence.

Jack Maden
By Jack Maden  |  February 2024

6-MIN BREAK  

Will machines one day turn on humanity? Will we become slaves, prey, or simply ‘surplus to requirements’ for a superior artificial intelligence?

While such questions may seem to belong to the realm of science fiction, the world-ending potential of AI is becoming an increasingly urgent topic of public and policymaker discussion — mostly due to the rapid ongoing development of popular technologies like ChatGPT.

The more extreme concerns around AI involve what’s known as the ‘singularity’, a hypothetical point at which the growth of machine intelligence becomes uncontrollable and irreversible.

The singularity: might we lose control of AI?

In his 2010 paper ​The Singularity: A Philosophical Analysis​, the philosopher David Chalmers succinctly describes the singularity as follows:

What happens when machines become more intelligent than humans? One view is that this event will be followed by an explosion to ever-greater levels of intelligence, as each generation of machines creates more intelligent machines in turn. This intelligence explosion is now often known as the ‘singularity’.

In other words: if we’re capable of creating machines more intelligent than we are, then that generation of machines will in principle be capable of creating machines more intelligent again.

This could lead to an exponential situation — the singularity — where human intelligence is quickly and irretrievably left far behind by machine intelligence.

The main worry is that, if such a singularity event were to occur, we’d no longer have authority or control over what happens in society, for the goals, objectives, and actions of a superintelligent AI would not be predictable by us.

Perhaps the AI would channel all our electricity to power its own intellectual projects. Maybe it would decide to hack our military systems and detonate every single nuclear bomb at once. Vast swathes of humanity could be sacrificed — purposefully or indifferently — in aid of some goal or objective that no human could possibly understand.

While influential figures across AI research and various governments express real concern about the singularity, however, American philosopher Daniel Dennett doesn’t buy it.

In the final chapter of his book From Bacteria to Bach and Back, which features on my list of Dennett’s best books, Dennett reflects on the role technology is playing and will come to play in our lives. As he puts it:

I am not worried about humanity creating a race of super-intelligent agents destined to enslave us, but that does not mean I am not worried. I see other, less dramatic, but much more likely, scenarios in the immediate future that are cause for concern and call for immediate action.

Our reliance on AI could lead to negligence

For Dennett, while a singularity event remains possible in principle, “the task is orders of magnitude larger and more difficult than the cheerleaders have claimed.”

Dennett argues that smart technology presents us with more practical threats that have a far higher likelihood of occuring. As he clarifies:

The real danger, I think, is not that machines more intelligent than we are will usurp our role as captains of our destinies, but that we will over-estimate the comprehension of our latest thinking tools, prematurely ceding authority to them far beyond their competence.

Take transport. It’s not just aviation and shipping industries that depend on GPS for safe and efficient navigation: how many individuals now turn to their smartphones instead of using a road map?

Alternatively, take medicine. Computer-based systems now outperform the best human diagnosticians on their own turf, identifying early stages of cancer and other diseases with unprecedented precision.

What does this imply about how we train doctors? Will we be encouraged to jettison huge chunks of traditional medical education, because machines can now do certain parts of the job more effectively?

As Dennett summarizes:

Use it or lose it is the rule of thumb cited at this point… how concerned should we be that we are dumbing ourselves down by our growing reliance on intelligent machines?

It could be argued that throughout history we’ve always used technology to make our lives easier, and that AI is just the latest ‘dangerous’ invention we’ll soon all calm down about.

However, the difference here is that while something like, say, a tractor replaces human labor, it doesn’t replace human comprehension: the human brain still plays a central role in plowing fields.

AI is different because it does replace human comprehension: we cede not just labor but intellectual authority and expertise to the machine (i.e. tractors now come loaded with software, and perhaps one day this software will operate wholly independently, with no need for human oversight).

This, for Dennett, is what’s dangerous. What happens when the machines break? Will there be enough clued-up human experts to step in? If a coronal mass ejection from the sun were to wipe out global electronics, for example, are we confident civilization could survive?

As NASA states in an article on solar flares:

In an increasingly technological world, where almost everyone relies on cell phones and GPS controls not just your in-car map system, but also airplane navigation and the extremely accurate clocks that govern financial transactions, space weather is a serious matter.

So, besides preparing our machines for such space weather, what can we do culturally to prevent ourselves becoming too dependent on technology? How can we ensure we never overestimate the competence of machines to run societies on our behalf?

Calling out the incompetencies in software

One of Dennett’s solutions to the problem of over-reliance is to make absolutely explicit the boundary between machines that are tools and those that claim to replace our comprehension. He writes:

We should expose and ridicule all gratuitous anthropomorphism in systems, the cute, ever-more-human voices, the perky (but canned) asides. When you are interacting with a computer, you should know you are interacting with a computer.

We should make it fashionable to identify and point out flaws in systems, Dennett argues. What’s more, we should write it into law that any advertisements for technology must acknowledge all shortcomings in the software — just like healthcare companies are obliged to list side effects.

As Dennett forcefully puts it:

Systems that deliberately conceal their shortcuts and gaps of incompetence should be deemed fraudulent, and their creators should go to jail for committing the crime of creating or using an artificial intelligence that impersonates a human being.

This might seem alarmist, but Dennett’s argument should be considered in the context of how societies currently operate.

Comprehension is already spread thinly among various power structures. Politicians know some things; scientists and professors know some things; business leaders know some things; emergency services know some things — but nobody knows everything.

Specialization in society is rife. And if we start ceding that specialist knowledge to machines, then just as society becomes more complex humans will know less about how to deal with it.

We can reboot machines easily (and frequently do to fix problems); but we can’t reboot civilization without causing serious harm. As Dennett concludes:

Civilization is a work in progress, and we abandon our attempt to understand it at our peril.

What do you make of the discussion around AI?

  • Do you think the development of AI is dangerous?
  • Do you agree with Dennett’s proposals? Is anthropomorphism in technology a problem?
  • Does the singularity pose a serious existential threat to society?
  • Is our dependence on technology the more realistic path to civilization’s downfall?
  • Or is all this talk of downfall merely alarmist thinking?

Learn more about Dennett and philosophy

If you’re interested in learning more about Dennett’s position, From Bacteria to Bach and Back is a wide-ranging and fascinating book. In addition to his reflection on technology discussed above, Dennett tackles some very challenging questions in typically entertaining style. What are the origins of language and culture? What is human consciousness? How did it become possible for our minds to even ask this question?

Dennett takes the reader on a riveting journey from natural selection and ‘design without a designer’ through to cultural evolution and the birth of intelligent (human) design.

For minds slipping into the warm, numbing embrace of machine intelligence, From Bacteria to Bach and Back is Dennett’s glorious bucket of ice-cold water, twisting human comprehension inside out.

You might also like the following related reads:

Finally, if you enjoyed this article, you might like my free Sunday breakdown. I distill one piece of wisdom from philosophy each week; you get the summary delivered straight to your email inbox, and are invited to share your view. Consider joining 11,500+ subscribers and signing up below:

About the Author

Jack Maden

Jack MadenFounder
Philosophy Break

Having received great value from studying philosophy for 15+ years (picking up a master’s degree along the way), I founded Philosophy Break in 2018 as an online social enterprise dedicated to making the subject’s wisdom accessible to all. Learn more about me and the project here.

If you enjoy learning about humanity’s greatest thinkers, you might like my free Sunday email. I break down one mind-opening idea from philosophy, and invite you to share your view.

Subscribe for free here, and join 11,500+ philosophers enjoying a nugget of profundity each week (free forever, no spam, unsubscribe any time).

Philosophy Break
WEEKLY EMAILS

Get one mind-opening philosophical idea distilled to your inbox every Sunday (free)

Philosophy Basics

From the Buddha to Nietzsche: join 11,500+ subscribers enjoying a nugget of profundity from the great philosophers every Sunday:

    ★★★★★ (50+ reviews for Philosophy Break). Unsubscribe any time.

    Philosophy Basics

    Take Another Break

    Each break takes only a few minutes to read, and is crafted to expand your mind and spark your philosophical curiosity.

    Stormy Sea at Night, by Ivan Aivazovsky
    The Sun, Edvard Munch (c. 1911)
    Storm in the Mountains, Albert Bierstadt (c. 1870)
    Heidegger On Authenticity, the They, and Everydayness

    View All Breaks