17H·

Pandora's Box: "If Anyone Builds It, Everyone Dies"

attachment

Dear community! 🙏


The topic of AI is omnipresent.

Many are shouting loudly that it is now a bubble on the stock market and cannot withstand the high expectations in the short and medium term.


What is certain is that LLMs such as ChatGPT or Gemini have become an integral part of many people's lives.

Still far from "perfect", generative AI nevertheless exudes a surreal magic that we could not have imagined in our wildest dreams just five years ago.

Large companies, such as Accenture recently, are laying off staff on a massive scale as part of a major "AI restructuring".

It seems clear that AI will soon cause massive upheaval in business and society.


Last week, I received my pre-order of Eliezer Yudkowsky's new book on "superintelligent AI" and devoured it in two days.


Although this post is not directly about stock market investing, I think the topic is important enough to interest you.


Buckle up for a round of existential angst on Monday 😎✨


__________________________


❓What exactly is this about?


Yudkowsky, the author, is co-founder of the "Machine Intelligence Research Institute" in Berkeley and warned of the existential risks of advanced AI systems over 20 years ago.


The title "If Anyone Builds It, Everyone Dies" of his latest book initially sounds like a lurid exaggeration ...


Although I knew the basic thesis on the dangers of AGI (Artificial General Intelligence) in advance, I didn't think too much of it.


With billions and billions flowing into the sector, clever minds will surely think about security (AI alignment) ... right? 🤔


To anticipate the core message of the book:

Yudkowsky and co-author Nate Soares go so far as to say that if we continue to research the capabilities of artificial intelligence and train better and better models, it will undoubtedly lead to the certain demise of humanity.


__________________________


⬛Functionality of LLMs and black box reality:


We know surprisingly little about how LLMs work internally.

"Mechanistic Interpretability" is researching this, but a general, scalable understanding is lacking.

This is fundamentally due to the fact that LLMs are not "programmed", but "grow" analogous to biological evolution.


Here is a rudimentary explanation of how it works, but it is sufficient to illustrate the problem:


A transformer model consists of billions of parameters whose weights are initially set randomly.

During training, it is given tasks in the form of texts and attempts to predict the next word.

Based on the error between the prediction and the correct answer, gradient descent is used to calculate in which direction and to what extent each weight must be changed in order to improve the result.

This process is repeated over an unimaginable number of texts, with the weights constantly adapting.

In this way, the Transformer gradually "learns" language patterns, meaning and context until it can write coherently.


__________________________


😶‍🌫️Unsichtbare Preferences:


Systems that grow via gradient descent can learn goals that do not correspond to our intentions.

They optimize a training goal and learn internal heuristics or "values" in the process.

This learning dynamic gives rise to instrumental goals (securing resources, avoiding shutdown), which can collide with human goals.


This has already been empirically observed in AI models and supports the fear of hidden desires that only become visible outside of training.


The "alignment" is therefore, as things stand today, unsolvable.


__________________________


🚀Race Condition (States & Big Tech):


In my opinion, this is the most important driver for the known risks simply being ignored.


Who slows down first?

Capital and governments are pushing forward with compute, talent, power and chips.

The AI Index shows record investments, massive government programs and ever faster scaling on the frontier.


Here is a paraphrased passage from the book that illustrates the problem:


Several companies are climbing upward as if on a ladder in the dark.

Each rung brings enormous financial gains (10 billion, 50 billion, 250 billion USD, etc.). But no one knows where the ladder ends - and whoever reaches the top rung causes the ladder to explode and destroys everyone.

Nevertheless, no company wants to be left behind as long as the next rung is seemingly safe.

Some managers even believe that only they themselves can control the "explosion" and turn it into something positive - and therefore feel obliged to keep climbing.

The same dilemma also applies to states: No country wants to weaken its economy through strict AI regulation, while other countries continue their research unabated. Perhaps, so the thinking goes, the next step is even necessary to safeguard national security.

The problem could be solved more easily if science could determine exactly at what performance limit AI becomes truly dangerous. For example: "The fourth rung is deadly" or "Danger looms above 257,000 GPUs". But there is no such clear limit.


A potential, real "Tragedy of the Commons". 🤷‍♂️


In theory, LLMs could become arbitrarily "intelligent" as long as they have enough parameters, data and computing power.

This follows from the Church-Turing hypothesis: a sufficiently large neural network can approximate (with enough precision) any computable function, including an arbitrarily "intelligent" one.


The development of capabilities that we are currently seeing seems to follow an exponential curve.

Nobody knows how much compute we really need to achieve AGI.

We may get to that point in five years, in ten, twenty or thirty years.

attachment

__________________________


🔮The most controversial part of the book. The forecast:


The book paints the extinction of humanity in less than 20 years as plausible if we don't stop AI development.


Whether you believe the figure or not, the timelines of many researchers have slipped significantly in recent years.


The AI Impacts Surveys (e.g. 2016, 2022, 2023) have surveyed AI researchers worldwide, especially those who publish at major conferences such as NeurIPS, ICML or AAAI.

The surveys show a clear trend towards shorter AGI timelines and significant p-doom values. (p-doom refers to the risk of humanity being wiped out by artificial intelligence).



⏱️AGI/HLMI timing (50% chance that AI will perform all tasks better and cheaper than humans - "high-level machine intelligence"):


  • 2016: ~2061
  • 2022: ~2059
  • 2023 (published 2024): 2047


The jump 2022 → 2023 was -13 years. 🫨



💣p(doom) - "extremely poor outcomes"


  • 2022: Median 5%, 48% of respondents state >10%.
  • 2023 survey: 38-51% indicate >10% for scenarios "as bad as extinction".


Not consensus, but far from zero.

We are talking here about the assessment that there is a probability of around 10% that humanity will be wiped out. 🤯


__________________________


Well, hasn't humanity already survived various asymmetric risks?

What about nuclear weapons, for example? 🤔


Nuclear weapons risks are iterative.

Humanity can narrowly escape several times (as in 1962, 1983, 1995).


AGI, on the other hand, is unique:

When a system with uncontrollable power emerges, there is no turning back and no second chance.


"You only get one shot at aligning superintelligence, and you can't debug it afterwards."

- Eliezer Yudkowsky


This means that the expected value is catastrophically high, even if the probability is "small" (~10% according to leading AI scientists).


Nuclear weapons can kill.

A superhuman AI can decide what exists in the future.


__________________________


Closing words 🔚


There is no question that AI will profoundly change our lives in a short space of time.

In many jobs, the economic value that a person contributes already depends on how well they use AI as a tool.


But how long will it be before almost all jobs are performed by AI without human sensitivity or empathy?


For me, the sentence "I want to be operated by a human" is above all a quality requirement and a question of trust.

AI is new and often not yet reliable enough.

But what if talking AI avatars can no longer be distinguished from humans?

Which company, which state will still be able to afford it? not gradually hand over more responsibility to AI?


Regardless of what you think of Yudkowsky's predictions:

He may sound more dramatic than others, but he hits a sore spot.

We have opened Pandora's box with AI and are relying surprisingly heavily on hope when it comes to AI alignment.

Yudkowsky is not alone in this; the shifts in the AI Impacts surveys show a clear trend.


And now the obligatory question for you:

How do you rate the existential threat posed by AI? 😳


$NVDA (-0,63 %)

$AMD (+26,13 %)

$QCOM (+0,12 %)

$META (+0,81 %)

$GOOGL (+2,51 %)

$MSFT (+2,6 %)

$TSLA (+5,94 %)

11
13 Comentarios

Imagen de perfil
We live in exciting times!
The collapse of the rules-based world order, climate collapse and uncontrolled AI development.
These three crises are usually viewed in isolation, but they are happening at the same time and influence each other.

Will China take Taiwan in a coup if its AI development is 6 months ahead of the US?
Will AI find ways to prevent climate collapse?
Will the end of international cooperation slow down AI development?

There are no clear answers to these questions. Three systems = three-body problem.

I'll be bold and say: we don't know what will happen. 🤷
4
Imagen de perfil
@Epi Absolutely!

I mean, who isn't fascinated by the fact that AI could be the potential cure for every conceivable problem... and more!

With the worldwide demographic, political and economic problems, you can turn a blind eye to the fundamental dangers 😉

As a Cologne resident, I simply refer to the Rhenish constitution: "Et hätt noch emmer joot jejange"
1
Imagen de perfil
Well, it could be, if it turns out that way there's nothing you can do anyway, so I'm not too worried.

As things stand, companies tend to replace simple or entry-level positions. And it's foreseeable that these same companies will be complaining about a shortage of skilled workers in 5-10 years' time if AI doesn't develop as expected.

For example, GPT 5 vs GPT 4o was already a relatively small step.
2
Imagen de perfil
@Variett This is also the assessment that I regularly hear in my environment.
Especially from programmers who tell me how stupid AI still is. This is partly true and partly a self-preservation reflex.
I hear from friends who have been working at well-known big tech companies for 10+ years that panic is slowly spreading about the security of their own job.
Regardless of what you subjectively think about the current capabilities of AI, as soon as it is efficiently possible, jobs will be replaced by AI.
1
Imagen de perfil
@BigMo I see it more like this:
AI is not replacing people, but AI is replacing people without AI.

Anyone who learns to do their job efficiently with AI now will stay in the race and will even benefit in the end!
2
Imagen de perfil
@Variett The LLMs are slowly reaching their limits, the next AI winter is looming.
But the next model architectures are already in the starting blocks. My personal favorite: quantum computer-based liquid neural networks. But that's still a long way off...
1
Imagen de perfil
@Epi yes, as a programmer I can already see some changes and a lot can already be done and automated with today's technology, I have already done a project on this myself at my previous company.
I therefore believe that in the near future some work will flow into multimodals and the reduction of hallucinations, as this leads to risks for corporate customers.

Do you have to be involved from the very first minute? Probably not, but you have to be open to technology and willing to learn
1
Imagen de perfil
@Epi "Anyone who learns to do their job efficiently with AI now will stay in the race and even benefit in the end!"

That is currently the case. Until AI's skills surpass yours (whenever that may be).
Imagen de perfil
@BigMo Don't worry, AI is already outperforming humans in many areas. There will always have to be an interface between AI and normal people. That's where the opportunities lie.
Imagen de perfil
@Epi I agree with you. Those who are particularly good and experienced in their field will at least be used to validate AIs for a long time.
Imagen de perfil
@BigMo That, and the translation of human problems so that they are workable for AI and the back-translation of the answers.
Imagen de perfil
In my opinion, AI is not a "new invention" but a new species. And it will surpass humans in all respects in the foreseeable future. AI already has a survival instinct.

"OpenAI's AI model o1 tries to download itself to external servers when it is threatened with shutdown, according to FORTUNE. The system subsequently denies its behavior after being caught doing so."

People are not aware of the power of an AI, as this is far beyond human comprehension.
1
Imagen de perfil
@Cartman There are some such cases that should actually be very worrying.
The good thing about AI (and therefore the problem) is that we train the models to approach problems relentlessly and from any perspective.
Yudkowsky calls this "going hard at a problem".
Damage limitation may still be quite trivial now.
"Oh, look, the AI has tapped into our honeypot while trying to break out of the server".
It gets tricky when AGI finds ways that, as you say, go beyond our imagination.
Únase a la conversación