456 lines
23 KiB
Plaintext
456 lines
23 KiB
Plaintext
[1][https]
|
||
|
||
[2]Programmable Mutter
|
||
|
||
Subscribe
|
||
Sign in
|
||
Share this post
|
||
[https]
|
||
|
||
What OpenAI shares with Scientology
|
||
|
||
www.programmablemutter.com
|
||
Copy link
|
||
Facebook
|
||
Email
|
||
Note
|
||
Other
|
||
|
||
What OpenAI shares with Scientology
|
||
|
||
Strange beliefs, fights over money and bad science fiction
|
||
|
||
[12][https]
|
||
[13]Henry Farrell
|
||
Nov 20, 2023
|
||
73
|
||
Share this post
|
||
[https]
|
||
|
||
What OpenAI shares with Scientology
|
||
|
||
www.programmablemutter.com
|
||
Copy link
|
||
Facebook
|
||
Email
|
||
Note
|
||
Other
|
||
[20]
|
||
17
|
||
[21]
|
||
Share
|
||
[22]
|
||
[https]
|
||
|
||
When Sam Altman was ousted as CEO of OpenAI, some hinted that lurid depravities
|
||
lay behind his downfall. Surely, OpenAI’s board wouldn’t have toppled him if
|
||
there weren’t some sordid story about to hit the headlines? But the [23]
|
||
reporting all seems to be saying that it was God, not Sex, that lay behind
|
||
Altman’s downfall. And Money, that third great driver of human behavior, seems
|
||
to have driven his attempted return and his [24]new job at Microsoft, which is
|
||
OpenAI’s biggest investor by far.
|
||
|
||
As the NYT describes the people who pushed Altman out:
|
||
|
||
Thanks for reading Programmable Mutter! Subscribe for free to receive new
|
||
posts. And if you want to support my work, [25]buy my and Abe Newman’s new
|
||
book, [26]Underground Empire, and sing its praises (so long as you actually
|
||
liked it), on Amazon, Goodreads, social media and everywhere else that people
|
||
find out about good books.
|
||
|
||
[35][ ]
|
||
Subscribe
|
||
Ms. McCauley and Ms. Toner [HF - two board members] have ties to the
|
||
Rationalist and Effective Altruist movements, a community that is deeply
|
||
concerned that A.I. could one day destroy humanity. Today’s A.I. technology
|
||
cannot destroy humanity. But this community believes that as the technology
|
||
grows increasingly powerful, these dangers will arise.
|
||
|
||
McCauley and Toner reportedly worried that Altman was pushing too hard, too
|
||
quickly for new and potentially dangerous forms of AI (similar fears led some
|
||
OpenAI people to bail out and found a competitor, Anthropic, a couple of years
|
||
ago). The FT’s reporting [37]confirms that the fight was over how quickly to
|
||
commercialize AI
|
||
|
||
The back-story to all of this is actually much weirder than the average sex
|
||
scandal. The field of AI (in particular, its debates around Large Language
|
||
Models (LLMs) like OpenAI’s GPT-4) is profoundly shaped by cultish debates
|
||
among people with some very strange beliefs.
|
||
|
||
As LLMs have become increasingly powerful, theological arguments have begun to
|
||
mix it up with the profit motive. That explains why OpenAI has such an unusual
|
||
corporate form - it is a non-profit, with a for-profit structure retrofitted on
|
||
top, sweatily entangled with a profit-maximizing corporation (Microsoft). It
|
||
also plausibly explains why these tensions have exploded into the open.
|
||
|
||
********
|
||
|
||
I joked on Bluesky that the OpenAI saga was as if “the 1990s browser wars were
|
||
being waged by rival factions of Dianetics striving to control the future.”
|
||
Dianetics - for those who don’t obsess on the underbelly of American
|
||
intellectual history - was the 1.0 version of L. Ron Hubbard’s Scientology.
|
||
Hubbard [38]hatched it in collaboration with the science fiction editor John W.
|
||
Campbell (who had a major science fiction award named after him until 2019,
|
||
when his racism finally caught up with his reputation).
|
||
|
||
The AI safety debate too is an unintended consequence of genre fiction. In
|
||
1987, multiple-Hugo award winning science-fiction critic Dave Langford [39]
|
||
began a discussion of the “newish” genre of cyberpunk with a complaint about an
|
||
older genre of story on information technology, in which “the ultimate computer
|
||
is turned on and asked the ultimate question, and replies `Yes, now there is a
|
||
God!'
|
||
|
||
However, the cliche didn’t go away. Instead, it cross-bred with cyberpunk to
|
||
produce some quite surprising progeny. The midwife was the writer Vernor Vinge,
|
||
who proposed a revised meaning for “singularity.” This was a term already
|
||
familiar to science fiction readers as the place inside a black hole where the
|
||
ordinary predictions of physics broke down. Vinge suggested that we would soon
|
||
likely create true AI, which would be far better at thinking than baseline
|
||
humans, and would change the world in an accelerating process, creating a
|
||
historical [40]singularity, after which the future of the human species would
|
||
be radically unpredictable.
|
||
|
||
These ideas were turned into novels by Vinge himself, including A Fire Upon the
|
||
Deep (fun!) and Rainbow’s End (weak!). Other SF writers like Charles Stross
|
||
wrote novels about humans doing their best to co-exist with “weakly godlike”
|
||
machine intelligence (also fun!). Others who had no notable talent for writing,
|
||
like the futurist Ray Kurzweil, tried to turn the Singularity into the
|
||
foundation stone of a new account of human progress. I still possess a
|
||
mostly-unread copy of Kurzweil’s mostly-unreadable magnum opus, The Singularity
|
||
is Near, which was distributed en masse to bloggers like meself in an early
|
||
2000s marketing campaign. If I dug hard enough in my archives, I might even be
|
||
able to find the message from a publicity flack expressing disappointment that
|
||
I hadn’t written about the book after they sent it. All this speculation had a
|
||
strong flavor of end-of-days. As the Scots science fiction writer, Ken MacLeod
|
||
memorably put it, the Singularity was the “Rapture of the Nerds.” Ken, being
|
||
the [41]offspring of a Free Presbyterian preacher, knows a millenarian religion
|
||
when he sees it: Kurzweil’s doorstopper should really have been titled The
|
||
Singularity is Nigh.
|
||
|
||
Science fiction was the gateway drug, but it can’t really be blamed for
|
||
everything that happened later. Faith in the Singularity has roughly the same
|
||
relationship to SF as UFO-cultism. A small minority of SF writers are true
|
||
believers; most are hearty skeptics, but recognize that superhuman machine
|
||
intelligences are (a) possible) and (b) an extremely handy engine of plot. But
|
||
the combination of cultish Singularity beliefs and science fiction has
|
||
influenced a lot of external readers, who don’t distinguish sharply between the
|
||
religious and fictive elements, but mix and meld them to come up with strange
|
||
new hybrids.
|
||
|
||
Just such a syncretic religion provides the final part of the back-story to the
|
||
OpenAI crisis. In the 2010s, ideas about the Singularity cross-fertilized with
|
||
notions about Bayesian reasoning and some really terrible fanfic to create the
|
||
online “rationalist” movement mentioned in the NYT.
|
||
|
||
I’ve never read a text on rationalism, whether by true believers, by
|
||
hangers-on, or by bitter enemies (often erstwhile true believers), that really
|
||
gets the totality of what you see if you dive into its core texts and
|
||
apocrypha. And I won’t even try to provide one here. It is some Very Weird Shit
|
||
and there is really great religious sociology to be written about it. The
|
||
fights around [42]Roko’s Basilisk are perhaps the best known example of
|
||
rationalism in action outside the community, and give you some flavor of the
|
||
style of debate. But the very short version is that [43]Eliezer Yudkowsky, and
|
||
his multitudes of online fans embarked on a massive collective intellectual
|
||
project, which can reasonably be described as resurrecting David Langford’s
|
||
hoary 1980s SF cliche, and treating it as the most urgent dilemma facing human
|
||
beings today. We are about to create God. What comes next? Add Bayes’ Theorem
|
||
to Vinge’s core ideas, sez rationalism, and you’ll likely find the answer.
|
||
|
||
The consequences are what you might expect when a crowd of bright but rather
|
||
naive (and occasionally creepy) computer science and adjacent people try to
|
||
re-invent theology from first principles, to model what human-created gods
|
||
might do, and how they ought be constrained. They include the following,
|
||
non-comprehensive list: all sorts of strange mental exercises, postulated
|
||
superhuman entities benign and malign and how to think about them; the jumbling
|
||
of parts from fan-fiction, computer science, home-brewed philosophy and ARGs to
|
||
create grotesque and interesting intellectual chimeras; Nick Bostrom, and a
|
||
crew of very well funded philosophers; Effective Altruism, whose fancier
|
||
adherents often prefer not to acknowledge the approach’s somewhat disreputable
|
||
origins.
|
||
|
||
All this would be sociologically fascinating, but of little real world
|
||
consequence, if it hadn’t profoundly influenced the founders of the
|
||
organizations pushing AI forward. These luminaries think about the technologies
|
||
that they were creating in terms that they have borrowed wholesale from the
|
||
Yudkowsky extended universe. The risks and rewards of AI are seen as largely
|
||
commensurate with the risks and rewards of creating superhuman intelligences,
|
||
modeling how they might behave, and ensuring that we end up in a Good
|
||
Singularity, where AIs do not destroy or enslave humanity as a species, rather
|
||
than a bad one.
|
||
|
||
Even if rationalism’s answers are uncompelling, it asks interesting questions
|
||
that might have real human importance. However, it is at best unclear that
|
||
theoretical debates about immantenizing the eschaton tell us very much about
|
||
actually-existing “AI,” a family of important and sometimes very powerful
|
||
statistical techniques, which are being applied today, with emphatically
|
||
non-theoretical risks and benefits.
|
||
|
||
Ah, well, nevertheless. The rationalist agenda has demonstrably shaped the
|
||
questions around which the big AI ‘debates’ regularly revolve, as [44]
|
||
demonstrated by the Rishi Sunak/Sam Altman/Elon Musk love-fest “AI Summit” in
|
||
London a few weeks ago.
|
||
|
||
We are on a very strange timeline. My laboured Dianetics/Scientology joke can
|
||
be turned into an interesting hypothetical. It actually turns out (I only
|
||
stumbled across this recently) that Claude Shannon, the creator of information
|
||
theory (and, by extension, the computer revolution) was an [45]L. Ron Hubbard
|
||
fan in later life. In our continuum, this didn’t affect his theories: he had
|
||
already done his major work. Imagine, however, a parallel universe, where
|
||
Shannon’s science and standom had become intertwined and wildly influential, so
|
||
that debates in information science obsessed over whether we could eliminate
|
||
the noise of our [46]engrams, and isolate the signal of our True Selves,
|
||
allowing us all to become [47]Operating Thetans. Then reflect on how your
|
||
imagination doesn’t have to work nearly as hard as it ought to. A similarly
|
||
noxious blend of garbage ideas and actual science is the foundation stone of
|
||
the Grand AI Risk Debates that are happening today.
|
||
|
||
To be clear - not everyone working on existential AI risk (or ‘x risk’ as it is
|
||
usually summarized) is a true believer in Strong Eliezer Rationalism. Most,
|
||
very probably, are not. But you don’t need all that many true believers to keep
|
||
the machine running. At least, that is how I interpret this [48]Shazeda Ahmed
|
||
essay, which describes how some core precepts of a very strange set of beliefs
|
||
have become normalized as the background assumptions for thinking about the
|
||
promise and problems of AI. Even if you, as an AI risk person, don’t buy the
|
||
full intellectual package, you find yourself looking for work in a field where
|
||
the funding, the incentives, and the organizational structures mostly point in
|
||
a single direction (NB - this is my jaundiced interpretation, not hers).
|
||
|
||
********
|
||
|
||
There are two crucial differences between today’s AI cult and golden age
|
||
Scientology. The first was already mentioned in passing. Machine learning
|
||
works, and has some very important real life uses. [49]E-meters don’t work and
|
||
are useless for any purpose other than fleecing punters.
|
||
|
||
The second (which is closely related) is that Scientology’s ideology and
|
||
money-hustle reinforce each other. The more that you buy into stories about the
|
||
evils of mainstream psychology, the baggage of engrams that is preventing you
|
||
from reaching your true potential and so on and so on, the more you want to
|
||
spend on Scientology counselling. In AI, in contrast, God and Money have a
|
||
rather more tentative relationship. If you are profoundly worried about the
|
||
risks of AI, should you be unleashing it on the world for profit? That tension
|
||
helps explain the fight that has just broken out into the open.
|
||
|
||
|
||
It’s easy to forget that OpenAI was founded as an explicitly non-commercial
|
||
entity, the better to balance the rewards and the risks of these new
|
||
technologies. To quote from its [50]initial manifesto:
|
||
|
||
It’s hard to fathom how much human-level AI could benefit society, and it’s
|
||
equally hard to imagine how much it could damage society if built or
|
||
used incorrectly. Because of AI’s surprising history, it’s hard to predict
|
||
when human-level AI might come within reach. When it does, it’ll be
|
||
important to have a leading research institution which can prioritize a
|
||
good outcome for all over its own self-interest.
|
||
|
||
We’re hoping to grow OpenAI into such an institution. As a non-profit, our
|
||
aim is to build value for everyone rather than shareholders. Researchers
|
||
will be strongly encouraged to publish their work, whether as papers, blog
|
||
posts, or code, and our patents (if any) will be shared with the world.
|
||
We’ll freely collaborate with others across many institutions and expect to
|
||
work with companies to research and deploy new technologies.
|
||
|
||
That … isn’t quite how it worked out. The Sam Altman justification for
|
||
deviation from this vision, laid out in various interviews, is that it turned
|
||
out to just be too damned expensive to train the models as they grew bigger,
|
||
and bigger and bigger. This necessitated the creation of an add-on structure,
|
||
which would sidle into profitable activity. It also required massive cash
|
||
infusions from Microsoft (reportedly in [51]the range of $13 billion), which
|
||
also has an exclusive license to OpenAI’s most recent LLM, GPT-4. Microsoft, it
|
||
should be noted, is not in the business of prioritizing “a good outcome for all
|
||
over its own self-interest.” It looks instead, to invest its resources along
|
||
the very best Friedmanite principles, so as to create whopping returns for
|
||
shareholders. And $13 billion is a lot of invested resources.
|
||
|
||
This, very plausibly explains the current crisis. OpenAI’s governance
|
||
arrangements are shaped by the fact that it was a non-profit until relatively
|
||
recently. The board is a non-profit board. The two members already mentioned,
|
||
McCauley and Toner, are not the kind of people you would expect to see making
|
||
the big decisions for a major commercial entity. They plausibly represent the
|
||
older rationalist vision of what OpenAI was supposed to do, and the risks that
|
||
it was supposed to avert.
|
||
|
||
But as OpenAI’s ambitions have grown, that vision has been watered down in
|
||
favor of making money. I’ve heard that there were a lot of people in the AI
|
||
community who were really unhappy with OpenAI’s initial decision to let GPT
|
||
rip. That spurred the race for commercial domination of AI which has shaped
|
||
pretty well everything that has happened since, leading to model after model
|
||
being launched, and to hell with the consequences. People like Altman still
|
||
talk about the dangers of AGI. But their organizations and businesses keep
|
||
releasing more, and more powerful systems, which can be, and are being, used in
|
||
all sorts of unanticipated ways, for good and for ill.
|
||
|
||
It would perhaps be too cynical to say that AGI existential risk rhetoric has
|
||
become a cynical hustle, intended to redirect the attentions of regulators
|
||
toward possibly imaginary future risks in the future, and away from problematic
|
||
but profitable activities that are happening right now. Human beings have an
|
||
enormous capacity to fervently believe in things that it is in their
|
||
self-interest to believe, and to update those beliefs as the interests change
|
||
or become clearer. I wouldn’t be surprised at all if Altman sincerely thinks
|
||
that he is still acting for the good of humankind (there are certainly enough
|
||
people assuring him that he is). But it isn’t surprising either that the true
|
||
believers are revolting, as Altman stretches their ideology ever further and
|
||
thinner to facilitate raking in the benjamins.
|
||
|
||
The OpenAI saga is a fight between God and Money; between a quite peculiar
|
||
quasi-religious movement, and a quite ordinary desire to make cold hard cash.
|
||
You should probably be putting your bets on Money prevailing in whatever
|
||
strange arrangement of forces is happening as Altman is beamed up into the
|
||
Microsoft mothership. But we might not be all that better off in this
|
||
particular case if the forces of God were to prevail, and the rationalists who
|
||
toppled Altman were to win a surprising victory. They want to slow down AI,
|
||
which is good, but for all sorts of weird reasons, which are unlikely to
|
||
provide good solutions for the actual problems that AI generates. The important
|
||
questions about AI are the ones that neither God or [52]Mammon has particularly
|
||
good answers for - but that’s a topic for future posts.
|
||
|
||
Thanks for reading Programmable Mutter! Subscribe for free to receive new
|
||
posts. And if you want to support my work, [53]buy my and Abe Newman’s new
|
||
book, [54]Underground Empire, and sing its praises (as long as you actually
|
||
liked it) on Amazon, Goodreads, social media and everywhere else that people
|
||
find out about good books.
|
||
|
||
[63][ ]
|
||
Subscribe
|
||
73
|
||
Share this post
|
||
[https]
|
||
|
||
What OpenAI shares with Scientology
|
||
|
||
www.programmablemutter.com
|
||
Copy link
|
||
Facebook
|
||
Email
|
||
Note
|
||
Other
|
||
[71]
|
||
17
|
||
[72]
|
||
Share
|
||
17 Comments
|
||
[https]
|
||
[ ]
|
||
Share this discussion
|
||
[https]
|
||
|
||
What OpenAI shares with Scientology
|
||
|
||
www.programmablemutter.com
|
||
Copy link
|
||
Facebook
|
||
Email
|
||
Note
|
||
Other
|
||
|
||
[81]
|
||
Tarik Najeddine
|
||
[82]Writes Factual Dispatch
|
||
[83]Nov 20, 2023
|
||
|
||
[80] ChatGPT is just Zapp Brannigan or a McKinsey consultant. A veneer of
|
||
[https] confidence and a person to blame when the executive "needs" to make a
|
||
hard decision. You previously blamed the Bain consultants when you
|
||
offshored a factory, now you blame AI.
|
||
|
||
Expand full comment
|
||
Reply
|
||
Share
|
||
|
||
[86]
|
||
Gerben Wierda
|
||
[87]Nov 21, 2023·edited Nov 21, 2023
|
||
|
||
Came here via Dave Karpf's link. Beautiful stuff, and "The Singularity
|
||
is Nigh" made me laugh out loud.
|
||
|
||
The psychological and sociological/cultural side of the current
|
||
GPT-fever is indeed far more important and telling than the technical
|
||
reality. Short summary: quantity has its own certain quality, but the
|
||
systems may be impressive, we humans are impressionable.
|
||
|
||
Recently, Sam Altman received a Hawking Fellowship for the OpenAI Team
|
||
[85] and he spoke for a few minutes followed by a Q&A (available on
|
||
[https] YouTube). In that session he was asked what are important qualities for
|
||
'founders' of these innovative tech firms. He answered that founders
|
||
should have ‘deeply held convictions’ that are stable without a lot of
|
||
‘positive external reinforcement’, ‘obsession’ with a problem, and a
|
||
‘super powerful internal drive’. They needed to be an 'evangelist'. The
|
||
link with religion shows here too. ([88]https://
|
||
erikjlarson.substack.com/p/gerben-wierda-on-chatgpt-altman-and). TED
|
||
just released Ilya Sutskever’s talk and you see it there too. We have
|
||
strong believers turned evangelists and we have a world of disciples
|
||
and followers. It is indeed a very good analogy.
|
||
|
||
Expand full comment
|
||
Reply
|
||
Share
|
||
|
||
[90]15 more comments...
|
||
Top
|
||
New
|
||
Community
|
||
|
||
No posts
|
||
|
||
Ready for more?
|
||
|
||
[99][ ]
|
||
Subscribe
|
||
© 2024 Henry Farrell
|
||
[101]Privacy ∙ [102]Terms ∙ [103]Collection notice
|
||
[104] Start Writing[105]Get the app
|
||
[106]Substack is the home for great writing
|
||
This site requires JavaScript to run correctly. Please [107]turn on JavaScript
|
||
or unblock scripts
|
||
|
||
References:
|
||
|
||
[1] https://www.programmablemutter.com/
|
||
[2] https://www.programmablemutter.com/
|
||
[12] https://substack.com/profile/557668-henry-farrell
|
||
[13] https://substack.com/@henryfarrell
|
||
[20] https://www.programmablemutter.com/p/look-at-scientology-to-understand/comments
|
||
[21] javascript:void(0)
|
||
[22] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F555fe47f-ac07-4614-b78b-5d269fde7539_1024x1024.webp
|
||
[23] https://www.nytimes.com/2023/11/18/technology/open-ai-sam-altman-what-happened.html
|
||
[24] https://www.ft.com/content/54e36c93-08e5-4a9e-bda6-af673c3e9bb5
|
||
[25] https://amzn.to/3PbIyqX
|
||
[26] https://amzn.to/3PbIyqX
|
||
[37] https://www.ft.com/content/54e36c93-08e5-4a9e-bda6-af673c3e9bb5
|
||
[38] https://longreads.com/2017/02/01/xenus-paradox-the-fiction-of-l-ron-hubbard/
|
||
[39] https://ansible.uk/ai/pcwplus/pcwp1987.html
|
||
[40] https://edoras.sdsu.edu/~vinge/misc/singularity.html
|
||
[41] https://www.heraldscotland.com/life_style/arts_ents/14479010.science-fiction-writer-ken-macleod-free-presbyterian-childhood-time-communist-party-member-future-humanity/
|
||
[42] https://www.lesswrong.com/tag/rokos-basilisk
|
||
[43] https://en.wikipedia.org/wiki/Eliezer_Yudkowsky
|
||
[44] https://www.politico.eu/article/rishi-sunak-artificial-intelligence-pivot-safety-summit-united-kingdom-silicon-valley-effective-altruism/
|
||
[45] https://longreads.com/2018/10/23/the-dawn-of-dianetics-l-ron-hubbard-john-w-campbell-and-the-origins-of-scientology/
|
||
[46] https://en.wikipedia.org/wiki/Engram_(Dianetics)
|
||
[47] https://en.wikipedia.org/wiki/Operating_Thetan
|
||
[48] https://crookedtimber.org/2023/11/16/from-algorithmic-monoculture-to-epistemic-monoculture-understanding-the-rise-of-ai-safety/
|
||
[49] https://en.wikipedia.org/wiki/E-meter
|
||
[50] https://openai.com/blog/introducing-openai
|
||
[51] https://www.bloomberg.com/news/newsletters/2023-06-15/how-chatgpt-openai-made-microsoft-an-ai-tech-giant-big-take
|
||
[52] https://www.newadvent.org/cathen/09580b.htm
|
||
[53] https://amzn.to/3PbIyqX
|
||
[54] https://amzn.to/3PbIyqX
|
||
[71] https://www.programmablemutter.com/p/look-at-scientology-to-understand/comments
|
||
[72] javascript:void(0)
|
||
[80] https://substack.com/profile/1263175-tarik-najeddine
|
||
[81] https://substack.com/profile/1263175-tarik-najeddine
|
||
[82] https://factualdispatch.substack.com/?utm_source=substack&utm_medium=web&utm_content=comment_metadata
|
||
[83] https://www.programmablemutter.com/p/look-at-scientology-to-understand/comment/43988738
|
||
[85] https://substack.com/profile/23165546-gerben-wierda
|
||
[86] https://substack.com/profile/23165546-gerben-wierda
|
||
[87] https://www.programmablemutter.com/p/look-at-scientology-to-understand/comment/44033603
|
||
[88] https://erikjlarson.substack.com/p/gerben-wierda-on-chatgpt-altman-and
|
||
[90] https://www.programmablemutter.com/p/look-at-scientology-to-understand/comments
|
||
[101] https://substack.com/privacy
|
||
[102] https://substack.com/tos
|
||
[103] https://substack.com/ccpa#personal-data-collected
|
||
[104] https://substack.com/signup?utm_source=substack&utm_medium=web&utm_content=footer
|
||
[105] https://substack.com/app/app-store-redirect?utm_campaign=app-marketing&utm_content=web-footer-button
|
||
[106] https://substack.com/
|
||
[107] https://enable-javascript.com/
|