Add november links

This commit is contained in:
David Eisinger
2024-10-30 23:40:20 -04:00
parent df81b457f8
commit 31ba49beea
13 changed files with 3529 additions and 6 deletions

View File

@@ -0,0 +1,426 @@
[1][https]
[2]One Useful Thing
SubscribeSign in
Share this post
[https]
Thinking Like an AI
www.oneusefulthing.org
Copy link
Facebook
Email
Note
Other
Thinking Like an AI
A little intuition can help
[13][https]
[14]Ethan Mollick
Oct 20, 2024
528
Share this post
[https]
Thinking Like an AI
www.oneusefulthing.org
Copy link
Facebook
Email
Note
Other
[21]
60
41
[22]
Share
This is my 100th post on this Substack, which got me thinking about how I could
summarize the many things I have written about how to use AI. I came to the
conclusion that [23]the advice in my book is still the advice I would give:
just use AI to do stuff that you do for work or fun, for about 10 hours, and
you will figure out a remarkable amount.
However, I do think having a little bit of intuition about the way Large
Language Models work can be helpful for understanding how to use it best. I
would ask my technical readers for their forgiveness, because I will simplify
here, but here are some clues for getting into the “mind” of an AI:
LLMs do next token prediction
Large Language Models are, ultimately, incredibly sophisticated autocomplete
systems. They use a vast model of human language to predict the next token in a
sentence. For models working with text, tokens are words or parts of words.
Many common words are single tokens, or tokens containing spaces, but other
words are broken into multiple tokens. For example, one tokenizer takes the 10
word sentence, “This breaks up words (even phantasmagorically long words) into
tokens” into 20 tokens.
[25]
[https]
When you give an AI a prompt, you are effectively asking it to predict the next
token that would come after the prompt. The AI then takes everything that has
been written before, runs it through a mathematical model of language, and
generates the probability of which token is likely to come next in the
sequence. For example, if I write “The best type of pet is a” the LLM predicts
that the most likely tokens to come next, based on its model of human language,
are either “dog”, “personal,” “subjective,” or “cat.” The most likely is
actually dog, but LLMs are generally set to include some randomness, which is
what makes LLM answers interesting, so it does not always pick the most likely
token (in most cases, even attempts to eliminate this randomness cannot remove
it entirely). Thus, I will often get “dog,” but I may get a different word
instead.
[26]
[https]
These are the actual probabilities from GPT-3.5, as are the other examples in
this post.
But these predictions take into account everything in the memory of the LLM
(more on memory in a bit), and even tiny changes can radically alter the
predictions of what token comes next. I created three examples with minor
changes on the original sentence. If I choose not to capitalize the first word,
the model now says that “dog” and “cat” are much more likely answers than they
were originally, and “fish” joins the top three. If I change the word “type” to
“kind” in the sentence, the probabilities of all the top tokens drop and I am
much more likely to get an exotic answer like “calm” or “bunny.” If I add an
extra space after the word “pet,” then “dog” isnt even in the top three
predicted tokens!
[27]
[https]
But the LLM does not just produce one token, instead, after each token, it now
looks at the entire original sentence plus the new token (“The best type of pet
is a dog”) and predicts the next token after that, and then uses that whole
sentence plus the next to make a prediction, and so on. It chains one token to
another like cars on a train. Current LLMs cant go back and change a token
that came before, they have to soldier on, adding word after word. This results
in a butterfly effect. If the first predicted token was the word “dog” than the
rest of the sentence will follow on like that, if it is “subjective” then you
will get an entirely different sentence. Any difference between the tokens in
two different answers will result in radically diverging responses.
[28]
[https]
The intuition: This helps explain why you may get very different answers than
someone else using the same AI, even if you ask exactly the same question. Tiny
differences in probabilities result in very different answers. It also gives
you a sense about why one of the biases that people worry about with AI is that
it may respond differently to people depending on their writing style, as the
probabilities for the next token may lead on the path to worse answers. Indeed,
[29]some of the early LLMs gave less accurate answers if you wrote in a less
educated way.
You can also see some of why hallucinations happen, and why they are so
pernicious. The AI is not pulling from a database, it is guessing the next word
based on statistical patterns in its training data. That means that what it
produces is not necessarily true (in fact, one of many surprises about LLMs are
how often they are right, given this), but, even when it provides false
information, it likely sounds plausible. That makes it hard to tell when it is
making things up.
It is also helpful to think about tokens to understand why AIs get stubborn
about a topic. If the first prediction is “dog” the AI is much more likely to
keep producing text about how great dogs are because those tokens are more
likely. However, if it is “subjective” it is less likely to give you an
opinion, even when you push it. Additionally, once the AI has written
something, it cannot go back, so it needs to justify (or explain or lie about)
that statement in the future. I like this example that [30]Rohit Krishnan [31]
shared, where you can see the AI makes an error, but then attempts to justify
the results.
[32]
[https]
The caveat: Saying “AI is just next-token prediction” is a bit of a joke
online, because it doesnt really help us understand why AI can produce such
seemingly creative, novel, and interesting results. If you have been reading my
posts for any length of time, you will realize that AI accomplishes impressive
outcomes that, intuitively, we would not expect from an autocomplete system.
[33]
[https]
Claude makes themed Excel formulas on demand and explains them in delightful
ways. Next token prediction is capable of lots of unexpected results.
LLMs make predictions based on their training data
Where does an LLM get the material on which it builds a model of language? From
the data it was trained on. Modern LLMs are trained over an incredibly vast set
of data, incorporating large amounts of the web and every free book or archive
possible (plus some archives that almost certainly contain copyrighted work).
The AI companies largely did not ask permission before using this information,
but leaving aside the legal and ethical concerns, it can be helpful to
conceptualize the training data.
The original [35]Pile dataset, which most of the major AI companies used for
training, is about 1/3 based on the internet, 1/3 on scientific papers, and the
rest divided up between books, coding, chats, and more. So, your intuition is
often a good guide - if you expect something was on the internet or in the
public domain, it is likely in the training data. But we can get a little more
granular. For example, [36]thanks to this study, we have a rough idea of which
fiction books appear most often in the training data for GPT-4, which largely
tracks the books most commonly found on the web (many of the top 20 are out of
copyright, with a couple notable exceptions of books that are much pirated).
[37]
[https]
Remember that LLMs use a statistical model of language, they do not pull from a
database. So the more common a piece of work is in the training data, the more
likely the AI is to “recall” that data accurately when prompted. You can see
this at work when I give it a sentence from the most fiction common book in its
training data - Alice in Wonderland. It gets the next sentence exactly right,
and you can see that almost every possible next token would continue along the
lines of the original passage.
[38]
[https]
Lets try something different, a passage from a fairly obscure mid-century
science fiction author, [39]Cordwainer Smith, with an unusual writing style in
part shaped by his time in China (he was Sun Yat-sens godson) and his
knowledge of multiple languages. One of his stories starts: Go back to An-fang,
the Peace Square at An-fang, the Beginning Place at An-fang, where all things
start. It then continues: Bright it was. Red square, dead square, clear square,
under a yellow sun. If I give the AI the first section, looking at the
probabilities, there is almost no chance that it will produce the correct next
word “Bright.” Instead, perhaps primed by the mythic language and the fact that
An-fang registers as potentially Chinese (it is actually a play on the German
word for beginning), it creates a passage about a religious journey.
[40]
[https]
The intuition: The fact that the LLM does not directly recall text would be
frustrating if you were trying to use an LLM like Google, but LLMs are not like
Google. They are capable of producing original material, and, even when they
attempt to give you Alice in Wonderland word-for-word, small differences will
randomly appear and eventually the stories will diverge. However, knowing what
is in the training data can help you in a number of ways.
First, it can help you understand what the AI is good at. Any document or
writing style that is common in its training data is likely something the AI is
very good at producing. But, more interestingly, it can help you think about
how to get more original work from the AI. By pushing it through your prompts
to a more unusual section of its probability space, you will get very different
answers than other people. Asking AI to write a memo in the style of [41]Walter
Pater will give you more interesting answers (and overwrought ones) than asking
for a professional memo, of which there are millions in the training data.
[42]
[https]
The caveat: Contrary to some people's beliefs, the AI is rarely producing
substantial text from its training data verbatim. The sentences the AI provides
are usually entirely novel, extrapolated from the language patterns it learned.
Occasionally, the model might reproduce a specific fact or phrase it memorized
from its training data, but more often, it's generalizing from learned patterns
to produce new content.
Outside of training, carefully crafted prompts can guide the model to produce
more original or task-specific content, demonstrating a capability known as
“in-context learning.” This allows LLMs to appear to learn new tasks within a
conversation, even though they're not actually updating their underlying model,
as you will see.
LLMs have a limited memory
Given how much we have discussed training, it may be surprising to learn that
AIs are not generally learning anything permanent from their conversations with
you. Training is usually a discrete event, not something that happens all the
time. If you have privacy features turned on, your chats are not being fed into
the training data at all, but, even if your data will be used for training, the
training process is not continuous. Instead, chats happen within what's called
a 'context window'. This context window is like the AI's short-term memory -
it's the amount of previous text the AI can consider when generating its next
response. As long as you stay in a single chat session and the conversation
fits inside the context window, the AI will keep track of what is happening,
but as soon as you start a new chat, the memories from the last one generally
do not carry over. You are starting fresh. The only exception is the limited
“memory” feature of ChatGPT, which notes down scattered facts about you in a
memory file and inserts those into the context window of every conversation.
Otherwise, the AI is not learning about you between chats.
Even as I write this, I know I will be getting comments from some people
arguing that I am wrong, along with descriptions of insights from the AI that
seem to violate this rule. People are often fooled because the AI is a very
good guesser, w[44]hich Simon Willison explains at length in his excellent post
on the topic of asking the AI for insights into yourself. It is worth reading.
The intuition: It can help to think about what the AI knows and doesnt know
about you. Do not expect deep insights based on information that the AI does
not have but do expect it to make up insightful-sounding things if you push it.
Knowing how memory works, you can also see why it can help to start a new chat
when the AI gets stuck, or you dont like where things are heading in a
conversation. Also, if you use ChatGPT, you may want to check out and[45] clean
up your memories every once in a while.
The caveat: The context windows of AIs are growing very long (Googles Gemini
can hold 2 million tokens in memory), and AI companies want the experience of
working with their models to feel personal. I expect we will see more tricks to
get AIs to remember things about you across conversations being implemented
soon.
All of this is only sort of helpful
We still do not have a solid answer about how these basic principles of how
LLMs work have come together to make a system that is [47]seemingly more
creative than most humans, that we enjoy speaking with, and which does a
surprisingly good job at tasks ranging from corporate strategy to medicine.
There is no manual that lists what AI does well or where it might mess up, and
we can only tell so much from the underlying technology itself.
Understanding token prediction, training data, and memory constraints gives us
a peek behind the curtain, but it doesn't fully explain the magic happening on
stage. That said, this knowledge can help you push AI in more interesting
directions. Want more original outputs? Try prompts that veer into less common
territory in the training data. Stuck in a conversational rut? Remember the
context window and start fresh.
But the real way to understand AI is to use it. A lot. For about 10 hours, just
do stuff with AI that you do for work or fun. Poke it, prod it, ask it weird
questions. See where it shines and where it stumbles. Your hands-on experience
will teach you more than any article ever could (even this long one). You'll
figure out a remarkable amount about how to use AI effectively, and you might
even surprise yourself with what you discover.
[56][ ]
Subscribe
[58]Share
528
Share this post
[https]
Thinking Like an AI
www.oneusefulthing.org
Copy link
Facebook
Email
Note
Other
[65]
60
41
[66]
Share
PreviousNext
Discussion about this post
Comments
Restacks
[https]
[ ]
[73]
Mickey Schafer
[74]Oct 20
Perfect timing! This will be the first post students read next semester
for a one-credit class called Prompting Curiosities 😊. I'm struggling
to find those 10 hours so embedding it into a class seemed like a fun
[72] way to get it done. Just me, 15 students, and the university's AI
[https] system which has most of the LLMs in 3-4 versions. We will start with
simple prompts across different LLMs, then as each finds their
favorite, they'll choose one thing as their final project and work on
it. All in all, it should produce at least 20 per person which will
help me understand these much better moving forward!
Expand full comment
Reply
Share
[76]2 replies
[78]
Clarke Pitts
[79]Oct 21Liked by Ethan Mollick
[77] An excellent essay, interesting and intelligible. Very little
[https] explanation about AI and LLM is as lucid.
Expand full comment
Reply
Share
[81]58 more comments...
Top
Latest
Discussions
No posts
Ready for more?
[94][ ]
Subscribe
© 2024 Ethan Mollick
[96]Privacy ∙ [97]Terms ∙ [98]Collection notice
[99] Start Writing[100]Get the app
[101]Substack is the home for great culture
Share
Copy link
Facebook
Email
Note
Other
This site requires JavaScript to run correctly. Please [108]turn on JavaScript
or unblock scripts
References:
[1] https://www.oneusefulthing.org/
[2] https://www.oneusefulthing.org/
[13] https://substack.com/profile/846835-ethan-mollick
[14] https://substack.com/@oneusefulthing
[21] https://www.oneusefulthing.org/p/thinking-like-an-ai/comments
[22] javascript:void(0)
[23] https://a.co/d/9onRd33
[25] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F805116f4-c2dc-4804-b277-253d14b2139d_1292x105.png
[26] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcfb74661-2025-4694-b0db-a96d2166865e_1098x711.png
[27] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623e802b-c122-4ef0-a667-6e429b09cc54_1992x504.png
[28] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff7f2a21-1252-474d-896d-d307dc88eea7_1255x837.png
[29] https://arxiv.org/pdf/2212.09251
[30] https://www.strangeloopcanon.com/
[31] https://x.com/krishnanrohit/status/1802747007838384382
[32] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc187f7b-6341-4ac9-b2e4-0c97d1eddef9_924x502.jpeg
[33] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd959adb9-d728-4e2f-b0f1-840b125ac9e0_1900x1126.png
[35] https://arxiv.org/abs/2101.00027
[36] https://arxiv.org/abs/2305.00118
[37] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedb0dd91-9b8e-468e-8c37-cdda8bd3db5c_1290x864.jpeg
[38] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3dc09899-6a1a-47b3-90b9-c23be78835f8_1504x429.png
[39] https://en.wikipedia.org/wiki/Cordwainer_Smith
[40] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc63b6bec-2dc7-48e4-8e71-ec056768ac96_1494x430.png
[41] https://en.wikipedia.org/wiki/Walter_Pater
[42] https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59a0b6a1-37ca-4447-8777-b94593809c4f_2025x1324.png
[44] https://simonwillison.net/2024/Oct/15/chatgpt-horoscopes/
[45] https://openai.com/index/memory-and-new-controls-for-chatgpt/
[47] https://docs.iza.org/dp17302.pdf
[58] https://www.oneusefulthing.org/p/thinking-like-an-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share
[65] https://www.oneusefulthing.org/p/thinking-like-an-ai/comments
[66] javascript:void(0)
[72] https://substack.com/profile/244712-mickey-schafer
[73] https://substack.com/profile/244712-mickey-schafer
[74] https://www.oneusefulthing.org/p/thinking-like-an-ai/comment/73352564
[76] https://www.oneusefulthing.org/p/thinking-like-an-ai/comment/73352564
[77] https://substack.com/profile/14800577-clarke-pitts
[78] https://substack.com/profile/14800577-clarke-pitts
[79] https://www.oneusefulthing.org/p/thinking-like-an-ai/comment/73452831
[81] https://www.oneusefulthing.org/p/thinking-like-an-ai/comments
[96] https://substack.com/privacy
[97] https://substack.com/tos
[98] https://substack.com/ccpa#personal-data-collected
[99] https://substack.com/signup?utm_source=substack&utm_medium=web&utm_content=footer
[100] https://substack.com/app/app-store-redirect?utm_campaign=app-marketing&utm_content=web-footer-button
[101] https://substack.com/
[108] https://enable-javascript.com/