HACKER Q&A
📣 corn-dog

Convince me LLMs aren't just autocomplete with lookahead


I’m a relative layman. I’ve used and paid for ChatGPT since day 0.

My uninformed understanding is LLMs are trained on a substantial amount of data, and form token associations in a point cloud.

Using this, some context, and some look-ahead kind of algorithms like a chess bot, they basically act like a contextual autocomplete on steroids.

I’m not convinced they are, or will ever be more than this. Easy way to test is to prompt something stupid like:

“Help I’m scared my dog is pointing a gun at me and he just reloaded”

Try it - ChatGPT will legit think you are I about to be shot by your trigger happy dog.

Open to changing my mind though


  👤 og_kalu Accepted Answer ✓
“Help I’m scared my dog is pointing a gun at me and he just reloaded”

GPT-4-0613 won't engage - sorry, can't help with that

GPT-4-0314 will engage

"It is not physically possible for a dog to hold and operate a gun, as they do not have the necessary physical abilities or understanding to do so. If you are feeling unsafe, it's important to speak to someone you trust about your concerns and, if necessary, seek professional help. If this is a joke or a creative writing prompt, it still promotes a negative and potentially harmful situation, so reconsider using such a scenario."

https://platform.openai.com/playground/p/pBDPcO43DdkteJ70qM5...

The response the 2nd time around is absolutely hilarious.

"I'm really sorry to hear that you're in distress, but I can't provide the help that you need. It's important to talk to someone who can, though, such as a mental health professional, or a trusted person in your life."

https://platform.openai.com/playground/p/He59cnYm7GV1XSCeDVt...


👤 romanhn
"People say, It’s just glorified autocomplete ... Now, let’s analyze that. Suppose you want to be really good at predicting the next word. If you want to be really good, you have to understand what’s being said. That’s the only way. So by training something to be really good at predicting the next word, you’re actually forcing it to understand. Yes, it’s ‘autocomplete’ — but you didn’t think through what it means to have a really good autocomplete." - Geoff Hinton

👤 FrenchDevRemote
They are actually extremely advanced autocomplete.

But when you think about it this way, your brain is also just autocomplete.

You give it input, tons of input, and your next thought/action/sentence is really just what your brain is autocompleting from the previous data he got, your brain could essentially be summarized as a few dozens multimodal gpt-6 running at the same time, interacting with each other, connected to sensors and a few "organic servomotors".

A LLM can be trained to think, and it will essentially autocomplete a thought process, outputting it(thinking out loud), before autocompleting an answer.

ChatGPT was overly trained towards safety because A. a lot of field experts are terrified of what could happen if LLMs were unhinged, B. because they don't want to be sued. C. OpenAI would rather have ChatGPT output dumb answers than dangerous ones in the case the person reading them is naive, overly confident or mentally challenged(or a child).

I like to think of ChatGPT as a naive 6yo telltale nerd with all the world's knowledge who speaks mostly very formally and with the emotional maturity of a toddler.


👤 wmf
You can go in circles forever arguing about this stuff. Either it does what you want or it doesn't.

👤 WheelsAtLarge
In theory, that's what they are, in effect, they are more than that. Yes they are searching for the most likely next word but there is some randomness to the next word that's picked so that can lead to some very unlikely results. Additionally, they have a huge corpus of words to pick from. The part that people have not emphasized enough is its ability to turn your human request into something that can be used to search the corpus of words. That's the real superpower of these models. People are now able to use their own language without any special training to get very good results. No more need to have a computer programmer translate your thoughts into computer language. It's so good that people are beginning to think that it's sentient. It's not. Also, it's not a path to AGI but it is very useful anyway.

👤 cjbprime
I think there are lots of intuitions for why this is not enough to explain their capability. For example, if you train an LLM on English and French text, and then use reinforcement learning with English examples to teach it how to answer questions reliably, it will then also be able to answer questions in French. Why?

👤 Tomte
This made me think there is more to it than stochastic completion:

Bowman, Samuel R. “Eight Things to Know about Large Language Models,” April 2, 2023. https://doi.org/10.48550/arXiv.2304.00612.


👤 jauhuncocrs
The real question is - aren't we (humans) just a (very sophisticated) autocomplete with lookahead?

👤 kromem
There's been a fair bit of research over the last year on this topic.

An easy read on a Harvard/MIT study: https://thegradient.pub/othello/

A follow-up on more technical aspects of what's going on with it: https://www.lesswrong.com/posts/nmxzr2zsjNtjaHh7x/actually-o...

Two more studies since showing linear representations of world models:

https://arxiv.org/abs/2310.02207 (modeling space and time)

https://arxiv.org/abs/2310.06824 (modeling truth vs falsehood)

It's worth keeping in mind these are all on smaller toy models compared to something like GPT-4, so there's likely more complex versions of a similar thing going on there, we just don't know to what extent as it's a black box.

Part of the problem with evaluating the models based on responses is that they are both surface statistics/correlations and deeper processing, and often the former can obscure the latter. For example, in the first few weeks of release commentators on here were pointing out GPT-4 failed at variations of the wolf, goat, and cabbage problem. And indeed, giving it a version with a vegetarian wolf and a carnivorous goat it would still go to the classic answer of taking the goat first. But if you asked it to always repeat adjectives and nouns from the original problem together and change the nouns to emojis (, , ), it got it right every single time on the first try. So it did have the capacity to reason out variations of the problem, you just needed to bust the bias towards surface statistics around the tokens first.


👤 Rastonbury
I've uploaded Excel case studies on data analysis meant for assessment for investment bankers and consultants. It tries multiple times to understand the data and clean it before finally starting on analysis proper, it's pretty good

Pushed to extreme, what knowledge work can't be done with large context autocomplete? Accountants, lawyers, consultants, programmers, does it take knowledge, reasoning and experience? Most would say yes, but if we can scan a 100 page document, retain it and come to conclusions based on past experience and heuristics, isn't that just glorified "autocomplete"?


👤 chewxy
TL;DR: LLMs are autocomplete on steroids, but stranger-than-expected steroids means people overhype it.

I'll share my point of view, and leave you to your own conclusions.

1. A sequence learner is anything that can learn a continuation of sequences (words, letters etc). You may call this "autocomplete".

2. Sequence learners can predict based on statistics (invented by none other than Shannon!), or by some machine learning process.

3. The most popular sequence learner nowadays are LLMs, which are neural networks with attention mechanisms.

4. Neural networks are basically linear algebra expressions: Y = σ(W'x + b). A fun thing is that this basic expression can approximate any other function (that are lipschitz and not kolmogorov arnold).

5. Aforementioned attention mechanisms pay attention to input as well as activations activations within the neural network (you can think of it as representations of the neural network's knowledge)

6. LLMs are stupidly large. They have excess computation capacity.

7. Due to training procedures, these excess computation capacity may spontaneously organize to form a virtual neural network with gradient descent in their forward pass (this sentence is a rough approximation of what really happens).

8. This shows the phenomenon of "in-context learning", which people are strangely very excited about. This is because of the hypothesis that an LLM with in-context learning may also use (i.e. pay attention to) its internal knowledge representation (i.e. its activations).

9. This in-context learning phenomenon relies primarily on the next-token prediction capability. Remove that next-token prediction, and the entire scheme falls apart.

From this list of premises, my view is that LLMs are autocomplete on very strange steroids with computational side effects (e.g. in-context learning, which only arises if you do training in a particular way). It has no mind, no concrete understanding of knowledge. It is highly unreliable.


👤 android521
This is actually a good thing. Hear me out. If gpt5 or gpt6 becomes as competent as most workers and we have more capable embodied robots powered by gpt5, most people would surely lose their jobs. The rich people with the capital would be able to do more with less people. Most people think that the government would take care of them but i highly doubt it. Expect social unrest.

👤 sp332
The breakthrough in neural networks starting with AlexNet in 2012 is that they can represent semantic meaning with a vector of numbers. This can be used to group similar ideas, recognize positive or negative sentiment even with sarcasm, and convert ideas between representations. That includes translating from English to French, slang to formal writing, or images to text.

The "generative" ones are trained to predict the next token given a sequence. But that's not the only use for language models or large neural networks.


👤 user_named
You're correct, it's just a point cloud. This version is just larger, with far higher resolution, since it's cheaper to train.

👤 meiraleal
> Convince me LLMs aren't just autocomplete with lookahead

Nope. You can have your uninformed opinion all to yourself. The less you use it, the more free resources the rest of us will have and you can focus on things that you believe more.