> "...it would sometimes regurgitate training data verbatim. That’s been patched in the years since..."
> "They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless."
This is totally misleading to anyone with less familiarity with how LLMs work. They are only programs in as much as they perform inference from a fixed, stored, statistical model. It turns out that treating them theoretically in the same way as other computer programs gives a poor representation of their behaviour.
This distinction is important, because no, "regurgitating data" is not something that was "patched out", like a bug in a computer program. The internal representations became more differentially private as newer (subtly different) training techniques were discovered. There is an objective metric by which one can measure this "plagiarism" in the theory, and it isn't nearly as simple as "copying" vs "not copying".
It's also still an ongoing issue and an active area of research, see [1] for example. It is impossible for the models to never "plagiarize" in the sense we think of while remaining useful. But humans repeat things verbatim too in little snippets, all the time. So there is some threshold where no-one seems to care anymore; think of it like the % threshold in something like Turnitin. That's the point that researchers would like to target.
Of course, this is separate from all of the ethical issues around training on data collected without explicit consent, and I would argue that's where the real issues lie.
> This is totally misleading to anyone with less familiarity with how LLMs work. They are only programs in as much as they perform inference from a fixed, stored, statistical model. It turns out that treating them theoretically in the same way as other computer programs gives a poor representation of their behaviour.
> But humans repeat things verbatim too in little snippets, all the time
Also, it's possible, although statistically improbable, for a human to generate the exact same thing another human generated (and copyrighted) without even knowing it.
At the frontier of science we have speculations, which until proper measurements become possible, are unknown to be true or false (or even unknown to be equivalent with other speculations etc. regardless of their being true or false, or truer or falser). Once settled we may call earlier but wrong speculations as "reasonable wrong guesses". In science it is important that these guesses or suspicions are communicated as it drives the design of future experiments.
I argue that more important that "eliminating hallucinations" is tracing the reason it is or was believed by some.
With source-aware training we can ask an LLM to give answers to a question (which may contradict each other), but to provide the training-source(s) justifying emission of each answer, instead of bluff it could emit multiple interpretations and go like:
> answer A: according to school of thought A the answer is that ... examples of authors and places in my training set are: author+title a1, a2, a3, ...
> answer B: according to author B: the answer to this question is ... which can be seen in articles b1, b2
> answer ...: ...
> answer F: although I can't find a single document explaining this, when I collate the observation x in x1, x2, x3; observation y in y1,y2, ... , observation z in z1, z2, ... then I conclude the following: ...
so it is clear which statements are sourced where, and which deductions are proper to the LLM.
Obviously few to none of the high profile LLM providers will do this any time soon, because when jurisdictions learn this is possible they will demand all models to be trained source-aware, so that they can remunerate the authors in their jurisdiction (and levy taxes on their income). What fraction of the income will then go to authors and what fraction to the LLM providers? If any jurisdiction would be first to enforce this, it would probably be the EU, but they don't do it yet. If models are trained in a different jurisdiction than the one levying taxes the academic in-group citation game will be extended to LLMs: a US LLM will have incentive to only cite US sources when multiple are available, and a EU trained LLM will prefer to selectively cite european sources, etc.
In addition to providing training sources, it's important to identify overlaps among the fragments used in the answer. For me, overlap doesn't mean simply identical expression, but conceptually identical.
We are much more likely to find conceptual overlap in code than in language and prose because Many of the problems we solve, as mathematicians say, reduce to previously solved problems, which IMO means substantially identical code.
A related question is how much change is necessary to a work of art, image, prose, or code for it to escape copyright? If we can characterize it and the LLM generates something that escapes copyright, I suggest the output should be excluded from future copyright or patent claims.
I wasn't aware of source-aware training, so thank you for the reference! It does seem a bit too good to be true; I believe in a system of tradeoffs so I feel like this must have an issue with reducing creativity. That's at first glance though, so I could be wrong.
No they're not. They're starving, struggling to find work and lamenting AI is eating their lunch. It's quite ironic that after complaining LLMs are plagiarism machines, the author thinks using them for translation is fine.
"LLMs are evil! Except when they're useful for me" I guess.
I think using translation software or AI images comes down to the ease of working with the system vrs a human. For example, I've created placeholder images for various pieces of work. The images didn't bring in any revenue; they were throwaway pieces, so I didn't give a shit. It wasn't worth paying a human for it.
Before I had AI-generated images, I either left out images from the work or used no-copyright clip art because, again, it wasn't worth arguing with or paying a human to do it.
When it came to diagrams, before Excalidraw, I would dust off my drafting skills, draw something on paper with colored pencils, take a picture of it, and use the picture as the diagram. In this case, I was willing to argue with and pay myself.
Simultaneously, if you hire human translators, you are likely to get machine translations. Maybe not often or overtly, but the translation industry has not been healthy for a while.
The industry is sick because everyone is looking for the lowest prices, but translators don't like machine translation. They don't want to just review the output, because actually doing the translation leads to better understanding of what they have to do.
>As a quick aside, I am not going to entertain the notion that LLMs are intelligent, for any value of “intelligent.” They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless. The rest of this essay will treat them as such. If you are already of the belief that the human mind can be reduced to token regurgitation, you can stop reading here. I’m not interested in philosophical thought experiments.
I can't imagine why someone would want to openly advertise that they're so closed minded. Everything after this paragraph is just anti-LLM ranting.
Because it's so entirely reductive and misunderstanding of where the technology has progressed. Hello world is s computer program. So it Microsoft Windows. New levels of "intelligence" unlock with greater complexity of a program.
Like look at our brains. We know decently well how a single neuron works. We can simulate a single one with "just a computer program". But clearly with enough layers some form of complexity can emerge, and at some level that complexity becomes intelligence.
You can't predict the next token in an arbitrary text unless you are highly intelligent and have a vast body of knowledge.
They're obviously intelligent in the way that we judge intelligence in humans: we pay attention to what they say. You ask them a question about an arbitrary subject, and they respond in the same way that an intelligent person would. If you don't consider that intelligence, then you have a fundamentally magical, unscientific view of what intelligence is.
To return to an analogy I used a couple of days ago ... birds can fly, planes can fly, ergo they are both flying things ... but they fly in completely different ways. So on the one hand (visible behavior) they are similar (or even the same), and on the other (physical mechanism) they are not similar at all.
Which one of these comparisons you want to use depends on context.
The same seems entirely possible for current LLMs. On the one hand they do something that visibly seems to to be the same as something humans do, but on the other it is possible that the way they do it entirely different. Just as with the bird/plane comparison, this has some implications when you start to dig deeper into capabilities (e.g. planes cannot fly anywhere near as slowly as birds, and birds cannot fly as fast as planes; birds have dramatically more maneuverability than planes, etc. etc).
So are LLMs intelligent in the same way humans are? Depends on your purpose in asking that question. Planes fly, but they are not birds.
I know you're arguing with someone else, but I think it is getting sidetracked.
Whether or not LLMs are intelligent (I think they are more intelligent than a cat, for instance, but less intelligent than a human) isn't my argument.
My argument is that complexity in and of itself doesn't yield intelligence. There's no proof of that. There are many things that are very very complex, but we would not put it on an intelligence scale.
A provocative aside in bad faith, anyway a completely minor point within the overall post, which some of the people he's telling to fuck off might have read
I disagree that the majority of it is anti-LLM ranting, there are several subtle points here that are grounded in realism. You should read on past the first bit if you're judging mainly from the initial (admittedly naive) first few paragraphs.
Not GP, but... the author said explicitly "if you believe X you should stop reading". So I did.
The X here is "that the human mind can be reduced to token regurgitation". I don't believe that exactly, and I don't believe that LLMs are conscious, but I do believe that what the human mind does when it "generates text" (i.e. writes essays, programs, etc) may not be all that different from what an LLM does. And that means that most of humans's creations are also the "plagiarism" in the same sense the author uses here, which makes his argument meaningless. You can't escape the philosophical discussion he says that he's not interested in if you want to talk about ethics.
Edit: I'd like to add that I believe that this also ties in to the heart of the philosophy of Open Source and Open Science... if we acknowledge that our creative output is 1% creative spark and 99% standing on the shoulders of Giants, then "openness" is a fundamental good, and "intellectual property" is at best a somewhat distasteful necessity that should be as limited as possible and at worst is outright theft, the real plagiarism.
It's more intellectually lazy to think boolean logic at a sufficient scale crosses some event horizon wherein its execution on mechanical gadgets called computers somehow adds up to intelligence beyond human understanding.
It is intellectually lazy to proclaim something to be impossible in the absence of evidence or proof. In the case of the statement made here, it is provably true that Boolean logic at sufficient scale can replicate "intelligence" of any arbitrary degree. It is also easy to show that this can be perceived as an "event horizon" since the measurements of model quality that humans typically like to use are so nonlinear that they are virtually step function-like.
Doesn't seem like you have proof of anything but it does appear that you have something that is very much like religious faith in an unforeseeable inevitability. Which is fine as far as religion is concerned but it's better to not pretend it's anything other than blind faith.
But if you really do have concrete proof of something then you'll have to spell it out better & explain how exactly it adds up to intelligence of such magnitude & scope that no one can make sense of it.
> "religious faith in an unforeseeable inevitability"
For reference, I work in academia, and my job is to find theoretical limitations of neural nets. If there was so much of a modicum of evidence to support the argument that "intelligence" cannot arise from sufficiently large systems, my colleagues and I would be utterly delighted and would be all over it.
Here are a couple of standard elements without getting into details:
1. Any "intelligent" agent can be modelled as a random map from environmental input to actions.
2. Any random map can be suitably well-approximated by a generative transformer. This is the universal approximation theorem. Universal approximation does not mean that models of a given class can be trained using data to achieve an arbitrary level of accuracy, however...
3. The neural scaling laws (first empirical, now more theoretically established under NTK-type assumptions), as a refinement of the double descent curve, assert that a neural network class can get arbitrarily close to an "entropy level" given sufficient scale. This theoretical level is so much smaller than any performance metric that humans can reach. Whether "sufficiently large" is outside of the range that is physically possible is a much longer discussion, but bets are that human levels are not out of reach (I don't like this, to be clear).
4. The nonlinearity of accuracy metrics comes from the fact that they are constructed from the intersection of a large number of weakly independent events. Think the CDF of a Beta random variable with parameters tending to infinity.
Look, I understand the scepticism, but from where I am, reality isn't leaning that way at the moment. I can't afford to think it isn't possible. I don't think you should either.
As I said previously, you are welcome to believe whatever you find most profitable for your circumstances but I don't find your heuristics convincing. If you do come up or stumble upon a concrete constructive proof that 100 trillion transistors in some suitable configuration will be sufficiently complex to be past the aforementioned event horizon then I'll admit your faith was not misplaced & I will reevaluate my reasons for remaining skeptical of Boolean arithmetic adding up to an incomprehensible kind of intelligence beyond anyone's understanding.
They are not similar. A LLM is a complex statistical machine. A brain is a highly complex neural network. A brain, is more similar the perceptron of some AMD CPUs that to a LLM.
[WARNING: seriously off-topic comment, I was triggered]
Here's what sea-lioned means to me:
I say something.
You accuse me of sea-lioning.
I have two choices: attempt to refute the sea-lioning, which becomes sea-lioning, or allowing your accusation to stand unchallenged, which appears to most people as a confirmation of some kind that I was sea-lioning.
It is a nuclear weapon launched at discussion. It isn't that it doesn't describe a phenomena that actually happens in the world. However, it is a response/accusation to which there is never any way to respond to that doesn't confirm the accusation, whether it was true or not.
It is also absolutely rooted in what appears to me to be a generational distinction: it seems that a bunch of younger people consider it to be a right to speak "in public" (i.e in any kind of online context where people who do not know you can read what you wrote) and expect to avoid a certain kind of response. Should that response arise? Various things will be said about the responder, including "sea-lioning".
My experience is that people who were online in the 80s and 90s find this expectation somewhere between humorous and ridiculous, and that people who went online somewhere after about 2005 do not.
Technologically, it seems to reflect a desire among many younger people for "private-public spaces". In the absence of any such actual systems really existing (at least from their POV), they believe they ought to be able to use very non-private public spaces (facebook, insta, and everything else under the rubric of "social media") as they wish to, rather than as the systems were designed. They are communicating with their friends and the fact that their conversations are visible is not significant. Thus, when a random stranger responds to their not-private-public remarks ... sea-lioning.
We used to have more systems that were sort-of-private-public spaces - mailing lists being the most obvious. I sympathize with a generation that clearly wants more of these sorts of spaces to communicate with friends, but I am not sympathetic to their insistence that corporate creations that are not just very-much-non-private-public spaces but also essentially revenue generators should work the way they want them to.
That’s not what sealioning is at all. The rest of your generational rant is whatever it is, but sealioning is pretty well defined: https://en.wikipedia.org/wiki/Sealioning
If I repeated asked you for data to support your generalizations (“which younger people? Do you have an example? Why 2005 and not 2010?”) without admitting outright that I disagreed with you, that would be sealioning.
If you are being accused of sealioning, and you have 1) stated your opinion and 2) are asking good-faith questions in an effort to actually understand, then you’re probably not doing it. OTOH if that happens a lot, you might be the problem without realizing it.
I know both the cartoon and the wikipage. My interpretation of what is in the cartoon and on that page differs from yours a little. The critical part of sealioning from my perspective is not "repeatedly asking for data to support the position, without outright admitting disagreement", because that is common in other contexts where it is applauded. This is how debates at scientific conferences (or in other contexts) happen, much of the time. It's how debate and discussion happen online sometimes, and is frequently seen as appropriate and even welcome.
The specific thing that the cartoon gets at is that the questioner was not invited into the conversation (c.f. the "You're in my house") frame. They take a position that they have a right to ask questions, when the other person/people involved did not invite them to be participants in the exchange at all. The people in the house do not consider their conversation about sealions to be public; the sealion does, and responds.
> I can't imagine why someone would want to openly advertise that they're so closed minded.
I would say the exact same about you, rejecting an absolutely accurate and factual statement like that as closed minded strikes me as the same as the people who insist that medical science is closed minded about crystals and magnets.
I can't imagine why someone would want to openly advertise they think LLMs are actual intelligence, unless they were in a position to benefit financially from the LLM hype train of course.
I have no financial position w.r.t. LLMs in any way that I am aware of (it is possible that some of the mutual funds I put money into have investments in companies that work with LLMs, but I know of no specifics there).
I am not ready to say that "LLMs are actual intelligence", and most of their publically visible uses seem to me to be somewhere between questionable and ridiculous.
Nevertheless, I retain a keen ... shall we call it anti-skepticism? ... that LLMs, by modelling language, may have accidentally modelled/created a much deeper understanding of the world than was ever anticipated.
I do not want LLMs to "succeed", I think a society in which they are common is a worse society than the one in which we lived 5 years ago (as bad as that was), but my curiosity is not abated by such feelings.
Cool, so clearly articulate the goal posts. What do LLMs have to do to convince you that they are intelligent? If the answer is there is no amount of evidence that can change your mind, then you're not arguing in good faith.
It’s maybe an ethical and identity problem for most people. The idea that something not grounded in biology has somewhat the same « quality of intelligence » as us is disturbing.
It rises so many uncomfortable questions like, should we accept to be dominated and governed by a higher intelligence, should we keep it « slave » or give it « deserved freedom ». Are those questions grounded in reality or intelligence is just decoupled from the realm of biology and we don’t have to consider them at all. Only biological « being » with emotions/qualia should be considered relevant as regards to intelligence which does not matter on its own but only if it embodies qualia ?
It’s very new and a total shift in paradigm of life it’s hard to ask people to be in good faith here
That’s the main problem isn’t it ? Because it does matter and there is consequences to that like, should you « unplug » from the grid an AI ? Should we erase the memories of AI ? We eat animals and forbid eating humans, why ? Could we let AI « eat » some of us like in the matrix ?
Should we consider it our equal or superior to us ? Should we give it the reigns of politics if it’s superior in decision making ?
Or maybe the premise is « given all the knowledge that exists coupled with a good algorithm, you look/are/have intelligence » ? In which case intelligence is worthless in a way. It’s just a characteristic, not a quality. Which makes AI fantastic tools and never our equal ?
Maybe, I don't know, not be based on a statistical model?
Come on. If you are actually entertaining the idea that LLMs can possibly be intelligent, you don't know how they work.
But to take your silly question seriously for a minute, maybe I might consider LLMs to be capable of intelligence if they were able to learn, if they were able to solve problems that they weren't explicitly trained for. For example, have an LLM read a bunch of books about the strategy of Go, then actually apply that knowledge to beat an experienced Go player who was deliberately playing unconventional, poor strategies like opening in the center. Since pretty much nobody opens their Go game in the center (the corners are far superior), the LLM's training data is NOT going to have a lot of Go openings where one player plays mostly in the center. At which point you'll see that the LLM isn't actually intelligent, because an intelligent being would have understood the concepts in the book that you should mostly play in the corners at first in order to build territory with the smallest number of moves. But when faced with unconventional moves that aren't found anywhere on the Internet, the LLM would just crash and burn.
That would be a good test of intelligence. Learning by reading books, and then being able to apply that knowledge to new situations where you can't just regurgitate the training material.
Have you seen the now-years-old transcripts of "ancient" LLMs inventing new languages with grammar and syntax structures completely different from our own?
Can we as a group agree to stop upvoting "AI is great" and "AI sucks" posts that don't make novel, meaningful arguments that provoke real thought? The plagiarism argument is thin and feels biased, the lock-in argument is counter to the market dynamics that are currently playing out, and in general the takes are just one dude's vibes.
I don't know, this one is a little novel. I've never seen the developer of a Buddhist meditation app discuss whether to use LLMs with a paragraph like:
> Pariyatti’s nonprofit mission, it should be noted, specifically incorporates a strict code of ethics, or sīla: not to kill, not to steal, not to engage in sexual misconduct, not to lie, and not to take intoxicants.
If you're already sold on the plagiarism narrative that big entertainment is trying to propagandize in order to get leverage against the tech companies, nothing I say is going to change your mind.
I don't really know what you mean by "big entertainment" trying to get leverage against tech companies. Tech companies are behemoths. Most of the artists I know fretting about AI don't earn half a junior engineer's salary. And this is coming from someone who is relatively bullish on AI. I just don't think the framing of "big entertainment" makes any sense at all.
> That view of humans - and LLMs - ignores the fact that when you combine large numbers of simple building blocks, you can get completely novel behavior.
I can bang smooth rocks to get sharper rocks; that doesn't make sharper rocks more intelligent. Makes them sharper, though.
Yes, that seems to hold for rocks. But that doesn’t shut down the original post’s premise, unless you hold the answer to what can and cannot be banged together to create emergent intelligence.
Extraordinary assumptions (i.e., AI is conscious) require extraordinary proof. If I told you my bag of words was sentient, I assume you'd need more proof than just my word.
We're talking about LLMs that you can talk to, and which for the most part respond more intelligently than perhaps 90% of HN users (or 90% of people anywhere).
I've talked to them; they aren't that impressive. Decent search engines (assuming they provide links), though. The part where their context windows get muddied and they switch tones and behave weirdly is pretty telling. I've never experienced that in 90% people on the internet.
They're insanely impressive. They know a huge amount about almost every subject, are able to write better than the vast majority of humans, understand and can translate between virtually every written language, understand context, and can answer almost any question you ask them intelligently. If you're not impressed by that, you've set the bar impossibly high. Five years ago, LLMs would have been considered magic.
> "...it would sometimes regurgitate training data verbatim. That’s been patched in the years since..."
> "They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless."
This is totally misleading to anyone with less familiarity with how LLMs work. They are only programs in as much as they perform inference from a fixed, stored, statistical model. It turns out that treating them theoretically in the same way as other computer programs gives a poor representation of their behaviour.
This distinction is important, because no, "regurgitating data" is not something that was "patched out", like a bug in a computer program. The internal representations became more differentially private as newer (subtly different) training techniques were discovered. There is an objective metric by which one can measure this "plagiarism" in the theory, and it isn't nearly as simple as "copying" vs "not copying".
It's also still an ongoing issue and an active area of research, see [1] for example. It is impossible for the models to never "plagiarize" in the sense we think of while remaining useful. But humans repeat things verbatim too in little snippets, all the time. So there is some threshold where no-one seems to care anymore; think of it like the % threshold in something like Turnitin. That's the point that researchers would like to target.
Of course, this is separate from all of the ethical issues around training on data collected without explicit consent, and I would argue that's where the real issues lie.
[1] https://arxiv.org/abs/2601.02671
> This is totally misleading to anyone with less familiarity with how LLMs work. They are only programs in as much as they perform inference from a fixed, stored, statistical model. It turns out that treating them theoretically in the same way as other computer programs gives a poor representation of their behaviour.
Can you share any reading on this?
The plagiarism by the models is only part of it. Perhaps it's in such small pieces that it becomes difficult to care. I'm not convinced.
The larger, and I'd argue more problematic, plagiarism is when people take this composite output of LLMs and pass it off as their own.
> But humans repeat things verbatim too in little snippets, all the time
Also, it's possible, although statistically improbable, for a human to generate the exact same thing another human generated (and copyrighted) without even knowing it.
To a large extent both "hallucinations" and "plagiarism" can be addressed with the same training method: source-aware training.
https://arxiv.org/abs/2404.01019
At the frontier of science we have speculations, which until proper measurements become possible, are unknown to be true or false (or even unknown to be equivalent with other speculations etc. regardless of their being true or false, or truer or falser). Once settled we may call earlier but wrong speculations as "reasonable wrong guesses". In science it is important that these guesses or suspicions are communicated as it drives the design of future experiments.
I argue that more important that "eliminating hallucinations" is tracing the reason it is or was believed by some.
With source-aware training we can ask an LLM to give answers to a question (which may contradict each other), but to provide the training-source(s) justifying emission of each answer, instead of bluff it could emit multiple interpretations and go like:
> answer A: according to school of thought A the answer is that ... examples of authors and places in my training set are: author+title a1, a2, a3, ...
> answer B: according to author B: the answer to this question is ... which can be seen in articles b1, b2
> answer ...: ...
> answer F: although I can't find a single document explaining this, when I collate the observation x in x1, x2, x3; observation y in y1,y2, ... , observation z in z1, z2, ... then I conclude the following: ...
so it is clear which statements are sourced where, and which deductions are proper to the LLM.
Obviously few to none of the high profile LLM providers will do this any time soon, because when jurisdictions learn this is possible they will demand all models to be trained source-aware, so that they can remunerate the authors in their jurisdiction (and levy taxes on their income). What fraction of the income will then go to authors and what fraction to the LLM providers? If any jurisdiction would be first to enforce this, it would probably be the EU, but they don't do it yet. If models are trained in a different jurisdiction than the one levying taxes the academic in-group citation game will be extended to LLMs: a US LLM will have incentive to only cite US sources when multiple are available, and a EU trained LLM will prefer to selectively cite european sources, etc.
In addition to providing training sources, it's important to identify overlaps among the fragments used in the answer. For me, overlap doesn't mean simply identical expression, but conceptually identical.
We are much more likely to find conceptual overlap in code than in language and prose because Many of the problems we solve, as mathematicians say, reduce to previously solved problems, which IMO means substantially identical code.
A related question is how much change is necessary to a work of art, image, prose, or code for it to escape copyright? If we can characterize it and the LLM generates something that escapes copyright, I suggest the output should be excluded from future copyright or patent claims.
I wasn't aware of source-aware training, so thank you for the reference! It does seem a bit too good to be true; I believe in a system of tradeoffs so I feel like this must have an issue with reducing creativity. That's at first glance though, so I could be wrong.
> Translators are busy
No they're not. They're starving, struggling to find work and lamenting AI is eating their lunch. It's quite ironic that after complaining LLMs are plagiarism machines, the author thinks using them for translation is fine.
"LLMs are evil! Except when they're useful for me" I guess.
I think using translation software or AI images comes down to the ease of working with the system vrs a human. For example, I've created placeholder images for various pieces of work. The images didn't bring in any revenue; they were throwaway pieces, so I didn't give a shit. It wasn't worth paying a human for it.
Before I had AI-generated images, I either left out images from the work or used no-copyright clip art because, again, it wasn't worth arguing with or paying a human to do it.
When it came to diagrams, before Excalidraw, I would dust off my drafting skills, draw something on paper with colored pencils, take a picture of it, and use the picture as the diagram. In this case, I was willing to argue with and pay myself.
Simultaneously, if you hire human translators, you are likely to get machine translations. Maybe not often or overtly, but the translation industry has not been healthy for a while.
The industry is sick because everyone is looking for the lowest prices, but translators don't like machine translation. They don't want to just review the output, because actually doing the translation leads to better understanding of what they have to do.
A machine translation might be useful just to get a bulk-mode translation of unfamiliar words and possibly idioms too. But then it's time for humans.
>As a quick aside, I am not going to entertain the notion that LLMs are intelligent, for any value of “intelligent.” They are robots. Programs. Fancy robots and big complicated programs, to be sure — but computer programs, nonetheless. The rest of this essay will treat them as such. If you are already of the belief that the human mind can be reduced to token regurgitation, you can stop reading here. I’m not interested in philosophical thought experiments.
I can't imagine why someone would want to openly advertise that they're so closed minded. Everything after this paragraph is just anti-LLM ranting.
What's wrong about the statement? The black box algorithm might have been generated by machine learning, but it's still a computer program in the end.
Because it's so entirely reductive and misunderstanding of where the technology has progressed. Hello world is s computer program. So it Microsoft Windows. New levels of "intelligence" unlock with greater complexity of a program.
Like look at our brains. We know decently well how a single neuron works. We can simulate a single one with "just a computer program". But clearly with enough layers some form of complexity can emerge, and at some level that complexity becomes intelligence.
> with enough layers some form of complexity can emerge, and at some level that complexity becomes intelligence.
It isn’t a given that complexity begets intelligence.
and it isn't a given that it doesn't, so maybe a little openness towards the possibility is warranted?
I’m open, but the comment I responded to asserted: “complexity becomes intelligence”, as if it is a fact. And it isn’t proven.
I said "intelligence can emerge" not that it will.
We have LLMs, which are obviously intelligent. How is it not proven?
There is no "obvious" about it, unless you define "intelligent" in a rather narrow (albeit Turing-esque) way.
The suspicion is that they are good at predicting next-token and not much else. This is still a research topic at this point, from my reading.
You can't predict the next token in an arbitrary text unless you are highly intelligent and have a vast body of knowledge.
They're obviously intelligent in the way that we judge intelligence in humans: we pay attention to what they say. You ask them a question about an arbitrary subject, and they respond in the same way that an intelligent person would. If you don't consider that intelligence, then you have a fundamentally magical, unscientific view of what intelligence is.
To return to an analogy I used a couple of days ago ... birds can fly, planes can fly, ergo they are both flying things ... but they fly in completely different ways. So on the one hand (visible behavior) they are similar (or even the same), and on the other (physical mechanism) they are not similar at all.
Which one of these comparisons you want to use depends on context.
The same seems entirely possible for current LLMs. On the one hand they do something that visibly seems to to be the same as something humans do, but on the other it is possible that the way they do it entirely different. Just as with the bird/plane comparison, this has some implications when you start to dig deeper into capabilities (e.g. planes cannot fly anywhere near as slowly as birds, and birds cannot fly as fast as planes; birds have dramatically more maneuverability than planes, etc. etc).
So are LLMs intelligent in the same way humans are? Depends on your purpose in asking that question. Planes fly, but they are not birds.
I know you're arguing with someone else, but I think it is getting sidetracked.
Whether or not LLMs are intelligent (I think they are more intelligent than a cat, for instance, but less intelligent than a human) isn't my argument.
My argument is that complexity in and of itself doesn't yield intelligence. There's no proof of that. There are many things that are very very complex, but we would not put it on an intelligence scale.
When has anyone ever said that every complex thing is intelligent?
It was implicated in another comment.
I said that complexity can lead to intelligence, not that it must.
But in the case of both biological and computer neurons, it is an empirical fact that complexity has led to intelligence.
A provocative aside in bad faith, anyway a completely minor point within the overall post, which some of the people he's telling to fuck off might have read
I disagree that the majority of it is anti-LLM ranting, there are several subtle points here that are grounded in realism. You should read on past the first bit if you're judging mainly from the initial (admittedly naive) first few paragraphs.
> You should read on past the first bit...
Not GP, but... the author said explicitly "if you believe X you should stop reading". So I did.
The X here is "that the human mind can be reduced to token regurgitation". I don't believe that exactly, and I don't believe that LLMs are conscious, but I do believe that what the human mind does when it "generates text" (i.e. writes essays, programs, etc) may not be all that different from what an LLM does. And that means that most of humans's creations are also the "plagiarism" in the same sense the author uses here, which makes his argument meaningless. You can't escape the philosophical discussion he says that he's not interested in if you want to talk about ethics.
Edit: I'd like to add that I believe that this also ties in to the heart of the philosophy of Open Source and Open Science... if we acknowledge that our creative output is 1% creative spark and 99% standing on the shoulders of Giants, then "openness" is a fundamental good, and "intellectual property" is at best a somewhat distasteful necessity that should be as limited as possible and at worst is outright theft, the real plagiarism.
So do you believe seahorse emoji exists?
I read the rest of it. It was intellectually lazy.
It's more intellectually lazy to think boolean logic at a sufficient scale crosses some event horizon wherein its execution on mechanical gadgets called computers somehow adds up to intelligence beyond human understanding.
It is intellectually lazy to proclaim something to be impossible in the absence of evidence or proof. In the case of the statement made here, it is provably true that Boolean logic at sufficient scale can replicate "intelligence" of any arbitrary degree. It is also easy to show that this can be perceived as an "event horizon" since the measurements of model quality that humans typically like to use are so nonlinear that they are virtually step function-like.
Doesn't seem like you have proof of anything but it does appear that you have something that is very much like religious faith in an unforeseeable inevitability. Which is fine as far as religion is concerned but it's better to not pretend it's anything other than blind faith.
But if you really do have concrete proof of something then you'll have to spell it out better & explain how exactly it adds up to intelligence of such magnitude & scope that no one can make sense of it.
> "religious faith in an unforeseeable inevitability"
For reference, I work in academia, and my job is to find theoretical limitations of neural nets. If there was so much of a modicum of evidence to support the argument that "intelligence" cannot arise from sufficiently large systems, my colleagues and I would be utterly delighted and would be all over it.
Here are a couple of standard elements without getting into details:
1. Any "intelligent" agent can be modelled as a random map from environmental input to actions.
2. Any random map can be suitably well-approximated by a generative transformer. This is the universal approximation theorem. Universal approximation does not mean that models of a given class can be trained using data to achieve an arbitrary level of accuracy, however...
3. The neural scaling laws (first empirical, now more theoretically established under NTK-type assumptions), as a refinement of the double descent curve, assert that a neural network class can get arbitrarily close to an "entropy level" given sufficient scale. This theoretical level is so much smaller than any performance metric that humans can reach. Whether "sufficiently large" is outside of the range that is physically possible is a much longer discussion, but bets are that human levels are not out of reach (I don't like this, to be clear).
4. The nonlinearity of accuracy metrics comes from the fact that they are constructed from the intersection of a large number of weakly independent events. Think the CDF of a Beta random variable with parameters tending to infinity.
Look, I understand the scepticism, but from where I am, reality isn't leaning that way at the moment. I can't afford to think it isn't possible. I don't think you should either.
As I said previously, you are welcome to believe whatever you find most profitable for your circumstances but I don't find your heuristics convincing. If you do come up or stumble upon a concrete constructive proof that 100 trillion transistors in some suitable configuration will be sufficiently complex to be past the aforementioned event horizon then I'll admit your faith was not misplaced & I will reevaluate my reasons for remaining skeptical of Boolean arithmetic adding up to an incomprehensible kind of intelligence beyond anyone's understanding.
It was actually much less anti LLM than I was expecting from the beginning.
But I agree that it is self limiting to not bother to consider the ways that LLM inference and human thinking might be similar (or not).
To me, they seem do a pretty reasonable emulation of single- threaded thinking.
They are not similar. A LLM is a complex statistical machine. A brain is a highly complex neural network. A brain, is more similar the perceptron of some AMD CPUs that to a LLM.
I would recommend investigating how contemporary LLMs actually work.
Possibly start with something like: https://transformer-circuits.pub/2025/attribution-graphs/bio...
> I can't imagine why someone would want to openly advertise that they're so closed minded.
It's not being closed-minded. It's not wanting to get sea-lioned to death by obnoxious people.
[WARNING: seriously off-topic comment, I was triggered]
Here's what sea-lioned means to me:
I say something.
You accuse me of sea-lioning.
I have two choices: attempt to refute the sea-lioning, which becomes sea-lioning, or allowing your accusation to stand unchallenged, which appears to most people as a confirmation of some kind that I was sea-lioning.
It is a nuclear weapon launched at discussion. It isn't that it doesn't describe a phenomena that actually happens in the world. However, it is a response/accusation to which there is never any way to respond to that doesn't confirm the accusation, whether it was true or not.
It is also absolutely rooted in what appears to me to be a generational distinction: it seems that a bunch of younger people consider it to be a right to speak "in public" (i.e in any kind of online context where people who do not know you can read what you wrote) and expect to avoid a certain kind of response. Should that response arise? Various things will be said about the responder, including "sea-lioning".
My experience is that people who were online in the 80s and 90s find this expectation somewhere between humorous and ridiculous, and that people who went online somewhere after about 2005 do not.
Technologically, it seems to reflect a desire among many younger people for "private-public spaces". In the absence of any such actual systems really existing (at least from their POV), they believe they ought to be able to use very non-private public spaces (facebook, insta, and everything else under the rubric of "social media") as they wish to, rather than as the systems were designed. They are communicating with their friends and the fact that their conversations are visible is not significant. Thus, when a random stranger responds to their not-private-public remarks ... sea-lioning.
We used to have more systems that were sort-of-private-public spaces - mailing lists being the most obvious. I sympathize with a generation that clearly wants more of these sorts of spaces to communicate with friends, but I am not sympathetic to their insistence that corporate creations that are not just very-much-non-private-public spaces but also essentially revenue generators should work the way they want them to.
That’s not what sealioning is at all. The rest of your generational rant is whatever it is, but sealioning is pretty well defined: https://en.wikipedia.org/wiki/Sealioning
If I repeated asked you for data to support your generalizations (“which younger people? Do you have an example? Why 2005 and not 2010?”) without admitting outright that I disagreed with you, that would be sealioning.
If you are being accused of sealioning, and you have 1) stated your opinion and 2) are asking good-faith questions in an effort to actually understand, then you’re probably not doing it. OTOH if that happens a lot, you might be the problem without realizing it.
I know both the cartoon and the wikipage. My interpretation of what is in the cartoon and on that page differs from yours a little. The critical part of sealioning from my perspective is not "repeatedly asking for data to support the position, without outright admitting disagreement", because that is common in other contexts where it is applauded. This is how debates at scientific conferences (or in other contexts) happen, much of the time. It's how debate and discussion happen online sometimes, and is frequently seen as appropriate and even welcome.
The specific thing that the cartoon gets at is that the questioner was not invited into the conversation (c.f. the "You're in my house") frame. They take a position that they have a right to ask questions, when the other person/people involved did not invite them to be participants in the exchange at all. The people in the house do not consider their conversation about sealions to be public; the sealion does, and responds.
> I can't imagine why someone would want to openly advertise that they're so closed minded.
I would say the exact same about you, rejecting an absolutely accurate and factual statement like that as closed minded strikes me as the same as the people who insist that medical science is closed minded about crystals and magnets.
I can't imagine why someone would want to openly advertise they think LLMs are actual intelligence, unless they were in a position to benefit financially from the LLM hype train of course.
I have no financial position w.r.t. LLMs in any way that I am aware of (it is possible that some of the mutual funds I put money into have investments in companies that work with LLMs, but I know of no specifics there).
I am not ready to say that "LLMs are actual intelligence", and most of their publically visible uses seem to me to be somewhere between questionable and ridiculous.
Nevertheless, I retain a keen ... shall we call it anti-skepticism? ... that LLMs, by modelling language, may have accidentally modelled/created a much deeper understanding of the world than was ever anticipated.
I do not want LLMs to "succeed", I think a society in which they are common is a worse society than the one in which we lived 5 years ago (as bad as that was), but my curiosity is not abated by such feelings.
Cool, so clearly articulate the goal posts. What do LLMs have to do to convince you that they are intelligent? If the answer is there is no amount of evidence that can change your mind, then you're not arguing in good faith.
It’s maybe an ethical and identity problem for most people. The idea that something not grounded in biology has somewhat the same « quality of intelligence » as us is disturbing. It rises so many uncomfortable questions like, should we accept to be dominated and governed by a higher intelligence, should we keep it « slave » or give it « deserved freedom ». Are those questions grounded in reality or intelligence is just decoupled from the realm of biology and we don’t have to consider them at all. Only biological « being » with emotions/qualia should be considered relevant as regards to intelligence which does not matter on its own but only if it embodies qualia ? It’s very new and a total shift in paradigm of life it’s hard to ask people to be in good faith here
But you don't and cannot know if qualia exist in a system, so how can that ever be a criteria for any kind of qualification?
That’s the main problem isn’t it ? Because it does matter and there is consequences to that like, should you « unplug » from the grid an AI ? Should we erase the memories of AI ? We eat animals and forbid eating humans, why ? Could we let AI « eat » some of us like in the matrix ?
Should we consider it our equal or superior to us ? Should we give it the reigns of politics if it’s superior in decision making ? Or maybe the premise is « given all the knowledge that exists coupled with a good algorithm, you look/are/have intelligence » ? In which case intelligence is worthless in a way. It’s just a characteristic, not a quality. Which makes AI fantastic tools and never our equal ?
Maybe, I don't know, not be based on a statistical model?
Come on. If you are actually entertaining the idea that LLMs can possibly be intelligent, you don't know how they work.
But to take your silly question seriously for a minute, maybe I might consider LLMs to be capable of intelligence if they were able to learn, if they were able to solve problems that they weren't explicitly trained for. For example, have an LLM read a bunch of books about the strategy of Go, then actually apply that knowledge to beat an experienced Go player who was deliberately playing unconventional, poor strategies like opening in the center. Since pretty much nobody opens their Go game in the center (the corners are far superior), the LLM's training data is NOT going to have a lot of Go openings where one player plays mostly in the center. At which point you'll see that the LLM isn't actually intelligent, because an intelligent being would have understood the concepts in the book that you should mostly play in the corners at first in order to build territory with the smallest number of moves. But when faced with unconventional moves that aren't found anywhere on the Internet, the LLM would just crash and burn.
That would be a good test of intelligence. Learning by reading books, and then being able to apply that knowledge to new situations where you can't just regurgitate the training material.
Have you seen the now-years-old transcripts of "ancient" LLMs inventing new languages with grammar and syntax structures completely different from our own?
> I can't imagine why someone would want to openly advertise that they're so closed minded.
Because humans often anthropomorphize completely inert things? E.g. a coffee machine or a bomb disposal robot.
So far whatever behavior LLMs have shown is basically fueled by Sci-Fi stories of how a robot should behave under such and such.
Can we as a group agree to stop upvoting "AI is great" and "AI sucks" posts that don't make novel, meaningful arguments that provoke real thought? The plagiarism argument is thin and feels biased, the lock-in argument is counter to the market dynamics that are currently playing out, and in general the takes are just one dude's vibes.
I don't know, this one is a little novel. I've never seen the developer of a Buddhist meditation app discuss whether to use LLMs with a paragraph like:
> Pariyatti’s nonprofit mission, it should be noted, specifically incorporates a strict code of ethics, or sīla: not to kill, not to steal, not to engage in sexual misconduct, not to lie, and not to take intoxicants.
Not a whole lot of Pali in most LLM editorials.
> not to engage in sexual misconduct
I must remember to add this quality guarantee to my own software projects.
My software projects are also uranium-free.
I dunno, I enjoyed reading about how the author personally feels about the act of working with them more then the whole "is this moral" part.
No
> The plagiarism argument is thin and feels biased
are you being serious with this one
If you're already sold on the plagiarism narrative that big entertainment is trying to propagandize in order to get leverage against the tech companies, nothing I say is going to change your mind.
Unless you're anti-copyright in general, there's a real issue here, which can easily be demonstrated by prompting LLMs to reproduce copyrighted work.
"Big entertainment" may be using that issue in ways you don't personally approve of, but that doesn't negate the issue.
I don't really know what you mean by "big entertainment" trying to get leverage against tech companies. Tech companies are behemoths. Most of the artists I know fretting about AI don't earn half a junior engineer's salary. And this is coming from someone who is relatively bullish on AI. I just don't think the framing of "big entertainment" makes any sense at all.
Give it up. Buddha would not approve.
And there will be more compute for the rest of us :)
> LLMs will always be plagiarism machines but in 40 years we might not care.
40 years?
Virtually nobody cares about this already... today.
(I'm not refuting the author's claim that LLMs are built on plagiarism, just noting how the world has collectively decided to turn a blind eye to it)
[flagged]
> That view of humans - and LLMs - ignores the fact that when you combine large numbers of simple building blocks, you can get completely novel behavior.
I can bang smooth rocks to get sharper rocks; that doesn't make sharper rocks more intelligent. Makes them sharper, though.
Which is to say, novel behavior != intelligence.
Yes, that seems to hold for rocks. But that doesn’t shut down the original post’s premise, unless you hold the answer to what can and cannot be banged together to create emergent intelligence.
Extraordinary assumptions (i.e., AI is conscious) require extraordinary proof. If I told you my bag of words was sentient, I assume you'd need more proof than just my word.
We're not talking about sharp rocks.
We're talking about LLMs that you can talk to, and which for the most part respond more intelligently than perhaps 90% of HN users (or 90% of people anywhere).
I've talked to them; they aren't that impressive. Decent search engines (assuming they provide links), though. The part where their context windows get muddied and they switch tones and behave weirdly is pretty telling. I've never experienced that in 90% people on the internet.
They're insanely impressive. They know a huge amount about almost every subject, are able to write better than the vast majority of humans, understand and can translate between virtually every written language, understand context, and can answer almost any question you ask them intelligently. If you're not impressed by that, you've set the bar impossibly high. Five years ago, LLMs would have been considered magic.
I stopped reading after "problem with LLMs is plagiarism"...
Too bad. You missed some interesting stuff. And I say that as someone who sees some of this very differently than the author.
Announcing that one line of the piece made you mad without providing any other thought is not very constructive.