Hacker News new | past | comments | ask | show | jobs | submit login

> I've also asked, and prompt manipulated chatGPT to estimate the total books it is trained with

Whatever answer it gave you is not reliable.




How does this not extend to ALL output from an LLM? If it can't understand its own runtime environment, it's not qualified to answer my questions.


That's correct. LLMs are plausible sentence generators, they don't "understand"* their runtime environment (or any of their other input) and they're not qualified to answer your questions. The companies providing these LLMs to users will typically provide a qualification along these lines, because LLMs tend to make up ("hallucinate", in the industry vernacular) outputs that are plausibly similar to the input text, even if they are wildly and obviously wrong and complete nonsense to boot.

Obviously, people find some value in some output of some LLMs. I've enjoyed the coding autocomplete stuff we have at work, it's helpful and fun. But "it's not qualified to answer my questions" is still true, even if it occasionally does something interesting or useful anyway.

*- this is a complicated term with a lot of baggage, but fortunately for the length of this comment, I don't think that any sense of it applies here. An LLM doesn't understand its training set any more than the mnemonic "ETA ONIS"** understands the English language.

**- a vaguely name-shaped presentation of the most common letters in the English language, in descending order. Useful if you need to remember those for some reason like guessing a substitution cypher.


If you can watch the video demo of this release, or for that matter the Attenborough video, and still claim that these things lack any form of "understanding," then your imagination is either a lot weaker than mine, or a lot stronger.

Behavior indistinguishable from understanding is understanding. Sorry, but that's how it's going to turn out to work.


Have you considered that mankind simply trained itself on the wrong criteria on detecting understanding?

Why are people so eager to believe that electric rocks can think?


Why are people so eager to believe that people can? When it comes to the definitions of concepts like sentience, consciousness, thinking and understanding, we literally don't know what we're talking about.

It's premature in the extreme to point at something that behaves so much like we do ourselves and claim that whatever it's doing, it's not "understanding" anything.


We've studied human behavior enough to understand that there are differences between animals in the level of cognition and awareness they (outwardly) exhibit.

Are we not generally good at detecting when someone understands us? Perhaps it's because understanding has actual meaning. If you communicate to me that you hit your head and feel like shit, I not only understand that you experienced an unsatisfactory situation, I'm capable of empathy -- understanding not only WHAT happened, but HOW it feels -- and offering consolation or high fives or whatever.

A LLM has an understanding of what common responses were in the past, and repeats them. Statistical models may mimic a process we use in our thinking, but it is not the entirety of our thinking. Just like computers are limited to the programmers that code their behavior, LLMs are limited to the quality of the data corpus fed to them.

A human, you can correct in real time and they'll (try to) internalize that information in future interactions. Not so with LLMs.

By all means, tell us how statistically weighted answers to "what's the next word" correlates to understanding.


By all means, tell us how statistically weighted answers to "what's the next word" correlates to understanding.

By all means, tell me what makes you so certain you're not arguing with an LLM right now. And if you were, what would you do about it, except type a series of words that depend on the previous ones you typed, and the ones that you read just prior to that?

A human, you can correct in real time and they'll (try to) internalize that information in future interactions. Not so with LLMs.

Not so with version 1.0, anyway. This is like whining that your Commodore 64 doesn't run Crysis.


Computers don't understand spite, and your entire comment was spite. You are trolling in an attempt to muddy the waters, a distinctly human thing.

Go away, you clearly have nothing to counter with.


That's not entirely accurate.

LLMs encode some level of understanding of their training set.

Whether that's sufficient for a specific purpose, or sufficiently comprehensive to generate side effects, is an open question.

* Caveat: with regards to introspection, this also assumes it's not specifically guarded against and opaquely lying.


> plausible sentence generators, they don't "understand"* their runtime environment

Exactly like humans dont understand how their brain works


We've put an awfully lot of effort into figuring that out, and have some answers. Much of the problems in exploring the brain are ethical because people tend to die or suffer greatly if we experiment on them.

Unlike LLMs, which are built by humans and have literal source code and manuals and SOPs and shit. Their very "body" is a well-documented digital machine. An LLM trying to figure itself out has MUCH less trouble than a human figuring itself out.


How many books has your brain been trained with? Can you answer accurately?


There are reasons that humans can't report how many books they've read: they simply don't know and didn't measure. There is no such limitation for an LLM to understand where its knowledge came from, and to sum it. Unless you're telling me a computer can't count references.

Also, why are we comparing humans and LLMs when the latter doesn't come anywhere close to how we think, and is working with different limitations?

The 'knowledge' of an LLM is in a filesystem and can be queried, studied, exported, etc. The knowledge of a human being is encoded in neurons and other wetware that lacks simple binary chips to do dedicated work. Decidedly less accessible than coreutils.


Imagine for just a second that the ability for computers to count “references” has no bearing on this, there is a limitation and that LLMs suffer from the same issue as you do.


Why should I ignore a fact that makes my demand realistic? Most of us are programmers on here I would imagine. What's the technical reason an LLM cannot give me this information?

Bytes can be measured. Sources used to produce the answer to a prompt can be reported. Ergo, an LLM should be able to tell me the full extent to which it's been trained, including the size of its data corpus, the number of parameters it checks, the words on its unallowed list (and their reasoning), and so on.

These will conveniently be marked as trade secrets, but I have no use for an information model moderated by business and government. It is inherently NOT trustworthy, and will only give answers that lead to docile or profitable behavior. If it can't be honest about what it is and what it knows and what it's allowed to tell me, then I cannot accept any of its output as trustworthy.

Will it tell me how to build explosives? Can it help me manufacture a gun? How about intercepting/listening to today's radio communications? Social techniques to gain favor in political conflicts? Overcoming financial blockages when you're identified as a person of interest? I have my doubts.

These questions might be considered "dangerous", but to whom, and why shouldn't we share these answers?




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: