I am live.

  • 0 Posts
  • 115 Comments
Joined 3 years ago
cake
Cake day: July 7th, 2023

help-circle






  • No. That is not what the analogy means. That is what you are choosing to extract from it because it supports the direction you want this exchange to go.

    The use of the word “regurgitate” carries a very specific implication. It suggests that LLMs retrieve and repeat stored information verbatim. That is not how they function. We both appear to agree on that point.

    LLMs do not rely on stored facts in the way the analogy implies. They generate outputs by modeling patterns in data, producing responses that are often novel rather than retrieved.

    Whether or not the model understands or comprehends the content is irrelevant to this distinction. Comprehension is not a requirement for the system to function. So yes, the analogy is overly simplistic and ignores the actual mechanism at work.

    To be precise: it does not matter that the model lacks awareness or understanding. It is still capable of analyzing patterns and generating new outputs from its training data. That is not regurgitation.

    Concisely as I can: llms do not regurgitate data, the analogy fails.








  • No. You’re not just wrong, you’re aggressively uninformed.

    By you repeating the same tired “AI is just regurgitating data” line makes it clear you don’t understand what you’re criticizing. Calling large language models “AI” the way you are doing it just exposes that you do not know what you are talking about. It is like a creationist smugly saying “orangutang” instead of “orangutan” and thinking they sound informed. You are not demonstrating insight. You are advertising ignorance.

    What you’re describing, reading a paragraph off Wikipedia, is literal retrieval. That is not how modern language models operate. They are not databases with a search bar attached. They are probabilistic systems trained to model patterns, structure, and relationships across massive datasets. When they generate a response, they are not pulling a stored paragraph. They are constructing output token by token based on learned representations.

    If it were just regurgitation, you would constantly see verbatim copies of training data. You do not. What you see instead is synthesis. Concepts are recombined, abstracted, and adapted to context. The system can explain the same idea multiple ways, shift tone, handle novel prompts, and connect ideas that were never explicitly paired in the source material. That is fundamentally different from reading something out loud.

    Your analogy fails because it assumes nothing is being transformed. In reality, transformation is the entire mechanism. Information is compressed into weights and then expanded into new outputs.

    Is it human intelligence. No. Is it perfect. No. But reducing it to “just reading Wikipedia out loud” is not skepticism. It is a basic failure to understand how the technology works.

    If you are going to criticize something, at least learn what it is first.







  • Yes, you are correct. Those of you who are concerned about this are not wrong to question it.

    However, the point that keeps being ignored is that laws like this have very limited enforceability when it comes to platforms like Linux and other open-source software.

    The reason is simple, anyone can modify the source code. There is no practical way to permanently embed restrictions like age verification into something that can be freely forked and redistributed. If a Linux distribution introduces age verification, a fork removing it will appear almost immediately. That is not hypothetical, that is how the open-source ecosystem functions.

    Even if you personally install a version that includes such a feature, it is often trivial to bypass or remove it through system-level access.

    Yes, the laws themselves are poorly conceived. They attempt to impose control in an environment that does not respond well to centralized regulation. But focusing on something like a birthday field in a Linux distribution misses the point. In that context, it is effectively meaningless and not something that warrants serious concern.