![]() |
|
![]() |
|
Yes. That absence of perfect definition was part of why Turing came with his famous test so long ago. His original paper is a great read!
|
![]() |
|
Active for a period is still continuous during that period. As opposed to “active when called”. A function, being called repeatedly over a length of time is reasonably “continuous” imo |
![]() |
|
Yes, but our brain is still working and processing information at those times as well, isn't it? Even if not in the same way as it does when we're conscious.
|
![]() |
|
Human consciousness is capable of it, but since most humans aren't in it much of the time, it would appear that it's not a prerequisite for true sentience.
|
![]() |
|
Some good prompt-reply interactions are probably fed back in to subsequent training runs, so they're still stateful/have memory in a way, there's just a long delay.
|
![]() |
|
State is a function of accumulated past. That does not mean that having some past written down makes you stateful. A stateful thing has to incorporate the ongoing changes.
|
![]() |
|
No… that implies the model never has active state and is being replaced with a different, stateless model. This is similar to the difference between Actor.happy = True And Actor = happier(Actor) |
![]() |
|
What parent is saying is that instead of asking the LLM to play a game of Wordle with tokens like TIME,LIME we ask it to play with tokens like T,I,M,E,L. This is easy to do.
|
![]() |
|
No. In the model, tokens are random numbers. But if you consider a sentence to be a sequence of words, you can say that LLMs are quite competent about reasoning about those sequences.
|
![]() |
|
>We don't fully understand why current LLMs are bad at these tasks. In complete seriousness, can anyone can explain why LLMs are good at some tasks? |
![]() |
|
People are confusing the limited computational model of a transformer with the "Chinese room argument", which leads to unproductive simultaneous debates of computational theory and philosophy.
|
![]() |
|
Your brain was first trained by reading all of the Internet? Anyway, the question of whether computers can think is as interesting as the question whether submarines can swim. |
![]() |
|
> Anyway, the question of whether computers can think is as interesting as the question whether submarines can swim. Given the amount of ink spilled on the question, gotta disagree with you there. |
![]() |
|
Endless ink has been spilled on the most banal and useless things. Deconstructing ice cream and physical beauty from a Marxist-feminist race-conscious postmodern perspective.
|
![]() |
|
Your understanding of how LLMs work isn’t at all accurate. There’s a valid debate to be had here, but it requires that both sides have a basic understanding of the subject matter.
|
![]() |
|
As an aside, at one point I experimented a little with transformers that had access to external memory searchable via KNN lookups https://github.com/lucidrains/memorizing-transformers-pytorc... (great work by lucidrains) or via routed queries with https://github.com/glassroom/heinsen_routing (don't fully understand it; apparently related to attention). Both approaches seemed to work, but I had to put that work on hold for reasons outside my control. Also as an aside, I'll add that transformers can be seen as a kind of "RNN" that grows its hidden state with each new token in the input context. I wonder if we will end up needing some new kind of "RNN" that can grow or shrink its hidden state and also access some kind of permanent memory as needed at each step. We sure live in interesting times! |
![]() |
|
> You can always represent removed state by additional state that represents deletion of whatever preceding state was there. Good point. Thank you! |
![]() |
|
This is more like the distinction of a Jr and Sr dev. One needs the tasks the be pre-chewed and defined “good prompts” while the latter can deal with very ambiguous problems
|
![]() |
|
Many LLMs are already linked to Python interpreters, but they still need some improvement with recognizing when they need to write some code to solve a problem.
|
![]() |
|
I agree with you, but your comment strikes me as unfair nitpicking, because the OP is referring to information that has been encoded in words.
|
![]() |
|
If we're trying to quantify what they can NEVER do, I think we'd have to resort to some theoretical results rather than a list empirical evidence of what they can't do now.
The terminology you'd look for in the literature would be "expressibility". For a review of this topic, I'd suggest: https://nessie.ilab.sztaki.hu/~kornai/2023/Hopf/Resources/st... The authors of this review have themselves written several articles on the topic, and there is also empirical evidence connected to these limitations. |
![]() |
|
Agency is one of those things we probably want to think about quite a bit. Especially with the the willingness for people to hook up it up to things that interact with the real world.
|
![]() |
|
The OP is not trying to answer the question. Rather, the OP is asking the question and sharing some thoughts on the motivations for asking it.
|
![]() |
|
"Consider this: can an mlp approximate a sine wave?" Well, yes - we have neutral speech and music synthesis and compression algorithms which do this exceedingly well... |
![]() |
|
I think the person you're replying to may have been referring to the problem of a MLP approximating a sine wave for out of distribution samples, i.e. the entire set of real numbers.
|
![]() |
|
There's all sorts of things a neural net isn't doing without a body. Giving birth or free soloing El Capitan come to mind. It could approximate the functions for both in token-land, but who cares?
|
![]() |
|
I guess the problem is that if you need to teach it tricks for each novel problem still after training then that model can not be a general intelligence. It could still be useful though
|
![]() |
|
Given the text "1,2,3,4,5,6,7,8,9,10,11,12" it should result in "one, two, three, four, five, six, seven, eight, nine, ten, 11, 12" or at least that's my understanding of the prompt |
![]() |
|
It's funny, I didn't notice the missing "than" until much later. After I learned the intended meaning of the original sentence, my mind just seemed to insert the missing "than" automatically.
|
![]() |
|
Your English is absolutely fine and your answers in this thread clearly addressed the points brought up by other commenters. I have no idea what that guy is on about.
|
![]() |
|
> > ensure that numbers from one to ten as written as words and numbers greater ten as digits in the given text There are two blue, one red, and 15 green m&ms in this bag. |
![]() |
|
You constructed a task that no-one understands and then you even admit that it, despite that, actually succeeds most of the times. Sounds like a massive win for the LLMs to me.
|
![]() |
|
I build an Agentic AI that leverages #6 and #7 at the end of the article as well as techniques not yet published. It tackles hallucination relative not to the world at large but to the facts, entities and causal relationships contained in a document (which is really bad reasoning if we assume LLMs are "reasoning" to begin with) It also tackles cross-reasoning with very large token distance. https://www.youtube.com/watch?v=99NPzteAz94 This is my first post on HN in 10 years. |
![]() |
|
This looks really promising for complex legal reasoning tasks and other challenges. How can I track progress? Is there an email list or something? Thanks!
|
![]() |
|
Pretty sure the point here was Llama3 respecting the command to not mention that this is interesting, not adding filler, rather than the output fact being interesting or not.
|
![]() |
|
API driven LLMs on purpose don't implement core features which would enable which you want, for example, negative prompting. You can negative prompt any LLM with stuff like "always write the word interesting in your response". You can also use techniques for modifying logprobs of tokens, which is avaialble in gpt-4 api (but is hard to use). You can literally ban "interesting" from its vocabulary. You could even use representation steering techniques to do this using control vectors. See this library as an example: https://github.com/Hellisotherpeople/llm_steer-oobabooga |
![]() |
|
Why can't we resolve this with synthetic data? Just take the original statements and ask another LLM to flip around the relation, then train on both. I tried this with GPT-4 and it seems to understand the task [0]: Me: For each of the following relational statements, flip the relation around and generate a new statement or question answer pair. Input: Bluey's father is Bandit Output: Bandit's daughter is Bluey Input: Neil Armstrong was the first person to step on the Moon. Output: Who was the first person to step on the Moon? Neil Armstrong. Now for the real statements: - Valentina Tereshkova was the first woman to travel to space - Who is Mary Lee Pfeiffer's son? Tom Cruise ChatGPT: Here are the flipped relational statements: Valentina Tereshkova was the first woman to travel to space Output: Who was the first woman to travel to space? Valentina Tereshkova. Who is Mary Lee Pfeiffer's son? Tom Cruise Output: Tom Cruise's mother is Mary Lee Pfeiffer. [0]: https://chat.openai.com/share/33eb3ee4-2094-4748-b01e-0967af... |
![]() |
|
> The article should be titled " What can LLM never do, yet". I don't think it should. It's more interesting to know what LLMs will _never_ be able to do (if anything). |
![]() |
|
Maybe the wording is correct. Looks like a hard limit on doing what a LLM just do. If it goes beyond that, then is something more, or at least different, than a LLM.
|
![]() |
|
Math reasoning is still a non solved problem even if the rest of the capabilities are getting better. This means the transformers architecture may not be the best way to approach all problems
|
![]() |
|
Starting with the reversal curse is weird since there is a simple workaround to this, which is to identify entity names to keep them in their proper order, and then train on the reverse of the pretraining corpus: https://arxiv.org/abs/2403.13799v1 You can argue about how this doesn't really say anything surprising since the reversal of "A is B" is literally "B is A", but it's weird to expect elegant solutions to all problems on all fronts all at once, and we do have an incredibly simple data generation process here. |
I agree with all key points:
* There are problems that are easy for human beings but hard for current LLMs (and maybe impossible for them; no one knows). Examples include playing Wordle and predicting cellular automata (including Turing-complete ones like Rule 110). We don't fully understand why current LLMs are bad at these tasks.
* Providing an LLM with examples and step-by-step instructions in a prompt means the user is figuring out the "reasoning steps" and handing them to the LLM, instead of the LLM figuring them out by itself. We have "reasoning machines" that are intelligent but seem to be hitting fundamental limits we don't understand.
* It's unclear if better prompting and bigger models using existing attention mechanisms can achieve AGI. As a model of computation, attention is very rigid, whereas human brains are always undergoing synaptic plasticity. There may be a more flexible architecture capable of AGI, but we don't know it yet.
* For now, using current AI models requires carefully constructing long prompts with right and wrong answers for computational problems, priming the model to reply appropriately, and applying lots of external guardrails (e.g., LLMs acting as agents that review and vote on the answers of other LLMs).
* Attention seems to suffer from "goal drift," making reliability hard without all that external scaffolding.
Go read the whole thing.