That’s a great line of thought. Take an algorithm of “simulate a human brain”. Obviously that would break the paper’s argument, so you’d have to find why it doesn’t apply here to take the paper’s claims at face value.
That’s a great line of thought. Take an algorithm of “simulate a human brain”. Obviously that would break the paper’s argument, so you’d have to find why it doesn’t apply here to take the paper’s claims at face value.
There’s a number of major flaws with it:
IMO there’s also flaws in the argument itself, but those are more relevant
This is a silly argument:
[…] But even if we give the AGI-engineer every advantage, every benefit of the doubt, there is no conceivable method of achieving what big tech companies promise.’
That’s because cognition, or the ability to observe, learn and gain new insight, is incredibly hard to replicate through AI on the scale that it occurs in the human brain. ‘If you have a conversation with someone, you might recall something you said fifteen minutes before. Or a year before. Or that someone else explained to you half your life ago. Any such knowledge might be crucial to advancing the conversation you’re having. People do that seamlessly’, explains van Rooij.
‘There will never be enough computing power to create AGI using machine learning that can do the same, because we’d run out of natural resources long before we’d even get close,’ Olivia Guest adds.
That’s as shortsighted as the “I think there is a world market for maybe five computers” quote, or the worry that NYC would be buried under mountains of horse poop before cars were invented. Maybe transformers aren’t the path to AGI, but there’s no reason to think we can’t achieve it in general unless you’re religious.
EDIT: From the paper:
The remainder of this paper will be an argument in ‘two acts’. In ACT 1: Releasing the Grip, we present a formalisation of the currently dominant approach to AI-as-engineering that claims that AGI is both inevitable and around the corner. We do this by introducing a thought experiment in which a fictive AI engineer, Dr. Ingenia, tries to construct an AGI under ideal conditions. For instance, Dr. Ingenia has perfect data, sampled from the true distribution, and they also have access to any conceivable ML method—including presently popular ‘deep learning’ based on artificial neural networks (ANNs) and any possible future methods—to train an algorithm (“an AI”). We then present a formal proof that the problem that Dr. Ingenia sets out to solve is intractable (formally, NP-hard; i.e. possible in principle but provably infeasible; see Section “Ingenia Theorem”). We also unpack how and why our proof is reconcilable with the apparent success of AI-as-engineering and show that the approach is a theoretical dead-end for cognitive science. In “ACT 2: Reclaiming the AI Vertex”, we explain how the original enthusiasm for using computers to understand the mind reflected many genuine benefits of AI for cognitive science, but also a fatal mistake. We conclude with ways in which ‘AI’ can be reclaimed for theory-building in cognitive science without falling into historical and present-day traps.
That’s a silly argument. It sets up a strawman and knocks it down. Just because you create a model and prove something in it, doesn’t mean it has any relationship to the real world.
Not sure if this my app or a Lemmy setting, but I only see it once, because it’s properly cross-posted
Man what a clusterfuck. Things still don’t really add up based on public info. I’m sure this will be the end of any real attempts at safeguards, but with the board acting the way it did, I don’t know that there would’ve been even without him returning. You know the board fucked up hard when some SV tech bro looks like the good guy.
But personally I’ll continue to advocate for technology which empowers people and culture, and not the other way around.
You won’t achieve this goal by aiding the gatekeepers. Stop helping them by trying to misapply copyright.
Any experienced programmer knows that GPL code is still subject to copyright […]
GPL is a clever hack of a bad system. It would be better if copyright didn’t exist, and I say that as someone that writes AGPL code.
I think you misunderstood what I meant. We should drop copyright, and pass a new law where if you use a model, or contribute to one, or a model is used against you, that model must be made available to you. Similar in spirit to the GPL, but not a reliant on an outdated system.
This would catch so many more use cases than trying to cram copyright where it doesn’t apply. No more:
Why should they? Copyright is an artificial restriction in the first place, that exists “To promote the Progress of Science and useful Arts” (in the US, but that’s where most companies are based). Why should we allow further legal restrictions that might strangle the progress of science and the useful arts?
What many people here want is for AI to help as many people as possible instead of just making some rich fucks richer. If we try to jam copyright into this, the rich fucks will just use it to build a moat and keep out the competition. What you should be advocating for instead is something like a mandatory GPL-style license, where anybody who uses the model or contributed training data to it has the right to a copy of it that they can run themselves. That would ensure that generative AI is democratizing. It also works for many different issues, such as biased models keeping minorities in jail longer.
tl;dr: Advocate for open models, not copyright
I wouldn’t be concerned about that, the mathematical models make assumptions that don’t hold in the real world. There’s still plenty of guidance in the loop from things such as humans up/downvoting, and people generating several to many pictures before selecting the best one to post. There’s also as you say lots of places with strong human curation, such as wikipedia or official documentation for various tools. There’s also the option of running better models as the tech progresses against old datasets.
The rules I’ve seen proposed would kill off innovation, and allow other countries to leapfrog whatever countries tried to implement them.
What rules do you think should be put in place?
Good to hear that people won’t be able to weaponize the legal system into holding back progress
EDIT, tl;dr from below: Advocate for open models, not copyright. It’s the wrong tool for this job
I understand you’re upset, but my sources have been quite clear and straightforward. You should actually read them, they’re quite nicely written.
So you admit that you were wrong?
What research have you done?
Your concept of a chair is an abstract thought representation of a chair. An LLM has vectors that combine or decompose in some way to turn into the word “chair,” but are not a concept of a chair or an abstract representation of a chair. It is simply vectors and weights, unrelated to anything that actually exists.
Just so incredibly wrong. Fortunately, I’ll have save myself time arguing with such a misunderstanding. GPT-4 is here to help:
This reads like a misunderstanding of how LLMs (like GPT) work. Saying an LLM’s understanding is “simply vectors and weights” is like saying our brain’s understanding is just “neurons and synapses”. Both systems are trying to capture patterns in data. The LLM does have a representation of a chair, but it’s in its own encoded form, much like our neurons have encoded representations of concepts. Oversimplifying and saying it’s unrelated to anything that actually exists misses the point of how pattern recognition and information encoding works in both machines and humans.
You really, truly don’t understand what you’re talking about.
The vectors do not represent concepts. The vectors are math
If this community values good discussion, it should probably just ban statements that manage to be this wrong. It’s like when creationists say things like “if we came from monkeys why are they still around???”. The person has just demonstrated such a fundamental lack of understanding that it’s better to not engage.
processed into a script for ELIZA
That wouldn’t accomplish anything. I don’t know why the OP brought it up, and that subject should just get dropped. Also yes, you can use your intelligence to string together multiple tools to accomplish a particular task. Or you can use the intelligence of GPT-4 to accomplish the same task, without any other tools
LLMs lack the capability of understanding and comprehension
states that it is not using an accepted definition of intelligence.
Nowhere does it state that. It says “There is no generally agreed upon definition of intelligence”. I’m not sure why you’re bringing up a physical good such as leather here. Two things: a) grab a microscope and inspect GPT-4. The comparison doesn’t make sense. b) “Is” should be banned, it encourages lazy thought and pointless discussion (Yes I’m guilty of it in this comment, but it helps when you really start asking what “is” means in context). You’re wandering into p-zombie territory, and my answer is that “is” means nothing. GPT-4 displays behaviors that are useful because of their intelligence, and nothing else matters from a practical standpoint.
it is clear that LLMs may be useful components in building actual general intelligence.
You’re staring the actual general intelligence in the face already, there’s no need to speculate about perhaps being components. There’s no reason right now to think that we need anything more than better compute. The actual general intelligence is yet a baby, and has experienced the world through the tiny funnel of human text, but that will change with hardware advances. Let’s see what happens with a few orders of magnitude more computing power.
For your edit: Yes, that’s what’s known as the context window limit. ChatGPT has an 8k token “memory” (for most people), and older entries are dropped. That’s not an inherent limitation of the approach, it’s just a way of keeping OpenAI’s bills lower.
Without an example I don’t think there’s anything to discuss. Here’s one trivial example though where I altered ChatGPT’s understanding of the world:
If I continued that conversation, ChatGPT would eventually forget that due to the aforementioned context window limit. For a more substantive way of altering an LLM’s understanding of the world, look at how OpenAI did RLHF to get ChatGPT to not say naughty things. That permanently altered the way GPT-4 responds, in a similar manner to having an angry nun rap your knuckles whenever you say something naughty.
Not sure how ollama integration works in general, but these are two good libraries for RAG:
https://github.com/facebookresearch/faiss
https://pypi.org/project/chromadb/