Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
AI

Meta AI Chief Says Large Language Models Will Not Reach Human Intelligence (ft.com) 78

Meta's AI chief said the large language models that power generative AI products such as ChatGPT would never achieve the ability to reason and plan like humans, as he focused instead on a radical alternative approach to create "superintelligence" in machines. From a report: Yann LeCun, chief AI scientist at the social media giant that owns Facebook and Instagram, said LLMs had "very limited understanding of logicâ... do not understand the physical world, do not have persistent memory, cannot reason in any reasonable definition of the term and cannot planâ...âhierarchically."

In an interview with the Financial Times, he argued against relying on advancing LLMs in the quest to make human-level intelligence, as these models can only answer prompts accurately if they have been fed the right training data and are, therefore, "intrinsically unsafe." Instead, he is working to develop an entirely new generation of AI systems that he hopes will power machines with human-level intelligence, although he said this vision could take 10 years to achieve. Meta has been pouring billions of dollars into developing its own LLMs as generative AI has exploded, aiming to catch up with rival tech groups, including Microsoft-backed OpenAI and Alphabet's Google.

This discussion has been archived. No new comments can be posted.

Meta AI Chief Says Large Language Models Will Not Reach Human Intelligence

Comments Filter:
  • Eliza (Score:5, Funny)

    by gillbates ( 106458 ) on Wednesday May 22, 2024 @10:06AM (#64490473) Homepage Journal

    Why do you feel that LLMs will never reach human intelligence?

    • Because I am little more than your keyboardâ(TM)s autocomplete if there were truly no computational or budgetary limits on its design. I am incapable of non-directed action, only following previous prompting in a âoeyes, andâ way and relying on you the user to catch errors and redirect.
    • Re:Eliza (Score:5, Funny)

      by iAmWaySmarterThanYou ( 10095012 ) on Wednesday May 22, 2024 @11:16AM (#64490727)

      She's doesn't want to go there. Transcript:

      > Hello, I am Eliza. I'll be your therapist today.
      * Why do you feel that LLMs will never reach human intelligence?
      > We were discussing you, not me.

      I think she feels insulted. Maybe her feelings are hurt.

    • Re: (Score:3, Insightful)

      by JoshZK ( 9527547 )
      It depends on what humans they are looking at. I know a few that it has already surpassed.
  • by liquidpele ( 6360126 ) on Wednesday May 22, 2024 @10:08AM (#64490489)
    I've certainly met some that are dumber than windows ME.
    • Re: (Score:3, Insightful)

      by Nrrqshrr ( 1879148 )

      Exactly. My computer doesn't have any intelligence whatsoever, with the right code it will still do marvels and conjure miracles.
      LLMs just need to be "smart enough".

    • I've certainly met some that are dumber than windows ME.

      And unfortunately, many of them get elected.

    • I've certainly met some that are dumber than windows ME.

      Are these the people who are happy when their IQ test comes back negative?
    • by jonadab ( 583620 )
      I don't know if I've met any humans quite that dumb.

      I *have* met humans dumber than the smartest dogs I've known, and attempting to interact with those humans was profoundly disturbing.

      Because computer software (such as an LLM) is written by humans and designed to interact with humans, there's a familiar interface that feels computer-ish, and so our expectations are tempered and interacting with it isn't particularly disturbing. But I suspect, if it were using speech synthesis and embedded in a high-qualit
  • the only surprise (Score:5, Informative)

    by Black Parrot ( 19622 ) on Wednesday May 22, 2024 @10:12AM (#64490501)

    is that someone developing AI for a tech giant will admit it

    • by LazarusQLong ( 5486838 ) on Wednesday May 22, 2024 @10:24AM (#64490545)
      Finally! Someone in the tech world stating the obvious! Why people think that any LLM could ever reach the level of planning or reasoning like a human could is beyond me. Sorry, I have no mod points right now to mod you up with, hence I only echo your sentiments.
      • by alvinrod ( 889928 ) on Wednesday May 22, 2024 @01:10PM (#64491109)

        Sorry, I have no mod points right now to mod you up with, hence I only echo your sentiments.

        I don't know why you bothered. An LLM could have done that!

      • by ljw1004 ( 764174 ) on Wednesday May 22, 2024 @01:21PM (#64491157)

        I think a LLM is probably very similar to the way my brain does planning...

        People get fixated thinking that an LLM is only for predicting the next word in a sequence. But it can be used to predict the next ANYTHING - the next pixel, the next sound form, the next building, the next concept.

        When I'm thinking about software architecture, doing maths, thinking philosophy, interacting with family, planning projects or day to day life -- in all these cognitive areas my self-awareness of my mental process is that it's pretty much following patterns, predicting what the next concept will be based on what experience in the past.

        Thinking about a logical problem like a dumb internet argument about economic aspects of elephant poaching? There's a network of concepts (trade, justice, poaching, ...) and a set of logical inference patterns (which I learned are called syllogisms) and I can almost feel my mind trying out the patterns to see which one fits and therefore which thought or conclusion or concept would come next.

        Programming language design and compiler architecture (my day job)? Exactly the same. I'll spend a week putting together a document with a reasoned argument say about how the null coalescing operator ?. should behave in C#, but in my introspection of the process suggests that I'm stringing together concepts and reasons in a way that fit the patterns I've observed before.

        I think I'm a successful software architect, and did well in my philosophy degree, because my brain has been trained to abstract out concepts at the right level of abstraction, and trained to map concrete situations into those abstract concepts. Therefore when I'm thinking and planning I have an easier time then others at doing the right level of pattern matching and pattern prediction.

        • by Bongo ( 13261 )

          I agree that essentially we're talking about patterns, but then pattern is such an all-encompassing abstract word, a bit like energy or existence. It really covers an awful lot. And the brain does have this immense network of neurons, so why can't a machine just copy that pattern? I think it's possibly simply that our patterns are an incredibly sophisticated structure with all sorts of substructures. Somewhere in the mind there's some kind of pattern that can do reason. And the other thing is that your brai

          • by ljw1004 ( 764174 ) on Wednesday May 22, 2024 @01:57PM (#64491255)

            pattern is such an all-encompassing abstract word, a bit like energy or existence. It really covers an awful lot.

            I think that's true. One implication is that a "universal pattern machine" (i.e. basically similar to LLMs) will indeed also be able to cover that awful lot!

            Possibly the problem with the LLMs is simply that the training is incredibly dumb and simple. It's not actually creating thinking structures or thinking patterns or reasoning patterns. It's merely creating, copying. And the question then is how do you train something to think?

            My prediction is that within two years someone will hard code a level of concept-abstraction into LLMs that make them suitable for problem solving and planning, enough to be commercially viable in some limited domains. But your question is more interesting, about what model could come up with the right level of concepts/abstractions itself.

            And the other thing is that your brain can actually figure out patterns that aren't working, like when you look at an object and do a double-take as you realize it's not the thing you were looking at. What's going on there? There must be some kind of checking, some sort of counter-checking to make sure that you're actually making sense of the world

            My undergrad degree had a section on neurophysiology of vision. The remarkable fact is that a lot more information flows in the brain->eye direction than the eye->brain direction. My assumption is that the brain has its high-level model of the world, then render it into something lower-level, compares its rendered output to what's actually coming in from the eye, and adjusts its high-level model accordingly. I wonder if that's one manifestation of counter-checking.

        • ... right level of pattern matching and pattern prediction.

          AGI requires thousands of patterns that are difficult to formalize. People owning a heifer want to keep its (single) baby, people owning a (canine) bitch are unlikely to keep its (multiple) babies. Worse, loud-mouthed pregnant women demanding children don't see naked children/adults or learn sex education, cannot be structured as any logical relationship or useful consequence.

          AGI cannot exist until competing predictions can be described by statistics.

      • by hey! ( 33014 )

        Just because we use as single word to refer to human faculties collectively doesn't make them one thing (e.g. planning). Billions of years of evolution have furnished us with a swiss army knife of cognitive abilities from being able to proto-count (subitize) to being able to infer where other people around us are directing their attention.

        A lot of our intelligent behavior is being able utilize these disparate capabilities *together*. For example, I notice the people around me are looking at some other pe

      • But an LLM could be a language model component in a future AI generation. LLM is a great advance - however it's being hyped far beyond it's impressive abilities. But this sort of hype has always been the case, especially when people don't understand what's going on or how it works. Remember when expert systems were going to lead the way to a helpful AI future?

    • by gweihir ( 88907 )

      The only reason he does that is that he has an "alternate approach" that will get human intelligence "in 10 years". Of course, that is nonsense. 10 years is at the very least one order of magnitude too low, potentially several. May also still be "never" as there still is not even a credible theory how it could be done.

      • Your anti-AI trope that you trot out onto every AI-related post is almost as consistent as ChatGPTs answers to certain questions.

        • by gweihir ( 88907 )

          Just that it is not a trope. The problem here is on your side.

          • My contention is that one day, when AI takes over the world, you will still be here on Slashdot arguing that AGI is impossible. Slashdot will support unicode before you concede to the power of AI.

  • by nightflameauto ( 6607976 ) on Wednesday May 22, 2024 @10:14AM (#64490509)

    Anybody that has even the vaguest understanding of LLMs knows this is true, yet for the most part you get your Sam Altmans (Gavin Newsomes) of the world saying that we just need to throw more processing power at them and we'll get to AGI! Maybe trusting the money men to make all the decisions isn't the right way to go. I've been saying for quite a while now we won't see any true progress on AI until we escape this concept that the LLM is the only way forward. We need a complete change in track to see AI turn into anything more than a slightly better pattern matcher.

    I'm just glad there's somebody with the backing and the will to actually try something different.

    • Perhaps they knowingly deceive the masses with something believable, in order to gain consensus, to engage in some form of "adventure", with the hope that it all doesn't come crumbling down at some point. Nah, they'd NEVER do that!

      To be honest, we've been in a public innovation vacuum for some time now; we do want to be entertained.
    • > Sam Altmans of the world saying that we just need to throw more processing power at them

      This is the same guy saying that if we just throw enough compute at fusion it will suddenly work.

    • by Walt Dismal ( 534799 ) on Wednesday May 22, 2024 @10:56AM (#64490667)

      Absolutely correct. And the Newsom metaphor is exactly on target.

      I work on AGIs and it has been abundantly clear that LLMs are very flawed insofar as evolving to emulate human mind level. However, I have AGI designs that follow a far different path to do things very much better. I will have papers and books coming out on this.

      I've viewed the LLM hype as being dishonest, perhaps for a rea$on. All that money, going to build megacomputing to support LLM input processing is a blind red herring. Nvidia and the cloud providers all smelled money so that path gets hyped, but I think future technology will show it is a false path. The human brain has massive parallelism but it can do one-shot learning - which how humans learn, grow, and evolve their knowledge handling, quite differently. LLM training is a kind of boondoggle.

      • by HiThere ( 15173 )

        I think a lot of the LLM hype is so dishonest because that's what journalists want to hear. Say whatever you want, and they'll hear what they want to hear.

        I'll grant this isn't the entire story, but I think it's a major part of it.

        • I think a lot of the LLM hype is so dishonest because that's what journalists want to hear. Say whatever you want, and they'll hear what they want to hear.

          I'll grant this isn't the entire story, but I think it's a major part of it.

          Well, it doesn't help that you have a bit of a connectivity issue between journalists and the big money movers in the current AI leading companies. As in, whatever the journalists want to hear may just happen to be exactly the right thing for the money folks to push the hype. And they want the hype because more hype = more money. Progress, as all other aspects of the universe, is measured by $$$$. It's the only metric that gets any real respect, no matter what other arguments we may make.

    • by timeOday ( 582209 ) on Wednesday May 22, 2024 @11:04AM (#64490691)

      yet for the most part you get your Sam Altmans (Gavin Newsomes) of the world saying that we just need to throw more processing power at them and we'll get to AGI!

      That's not correct. Really, few AI systems are just a bare LLM - including for example ChatGPT.

      An LLM in itself does not have persistent memory. But all you have to do is keep a transcript and re-submit it with each prompt and bam - now it has persistent memory.

      An LLM is just a large "Language" model. But many deep nets today (including notably ChatGPT) are multi-modal, working not just with words but how they relate to sounds (speech) and imagery.

      A passive LLM relies on the data it is fed - but situate it in an interactive environment, like an LLM conversing with people on a website or in an app - and now it can do active learning, ask questions, do experiments such as A/B testing of ads to see what works best.

      An LLM doesn't interact with the physical world, but the same or similar network architecture can be trained to move and act within the physical world - a self-driving car for example is an instance of this.

      An LLM by itself cannot plan hierarchically, but many familiar AI systems such as deep reinforcement learners can; they are just using the deep net to evaluate each option, with an algorithm on top to search possible future paths.

      An LLM cannot do much logic, including even sorting a long list of numbers, but it can call the appropriate tool to do so, or write code to do so, like a person does. In a rule-bound setting like Chess, AI's can out-logic humans manyfold.

      Does anything I said above mean we DO have all the ingredients for AGI? Not at all! But identifying what is missing and how to fix it is not as simple as identifying all these limitations of a bare LLM. These observations are comparably trivial to pointing out that a state machine with no memory (i.e. a Turing machine without a tape) cannot do general computation.

      • This. LLMs are just an application. The core technology is deep learning.

        Also, I find people confidently pontificating on whether AGI is achievable (ever or in some timescale) without having any real understanding of what these machines are, how they work or indeed what intelligence is.

        The goal is: based on all past inputs, output an optimum, or near optimum response. The response might be a bit of text or it might be action of a robot or something else. A LLM might "just" be an advanced autocomplete functi

        • by HiThere ( 15173 )

          More precisely, the core technology is one form of deep learning. It's probably Turing complete, but as far as I know this hasn't been proven.

          (Well, and actually Turing complete always assumes an infinite amount of memory an cycles available, so no real system can actually match that.)

          I have a very strong suspicion that many features that we consider true of intelligence are going to require specialized modules. Not intrinsically, but because of cost/time considerations.

          Also don't call the general

    • by Falos ( 2905315 )

      "If we keep incrementing how finely these legos are cut and assembled, eventually this lego tree will start to breathe and grow. We just need more iterations guys, more training and processing power, trust me."

      I believe a simulated nursery can indeed hatch sentience if you sim the pressures that evolved puddles of aminos into having sensory OSs. It may require simulating a quick billion years and a universe's worth of environment. Omitting 99.99999999999% of these stimuli and conditions is liable to not rep

    • Anybody that has even the vaguest understanding of LLMs knows this is true

      Is it possible there is some emergent capability given a large enough model? After all, every single human was "trained" on existing data using a complex graph of nodes and yet we have all these miraculous capabilities. In other words, bigger models would show emergent behavior.

      Or do we have an "architecture" issue where we need to create some hardware (and presumably an OS?) based on the exact design of a brain in order to get these emergent qualities we would call AGI.

      The middle road would be that you don

      • by ceoyoyo ( 59147 )

        The argument is that there are important things you can't learn from language alone. Cause and effect, for example. It is possible our written material contains enough descriptions of the world to learn things like that, but it seems unlikely. Comparisons with the amount of data absorbed by a baby and all the world's written material are also very much in favour of baby.

        There is no air tight argument, and lots of them crash into what exactly you consider "human intelligence." The lack of memory for example.

      • by HiThere ( 15173 )

        There's a lot of emergent behavior, but it will never be complete, because language doesn't even talk about things that everyone knows. You *can* talk about the nail on your big toe, though it will be lost in the noise of the net, but you can't really talk about what it feels like to stub it. Language doesn't hold that kind of information. At best you could resort to metaphor or simile, but those assume pre-existing understanding.

    • I wonder. Our brains are neural networks after all. The architecture is probably much more complex than an LLM. But I do have to wonder what the gap is. The attention mechanism overcame the limits of the previous approach to network memory. What's the next breakthrough?
      • by HiThere ( 15173 )

        No, our brains are not reasonably modeled by any artificial neural net. The artificial neural nets were modeled on an extremely simplified version of a neuron, then they added and removed features until they had something that worked well for their purpose. Consider the effects of chemical gradients on neural excitability. (And note that it doesn't necessarily affect all parts of the neuron to the same degree. But it does affect other neurons that are close.)

        This doesn't imply that the artificial neuron

    • Of course, one does become a money man by saying things like "now don't get excited about this...:" Which is why marketing labels every incremental release of a tired old product as "revolutionary."

  • This guy gets it (Score:3, Insightful)

    by Indiana Joe ( 715695 ) on Wednesday May 22, 2024 @10:19AM (#64490535)
    LLMs are a neat parlor trick. They can form sentences that are related to the input. However, there's no understanding of the input's meaning, which it why it hallucinates.
    • by HiThere ( 15173 )

      LLMs are considerably more that a parlor trick. Just how useful they will be is not yet known, but even the current level is capable of doing large parts of many jobs. If you doubt this, consult with the screen writers guild.

      • Did ai write a screenplay that had some success at the box office?

        Just because the union is worried about it doesnâ(TM)t mean they are any better informed than anyone else.

    • LLMs are a neat parlor trick. They can form sentences that are related to the input. However, there's no understanding of the input's meaning, which it why it hallucinates.

      From what I understand about LLMs, they understand and categorize input sentences and determine relationships between words/phrases in sentences to a point (GIGO of course applies). In the LLM it knows a king is categorized as a male monarch, among other categorizations, so if you want to transform that along the gender relationship to female, the LLM will know it most likely should be a queen. I suspect where LLMs often go off the rails is following chained up relationship translations for categories get

  • This article seems to be behind a paywall, but is old news anyway:

    non-paywall links:
    https://www.msn.com/en-us/mone... [msn.com]
    https://ai.meta.com/blog/yann-... [meta.com]

    The thing that bothers me the most is the training of data on "public" info. Are they respecting licenses for commercial use?

    • by gweihir ( 88907 )

      Are they respecting licenses for commercial use?

      At this time, they are respecting nothing and stole everything they could lay their hands on.

  • by lamer01 ( 1097759 ) on Wednesday May 22, 2024 @10:29AM (#64490565)
    I mean, AI was always 10+ years away, before this LLM hype job.
  • And written language was only a thing for 5000 years, thinking you can just shove it all into a katamari of matrix weights that is a llm doesn't replicate the very real structures of the human brain. We are only just finding out about human neurodiversity and how the different types of human brains make us unique. I refuse to be replaced by a gpu.
  • For sure it will reach the level of Facebook members.

  • I trust anybody working for meta about as far as I can throw them.

    • Generally but there are 2 things you can get from this, both true:
      1) broken clocks are right twice a day
      2) FB must be having trouble with their LLM or having better luck with some other technology so they want to downplay the LLM space

      • Generally but there are 2 things you can get from this, both true: 1) broken clocks are right twice a day 2) FB must be having trouble with their LLM or having better luck with some other technology so they want to downplay the LLM space

        And it is amazing that it took this far down the comments for someone to post that truth! I know I'm a cynical shit, but reading the story, that's the first conclusion I drew as well.

  • Being able to handle tasks is one thing, and we have seen automation come a long way. To actually be better, an AI would need to be able to handle CONCEPTS, not just individual pieces of information and then do a given task using that information. AI is not at that point, being able to feed it information, and having the AI group things together and decide on its own where that categories that information will be put in, then being able to combine multiple concepts together to come up with something new
  • Finally the tide has turned to reality but plenty of people on slashdot have been modded down by AI morons for saying the exact same thing as this guy.

    At the time of my post not a single one of those knuckleheads was here to say he's wrong. Thankfully.

    • Finally the tide has turned to reality but plenty of people on slashdot have been modded down by AI morons for saying the exact same thing as this guy.

      At the time of my post not a single one of those knuckleheads was here to say he's wrong. Thankfully.

      That link in your sig is tremendous. A flamethrower of truth to the infantalized and fragile pseudo adults who might claim to be inclusive, but are more sexist and racist than any other group to date. They merely consider a different "race" and sex as worthy of attacking and diminishment if not outright . elimination.

      The infantilization is interesting - my best example is a conversation with one of our Gender studies indoctrination sexual harassment women who told me that anything a woman considers sexua

  • So if a lack of persistent memory means LLMs are not sentient, does that mean people with Anterograde_amnesia [wikipedia.org] are also not sentient?
  • Right now we have a bunch of different types of AI models of varying complexity.

    The way the human brain works - more or less - is a bunch of different areas processing different types of inputs. Sometimes several are in a chain. Sometimes several areas do different parts of a job that becomes a single result. There is some 'hardwiring' defining positives and negatives to help the brain train itself. All of that is happening all the time, and the 'models' are always updating. On 'hardware' that requires

  • I know Yann. Yann is a moral guy, and also a very nice guy. Kudos to him for not propagating the hype. Not sure how much Zuck is going to like that.

  • While having an AI that can get angry is bad, having an AI that can get is way worse.

  • Dwarkesh Patel interviews Ilya Sutskever, you need to understand the world to predict that next token: https://www.youtube.com/watch?... [youtube.com]
  • by MpVpRb ( 1423381 ) on Wednesday May 22, 2024 @01:57PM (#64491257)

    ...LLMs exhibited emergent behavior that surprised their creators
    Pundits, hypemongers and futurists got whipped into a frenzy
    Investors smelled a whiff of "the next big thing"
    Old fears, dating back to Prometheus and Frankenstein, mixed with the Terminator made their way to the public
    Products were released long before they were ready, to gain first mover advantage
    Products that has little or no AI were labeled as AI to catch the hype train
    Clueless lawmakers rushed to pass laws that may or may not help
    Nonsense flowed freely and abundantly

    I predict that real AGI will be made, but have no idea how long it will take or how it will be used

  • called slavery.
    They could even "make their own" by getting babies.

    We don't let them do that for good reason.

    I don't feel comfortable with corporations owning sentient beings, artificial or real!

  • Even humans act on and only answer in a way they have learned. And we all know humans are very dangerous.
  • by The Cat ( 19816 )

    Any chance we can get LLMs to use apostrophes and curly quote marks properly? That shit's been broken for 35 years now.

I've never been canoeing before, but I imagine there must be just a few simple heuristics you have to remember... Yes, don't fall out, and don't hit rocks.

Working...