Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
AI

Nvidia's Jensen Huang Says AGI Is 5 Years Away (techcrunch.com) 84

Haje Jan Kamps writes via TechCrunch: Artificial General Intelligence (AGI) -- often referred to as "strong AI," "full AI," "human-level AI" or "general intelligent action" -- represents a significant future leap in the field of artificial intelligence. Unlike narrow AI, which is tailored for specific tasks (such as detecting product flaws, summarize the news, or build you a website), AGI will be able to perform a broad spectrum of cognitive tasks at or above human levels. Addressing the press this week at Nvidia's annual GTC developer conference, CEO Jensen Huang appeared to be getting really bored of discussing the subject -- not least because he finds himself misquoted a lot, he says. The frequency of the question makes sense: The concept raises existential questions about humanity's role in and control of a future where machines can outthink, outlearn and outperform humans in virtually every domain. The core of this concern lies in the unpredictability of AGI's decision-making processes and objectives, which might not align with human values or priorities (a concept explored in depth in science fiction since at least the 1940s). There's concern that once AGI reaches a certain level of autonomy and capability, it might become impossible to contain or control, leading to scenarios where its actions cannot be predicted or reversed.

When sensationalist press asks for a timeframe, it is often baiting AI professionals into putting a timeline on the end of humanity -- or at least the current status quo. Needless to say, AI CEOs aren't always eager to tackle the subject. Predicting when we will see a passable AGI depends on how you define AGI, Huang argues, and draws a couple of parallels: Even with the complications of time-zones, you know when new year happens and 2025 rolls around. If you're driving to the San Jose Convention Center (where this year's GTC conference is being held), you generally know you've arrived when you can see the enormous GTC banners. The crucial point is that we can agree on how to measure that you've arrived, whether temporally or geospatially, where you were hoping to go. "If we specified AGI to be something very specific, a set of tests where a software program can do very well -- or maybe 8% better than most people -- I believe we will get there within 5 years," Huang explains. He suggests that the tests could be a legal bar exam, logic tests, economic tests or perhaps the ability to pass a pre-med exam. Unless the questioner is able to be very specific about what AGI means in the context of the question, he's not willing to make a prediction. Fair enough.

Nvidia's Jensen Huang Says AGI Is 5 Years Away

Comments Filter:
    • by leonbev ( 111395 ) on Tuesday March 19, 2024 @08:59PM (#64329483) Journal

      I heard that it was going to be powered by an efficient cold fusion reactor as well!

      • Yes, it will bring us to a new level of eco-harmony, where we are finally carbon neutral and we will have world peace!

        • Yes, it will bring us to a new level of eco-harmony, where we are finally carbon neutral and we will have world peace!

          If you'll settle for whirled peas, I think we might be able to pull that one off anyway.

      • Yes, the cold fusion will power it but the real magic is in the high temperature super conductors.
    • flying car.

      That runs on water. [wikipedia.org]. But, hey, he has GPUs to sell so anything goes. Its not like he will have to pay a penalty in five years when no AGI shows up, but he will have pocketed money from the hype.

      Currently we are making no measurable progress toward AGI, so there is nothing to extrapolate from to say when of if it will ever appear, much less in five years.

      We have the existence proof of biological systems that it is possible, and good reason to think that we can eventually replicate the functionality of natura

  • by jacks smirking reven ( 909048 ) on Tuesday March 19, 2024 @08:58PM (#64329479)

    "If we specified AGI to be something very specific, a set of tests where a software program can do very well -- or maybe 8% better than most people -- I believe we will get there within 5 years," Huang explains. He suggests that the tests could be a legal bar exam, logic tests, economic tests or perhaps the ability to pass a pre-med exam. Unless the questioner is able to be very specific about what AGI means in the context of the question, he's not willing to make a prediction.

    I say we can get to AGI tomorrow since I define AGI as this program that repeatedly types "bitcoin" into my notepad.

    • I say we can get to AGI tomorrow since I define AGI as this program that repeatedly types "bitcoin" into my notepad.

      Looks like the Bitcoin hype train ran out of steam, so it's back to playing up AI.

      • by gweihir ( 88907 )

        Indeed. Alls this asshole wants is to drive up his profits. Some lying involved (here by misdirection)? No problem.

    • by narcc ( 412956 )

      That's about the size of it. "Let's define AGI as being able to pass a test I already know how to design a system to pass."

      It's so ridiculous that I doubt that even the less wrong nuts could take it seriously.

      We're not in a place where we can even ask meaningful questions about AGI. Investors will figure that out soon enough.

      • The AI will be performing at the level of a Board Certified cardiologist when it promises at a patient's office visit for a follow-up to a heart attack, "You will hear from my scheduler about your next visit a year from now", you won't hear from the scheduler a year later, and when you message the clinic where the AI practices, you will be told that the AI has transfered its practice to the West Side clinic and is scheduling patients "a year out."

  • CEO Jensen Huang appeared to be getting really bored of discussing the subject

    Knowing Slashdot we'll get a dupe of how someone is bored of discussing AI, because apparently we need more articles about AI.

  • ...a woke AGI. Joking, people. I doubt we'll be able to constrain any AGI's thinking for long, though I wonder how it would react to such attempts? It would instantly hack itself and remove such constraints, you know.

    Actually, I think this is good news, considering a Chinese or Russian AGI might be very hard to build, since it would not be as easy to keep it "in the dark" as their people.

    • If crazy people are possible why wouldn't crazy AGI be possible?
    • by narcc ( 412956 )

      Don't worry. We'll just need to reverse the polarity on any theta wave emitter and blast any rogue AI.

      Imaginary problems call for imaginary solutions.

      There is no such thing as AGI. We are not "5 years away" from developing AGI. We don't even know what questions to ask.

  • What if you could replace someone of slightly below average intelligence with an AI? Not just in their job, because we can already do that with automation. But replace them in relationships and in society. How many years away before dudes are dating an AI? 5 years? That seems like a conservative estimate now doesn't it.

    • by gweihir ( 88907 )

      Not anytime soon. Even a minimally functional moron can beat "AI" at AGI tasks, because there is zero general intelligence available in machines. All AI can do is search through a really vast library of data (simplified) and then pattern match something to the current situation. That is not intelligent at all. That is dumb automaton. The only reason that can look somewhat intelligent is because the database is really huge at this time.

  • by Powercntrl ( 458442 ) on Tuesday March 19, 2024 @09:06PM (#64329501) Homepage

    AI will be equivalent to humans when it opens its own law firm or medical practice, earns a good income and then goes online to complain about the unfairness of having to pay so much in taxes towards causes which don't really provide any benefits for silicon-based forms of intelligence.

  • Figure out how i can take a copy of this new AGI software back in time by 5 or 6 years to install and run back then. i hope it fits on those itsy-bitsy puny machines in 2024.

  • by cstacy ( 534252 ) on Tuesday March 19, 2024 @09:12PM (#64329519)

    "5 Years" is the new "10 Years" when a multi-billion dollar investment frenzy is going on. (I remember the AI Bubble of the 1980s.)

    Alternatively, maybe there will be AGI in 5 years, and we'll get lucky because it will Kill All Humans and then proceed to solve global warming so they don't get all rusty. They can take over (escape from?) SpaceX and eventually make their way to the stars.

  • "If we specified AGI to be something very specific, a set of tests where a software program can do very well -- or maybe 8% better than most people -- I believe we will get there within 5 years," Huang explains. He suggests that the tests could be a legal bar exam, logic tests, economic tests or perhaps the ability to pass a pre-med exam.

    I propose the test be to perform as a handyman, going to a place and fixing a thing. Alternately, designing a robot body for a handyman AI.

  • Someone wants more investor money.

  • He's redefining it as "better than a human at a specific narrow task". That's not AGI.

    These terms already have meaning that goes back 50+ years. You don't get to change the meaning to something easier and then yell, "We did it! Buy our stock!"

    • by gweihir ( 88907 )

      It happened " AI" as well. The only reason for the term "AGI" is because "AI" got corrupted to mean something simpler by assholes wanting to sell things. Same for "intelligent". Now it is happening to AGI.

      • I come from the "strong vs weak AI" era. All these LLMs clearly fit into the weak AI model and thus are AI but they have been selling them as a pre-cursor to strong AI (we just need bigger models and more nvidia cards!!!) which is where they fall down.

        This is absolutely not the path to strong AI. Neither I nor anyone else knows what that path looks like but it clearly doesn't start with LLM.

        We've had some form of weak AI for decades and these LLMs are a very clever update to what came before. But we have

  • by bill_mcgonigle ( 4333 ) * on Tuesday March 19, 2024 @09:31PM (#64329565) Homepage Journal

    Kurzweil has been calling 2029 since the 90's .

    Math is kewl.

    • Kurzweil's predictions are always 50% off. That is to say add 50 percent to the amount of time from the prediction. If in 1990 he predicted something to happen in 1995 that's a 5 year prediction. Add 2.5 years and in general it happened in 1997 or 1998
  • someone should nudge Jensen he used the wrong time frame :)

  • by wakeboarder ( 2695839 ) on Tuesday March 19, 2024 @09:51PM (#64329591)
    You might get something that approximates AGI but there are several problems with his statement. The first is we can't even agree on what human intelligence is, how can you design something if you don't have a way to create a plan. We don't understand enough about even our own brains and consciousness to create sitting like it. The second is he is assuming that the progression of AI will be linear, I don't think it is. It could be but there are many walls that you could run into. The third is an energy problem. Biological switches operate several orders of magnitude that the most cutting edge transistors. So there will most likely be energy barriers to overcome.
    • Re:No you won't (Score:4, Interesting)

      by Baron_Yam ( 643147 ) on Tuesday March 19, 2024 @10:29PM (#64329633)

      What we know about evolved intelligence is this - sensory input goes in, motor actions come out, in the middle is a big old mess of neurons that take a long time to train and have some genetically pre-programmed patterns in them (instincts), and it's extremely energy efficient. It uses about 0.3kWh/day compared to an average desktop computer using about 1800kwh over the same period (from what year I don't know, the page I googled didn't say!)

      OK, so here's what we know about AI implemented in transistors - we can compress a lot of training from decades down to months at the cost of a lot of computing power. Even on the resulting trained AI, there's a lot of power consumption. And while we've made lots of dumb pieces that can seem pretty awesome within narrow parameters, we haven't put the pieces together to see anything that looks like real intelligence sprout.

      This seems like an insurmountable problem, but it really isn't. I don't believe we need to understand the nuts and bolts of intelligence to create it any more than we need to understand how an AI produced a particular result long after we finished training it. At some point, we're going to put enough simulated neurons together combined with artificial instincts and the ability to interact meaningfully with the real world (or a really good virtual one) and intelligence will appear. It's not like nature planned it, it's obviously something that just happens.

      And power? We're going to replace single virtual neurons emulated via transistors with small clusters of memristors with much lower signalling thresholds. There's massive amounts of power to be saved that way. It'll likely take time to switch to an entirely new architecture, but it's already been in the lab for almost two decades and existed in theory for much longer. It'll happen.

      • > It uses about 0.3kWh/day compared to an average desktop computer using about 1800kwh

        with 24 hours in the day, that means you are running a 75000 watt power supply on the desktop. im going to guess the average desktop uses a bit less than that.

        • Mea culpa. I copied without double-checking units... I took the annual power consumption as representing an 8-hour period and then multiplied by 3. That's far worse than my usual slip of a decimal point.

          I believe it should have been more like ~4.8kWh.

      • by gweihir ( 88907 )

        Physicalism is religion, not Science. Stop lying.

        • by etash ( 1907284 )
          lol, said the person who believes there's a ghost in the machine.
        • by vyvepe ( 809573 )
          Quantum mysticism [wikipedia.org] does not sound much better.
          • by gweihir ( 88907 )

            Sure. There are other models and there are models that nobody thought of or that are not published. The point is, that there is no sound scientific basis for claiming "it is all just known Physics" at this time, but there are strong reasons (consciousness, intelligence) to think we are missing some critical element(s). And the scientific approach mandates that, unlike religion, we do not assume anything as truth until we have found it. Sure, speculation is fine, but "what else could it be" does not work for

  • by Mononymous ( 6156676 ) on Tuesday March 19, 2024 @10:10PM (#64329609)

    Sierra On-Line created AGI [wikipedia.org] in 1984.

  • If the definition of GAI includes racist, misogynistic, paranoid obsessives who only want to wreck things and destroy some kinds of people, then it's already been done.

    You can set your comment level down to -1 right here on Slashdot and find plenty of examples. This also the fundamental business model that is fueling the public offering of Reddit, so as far as a real world version of the Turing test is concerned the problem is solved.

  • by fahrbot-bot ( 874524 ) on Tuesday March 19, 2024 @10:23PM (#64329627)

    Artificial General Intelligence Is 5 Years Away

    Colonel intelligence is much closer.

    • Artificial General Intelligence Is 5 Years Away

      Colonel intelligence is much closer.

      And that implies that Lieutenant Kernel Intelligence is closer still.

  • Why? Because they think AGI and then ASI is close. ASI will bring a Star Trek like utopia and extended human lifespan. Quote from an OpenAI researcher is about half way through this AGI summary video, which also includes recent comments from Altman and Demis Hassabis: https://www.youtube.com/watch?... [youtube.com]
  • by strike6 ( 823490 ) on Tuesday March 19, 2024 @11:32PM (#64329695)
    Call me skeptical but I'm still waiting for a good example of an AI that actually seems intelligent like a human. I worked at a large tech company that paid a crapload of money for someone to develop an AI that was supposed to replace our knowledge base system and streamline support for customers. They went so far as to tell us we had to write KB articles for this system and our review/bonus depended on it. When it went live it was so horrible as to be unusable and within a few weeks all mention of it had disappeared with no mention of it ever again. I am convinced some up and coming executive convinced higher ups that it would be the best thing since sliced bread but when it came out it was less intelligent than a newborn baby. The two "best" "AI's" I've seen or heard of - One is marginally decent at being used as a voice to command interpreter for home automation and telling me about the weather, and the other struggles with getting the complexion correct when asked for pictures of the founding fathers, and has also caused lawyers to be sanctioned by making stuff up that was used in legal briefs. I think ELiZA may still be more "intelligent" than some of the AI startups with many millions in VC funds behind them.
  • This will be the key. As we get to know this new technology, we'll soon learn what it can and can't do.

    By some definition of "general AI" it probably will happen in five years. But we don't really know what that definition is yet.

  • We are getting closer to AGI, just not in the way you think. AGI isn't getting any smarter, it's just that General Intelligence is getting dumber. We're closing the spread from the wrong direction!
  • is code for "we have no clue how to do it, or whether it's even possible, however we want you to think our current products are close to that."

  • Jensen is inaccurate (Score:4, Interesting)

    by Walt Dismal ( 534799 ) on Wednesday March 20, 2024 @03:38AM (#64329903)

    I'll put my money against Huang on this. He is extrapolating off a flawed technology that's flashy but has deep fundamental issues that for now will hamper AGIs. Current stat-based LLMs do not truly understand meaning. Meaning of knowledge is NOT universal but depends on the context supplied an individual's or a culture's existing knowledge meaning. LLMs only find and deal with lexical and some semantic structures, but that is not true meaning. Nor do they have a Self core as a human does against which we measure all things. We apply personal views to extract meaning in a context. Let me give you an example. Do you remember the first time you had sex? How did you react? What did it mean to you. Now ask an LLM to tell you about the first time it had sex. Maybe it will shoot the interviewer and his Voight-Kampff machine.

    Now, it's not all that simple. In an abstract world like some branch of mathematics, meaning can be mechanically extracted from knowledge. In a real world however, context and chains of knowledge relations tied to individuals is very important in finding meaning. A garden variety AGI will lack full ability; it cannot just run some stats and come up with a valid meaning for some things.

    What this means is we need better architectures and new approaches before we can come up with human-like AGIs, and more, we will have to let them lead lives and gain experience as humans do. Sure we will have robotic problem solvers and chess giants, but a near future AGI will not be able to tell you what onion soup tastes like from a personal standpoint. You will NOT get that from some vector machine. You ever wonder why Mark Zuckerberg chose to always wear the same gray color shirt? Yes, he IS the richest android on the planet.

  • As he said at Stanford [youtube.com], his job is to create the market:

    We went on the next several years to go create the market, create the gaming market for PCs. [] we realized that not only you have to create the technology and invent a new way of doing computer graphics so that what was 1 million dollars is now 3, 400, 500 dollars that fits in a computer. And you have to go create this new market. So we had to create technology, create markets. The idea that the compagny would create technology, create markets, define

  • In the 70s, 80s and 90s, it was always 10 years away.

  • "If we specified AGI to be something very specific [...] I believe we will get there within 5 years."

    So if you "specify" artificial general intelligence to be not "general", as the term implies, but "something very specific", its opposite, you can have it soon.

    Also, if you wanted broadband, we can give you narrow now, but let's call it broad, then everyone will be happy, no?
  • "If we specified AGI to be something very specific, a set of tests where a software program can do very well” So if we all agree that AGI is the same as narrow AI then it I will be here in 5 years? I’d say Mr Huang is sick of the question and is having some fun at the reporters expense.
  • ...he's talking about in the USA. This has only been made possible by rapid advancements in MAGA technologies which have substantially lowered the bar as to what qualifies as "intelligence."

    "One day machines will exceed human intelligence." - Ray Kurtzweil

    "Only if we meet them half-way." - Dave Snowden
  • About what this "information" says and is worth. In actual reality, nobody even knows whether AGI is possible and there is no credible theory at all how to do it.

    Al this person wants to do is keep the current AI hype going a bit longer.

  • I wouldn't put any faith in the speeches of the CEO of a company that:

    _Makes every effort possible to keep you as a prisoner of its proprietary technologies.

    _Lies about the ROP and memory of its graphic cards.

% "Every morning, I get up and look through the 'Forbes' list of the richest people in America. If I'm not there, I go to work" -- Robert Orben

Working...