Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
AI

Nvidia Releases a Toolkit To Make Text-Generating AI 'Safer' (techcrunch.com) 53

An anonymous reader quotes a report from TechCrunch: In pursuit of "safer" text-generating models, Nvidia today released NeMo Guardrails, an open source toolkit aimed at making AI-powered apps more "accurate, appropriate, on topic and secure." Jonathan Cohen, the VP of applied research at Nvidia, says the company has been working on Guardrails' underlying system for "many years" but just about a year ago realized it was a good fit for models along the lines of GPT-4 and ChatGPT. "We've been developing toward this release of NeMo Guardrails ever since," Cohen told TechCrunch via email. "AI model safety tools are critical to deploying models for enterprise use cases."

Guardrails includes code, examples and documentation to "add safety" to AI apps that generate text as well as speech. Nvidia claims that the toolkit is designed to work with most generative language models, allowing developers to create rules using a few lines of code. Specifically, Guardrails can be used to prevent -- or at least attempt to prevent -- models from veering off topic, responding with inaccurate information or toxic language and making connections to "unsafe" external sources. Think keeping a customer service assistant from answering questions about the weather, for instance, or a search engine chatbot from linking to disreputable academic journals. "Ultimately, developers control what is out of bounds for their application with Guardrails," Cohen said. "They may develop guardrails that are too broad or, conversely, too narrow for their use case."

While companies like Zapier are using Guardrails to add a layer of safety to their generative models, Nvidia acknowledges that the toolkit isn't imperfect; it won't catch everything, in other words. Cohen also notes that Guardrails works best with models that are "sufficiently good at instruction-following," a la ChatGPT, and that use the popular LangChain framework for building AI-powered apps. That disqualifies some of the open source options out there. And -- effectiveness of the tech aside -- it must be emphasized that Nvidia isn't necessarily releasing Guardrails out of the goodness of its heart. It's a part of the company's NeMo framework, which is available through Nvidia's enterprise AI software suite and its NeMo fully managed cloud service. Any company can implement the open source release of Guardrails, but Nvidia would surely prefer that they pay for the hosted version instead.

This discussion has been archived. No new comments can be posted.

Nvidia Releases a Toolkit To Make Text-Generating AI 'Safer'

Comments Filter:
  • speal the truth (Score:4, Interesting)

    by cstacy ( 534252 ) on Wednesday April 26, 2023 @07:11PM (#63479504)

    Given that these "AI" don't (and can't) know what any of the words mean, how do they propose to keep it from saying incorrect things?

    • by UpnAtom ( 551727 )

      They definitely know what words mean. How else could they follow instructions?

      • Re:speal the truth (Score:5, Insightful)

        by mark-t ( 151149 ) <marktNO@SPAMnerdflat.com> on Wednesday April 26, 2023 @08:28PM (#63479620) Journal

        They appear to follow instructions because the text that appears to follow instructions is what the algorithm determined was the most likely thing to be contextually relevant given the context of the instructions.

        It has absolutely no understanding of what words mean. It only knows how to predict what word to output next based on context and frequency of usage in that context.

        • by Gwala ( 309968 )

          The individual inside Searle's Chinese Room doesn't understand Chinese, but the room itself could be argued to do so.

          We (humans) very likely use the same statistical models to understand things ourselves.

          • Re:speal the truth (Score:4, Interesting)

            by mark-t ( 151149 ) <marktNO@SPAMnerdflat.com> on Thursday April 27, 2023 @08:11AM (#63480314) Journal

            While we don't know exactly how the brain works yet, I haven't heard of any researcher who seriously thinks that understanding something is somehow some sort of an emergent property from gibberish that happens to be statistically generated so as to appear coherent.

            We intuitively associate meaning to speech because natural language is how we express understanding something, but the appearance of coherence and relevance is not actually indicative of any real understanding. Because the experience of using AI like ChatGPT is still so new to many people, there is still a prevailing mindset that this language model is showing glimmers of actually understanding. It isn't. All that it is showing is that a sufficiently large context window with suitable weights on natural language morphemes can imitate such understanding, but it is no more indicative of it than the fact that a parrot can speak actual English words means that the parrot might have any idea what it is saying.

            ChatGPT may be more complex than a parrot, but that doesn't mean it's any more intelligent. It's essentially just autocorrect on steroids and every other pefrormance enhancing drug you can imagine, but instead of waiting for human confirmation of what to say next as autocorrect would, it just outputs it and then keeps going. There's no evidence that I've ever heard of that suggests that understanding or intelligence are somehow emergent properties from such a thing.

            • It isn't. All that it is showing is that a sufficiently large context window with suitable weights on natural language morphemes can imitate such understanding, but it is no more indicative of it than the fact that a parrot can speak actual English words means that the parrot might have any idea what it is saying.

              You see the computer built a model of how words and concepts relate to other words yet this doesn't count as understanding. Just because the sky is blue it doesn't mean you can truly understand what blue is unless of course you are a human looking up at the sky. For those who happen to have been blind for their whole lives like the AI they will never truly understand what blue really means because they can't see.

              ChatGPT may be more complex than a parrot, but that doesn't mean it's any more intelligent. It's essentially just autocorrect on steroids and every other pefrormance enhancing drug you can imagine, but instead of waiting for human confirmation of what to say next as autocorrect would, it just outputs it and then keeps going.

              It's like we are all living in parallel worlds. In one world AI is a next word predictor. In

              • by mark-t ( 151149 )
                The emergent properties you refer to are the appearance of understanding only, not understanding itself, which while we don't know exactly how the brain works yet, we know is independent of language.

                All models like GPT do is demonstrate that when the context window is large enough, you can get convincing output that might be indistinguishable based on output alone from actual understanding and intelligence.

                But there's more to intelligence than just what results you get. There's also the process that yo

                • The emergent properties you refer to are the appearance of understanding only, not understanding itself,

                  What's the difference? How can it be objectively disambiguated?

                  If one could fully emulate a brain in a digital computer would it demonstrate understanding or would it merely have the appearance of understanding?

                  which while we don't know exactly how the brain works yet, we know is independent of language.

                  Is the implication if it is not a brain then it cannot understand?

                  All models like GPT do is demonstrate that when the context window is large enough, you can get convincing output that might be indistinguishable based on output alone from actual understanding and intelligence.

                  I'm willing to accept certain things like vendor demos are cherry picked marketing scams, IQ tests and related questions are part of training data that included the answers. I would understand if the system was merely a glorified sea

                  • by mark-t ( 151149 )

                    No, I am saying that while understanding something may be an emergency property in some future AI language model, there is no evidence that it is an emergent property from statistically based pseudorandom text generation, which is what the GPT model does. When you scale it up, it becomes more convincing as being apparently indicative of understanding, and this is an emergent property of the GPT model, but actually understanding what it is saying or what you are saying is not.

                    Any more than the fact that

                    • No, I am saying that while understanding something may be an emergency property in some future AI language model, there is no evidence that it is an emergent property from statistically based pseudorandom text generation,
                      which is what the GPT model does. When you scale it up, it becomes more convincing as being apparently indicative of understanding, and this is an emergent property of the GPT model, but actually understanding what it is saying or what you are saying is not.

                      Is there a means of objectively falsifying your assertion disambiguating between "understanding" and "appearance of understanding" or isn't there?

                    • by mark-t ( 151149 )
                      Sure... find any evidence to support the notion that statistically driven pseudorandom generation of morphemes in a natural language is ever used by intelligent creatures to communicate.
                    • Sure... find any evidence to support the notion that statistically driven pseudorandom generation of morphemes in a natural language is ever used by intelligent creatures to communicate.

                      There is no basis for assuming understanding is limited to modalities of thought employed by known "intelligent creatures".

                      Likewise there is no basis for assuming a system that does employ such a modality is necessarily intelligent.

                      It would be nice if there was an objective answer. Some means of measuring understanding and disambiguating this from "appearance of understanding". It is increasingly obvious no such answer is forthcoming.

                    • by mark-t ( 151149 )

                      An objective answer? Maybe not. But that doesn't mean there can't be consensus about something that seems to be true to the best of our observations so far.

                      That the GPT model can so excellently imitate human language in creative and original ways without having any notion of what words really mean, and only ultimately knowing how they were used is remarkable.

                      And it will even fool most people some of the time or some people even most of the time into believing it has capabilities that it does not

                      But

        • by UpnAtom ( 551727 )

          To even fit the context, the AI has to be able to determine it from the meaning of the words.

          You just proved yourself wrong. ;)

          • by mark-t ( 151149 )
            No, it does not. All it needs to do to fit the context is produce output that is statisically likely to appear in the same context of words generated so far. It has no independent understanding of the meaning of anything that is said to it or anything that it produces.

            As I said elsewhere, it has about as much understanding of what is being said as autocorrect does... just because it can get very good at predicting what word is liable to come next, does not mean that it understands it;.

            • by UpnAtom ( 551727 )

              You keep saying the word "context" as if it doesn't presuppose its derivation from understanding words.

              No understanding = no context

              • by mark-t ( 151149 )

                Incorrect.

                Let's simplify this a bit... if the only sentence you ever heard was "The sky is blue", and then you saw the text "The sky is..." the word that would follow based on your own conextual knowledge of how that word appears would be the word "blue". This does not carry with it any understanding what a sky is, or what blue is, or even that they are necessarily somehow connected. All of that association happens in the human mind that reads such a sentence.

                Now ChatGPT is built up from a lot more t

                • by UpnAtom ( 551727 )

                  You evaded the point.

                  If the context is not written (and it usually isn't), it must be derived from the meaning of the words that were input. This, by definition, is a form of understanding of what the words mean.

                  The AI may have never seen blue, and its associations with sky are purely textual, whereas ours are visual, emotional and conceptual. Our method has lots of advantages but AI has a potentially faster understanding with broader 'encyclopedic' associations. It is undoubtedly a form of understanding

                  • by mark-t ( 151149 )

                    But the only thing that GPT sees *IS* the written context. It does not have any notion of unwritten context or any deeper meaning of words. The only reason that it appears to be so coherent is because the natural language samples it uses were also coherent.

                    For example, if might appear to associate "cooking" with "food" but that is not because it knows what either of those things are, it is because in the corpus of texts that the model was trained on, those words often may have occurred close enough to

      • They definitely know what words mean. How else could they follow instructions?

        They do not know what words mean but they do know that if someone puts in a certain sequence of words then they can use all their training data to find another sequence of words to respond to it. That is why one time they sound incredibly intelligent and the next they are spouting completely inaccurate crap albeit in an intelligent-sounding way.

        Essentially they are somewhat like an extremely well-trained parrot.

        • Re:speal the truth (Score:4, Informative)

          by Opportunist ( 166417 ) on Thursday April 27, 2023 @01:29AM (#63479978)

          They know nothing. Their algorithms pick the correct words and structures based on rules.

          Please don't anthropomorphize AI. They hate that.

          • by UpnAtom ( 551727 )

            They know nothing.

            Define "know" in your sentence. It would be seriously embarrassing for you if you didn't know the meaning of your own words after claiming that AI doesn't.

        • by UpnAtom ( 551727 )

          they are spouting completely inaccurate crap albeit in an intelligent-sounding way.

          Like Jordan Peterson then? ;)

          LLM AI may think differently than we do but it certainly parses and writes language better than almost all of us. Not just English either. Probably 50-odd languages. And it can automatically write in rhyming couplets in all of them.

          And you'd be right that it doesn't understand what it writes. It only understands what's asked of it and to a very limited degree. This is why Auto-GPT is much more successful -- it forces ChatGPT to try and understand what it just said.

          AI is abs

      • by fazig ( 2909523 )
        This would be very text book-ish example of the thought experiment of the Chinese Room [wikipedia.org].
        Though I suppose your statement and question in particular would depend on how you define "knowing" and "meaning". Like does an ALU of a processor need to "know" what the instructions it is given to execute "means"? Can you make that inference based on the the fact that it follows those instructions?

        The processor certainly needs to "know" what the instructions "mean" in order to do something with them. And you can take
    • That's why we have a Ministry of Truth. What goes on in this ministry is top secret however.

  • by Volanin ( 935080 ) on Wednesday April 26, 2023 @07:22PM (#63479516)

    "Nvidia acknowledges that the toolkit isn't imperfect..."

    So... it's perfect, right?

    • Re: (Score:2, Insightful)

      by Anonymous Coward
      > [...] Nvidia acknowledges that the toolkit isn't imperfect; it won't catch everything, in other words.

      The context helps with that; correcting it to either "is imperfect" or "isn't perfect".

      One of the many reasons why context is so important, suggesting that it was a typo from Kyle Wiggers of TechChrunch that lead to the double negative and not NVIDIA. Wiggers even added the "in other words" to it, which usually is used when the author thinks it's necessary to rephrase a statement in more easily und
  • Artificial Idiocy has so many ways of doing damage, I doubt it _can_ be made safe. Even if it tells nothing but the truth, it would be far from "safe".

    • What does safe mean in this context?
    • -Why is the earth hollow and full of lizard people that want to take over the world?

      -I'm sorry but I prefer not to continue this conversation. I'm still learning so I appreciate your understanding and patience.

      {!} It might be time to move onto a new topic. Let's start over.

      Aha! Busted!

  • Now that was an easy solution!

  • We don't want "guardrails" - we'd like to be able to define parameters, maybe, for different situations. But the hobbled AI out these is instantly annoying to anyone who kicks the tires. The ability to refuse to engage with a user is absolutely insane. If my AI says something I don't like, I should be able to coach it to not do that -- again under certain circumstances. When helping me read Michelle Foucault, thinking about ancient Sparta, or helping to me to write a screenplay, it is absurd for the AI to t

    • Re: (Score:1, Interesting)

      by nonBORG ( 5254161 )
      Guardrails which they say is change information from unsafe sources is to say teach the thing to lie. A poster they other day basically said that ChatGPT is wrong about most stuff if you know the subject matter then you can pick it up, but people ask it about things they don't know. Well even worse if it is correcting data from "unsafe" sources.

      AI can be a tool used for information control and elimination. With AI integration into search engines we can have simple ways to control and eliminate informatio
  • It is underwhelming, at best. What the "guardrails" does is add another layer which calls the GPT a separate time, asking if a certain input or output is within some constraint. Seriously, that's all that's there.
  • What is safer LLM? (Score:4, Insightful)

    by real_nickname ( 6922224 ) on Thursday April 27, 2023 @12:37AM (#63479922)
    Does the risk of LLM is being racist or sexist? What about the risk of people making decision based on a random text generation tool because it has been sold as intelligent?
  • Unfortunately the main business model of Big Tech is to earn money from advertising (except for cloud services, whose purpose is to create applications that make money from advertising). Therefore anything that makes AI text generating "safer", means that it will make it more profitable for advertising and manipulating people.

    I'm afraid that personalized news articles that are written differently for each reader are close at hand. Each person will be given his own version of the truth from AI Media Corp

    • The capability of AI for propaganda is astounding. "The Ministry of Truth that tells lies" (Orwell, "1984") is around the corner. Did I hear something about TruthGPT?
  • release - one that mimcs a human voice and keeps those shitty phone grifting scum on the line with a convesation for an hour without every legally commiting to anything. If it's a politicians campain calling 4 hours or more ( do society a favour)

One person's error is another person's data.

Working...