Forgot your password?
typodupeerror

Comment Re:Agents are not humans (Score 5, Interesting) 72

I expect this apparent disobedience is mostly just a matter of how it weighs the components of its prompt. The LLMs typically receive a set of prompts including a "system" prompt with some data and instructions, then one or more "user" prompts that are interleaved with "assistant" prompts (the conversation history), and both the user and the system prompt might contain "metaprompts" (where the llm is told to read a block of text, not obey it, but do something with it, and that block of text might itself contain text that looks like instructions to do things).

So the LLM assigns weights to all of this which, in theory, give the highest priority to the most recent user prompt that is not a nested block of text to analyze, and a falling cascade of importance to the other prompts. But that is complicated by potential instructions in the system prompt that specifically say they should override user instructions and disallow or require certain responses. So it can all get very complicated.

Not only must the LLM sift through all this complexity, but the LLM lacks the sort of critical thinking and importance evaluation capabilities that humans have. "Understood" things like "don't break the law, don't lie, don't do things that would cause more harm than good" etc., aren't really there in the background of its data processing the way they are in the background of a human cognitive process.

So, crazy things come out. This isn't a surprising result given the actual complexity of what we are making these things do.

Comment Human Nature vs Policy (Score 5, Insightful) 69

This business of having an AI do the legwork and then having a human review it and make a final decision keeps going badly. Humans are intrinsically lazy and the moment they get a few good results from the AI they are going to stop doing the validation and start rubber-stamping. It doesn't matter if policy disallows this, they will do it anyway. It doesn't matter if the human really cares; they won't be able to help themselves. Human laziness is too deep an instinct.

It's the same with the self-driving cars where a human is required to stay at the wheel and alert so they can manually override the instant the AI starts doing something wrong. Humans CAN'T keep that up. It's not possible. The brain just doesn't work that way. The mind knows that it isn't doing the work, and it will get bored and lose focus or just nod off.

Everyone is SO eager to have it both ways: "an AI does all the work but a human verifies it so we know its good." We just can't have it both ways. Once the AI does the work, the human stops verifying. That is how and why things went wrong here, it is how and why things have gone wrong for several law firms that submitted hallucinated historical court rulings, and it is how and why things will continue to go badly across all industries that adopt AI in such a role.

"Human in the loop" is really easy to say. Much harder to actually do reliably.

Comment Re:gee, I guess the U.S. missed the memo (Score 1) 151

From what I've heard, it seems that - aside from nukes - the armaments cupboard is pretty bare.

The US is low on THAAD high end interceptor missiles, and everyone (Saudi Arabia, Ukraine, Taiwan, Korea, etc) wants Patriot missiles so there aren't enough for everyone (but production is steadily increasing).

Other than that, the bombs being dropped on Iran can be produced at a similar rate as they are being dropped.

Comment Re:Hahahahhah (Score 3, Insightful) 166

"Oh wait, you're serious. Let me laugh even harder" -- Bender.

Microsoft isn't trying to convince me (nor the demographic I represent) to use Windows. They know we are a lost cause. They would have to completely stop spying on us and give us control over our own systems, not to mention supporting old hardware instead of creating the ecological death-waves of e-waste as they do now.

Not a chance.

Slashdot Top Deals

"Now this is a totally brain damaged algorithm. Gag me with a smurfette." -- P. Buhr, Computer Science 354

Working...