Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror

Comment Re:RAG (Score 1) 5

This sort of system is only useful because LLMs are limited. If they can be told to farm certain categories of step to middleware, then when they encounter such a step, they should farm out the request. I've found, with trying engineering problems, that LLMs consume a lot of steps finding out what to collect, with a risk of hallucination. That's exactly the sort of thing that can be farmed out.

According to both Claude and ChatGPT, that sort of process is the focus of a lot of research, right now, although apparently it's not actually been tried with reasoners.

Comment Re:RAG (Score 1) 5

Yeah, that would count.

Specifically, with something like Protege, I can define how things relate, so I could set up an ontology of cameras, with digital and film as subtypes, where lenses are a component of cameras, films are a component of film cameras, and pixel count is a property of digital cameras.

The reasoner could then tell you about how the bits relate, but a SPARQL search could also search for all records in a database pertaining specifically to any of these parameters.

At least some search engines let you do this kind of search. On those, you can specifically say "I want to find all pages referencing a book where this is specifically identified as the title, it isn't referencing something else, and that specifically references an email address".

So, in principle, nothing would stop Claude or ChatGPT saying "I need to find out about relationships involving pizza and cheese", and the reasoner could tell it.

That means if you're designing a new project, you don't use any tokens for stuff that's project-related but not important right now, and you use one step no matter how indirect the relationship.

This would seem to greatly reduce hallucination risks and keeps stuff focussed.

What you're suggesting can bolt directly onto this, so Protege acts as a relationship manager and your add-on improves memory.

This triplet would seem to turn AI from a fun toy into a very powerful system.

User Journal

Journal Journal: Question: Can you use Semantic Reasoners with LLMs 5

There are ontology editors, such as Protege, where there are a slew of logical reasoners that can tell you how information relates. This is a well-known weakness in LLMs, which know about statistical patterns but have no awareness of logical connections.

Comment Huh. (Score 3) 40

Why are they monitoring syscalls?

The correct solution is surely to use the Linux Kernel Security Module mechanism, as you can then monitor system functions, regardless of how they are accessed. All system functions, not just the ones that have provision for tracepoints.

For something like security software, you want the greatest flexibility for the least effort, and Linux allows you to do just that.

Because it's fine-grained, security companies can then pick and choose what to regard or disregard, giving them plenty of scope for varying the level of detail. And because the LSM allows services yo be denied, there's a easy way for the software to stop certain attacks.

But I guess that the obvious and most functional approach would mean that the vendors would have to write a worthwhile product.

Submission + - Whistleblower reports terrible things due to DOGE (youtube.com) 9

echo123 writes: NLRB employee Daniel Berulis reports on CNN that within 15 minutes of DOGE staff receiving new accounts with access to highly sensitive Department of Labor (DoL) data, someone within Russia logged in with the correct username and password over 20 times, but were rejected by location-related conditional access policies. Additionally a traffic spike of 10Gb of data exiting DoL was witnessed which is highly unusual activity at anytime.

Also, DOGE is using Starlink to exfiltrate data, and Starlink is known to be hacked by Russia.

He also reports this activity is not limited to the DoL, it has been witnessed across the government I.T. infrastructure, and that sensitive databases have recently been exposed to the open internet.

Daniel Berulis also received a clear message to stop looking. Part of the package he received included drone footage of him walking his dog.

Fast forward to 4min 15seconds if you're in a hurry.

= = =

Via Reuters

Berulis alleged in the affidavit that there are attempted logins to NLRB systems from an IP address in Russia in the days after DOGE accessed the systems. He told Reuters Tuesday that the attempted logins apparently included correct username and password combinations but were rejected by location-related conditional access policies.

Berulis' affidavit said that an effort by him and his colleague to formally investigate and alert the Cybersecurity and Infrastructure Security Agency (CISA) was disrupted by higher-ups without explanation.

As he and his colleagues prepared to pass information they'd gathered to CISA he received a threatening note taped to the door of his home with photographs of him walking in his neighborhood taken via drone, Andrew Bakaj, Whistleblower Aid's chief legal counsel, said in his submission to Cotton and Warner.

"Unlike any other time previously, there is this fear to speak out because of reprisal," Berulis told Reuters. "We're seeing data that is traditionally safeguarded with the highest standards in the United States government being taken and the people that do try to stop it from happening, the people that are saying no, they're being removed one by one."

via NPR

The top Democrat on the House Oversight Committee is calling for an investigation into DOGE's access to the National Labor Relations Board following exclusive NPR reporting on sensitive data being removed from the agency.

Ranking Member Gerry Connolly, D-Va., sent a letter Tuesday to acting Inspector General at the Department of Labor Luiz Santos and Ruth Blevins, inspector general at the NLRB, expressing concern that DOGE "may be engaged in technological malfeasance and illegal activity."

"According to NPR and whistleblower disclosures obtained by Committee Democrats, individuals associated with DOGE have attempted to exfiltrate and alter data while also using high-level systems access to remove sensitive information—quite possibly including corporate secrets and details of union activities," Connolly wrote in a letter first shared with NPR. "I also understand that these individuals have attempted to conceal their activities, obstruct oversight, and shield themselves from accountability."


Comment Re:They are going from 4.5 to 4.1? (Score 1) 13

Since R1 has good reasoning, but no real breadth, and is open source, the logical thing would be to modify R1 to pre-digest inputs and create an optimised input to 4.1. The logic there would be that people generally won't provide prompts ideally suited to how LLMs work, so LLM processing will always be worse than it could be.

R1 should, however, be ample for preprocessing inputs to make them more LLM-friendly.

Comment Re:Hmmm. (Score 1) 38

Gemini struggles. I've repeated the runs on Gemini and can now report it says files were truncated two times in three. After reading in the files, it struggles with prompts - sometimes erroring, sometimes repeating a previous response, sometimes giving part of an answer.

When it gives some sort of answer, further probing shows it failed to consider most of the information available. In fairness, it says it is under development. Handling complex data sources clearly needs work.

But precisely because that is its answer, I must conclude it is not ready yet.

I would be absolutely happy if any of these AI companies were to use the files as a way to stress-test, but I'm sure they've better stress tests already.

Still, it would be very unfair of me to argue a need for improvement if I were to insist on not providing either evidence or a means of testing.

Slashdot Top Deals

A meeting is an event at which the minutes are kept and the hours are lost.

Working...