Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror

Comment Just shoddy... (Score 4, Interesting) 91

What seems most depressing about this isn't the fact that the bot is stupid; but that something about 'AI' seems to have caused people who should have known better to just ignore precautions that are old, simple, and relatively obvious.

It remains unclear whether you can solve the bots being stupid problem even in principle; but it's not like computing has never dealt with actors that either need to be saved from themselves or are likely malicious before; and between running more than a few web servers, building a browser, and slapping together an OS it's not like Google doesn't have people who know that stuff on payroll who know about that sort of thing.

In this case, the bot being a moron would have been a non-issue if it had simply been confined to running shell commands inside the project directory(which is presumably under version control, so worst case you just roll back); not above it where it can hose the entire drive.

There just seems to be something cursed about 'AI' products, not sure if it's the rush to market or if mediocre people are most fascinated with the tool, that invites really sloppy, heedless, lazy, failure to care about useful, mature, relatively simple mitigations for the well known(if not particularly well understood) faults of the 'AI' behavior itself.

Comment Re:Only part of the story... (Score 1) 123

What always puzzled me about Intel's...more peripheral...activities is that they seemed to fall into a weird, unhelpful, gap between 'doing some VC with the Xeon money; rather than just parking it in investments one notch riskier than savings accounts' and 'strategic additions to the core product'; which normally meant that the non-core stuff had limited synergies with intel systems; and had the risks associated with being a relatively minor program at a big company with a more profitable division; and thus subject to being coopted or killed at any time.

Seemed to happen both with internal projects and with acquisitions. Intel buys Altera because, um, FPGAs are cool and useful and it will 'accelerate innovation' if Intel is putting the PCIe-connected FPGA on the CPU's PCIe root complex rather than a 3rd party vendor doing it? Or something? Even at the tech demo level I'm not sure we even saw a single instance of an FPGA being put on the same package as a CPU(despite 'foveros' also being the advanced-packaging hotness that Intel assured us would make gluing IP blocks together easy and awesome). They just sort of bought them and churned them without any apparent integration. No 'FPGA with big fuck-off memory controller or PCIe root we borrowed from a xeon' type part. No 'Intel QuickAssist Technology now includes programmable FPGA blocks on select parts' CPUs or NICs. Just sort of Intel sells Altera stuff now.

On the network side, Intel just kind of did nothing with and then killed off both the internal Omni-path(good thing it didn't turn out that having an HPC focused interconnect you could run straight from your compute die would have been handy in the future...luckily NVlink never amounted to much...) and the stuff they bought from Barefoot; and at this point barely seems to ship NICs without fairly serious issues. I'm not even counting Lantiq; which they seem to have basically just spent 5 years passing on to Maxlinear with minimal effect; unless that one was somehow related to that period where they sold cable modem chipsets that really sucked. It's honestly downright weird how bad the news seems to be for anything that intel dabbles in that isn't the core business.

Comment Re:Quality Work Can't Be Rushed (Score 1) 123

Not delivering on schedule is absolutely a symptom; it's just a somewhat diagnostically tricky one since the failure can come from several directions; and 'success' can be generated by gaming the system in several places, as well as by successful execution.

In the 'ideal' case things mostly happening on schedule is a good sign because it means both that the people doing the doing are productive and reliable and the people trying to plan have a decent sense(whether personally, or by knowing what they don't know and where they can get an honest assessment and doing so) of how long things are going to take; whether there's something useful that can be added or whether forcing some mythical man-month on the people already working on it would just be a burden; keeping an eye on whether there's anything in the critical path that is going to disrupt a bunch of other projects, and so on.

If you start losing your grip on the schedule, that fact alone doesn't tell you whether your execution is dysfunctional or your planners are delusional, or some combination of the two; but it's not a good sign. Unhelpfully, the relationship between how visibly the gantt charts are perturbed and how big a problem there is is non-obvious(a company whose execution is robust but whose planners live in a world of vibes-based theatre and one whose execution is dysfunctional and crumbling and whose planners are reusing estimates from the time before the rot set in might blow a roughly equal number of deadlines; despite one having mostly a fluff problem and one probably being in terminal decline); but it's never a good sign.

Comment Re:What's old is new again (Score 1) 43

That wasn't *all* I said, but it is apparently as far as you read. But let's stay there for now. You apparently disagree with this, whnich means that you think that LLMs are the only kind of AI that there is, and that language models can be trained to do things like design rocket engines.

Comment Re:What's old is new again (Score 5, Informative) 43

Here's where the summary goes wrong:

Artificial intelligence is one type of technology that has begun to provide some of these necessary breakthroughs.

Artificial Intelligence is in fact many kinds of technologies. People conflate LLMs with the whole thing because its the first kind of AI that an average person with no technical knowledge could use after a fashion.

But nobody is going to design a new rocket engine in ChatGPT. They're going to use some other kind of AI that work on problems on processes that the average person can't even conceive of -- like design optimization where there are potentially hundreds of parameters to tweak. Some of the underlying technology may have similarities -- like "neural nets" , which are just collections of mathematical matrices that encoded likelihoods underneath, not realistic models of biological neural systems. It shouldn't be surprising that a collection of matrices containing parameters describing weighted relations between features should have a wide variety of applications. That's just math; it's just sexier to call it "AI".

Comment Re:Seems reasonable (Score 2) 24

It seems reasonable; but also like something that should really spook the customers.

It seems to be generally accepted that junior devs start out as more of an investment than a genuine aid to productivity; so you try to pick the ones that seem sharp and with it, put some time into them, and treat them OK enough that they at least stick around long enough to become valuable and do some work for you.

If that dynamic is now being played out with someone else's bots, you are now making that investment in something that is less likely to leave, whatever as-a-service you are paying for will continue to take your money; but which is much more likely to have a price regularly and aggressively adjusted based on its perceived capabilities; and have whatever it learned from you immediately cloned out to every other customer.

Sort of a hybrid of the 'cloud' we-abstract-the-details arrangement and the 'body shop' we-provision-fungible-labor-units arrangement.

Some customers presumably won't care much; sort of the way people who use Wix because it's more professional than only having your business on Facebook don't exactly consider web design or site reliability to be relevant competencies; their choice is mostly going to be between pure off the shelf software and maybe-the-vibe-coded-stuff-is-good-enough; but if your operation depends in any way on your comparative ability to build software Amazon is basically telling you that any of the commercial offerings are actively process-mining you out of that advantage as fast as the wobbly state of the tech allows.

Comment Re:Blaming a single cause (Score 1) 87

we still don't know why this particular civilization disappeared without a record of what happened.

We do have records. We just can't read them. The Harappan language has never been deciphered. There are about 5000 inscriptions known.

That droughts led to the end of Indus Valley Civilization has been surmised for decades, this study provided a much more detailed account of the process.

For people to settle in "untouched tracts of land" you need to have water to irrigate it. Large empty areas on Earth require water for them to be "tamed".

Comment Well, that answers my question... (Score 4, Insightful) 38

So the 'hyperloop' people have a cool website; while the 'train' people are just plain getting on with building stuff; whether conventional or the now-quarter-century-ish old maglev option.

Looks like someone signed up for another round of 'faff with apps vs. offshoring our entire high tech supply chain' and hoped it would work better this time.

And some dumbass 'managing director' is telling us that a gigantic safety-critical vacuum system is 'not effected by strikes'; more or less because he has no idea what the maintenance and operations would involve? Truly a joke telling itself.

Slashdot Top Deals

How can you do 'New Math' problems with an 'Old Math' mind? -- Charles Schulz

Working...