Forgot your password?
typodupeerror

Submission + - Has Slashdot Become More Ads Than "News for Nerds, Stuff That Matters"? 2

FictionPimp writes: Load Slashdot's front page today without an ad blocker and count what you see before scrolling.

Above the fold, there are 6 distinct ad placements: a full-width Retool banner just below the navigation, a MongoDB Atlas inline banner styled to look like a site notice sitting directly above the first story, two sidebar ad units (one for a game dev course bundle, one for business software comparison), a "Sponsored Content" slot beginning to appear at the bottom edge, and a sticky MongoDB footer bar fixed to the bottom of the screen. MongoDB alone holds two simultaneous placements on the same page load. The ratio is 6 ads to 2 stories before you even scroll.

Slashdot has carried the tagline "News for nerds, stuff that matters" since Rob Malda was running the site out of a college dorm in 1997. It is now owned by Slashdot Media, the same parent as SourceForge, and the nav bar includes a "Thought Leadership" section, which is industry parlance for paid editorial content.

None of this is unique to Slashdot. Display advertising is how independent tech publications survive. But there is a meaningful difference between ads that share a page with content and ads that outnumber and surround the content, with some of them actively designed to look like part of the editorial feed.

The question for the Slashdot community: at what point does the original promise of the site, a curated community-moderated signal in a noisy web, get buried under the noise it was supposed to filter? Should the site be rebranded: "Ads for Nerds, News if we can fit it in"?

Comment Re:You can lead a bot to solder.. (Score 1) 61

AI cannot extrapolate, per construction. They can interpolate better than anyone, but as soon as you leave its dataset, it has no clue anymore.

You can feed the best AI a trillion photos of cats, if none of them included a black cat, it will be fundamentally unable to tell you that a picture of a black cat contains a cat.

The illusion that it can extrapolate comes from the fact that these models are fed with humongous amounts of data, so even just interpolating is still mostly good enough as you won't go near the edge of the dataset.

An artificial algorithmic "mind" knows what it knows.
A human mind can conceive that there are things it does not know.

So AI can sort through and retrieve an item and that item's entire ranked adjacent items, faster and more reliably than any human brain. But it cannot break its own rules. You can ask AI to generate an image of a cat with nine tails and zebra striped fur, and it can do that because you - the human mind - prompted it to invoke those specific rules (cat, nine, tails, zebra, fur) and combine them in specific ways. But if you - a human mind - paint a photorealistic image of a cat with nine tails and zebra striped fur wearing a ballet tutu, a top hat, and a monocle, swimming past a coral reef, then present that image to an AI program and request it identify the thing in the picture, it might return anything. And, whatever it outputs as its answer, if you follow up to input "Are you sure that's correct? It has a tutu like a ballerina, stripes like a zebra, top hat like a victorian gentleman, (etc.)" it is very likely to output a completely different guess, or (if programmed by human minds to do so) will output that it can't be sure.

Meanwhile, every (non-neurologically damaged) human being over the age of 5 will immediately say "a cat" or something like "a cat with zebra fur and a lot of tails and a tutu and...". If you follow up to ask, "Are you sure that's correct? Maybe it's something else" they will say "It's a weird cat, but yeah it's a cat" - even though there has never existed such a cat in the history of the world, and thus from a taxonomic, scientific, reality sense that image does not contain a cat.

The human mind has capacity for unknowns, so it can instantly expand to create space where that completely-not-a-real-thing still is a cat.

The intriguing "magic" is that you could train a current-gen AI to recognize that image as a cat. IF you first created tens of thousands of images of equally bizarre cats, mixed them with images of equally bizarre dinosaurs, giraffes, cactuses, etc., then had thousands of human minds click CAPTCHA tests on the images, then added that data to the AI. That is, the AI could eventually have a high accuracy rate on bizarrely-visualized cats if subjected to thousands/millions of instances of cats. Meanwhile, a 5 year old human who may have only seen a small handful of cats in their entire life, and never ever one wearing a tutu/hat/monocle/stripes while swimming in the ocean, will instantly and with total certainty know "that's a cat".

That's one reason I scoff at the claims we're right around the corner from (or already across the line to) human-level general intelligence if we just (n+1) faster. Every big LLM out there has "read" more text than a random sample of a billion human beings. Human understanding is NOT just about adding more instances of a thing. There's something else (and a lot of it) going on that makes human conscious cognition what it is.

What's even more provocative to me is watching what happens when your toddler learns what a thing is. They become obsessed with it. For the next several months, every time a cat (physical or an image) comes within view, they point and excitedly say "Cat!!!" "DaDa! Cat!!!" Same thing for every airplane in the sky, every bird at the window. They are hungry for ideas to fill their capacity. The human animal inherently LOVES acquiring concepts. We crave them. They pleasure us.

It's why elementary school kids love corny jokes. A joke takes a concept, then jumps from that to a different, unexpected concept. The brain gets the pleasure of knowing the first concept and is following along, then gets a hit of pleasure from suddenly having a new concept thrust into the concept space and recognizing both the nature of the connection and the jump. The more apparent distance between the two concepts, the more pleasure we derive from it. Almost all kid jokes (and a huge percentage of grownup humor) takes the form:
1) Banal context followed by weird incongruous situation. Knock knock. Who's there? Me. Me who. Meow!
OR
2) Weird context followed by banal situation. Why did the chicken cross the road? To get to the other side!

5 year olds absolutely love jokes like #1. The reason is because of that pleasure hit from making a new connection. Until they first hear the joke, the word "me" has only been used to refer to the self. The word "meow" has only been used to refer to the sound of a cat. The concepts of "self" and "sound of a cat" have zero conceptual connection. That is, they have an infinite distance between them, semantically. They only connect through an accident of the particular phonetics of English and a few other languages where both words start with the same syllable. And kids get that. They don't know the word onomatopoeia, but they already understand and know how to work with the concept of onomatopoeia.

You can melt every ounce of ore in the world and refine every drop of oil in order to produce 60 trillion GPUs and then use every centimeter of land to create one massive datacenter, and the absolute best of our current AI models will still not get the joke. They will have less understanding and less concept-manipulation capacity than a 5 year old. It's all still just programmatic I/O and logic gating at this point.

Comment This story is why I believe IP must be abolished (Score 1, Offtopic) 86

Global mergers/acquisitions heading toward all technology being the proprietary possession of a small coterie of Barons.
You will be technically "free" to live as you wish, the same way the poor schmucks who took the Red Pill exited the Matrix and were free-- to scrabble out their depressing MRE-hardtack existence hiding in the dark crags of subterranean tunnels.

You will, on paper, philosophically speaking, have all your freedoms, but they won't be worth exercising. Because if you want access to the productivity, the food, the space, the medical care, the communication, the platforms, the government services... access to anything at all, everything on the planet is behind some kind of paywall.
16 tons, and what do you get?

1) Convert the planet's entire socioeconomic infrastructure into dependence on ComputeAsAService.
2) Turn compute access into currency.
3) ???
4) Profit!

(Where "???" means "breaking the placenta becomes a shrink-wrap TOC whereby you implicitly agree at birth to become a wholly owned subsidiary of Weyland-Yutani Corp". And "Profit!" means "Sweep the table. It's yours in perpetuity. Only thing left is cloning and mind downloads, then it's Godhood for one ultimate winner of the Game Of Clones".)

Do you want to live to see the world of "Alien : Earth" series? Because this path is how you get the world of the "Alien : Earth" series. Or Altered Carbon. Or a few thousand other SF stories over the years.

Comment This is why IP must be abolished (Score 1) 2

Global mergers/acquisitions heading toward all technology being the proprietary possession of a small coterie of Barons.
You will be technically "free" to live as you wish, the same way the poor schmucks who took the Red Pill exited the Matrix and were free to scrabble out their depressing MRE-hardtack existence hiding in the dark crags of subterranean tunnels.

You will have all your freedoms, but they won't be worth anything. Because if you want access to the productivity, the food, the space, the medical care, the communication, the platforms, the government services... access to anything at all, everything on the planet is behind some kind of paywall.
16 tons, and what do you get?

1) Convert the planet's entire socioeconomic infrastructure into dependence on ComputeAsAService.
2) Turn compute access into currency.
3) ???
4) Profit!

(Where "???" means "breaking the placenta becomes a shrink-wrap TOC whereby you implicitly agree at birth to become a wholly owned subsidiary of Weyland-Yutani Corp". And "Profit!" means "Sweep the table. It's yours in perpetuity. Only thing left is cloning and mind downloads, then it's Godhood for one ultimate winner of the Game Of Clones".)

Do you want to live to see the world of "Alien : Earth" series? Because this path is how you get the world of the "Alien : Earth" series. Or Altered Carbon. Or a few thousand other SF stories over the years.

Submission + - Silicon Valley is buzzing about this new idea: AI compute as compensation (businessinsider.com) 2

sziring writes: Silicon Valley has long competed for talent with ever-richer pay packages built around salary, bonus, and equity. Now, a fourth line item is creeping into the mix: AI inference.

As generative AI tools become embedded in software development, the cost of running the underlying models â" known as inference â" is emerging as a productivity driver and a budget line that finance chiefs can't ignore.

Software engineers and AI researchers inside tech companies have already been jousting for access to GPUs, with this AI compute capacity being carefully parceled out based on which projects are most important. Now, some tech job candidates have begun asking about what AI compute budget they will have access to if they decide to join.

"I am increasingly asked during candidate interviews how much dedicated inference compute they will have to build with Codex," Thibault Sottiaux, engineering lead at OpenAI's Codex, the startup's AI coding service, wrote on X recently.

Comment "It is hoped" by... ? (Score 1) 30

It is hoped that remote robotic surgery could spare future patients the "vast expense and inconvenience" of travelling for treatment, and help deliver better healthcare to people in more remote locations.

Whoever out there is hoping the above, clearly hasn't paid attention to how products like this are brought to market.

To be accurate, the sentence should read:

It is hoped - by hospital corporations, med tech manufacturers, insurers, and all the corporations who supply/service or invest in the business of medical care - that remote robotic surgery could spare future investors the vast expense and inconvenience of employing enough humans for each city/region to have a set of people and facilities to provide common medical care; now they can source their surgeons through a rotation of countries with the lowest-billable-hour wages, and help deliver better profits to equity investment firms.

Comment Re: AI can find it faster (Score 1) 54

As long as you aren't using any locally identifying language or specific slang then I doubt you can tell one person definitely from another via word choice alone. At least not within a sea of multiple accounts. If you were given two pieces of text and had to surmise if they were the same person sure but one to millions without a specific phrase that's unique to the author I'd be surprised.

People do not create whole sentences de novo. We each have tendencies to re-use phrases and constructs, as well as tendencies for how we type sentences -- including this one which contains multiple clauses, commas, an em dash, a hyphenated "re-use" instead of "reuse", I wrote "de novo" instead of ex nihilo or ab chao or out of nowhere or from scratch, and so on.

and so on
and whatnot
on and on
yadda yadda
etc.
etc
etcetera
&c

There are lots of possible variations for every choice of word and punctuation; enough variety to develop fingerprinting based on emergent clusters of those variations.

Your scenario of one-to-one matching versus one-to-a-million matching is technically correct, but doesn't reflect the reality. Yes, it would be potentially difficult to firmly prove the author of any one single anonymous post with no context. But to determine who has control of an anonymous account based on that anon account's wording patterns, the pattern of when they post, how frequently, which topics do they post on, even what opinions they express.... all of that is getting easier and easier to do with automation. The future we're building will be 100% doxxable from the ground up.

Comment Re:Adverts and films? (Score 0) 96

This is the best possible outcome. You cant just fire people and replace them with AI if you want IP. If I make an advertisement and want copyright it'll need at least SOME work done by humans. If you want music you can own and generate royalties from, it'll need humans in there somewhere. This protects human labour while still letting these tools be out there to *assist*.

I'd make an even stronger argument -- not only is this "the best possible outcome", it is the only plausible outcome.
IP doesn't exist in. Copyright doesn't exist. Unlike the right to control your body or the right to free thought or the right to defend yourself, there is nothing about IP law that is natural, tangible, or inherent to existing. IP law is a whole-cloth invention. Governments conjure and construct IP protections, justified by one and only one purpose -- to provide extra ways for humans to earn money with which to feed, clothe, and shelter themselves.

Any application of IP law which does not give a piece of IP into the control of a human or group of humans for the above purpose, is immediately and irredeemably illegitimate, is outside the justifying purpose for which the fabricated construct of "Intellectual Property" exists.

IP isn't real.
We created it.
We can un-create it at any time.

If non-humans can have IP, and indeed if IP becomes actively hostile to its purpose, it should be nullified.

Comment Re: Slashslop (Score 2) 82

Test using âoeb i z xâ (remove spaces) instead. This is the name of slashdot's corporate owner.

I tried. Interesting... all references to that string are censored, regardless of whether the overall post is positive or negative. Merely mentioning the name triggers the censor.
With a filter that aggressive, I suspect that by trying to evade the content block you're probably gonna get Bixzslopped into next Tuesday.

Comment Re:Please don't use Paramount+ Platform (Score 2) 55

Really hoping they don't use the Paramount+ platform going forward. It is by far the worst performing streaming application I have experienced. I have not used HBO Max, but it cannot be as bad as Paramount+. They should hire some programmers and make a proper application that runs quickly on the devices that run it. There should be nothing too complex about browsing lists and surfacing data quickly. Paramount+ is a failure and really frustrating to use compared to every other service, Netflix and Apple+ being the best in my opinion.

When you say "platform" what do you mean?
The server uptime/latency?
Their Android app?
Their iOS app?
Their website?
The account/content settings available to users?

Comment Re:"Deterministic" (Score 1) 28

More seriously, even if humans are entirely deterministic I don't think we could function under that belief system. How could you jail a murderer for their crime if they had no ability to choose to do otherwise? I suppose you could because you believe you also have no choice but to convict them despite that conclusion being logically absurd. Any human that tried actually living by that premise would have to be exceptionally detached or would drive themselves mad.

Despite the increasing popular support for thoughtcrime over the past several decades, it has never been a valid approach. Morality is a terribly illogical thing to base crime and punishment on. When a neighbor's dog digs into your backyard and mauls your toddler, you do not need to be "exceptionally detached" or "driven mad" to insist that the dog be euthanized or permanently caged, despite the fact that the dog has zero moral culpability and had zero ability to choose otherwise. You identify the threat and you disarm, relocate, or destroy the threat.

Predicating crime and punishment on the moral culpability that arises from the alleged existence of "free will" is foolish, and is the direct progenitor of nonsense like the "gay panic" defense or Lorena Bobbit's acquittal for a crime she freely (ha ha) admitted to committing. There should never be any such thing as "Not Guilty by reason of temporary insanity" unless you believe in fairy tales like angels and demons and gods meddling in human minds. There should only be "Guilty and disordered". We are cognitive-behavioral engines; we are who we are, and we do who we are, and we are what we do. The "insanity" is your mind/body being exactly what it is. Insanity isn't some quantum subparticle ephemera that appears ex nihilo and then just as suddenly winks out once the math is balanced. The "insanity" wasn't temporary, even though the circumstances which triggered your behavior may be. Society has every right to disarm, incarcerate, or destroy you for your own damaging/destructive behavior, regardless of whether you did it freely or out of neurohormonal compulsion. Your child is not any less mauled/dead because the dog has no moral agency.

Submission + - Long Before Tech CEOs Turned to Layoffs to Cover AI Expenses, There Was WorldCom

theodp writes: Jeopardy time. A. This company spurred CEOs to make huge speculative capital expenditures based on wild unverified claims of future demand, resulting in the layoffs of tens of thousands of workers to reduce the resulting expenses, harming their core businesses. Q. What is OpenAI?

Sorry, the correct response is, "What is WorldCom?" In 2002, WorldCom, the second largest long-distance company in the U.S., entered Chapter 11 bankruptcy after disclosing accounting fraud that eventually totaled $11 billion, the biggest ever at the time. CEO Bernard Ebbers was subsequently sentenced to 25 years in prison. CNBC reported that an employee of WorldCom’s Internet service provider UUNet set off a frenzy of speculative investment and infrastructure overbuild after he used Excel to create a best-case scenario model for the Internet’s growth that suggested in the best of all possible worlds, Internet traffic would double every 100 days, a scenario that would greatly benefit WorldCom, whose lines would carry it. Despite no evidence to support it, WorldCom’s lie became an immutable law and businesses around the world made important decisions based on the belief that traffic was doubling every 100 days.

"For some period of time I can recall that we were backfilling that expectation with laying cables, something like 2,200 miles of cable an hour,” AT&T CEO Michael Armstrong said. “Think of all the companies that went out of business that assumed that that was real.” Armstrong and former Sprint CEO Bill Esrey struggled for years to understand how WorldCom could beat them so handily. “We would look at the conduct of WorldCom in terms of their pricing, revenue growth, margins, in terms of their cost structure ... and the price leader almost every quarter was WorldCom,” Armstrong said. Added Esrey, “We couldn’t figure out how they were pricing as aggressively as they were. ... How could they be so efficient in their costs and expenses?”

AT&T and Sprint began cutting jobs to push down their costs to WorldCom’s level. “The market said what a marvelous management job WorldCom was doing and they would look over to AT&T and say, ‘these guys aren’t keeping up.’ So, my shareholders were hurt. We laid off tens of thousands of employees in an accelerated fashion [in a futile effort to match WorldCom's phantom profits] and I think the industry was hurt,” Armstrong says. “It just wrecked the whole industry,” says Esrey.

Comment Re:Reading of fiction does not tell much (Score 1) 73

People reading (potentially lots of) fictional books may prove that they are able to read, but that does not tell us whether (1) they have good reading comprehension and could learn from non-fictional texts, it also does not tell us whether (2) they are able to communicate actively. And it is (1) and (2) that I am most concerned with in young colleagues, with a recent addition of (3) have they outsourced thinking to LLMs? I really would not mind working with people who never read a single book, as long as they have been able and willing to read non-fiction texts (on whatever medium), comprehended them, and are able to apply and communicate what they learned from them.

A more concerning symptom than "not reading books", from my observation, is the immense popularity of few-seconds-short videos. Those videos are entirely incapable of transporting any meaningful amount of information, and my suspicion is that those consuming such videos en masse have developed severe attention span deficits.

Speaking of poor communication comprehension, who modded this appropriate, on-topic, mild, reasoned comment as Troll?

It raises a perfectly legitimate question. I personally know several people who do not read fiction for entertainment at all, but are highly communicative critical thinkers who have written non-fiction books in their field. And on the other end, I have worked with many people who are always reading entertaining fiction, yet cannot compose structured explanatory documents or create clear instructions for a process.

Slashdot Top Deals

I don't do it for the money. -- Donald Trump, Art of the Deal

Working...