Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror

Comment Re:This is how it should be (Score 1) 14

My guess is phones and silicon will improve to make 4b on mobile by 2030. If not then those requests will get forwarded to xyz cloud service. I can see a world where 7-12b cloud models are ad supported free tier and you either pay or self host 70-600b yourself. I expect processing requirements to drop by half due to whatever breakthrough comes next and then there's a long tail of improvement after that. Token verification was a major improvement.

Comment Re:Outrage Fatigue is also a factor (Score 1) 180

Yep I login to some of these sites and I see 3-4 tangentally-related things that are clearly algorithm rage-bait, designed to drive user interaction. I'm so tired of this. I login for 10-15 minutes a handful of times a week to check in on friends and family for updates, and then promptly uninstall the app.

Comment Re:But who will train the AI? (Score 1) 73

Probably in 3-10 years there will be a handful of open, legally copyright free training sets anyone can use to train their own ~600b class model with whatever architecture is current state of the art. Researchers are already putting together 7b training sets like this. And LLMs can use tools like search now so they won't always need the most up to date news or info - they can use tools for that instead. Most of finance is analysis of documents, which LLMs have been excellent at for a while now.

Comment Re:This is how it should be (Score 1) 14

Google announced roughly the same thing, on device models for phones a couple weeks ago at their developer conference. The 1b model is fine for basic tasks like turning on lights, checking email, social media notifications etc and runs ok on midrange phone hardware. The 4b model technically runs but it's borderline unusable speed but it can answer questions like "how does a microwave work?" with moderate accuracy at a semi-scientific level which is impressive. I suspect most devices will be able to run a 1b and by the end of the decade most everything will run a 4b model at least at talking speed. There's a concept that all AI processing will be done in the datacenter, I suspect 80%+ of consumer LLM will happen on the device, and more complex tasks will get routed to the cloud. For a lot of end users (high school students, etc) 98%+ of requests will be on-device.

Comment Seems to fall apart above 200 LOC (Score 1) 71

LLMs are really good at stuff, better and faster than humans, as long as the complexity isn't much more than ~200 LOC (lines of code). 250-300 LOC and things start falling apart quickly. Sometimes (1/50) you'll get ready and it'll pop out 400 LOC without major errors but that seems to be the absolute limit of the current statistical model family everyone is using.
 
LLMs are really good at analyzing and summarizing text though, it has no problem analyzing 20-30 page PDFs of economic or financial data.
 
But yeah there was this idea that if you just kept training on bigger datasets for longer eventually you'd just arrive at AGI and it's pretty obvious via many many research papers that the error limit right now is ~1% and getting below that is really really dang hard. We're going to need a new breakthrough to get the ball further down the field.

Comment Putting someone more capable in charge of AI (Score 4, Informative) 16

Microsoft CFO was publicly complaining during last month's earnings call that the current Microsoft AI chief wasn't cutting it, so they're putting it under this guy instead. Who knows how long it will last; they'll keep reassigning it until it becomes a profitable, or path-to-profitability. Right now Microsoft is just barely ahead of Apple in public AI success, despite being an early partner with OpenAI.

Comment Re:Ouch. (Score 1) 40

The big issue is they wiped out their ci/cd and more importantly their entire AWS platform the system runs on. Sure you can probably re-deploy the code, but if it's not infrastructure as code then a lot of that stuff is bespoke hand-wired stuff and there's probably no backups for the config files etc. Maaaaybe you can deploy everything on a giant EC2 instance but you still lost all your database backups which means all your customers now have to create new logins, you need to create a new account with your credit card processor (if you can even get them on the phone) test everything.... I sorta specialize in business continuity having something like terraform or other infra as code helps a lot, and we do monthly or at least quarterly wargames where we restore develop from scratch to ensure there's been no drift in the interim. Quarterly end to end test of your backups isn't perfect but it's unlikely you'd be down for more than a couple of days for all but the largest enterprise setups. With something like terraform and kubernetes you ought to go from cold dark black to up and running inside of 12-18 hours, depending on where your offsite backups live.

Comment Re:Salt the Internet (Score 1) 88

Evaluation is a lot less computationally complex than generation; I would expect anything that gets scraped by a larger AI company goes through at least a two pass system 1) is this even relevant/compiling code? 2) does this meet a minimum standard? and probably for largest companies building a library of quality content 3) is this high enough quality to include in the library for future training? Steps 1/2 can probably be done with something like a 30b or 70b model fairly quickly and step 3 is probably evaluated by a 300-600b model. Evaluation is something like 10x faster than token generation so it's pretty straightforward.

Comment Re:Lack of imagination (Score 1) 61

The movie studios (particularly 900lb gorilla behemoths like Disney, and whatever the name of the company that owns HBO is this week) already own the copyright on enough content they don't need to train outside of their own copyright library. LLM summary of script -> Script of trailer -> shot by shot description of trailer -> 2-5 second "video clips" -> human or machine stitches them together, throws on a backing track is definitely something they can do/train on using their existing library. Since that kind of tooling isn't for export to customer facing they can train their model on whatever they want (even content outside of their own library) so long as they don't sell it.

Comment Re:Post office (Score 1) 41

US Post code + route number is pretty adequate with human mail carriers. I see my postman IRL probably 2-3x a month and he knows who I am, if you send a letter to my post code + route number with the name smudged illegibly it will almost certainly get delivered to the correct house. Not all mail carriers are as social as mine though.

Comment Pretty average executive compensation (Score 1) 33

Average starting executive compensation is north of $300k usually with like a 30% bonus, and goes up from there. 700k sounds like a lot but they've probably been there more than 5 years and what they've built is cash flow positive. If you include health insurance, expenses account, company car etc you could easily tack on another $200k to the total comp.

Comment Also why I'm not interested in using their AI prod (Score 1) 37

I mostly use google for it's search, email, maps stuff. Originally the tradeoff was I gave you info, and you gave me personalized results, and we paid for it with ads. Now the results are terrible generic stuff. I can just use a chatgpt type product and get similar answers for a nominal monthly fee and no ads. Why would I ever go back to google at this point? My search volume with google is probably a third of a quarter of what it used to be.
 
I don't see much value in using google's AI product unless it's phenomenally better, and there's an ad-free option. Until then, I'll continue using every AI product but google's.

Slashdot Top Deals

I have a very small mind and must live with it. -- E. Dijkstra

Working...