Ok you hit your coffee limit early in the day. Sodium battery technology is here, finally, and solar and wind have been mature for a long time. There's no reason most countries can't go fully electric.
I think the idea is to keep shitting out 4.X releases until openai releases GPT-6, at which point anthropic will release whatever version of "mythos" they have mostly working that day as Opus 5.0 on the same day, or the next. They'll probably claim it's so operful they're skipping 5 and just calling it Opus 6 for marketing reasons
This is less and less true as time goes on, and only for either heavily parallelized processes, processes that require ENORMOUS amounts of memory, or processes that require extreme amounts of GPU. Even then, the GPU gap is getting smaller if you're not talking about gaming.
I compile Unreal Engine for my job. I do it a lot. My M1 Mac Mini, 5 years old, keeps up astonishingly well with my work-issued i9. I'm sure an M5 would blow it out of the water.
I'm sure there are workloads where what you're saying is true, but I don't think you can make this as a broad claim anymore. M-series chips aren't low-powered chips, they're higher efficiency chips. They do more with less. I don't look at benchmarks from Apple (or Intel), either--I only pay attention to benchmarks from 3rd parties, and they're still quite favourable. Aside from gaming--where we ABSOLUTELY optimize the hell out of things, and tune things specifically to hardware as best as we can--most general purpose software that runs on both runs within a totally acceptable margin either way.
Everyone has to stop what they're doing for an entire day, travel to the training center, which costs money, they have to rent the training center, which costs money, they have to pay the training person to present the training materials, which costs money, and they have to develop the training materials/course, which costs money.
And then the next day is going to be complete chaos, because the training materials were developed against v0.7 of the software, and everyone is using v1.3 of the software, and nothing will get done for, minimum, 2 days, and you won't actually be at the same level of effectiveness for 3-6 months, and in some edge cases, 18-24 months, possibly longer.
I was at one company, and the 80 year old lady, Wanda, who ran payroll, worked on a specialty windows 95 computer, because she could not be retrained and didn't want to learn new software. And nobody messes with the payroll lady. This was in ~2014. I just looked her up looks like she passed away finally, probably still using Windows 95 all the way until 2023 bless her.
Anyways TL;DR for digital soveignry $2000/user is a trivial amount especially as a one time cost
In the Mac space, a lot of malware just gives up if it detects Little Snitch is running. It's a really effective tool, and honestly, the cost is so minimal compared to the value. I've been running it for years, and I appreciate how much junk traffic I can block.
If one of the six hydrocoptic marzel vanes fails and is no longer mounted to the lunar wane shaft, side fumbling will no longer be effectively prevented!
Yeah I'm kind of done feeding Google every scrap of information about me for advertising purposes. I switched to private search and email last year for about $100/year and don't miss them at all.
output is about 6 tokens/s with 16k context window i'm not having any issues since it went live this afternoon. it's not sparkling like opus 4.5/6 but gets the job done
i generally send it a voice note via telegram while driving and then check back in like 1-2 min, or it is sending me a reminder about something on our shared calendar. it's still faster than texting my buddy about making plans for this weekend or whatever.
I'm using a $200 used ~5 year old (from the ebay listing) HP EliteDesk 805 G6 DM Desktop Ryzen 5 PRO 4650GE 3.3GHz 32GB RAM 512GB SSD WiFi in cpu mode... you don't need a gpu to run single user local LLM... just a bunch of ram. This isn't 2022 anymore
It's about 5 tokens/second which is totally fine for an async assistant. 20 tokens/second is about the lower limit for usable in realtime. You can also set it up to use a smaller model for quick questions (what are the next 6 items on my calendar/to-do list?) and drop through to the bigger slower model for harder questions (can you add this feature to my internal ticketing system and redeploy?)
I ordered 64gb of ram about an hour ago and i'm planning on running either qwen 35B-A3B 8 bit or 122B-A10B 3 bit in fully offline mode.
>the actual cost of 'running the AI.'
is a fixed $200 cost (ram upgrade) + electricity
I cancelled my subscription overnight, and I'm using the free credits they gave me to wrap up some things and transition away. I am not going to be locked into someone's walled garden again.
"Michael Marshall Smith's 1996 novel Spares, in which the hero liberates intelligent clones from a "spare farm", was optioned by DreamWorks in the late 1990s, but was never made. It remains unclear if the story inspired The Island, so Marshall Smith did not consider it worthwhile to pursue legal action over the similarities."
Anyway, we're all saying the same thing here. This is all Torment Nexus stuff. We know how this ends.
Whenever people agree with me, I always think I must be wrong. - Oscar Wilde