Comment Re:Test run (Score 1) 7
Haha, no. Humans' collars are already here, we just keep them in our pockets and pay for them ourselves.
Haha, no. Humans' collars are already here, we just keep them in our pockets and pay for them ourselves.
I think you'd be hard-pressed to find a company that doesn't hide things. The difference between Anthropic and (insert company here) is only that Anthropic leaked their source code, so now we can see what they kept hidden.
output is about 6 tokens/s with 16k context window i'm not having any issues since it went live this afternoon. it's not sparkling like opus 4.5/6 but gets the job done
i generally send it a voice note via telegram while driving and then check back in like 1-2 min, or it is sending me a reminder about something on our shared calendar. it's still faster than texting my buddy about making plans for this weekend or whatever.
I'm using a $200 used ~5 year old (from the ebay listing) HP EliteDesk 805 G6 DM Desktop Ryzen 5 PRO 4650GE 3.3GHz 32GB RAM 512GB SSD WiFi in cpu mode... you don't need a gpu to run single user local LLM... just a bunch of ram. This isn't 2022 anymore
It's about 5 tokens/second which is totally fine for an async assistant. 20 tokens/second is about the lower limit for usable in realtime. You can also set it up to use a smaller model for quick questions (what are the next 6 items on my calendar/to-do list?) and drop through to the bigger slower model for harder questions (can you add this feature to my internal ticketing system and redeploy?)
I ordered 64gb of ram about an hour ago and i'm planning on running either qwen 35B-A3B 8 bit or 122B-A10B 3 bit in fully offline mode.
>the actual cost of 'running the AI.'
is a fixed $200 cost (ram upgrade) + electricity
I cancelled my subscription overnight, and I'm using the free credits they gave me to wrap up some things and transition away. I am not going to be locked into someone's walled garden again.
If you can't figure out for yourself what's wrong with ordering large numbers of men to their deaths, then I won't be able to explain it to you.
Yes there is, it's hardware and driver version dependent. It's far more efficient to just do the compilation in the background than to keep a precompiled version for each game for each combination of hardware and driver, x2 once for Vulkan and once for DirectX for games which support both.
They could take that one step further: once your computer has compiled the appropriate shader for its particular combination of hardware/driver/etc, the game could upload that particular shader to a repository, so that the next install with the exact same combination of conditions could just download it instead of having to duplicate the work. I imagine there are a lot of people out there running functionally identical systems that would benefit.
I suppose they don't do that because they don't trust people not to repurpose the mechanism as a malware vector, or something.
U scared bro?
He's probably not scared enough. Anyone old enough to remember Vietnam knows how the song goes from here. "We must throw another batch of American men into the meat grinder, otherwise the lives of the previous batch will have been sacrificed in vain", and repeat ad infinitum.
Turning to AIs for a ruling on who is human and who is not is an interesting inversion of the Turing test.
It could be that. What it definitely was, though, is that Sony thought they could make more money selling to data centers than to the public.
Antiprotons, the forbidden PopRocks
I don't like ads either, but I do like that they (at least for now) have a paid tier with no ads. If there was an option to use google services at some paid tier, without being part of their ad network, I'd probably pay it. But there isn't and llm is as good as search these days (in many cases anyways) so I'm happy to jump ship. Piss off, google.
The "cutting edge" is getting rather dull. -- Andy Purshottam