In my experience, it just ends up requiring replacement of the starter battery more often, and the batteries required are larger, AGM type batteries that tend to be more expensive than smaller/flooded lead acid type batteries, (if you care about battery cost). Now, I suppose, if you sit in traffic a lot, and don't need A/C or heat too much, don't have a lot of accessories that are on a lot, (like heated or cooled seats, radio, etc.), and are absolutely wanting to save that last $2/month on gas, then go ahead. All I would ask is that the button *remember* the setting, between on/off ignition cycles, so if I turn it off, it just stays Off. I have installed a few harness adapters that connect to the harness going to the start/stop button so that it can 'remember' it's setting and only enables if the button is pressed to enable it back on. That is how it should have been from the factory.
I love how someone expressing their opinions is slowly moderated down to Troll. Heaven forbid that someone don't agree with you.
No problem for the AI hyper scalers...they'll just do some 100 trillion dollar deals back and forth and it all will work out.
I genuinely don't think the circle-jerk deals are the real force behind this. Microsoft "investing" in OpenAI doesn't inherently create datacenter demand. OpenAI "investing" in nVidia doesn't either. What does create demand is all the people asking CoPilot to summarize an e-mail for them. Or to write a paragraph of code they could Google the results of. Or create a picture of their wife's head on a walrus. The load is what's screwing us. All the useless, pointless, wasteful load.
Agree. That could be solved by the AI-bros actually charging for the services, which will have to happen someday. They will have to start making actual profit on this shit, someday. Although, I'd also think that a lot of the useless shit people ask AI to do could be handled by local models, which would be cheaper to run. I would think summarizing an email should certainly not require 10 top of the line GPUs to do by now.
I hope you do not plan to give that code any publicly available interfaces. Because LLMs are absolute crap at writing secure code. Probably because most people that proudly put their code on the net as "examples" are too.
If I was going to give a public interface to anything, it's going to go through the same process that I have been using for a number of years now - a templated design that has been tested/pentested/tested again for holes. And these days, it will likely be behind an AWS api gateway as well if we're talking about internet-facing apis. It won't be some EC2 instance with a public ip and a sec group of 80 -> 0.0.0.0/0.
I should have added in my original post that I feel AI tools can be very useful in the hands of trained and experienced developers. The CEO that vibe-codes some application over a weekend and then wants it pushed to prod at his company on Monday is not what I am talking about here. We have a group now at my work that has vibe-coded a very complex, very expensive to run app with an insanely complex DB structure that will be unmaintainable. I have already made my concerns known about that particular project, and it will not go to prod until the people that want to assume liability for it sign off. It is an MCP implementation that has security issues all through it. Yes, that shit isn't going to prod until it has been eval'd by competent engineers and/or someone wants to take the fall for it.
So, what I am saying, is that I agree with you. AI has it's place.
This "I don't understand what money is" ranting is rather sad to see.
There is no cap on money, because money represents only one thing. Trust in value. It has no value in of itself. This is why "gold standard" societies either gave it up or failed after hitting large growth spikes. Because if trust in value is tied to a specific metal, you need enough supply to expand the value of medium of exchange to keep up with societal trends. It's what killed Spanish Empire for example. They grew so fast, they couldn't mint enough coins to keep trade lubricated.
Same rule applies here. The money merely represents trust in value of whatever is being made. If what is being made is valuable, amount of money will expand to match that value so that trade remains possible for the entire value that is created and is in use.
And if AI goes as hard as it seems to be going in terms of value generation, we have a good chance of seeing value of things created by humanity go into quadrillions of today's dollars in our lifetimes.
If we were talking about durable goods like homes and swiss watches and true infrastructure, I would agree with you. But we're talking about GPUs and software that changes by the hour. The GPUs lose most of their value within 24 months, use enormous amounts of power, and the software is full of holes and bugs. The value generation isn't there, not just because of this, but because a lot of that hardware is being used to generate cat videos and youtube vids full of misinformation/propaganda.
So, No, I don't agree with your assessment.
People on slashdot who are luddites are hilarious to me. AI is the next stage of human evolution (as soon as we can integrate it into our brains), and yet they resist. Reminds me of how VR/AR is a logical step to cybernetics, and yet they resist. The real beta tests for that, ghost in the shell cybernetic utopia future, were google glass etc, but the beta testers were called glassholes, when all they were, were visionaries who were a few decades too early to a future that is coming.
No. These people were glassholes because the only thing they enabled was recording people for the purpose of large companies somehow monetizing it.
One is not a luddite for shunning shit tech.
Nondeterminism means never having to say you are wrong.