Comment Re:That's not very hard (Score 2) 39
Not sure if you have explored it very much but LM Studio and ComfyUI are huge ways you can run your own models. ComfyUI is built out a bit more for training and while it has more of focus on text-to-image workflows, you can configure it pretty well to work with LLMs.
Many of the models are quite light-weight and a decent gaming system can keep them in memory though quantization is generally needed a bit to help. Quanitization seems very necessary more with image models but in the case of code, you likely could get down a general token/syntax to significantly reduce the model's size. One of the new image models ERNIE uses a very interesting approach for this. ERNIE is basically the next generation of image-to-text models that focus on generating images that are text heavy (e.g. a poster or a flyer). I have only played with it a bit locally with a bit disappointing results but I think in part that's because I still don't fully understand it's architecture. Diffusion transformers work with a set of tokens from an LM, some input image, and then some parameters. However, some of these models like Flux2 require very detail oriented prompts to generate a good set of tokens for generation. ERNIE uses a "prompt enhancer" that takes a second LM, runs the initial simple prompt through it and generates an enhanced text prompt which is then passed to the main LM and used with the DiT.
There are tons of fun experiments in the community regarding this. Discussions about obliterating the filter that limit responses from an LM, about retraining DiTs to handle censored materials, removing the need for variational autoencoders (VAEs) by generating images in pixel space, etc.
I mean I think there is a lot to discuss and be concerned with about AI. However, I also think most of the things people are pushing about using AI are basically just the generic dogfood AI. There are some pretty cool communities out there really exploring interesting topics and they are mostly all OSS.