by philk10 on 3/11/25, 1:41 PM with 18 comments
by simonw on 3/11/25, 3:29 PM
On macOS it's worth investigating the MLX ecosystem. The easiest way to do that right now is using LM Studio (free but proprietary), or you can run the MLX libraries directly in Python. I have a plugin for my LLM CLI tool that uses MLX here: https://simonwillison.net/2025/Feb/15/llm-mlx/
by filoeleven on 3/11/25, 4:22 PM
Upon getting a model up and running though, I quickly realized that I really have no idea what to use it for.
by tegiddrone on 3/11/25, 3:06 PM
by smjburton on 3/11/25, 4:44 PM
Anyone experimented with local LLMs and compared the output from ChatGPT or Claude? The article mentions that they use local LLMs when they're not overly concerned with the quality or response time, but what are some other limitations or differences to running these models locally?
by FloatArtifact on 3/11/25, 4:26 PM
End user experience should start by selecting the models of interest to run and output hardware builds with price tracking for components.
by adultSwim on 3/11/25, 4:26 PM
by zoogeny on 3/11/25, 9:29 PM
I'm wondering about multi-modal models, or generative models (like image diffusion models). For example, I was wondering about noise removal from audio files and how hard it would be to find open models that could be fine tuned for that purpose and how easy they would be to run locally.