Hacker News with Generative AI: Local Development

Llama.cpp guide – Running LLMs locally on any hardware, from scratch (steelph0enix.github.io)
…and it’s pretty fun.
Ask HN: What is your local LLM setup? (ycombinator.com)
Ask HN: What is your local LLM setup?
Directly run and investigate Llama models locally (github.com/anordin95)
Run and explore Llama models locally with minimal dependencies on CPU
Build a quick Local code intelligence using Ollama with Rust (bosun.ai)
Now We Know What Local-First Means (and It's Not What You Think) (docnode.dev)
Ask HN: What do you use local LLMs for? (ycombinator.com)