The Beep Boop is coming from inside the house!
"With a decent local GPU and some free open source software like ollama and open-webui you can try 'open source' LLM models like Meta's llama, Mistral AI's mistral, or Alibaba's qwen entirely offline." Lenie Clarke has a big ol' post on Homebrew LLMs and Open Source Models