pull down to refresh

v2 of my reply from #348300
  • i've switched from chatgpt to claude, i still occasionally mix it up with chatgpt, specially for non-coding tasks
  • i started using perplexity, both for research and occasional data collection for something I'm building
  • I use bolt.new for building frontends (occasionally I expose backend somewhere publicly so i can have live backend in bolt.new when it develops frontend)
  • cursor is my main ide now, tho i'll occasionally use copilot instead of their llm integration just to see how it performs
I run a tiny model locally for my terminal llm for quick "i forgot how this bash command should look exactly" lookups, I also have a cloud gpu that runs the same stack but better (slightly) gpu than my local instance so I can have the same setup when i'm away from my battlestation and just use ssh port forwarding to access ollama on my cloud gpu instance from my laptop
I use to have access to A100 so I'd run larger models locally and use open-webui, now i'm experimenting with tiny models like qwen2.5 0.5b or 3b
interesting, will check out bolt