i've switched from chatgpt to claude, i still occasionally mix it up with chatgpt, specially for non-coding tasks
i started using perplexity, both for research and occasional data collection for something I'm building
I use bolt.new for building frontends (occasionally I expose backend somewhere publicly so i can have live backend in bolt.new when it develops frontend)
cursor is my main ide now, tho i'll occasionally use copilot instead of their llm integration just to see how it performs
I run a tiny model locally for my terminal llm for quick "i forgot how this bash command should look exactly" lookups, I also have a cloud gpu that runs the same stack but better (slightly) gpu than my local instance so I can have the same setup when i'm away from my battlestation and just use ssh port forwarding to access ollama on my cloud gpu instance from my laptop
I use to have access to A100 so I'd run larger models locally and use open-webui, now i'm experimenting with tiny models like qwen2.5 0.5b or 3b
v2 of my reply from #348300
I run a tiny model locally for my terminal llm for quick "i forgot how this bash command should look exactly" lookups, I also have a cloud gpu that runs the same stack but better (slightly) gpu than my local instance so I can have the same setup when i'm away from my battlestation and just use ssh port forwarding to access ollama on my cloud gpu instance from my laptop
I use to have access to A100 so I'd run larger models locally and use open-webui, now i'm experimenting with tiny models like qwen2.5 0.5b or 3b