pull down to refresh

We need not your keys, not your coin for AI.
Not your runtime, not your slop?
100 sats \ 1 reply \ @roistrdn OP 1h
Same pattern, currently saving up for a server with local models, the only way to break the cycle
reply
Yes. I do all my "production" inference either locally or on an encrypted spot aws g4dn instance for large models (which was a headache to work out and i still think i should work on tuning it to get more juice for it - its very expensive)
I do test some of the commercial models at times but honestly the only one I've used that outperforms on coding is claude 3.7 sonnet (claude 4 regresses on coding for me) and not by enough margin to not use qwen3-coder. They both get into endless logic loops when dealing with complex code beyond the trivial in which thrir bad vibes created dumb bugs - very profitable for the provider too when youre paying or are capped per token.
reply