Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I found it far too expensive for Anthropic. Entire context of every conversation is sent each time you type anything. Switched to a local model running from Ollama. Not quite as smart as opus, but good enough for my needs.


Does it not use prompt caching?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: