Hacker Newsnew | past | comments | ask | show | jobs | submit | technotony's commentslogin

Not only that but they seem to have cut my plan ability to use Sonnet too. I have a routine that used to use about 40% of my 5 hour max plan tokens, then since yesterday it gets stopped because it uses the whole 100%. Anyone else experience this?

yeah it seems like sonnet 4.6 burns thru tokens crazy fast. I did one prompt, sonnet misunderstood it as 'generate an image of this' and used all of my free tokens.

Maybe but having it search first to load the context with relevant information sure gets better results

It's the other way around. Cashier's spend their 4 percent, where's the lawyers probably save it. Though of course median salary for the two categories means 4 percent change is different in absolute dollars


Why? What's the incentive/value to commenting here with AI?


If you control a bunch of established accounts, you can use them to either shill for products, or upvote certain topics.


HN is a pretty influencial forum. Lots of tech journalists in mainstream media use it to get a pulse on what the SV/VC/Startup/BigTech crowd and adjacencies are talking about.


- Spam a product/service

- Generate age so spamming a product/service is easier and the account appears more trustworthy

- Influence discussions in a particular direction for monetary gain, i.e. "I got rich on bitcoin, you'd be crazy not to invest".

- Influence discussions in a particular direction for political gain, i.e. "I went to Xinjiang and the Uyghurs couldn't be happier!"


what are your impressions?


I got Codex CLI running against it and was sadly very unimpressed - it got stuck in a loop running "ls" for some reason when I asked it to create a new file.


You probably have seen it by now, but there was a llama.cpp issue that was fixed earlier today(?) to avoid looping and other sub-par results. Need to update llama-server as well as redownload the GGUFs (for certain quants).

https://old.reddit.com/r/unsloth/comments/1qvt6qy/qwen3coder...


I hadn't seen that, thanks very much!


Yes sadly that sometimes happens - the issue is Codex CLI / Claude Code were designed for GPT / Claude models specifically, so it'll be hard for OSS models directly to utilize the full spec / tools etc, and might get loops sometimes - I would maybe try the MXFP4_MOE quant to see if it helps, and maybe try Qwen CLI (was planning to make a guide for it as well)

I guess until we see the day OSS models truly utilize Codex / CC very well, then local models will really take off


I would recommend you fiddle with the repeat penalty flags. I use local models often, and almost all I've tried needed that to prevent loops.

I'd also recommend dropping temperature down to 0. Any high temperature value feels like instructing the model "copy this homework from me but don't make it obvious".


How would a USA civil be the best scenario globally? Who knows what wars that would trigger globally


It would also help with sites that can't handle the hacker News traffic load. Happens all the time


Interesting application. Can you share more about your stack and how you are approaching that build?


Checkout will happen directly in the app, and yes they will collect a fee on it.


Not true of Netflix, founder came from PayPal. Apple required founder to leave and learn with a bunch of other companies like Pixar and next.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: