Hacker Newsnew | past | comments | ask | show | jobs | submit | schneehertz's commentslogin

This price is high even because of the current shortage of inference cards available to DeepSeek; they claimed in their press release that once the Ascend 950 computing cards are launched in the second half of the year, the price of the Pro version will drop significantly

In six month deepseek won't be sota anymore und usage will be wayyyy down.

Only comparing on SOTA scores (ignoring price etc.) is like choosing your daily-driver by looking at who makes the fastest sports-car...

The constant improvements of SOTA are the main thing keeping the investment machine running. We can't really remove training costs from inference costs, because a bunch of the funding and loans for the inference hardware only exists because the promises the continuous training (tries to) provides.

Not really. SOTA vs non SOTA is "can I get my coding work actually done today" vs. "this can do customer support chat"

It is like car vs. kick scooter.


It really isn't. We get coding work actually done today on Opus 4.5. That's not SOTA any more, and anything proximate to that level, even quite loosely, is genuinely useful.

OK we are in Opus 4.5 is not SOTA. Right by that definition .... yes you are right.

I mean its almost halve a year, i think that counts ?

Time wise you are correct.

> "can I get my coding work actually done today" vs. "this can do customer support chat"

I think you need to define "can get coding work done" for this to make sense. Ive been using GPT-3 back-then for basic scripts, does that count ? Or only Claude-Code ?

I also think this is a false dichotomy, if you look at the Project Vend project or Vending-Bench, customer support etc. is at no means trivial. (Old but great story https://www.businessinsider.com/car-dealership-chevrolet-cha...)


This, I have been doing my side hustle code with open code an 3.2 reasoner and it is way better than what I have at day job with copilot and whatever models are there.

Copilot is a bad harness that perverts the productivity of models like GPT 5.5.

Tell me more please!

A huge proportion of those scores are gamed anyways. Use whatever works for you at the price and availability you can afford

Or there will be DSv4.1/2/3 ;)

Definitely something in this realm, they call the models "preview" at a bunch of different points in the paper.

What im really hoping is for a double-punch like with V3 -> R1


Well, if they distilled once…

This is not an official page; deepseek4.hk is not a domain owned by deepseek.


Generating a 4096x4096 image with gemini-3.1-flash-image-preview consumes 2,520 tokens, which is equivalent to $0.151 per image.

Generating a 3840x2160 image with gpt-image-2 consumes 13,342 tokens, which is equivalent to $0.4 per image.

This model is more than twice as expensive as Gemini.


this is apples to oranges, the flash version version a full version

this thing is like 5x better than flash at fine grain detail


Google's naming might be misleading, currently 3.1 flash image outperforms the available pro version (3.0 pro) on most benchmarks: https://deepmind.google/models/model-cards/gemini-3-1-flash-...

.40 cents for high quality output is insanely cheap

it is only going to get cheaper


> .40 cents

Warning: Verizon math ahead.


In case anyone is unfamiliar with one of the most infuriating phone calls of all time: https://www.youtube.com/watch?v=MShv_74FNWU

lol, noted thanks!

You people keep saying this and token prices keep doubling. The cope of the gambler is truly one to marvel.

Misleading conclusion.

This model is 8 times cheaper than Gemini for 1K images. Gemini is extremely overpriced.

1K image with Gemini is roughly $0.08 and only $0.01 with GPT Image.


In fact, you need to pay regardless of whether the output includes reasoning tokens or not

Vue provides a computed feature that acts as a buffer layer between the view and the state, so the view and the state are not necessarily strictly bound


Computed is only a concept because you need a band-aid for that lack of separation. You can read more about my efforts ditching Vue here:

https://blog.nestful.app/s/the-tech-behind-nestful


Wouldn't it be better to use verified medicine? Even if a patient could see a doctor at any time, the doctor would not prescribe such unverified peptides to the patient


Container ships + modular weapons are too crazy


No matter how large the library, it won't include WhatsApp's API


san check, 1d10


Oh, bro, you're practically living in 1689


That's the dream.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: