Hacker Newsnew | past | comments | ask | show | jobs | submit | usef-'s commentslogin

If this were true, we're essentially saying that no one tried to scan vulnerabilities using existing models, despite vulnerabilities being extremely lucrative and a large professional industry. Vulnerability research has been one of the single most talked about risks of powerful AI so it wasn't exactly a novel concept, either.

If it is true that existing models can do this, it would imply that LLMs are being under marketed, not over marketed, since industry didn't think this was worth trying previously(?). Which I suspect is not the opinion of HN upvoters here.


I use the models to look for vulnerabilities all the time. I find stuff often. Have I tried to do build a new harness, or develop more sophisticated techniques? No. I suspect there are some spending lots of tokens developing more sophisticated strategies, in the same way software engineers are seeking magical one-shot harnesses.

...The absolute last thing I'd want to do is feed AI companies my proprietary codebase. Which is exactly what using these things to scan for vulns requires. You want to hand me the weights, and let me set up the hardware to run and serve the thing in my network boundary with no calling home to you? That'd be one thing. Literally handing you the family jewels? Hell no. Not with the non-existence of professional discretion demonstrated by the tech industry. No way, no how.

To be honest, this just sounds like a ploy to get their hands on more training data through fear. Not buying it, and they clearly ain't interested in selling in good faith either. So DoA from my point-of-view anyways.


I don’t think these companies are hurting for access to code.

The proverbial "50B" is investment in next year's model. The current model cost under "30B", and therefore "is profitable". It is a bet on scaling, yes, but that's been common throughout the industry (see, eg, Amazon not being profitable for many years but building infrastructure)

Also see the Dario interview with Dwarkesh:

> If every year we predict exactly what the demand is going to be, we’ll be profitable every year. Because spending 50% of your compute on research, roughly, plus a gross margin that’s higher than 50% and correct demand prediction leads to profit. That’s the profitable business model that I think is kind of there, but obscured by these building ahead and prediction errors.

(a lot more at the link)

https://www.dwarkesh.com/p/dario-amodei-2?open=false#%C2%A70...


Except the rumors are they subsidize even the inference, not that they have capex in training.

The maths shows inference is very profitable. Look at how Google/AWS/Azure change the same rates as Anthropic does for running Claude models.

You're missing the forest for the trees. Per-token pricing is irrelevant when you're just trying to get shit done. I pay 20 bucks a month for OpenAI, but I use likely $200+ a month of tokens just on the coding (and I'm just looking at the raw tokens, this is ignoring all the harnessing on their end). Even OpenAI has said that they're losing money on the 200-dollar subscriptions[1]. This is not a viable business model. Why do you think they are introducing ads this year[2]?

[1] https://fortune.com/2025/01/07/sam-altman-openai-chatgpt-pro...

[2] https://openai.com/index/testing-ads-in-chatgpt/


Maybe he's comparing the renting price of a bare metal server on its own, and doesn't realise how drastically cheaper they are to batch together for an API provider.

Btw, it doesn't need to be actively coordinated for this to happen.

Building architectural styles used to be per city and now buildings look roughly the same worldwide. Style is dependent on the year built not the location.

Because every architect is "reading the same magazine" worldwide now that the internet exists, rather than debating in their own city.

Similar monoculture of global thought is happening in all fields.


> Similar monoculture of global thought is happening in all fields.

Thereby removing yet more interesting things to see in the world through the spread of hyper-optimized inoffensive blandness. In the same way that restaurants are slowly turning into the same set of grey boxes with little of note distinguishing each.


> interesting things to see in the world

I mean, kinda the least of our worries in this thread, no? Restaurants and tourism??


I often wondered if this is timezone related. Those of us awake during the quieter times might see fewer issues?


I’m in Western Europe.


Where did they say this?


Not Windows: Operating systems. We did get more capable operating systems. The point of the quote is "this is the worst the SOTA will ever be".

If Windows XP were fully supported today I still wouldn't use it, personally, despite having respect for it in its era. The core technology of how, eg OS sandboxing, security, memory, driver etc stacks are implemented have vastly improved in newer OSes.


You're just moving the goal posts unfortunately. The point is that positive progress is never actually guaranteed.


Of course not. But I believe your Windows example was implying fundamental tech got worse.

The original "worst" quote is implying SOTA either stays the same (we keep using the same model) or gets better.

People have been predicting that progress will halt for many years now, just like the many years of Moore's law. By all indications AI labs are not running short of ideas yet (even judging purely by externally-visible papers being published and model releases this week).

We're not even throwing all of what is possible on current hardware technology at the issue (see the recent demonstration chips fabbed specifically for LLMs, rather than general purpose, doing 14k tokens/s). It's true that we may hit a fundamental limit with current architectures, but there's no indication that current architectures are at a limit yet.


Have you hit that? I thought it was only in extreme cases when Claude felt uncomfortable, like awful heavy psychological coercion. They wanted Claude not to be forced to reply endlessly.


I don't think that's an apt metaphor. You bought one general water supply, like an API user. If they sold a "no baths" cheaper option I'd be fine with them banning baths to those customers.

Google's API does let you use any client.

The gemini/antigravity clients are a different (subscription) service. When you reverse engineer the clients and use their internal auth/apis you will typically have very different access patterns to other clients (eg: not using prompt caching), and this is likely showing up in their metrics.

This isn't unusual. A bottomless drink at a restaurant has restrictions: it's for you to drink, not to pass around to others at the table (unless they buy one too). You can't pour it into bottles to take large quantities home, etc. And it's priced accordingly: if sharing/bottling was allowed the price would have to increase.


The irony of an ex-Google engineer coining Hyrum’s Law (https://www.hyrumslaw.com/)


There is a reality that when they control the client it can be significantly cheaper for them to run: the Claude code creator has mentioned that the client was carefully designed to maximise prompt caching. If you use a different client, your usage patterns can be different and it may cost them significantly more to serve you.

This isn't a sudden change, either: they were always up-front that subscriptions are for their own clients/apps, and API is for external clients. They don't document the internal client API/auth (people extracted it).

I think a more valid complaint might be "The API costs too much" if you prefer alternative clients. But all providers are quite short on compute at the moment from what I hear, and they're likely prioritising what they subsidise.


It reminds me of the net neutrality debate from a decade ago. I'm not American but I remember the discord and online hate towards Ajit Pai when they were repealing it.

On one side you had the argument that repealing net neutrality would mean you can save money on your internet bill by only paying for access to what you use. On the other, you had the argument that it would just enable companies to milk you for even more profit and throttle your connection as they see fit.

IMO we need 'net neutrality' for LLM clients. I feel like AI companies are hypocrites for talking about safety all the time, but want us to only use their LLMs in the way they intend. They're saying we're all going to be replaced by AI in 12 months, and we have to use their tools to survive, right?

Yann LeCun recently warned that the AI coming out of China is trending towards being more open than the American alternative. If it continues like this, I can see programmers being pushed towards Chinese models. Is that what the US government wants?


Use of Chinese models: If I had not got a discount for signing up for a full year of Gemini AI Pro for something like $14/month, I might have started just using a Chinese chat model for things where privacy is not an issue. Ironic that I am now paying for both Gemini AI Plus and also $20/month for Ollama Cloud (as a super easy way to experiment with many open models). I am also paying Proton $10/month to use their handy lumo+ private chat service built on Mistral models. I feel like I am spending too much money but I don’t want to feel locked into just a few vendors, and to be honest it is fun having alternatives. A year ago I used APIs for Chinese models (and Mistral in France) and the cost was really low.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: