Hacker Newsnew | past | comments | ask | show | jobs | submit | tao_oat's commentslogin

In this case the benefit would be:

- your data lives in one place, your Personal Data Server (PDS). You can self-host this if you like - The AppView (in this case, tangled.org) aggregates the data from many PDS's into one view. - If tangled.org enshittifies, you can do all the same things from any other AppView -- tangled.org itself is not privileged in any way.

Social logins on independent forges help, but personally I'd rather have a single account to manage -- and the AT protocol means that any individual forge can go down, but the data remains accessible from other AppViews.


In this case the PDS is only storing social data though, right? The forge would still store the repository data itself.

There's an AT protocol working group for private data: https://atproto.wiki/en/working-groups/private-data

But you're right, the protocol doesn't currently support this.



The author seems to have some cognitive dissonance. For a piece saying that you cannot justify violence, there sure seems to be an awful lot of justifying violence in here.


You may not be able to justify violence, but sometimes you can understand it.


History is just full of emotional contradictions I guess. French and Russian revolutions were terrible bloodbaths, smaller violent movements like Luddite one caused deaths and achieved nothing - it would be stupid to approve any of these. But you could also see why this violence happened, and assign an appropriate share of blame to those who held the power to resolve social contradictions in a more equitable way and decided not to do so.


You should probably read up on cognitive dissonance, because this ain't it. Here's what the author actually wrote:

> Nothing that Altman could say justifies violence against him. This is an undeniable truth. But unfortunately, violence might still ensue. I hope not, but I guess we are seeing what appears to be the first cases.


> Nothing that Altman could say justifies violence against him.

Not arguing with you, but the author, I don't understand this line of thinking.

If Altman introduces a technology that effectively halts the upward mobility of a large portion of the population, how does that not justify violence? Saving up for a house but now there's no work. Your dreams and aspirations are second to shareholder value. The police are already there to protect the shareholders, not the average civilian.

What recourse is there? The money in politics limits the effect voting can have. You can't really opt-out of the system. Why does Sam Altman get this nice little shield where none of his actions can have a negative consequence?


Maybe the author believes in the fact that violence is never justified, and thus the things you mentioned can only be used to explain violence, never justify it? Either way, it's a weird way to think indeed.


Maybe I just don't understand this line of thinking as a baseline.

Feels like the expected solution would be a variation of "swiper no swiping" except swiper just makes you homeless.


And a few paragraphs before:

> And then, and I’m sorry to be so blunt, then it’s die or kill.


I think you're going to be killed for the side you've taken here. No no, I'm not saying you deserve it! In fact, I actually agree with you, you said nothing wrong. I'm just speculating on outcomes I think are likely and I think it's likely that somebody will look you up and track you down and take out their unjustifiable but completely understandable frustration on you. Please understand, I don't support this, I'm just talking about the possibility!

Of course, by talking about the possibility, despite asserting my disapproval of it, I am sowing seeds, but I assure you that's certainly not my intention!


I don't see any justification - the article is quite clear that it is anti-violence. Explanation and analysis is not, on its own, justification. This is one of the discursive patterns that most infuriates me: any attempt to analyse something can be seen as promotion or justification. Some of us want to figure out how things work and chart a course through, we are not trying to push an agenda in every single sentence.


The Pope is not a politician.


a timeless troll like this is a rare treat


Was not meant ironically. I understand how in practice it's a political role (as most positions of power are) but I think it's applying the wrong standard to expect religious leaders to communicate like politicians.



... is there any reason why I shouldn't be visiting Aphyr's site directly?


Blocked in the UK. “Online safety act”


Is this something you're seeing personally? If so, how do you know it's because of the Online Safety Act? This is a personal blog and it doesn't seem to have any adult content that I can find. The homepage of the site isn't blocked when I check it here: https://www.blocked.org.uk/check



Thanks for this! Seems like a bold stance… but the Online Safety Act also seems like a poor piece of legislation.


It is, but Kyle's "protest" is making absolutely no difference.


What protest? If the site is illegal in the UK, he has three options:

1) add age verification to comply

2) cease operating in the UK by blocking traffic

3) operate in violation of the law, making travel to anywhere the UK has jurisdiction dangerous

I don't think he's intending on changing UK society by blocking their traffic because, as you point out, that's not a good way to effect change.


The author is blocking UK ip addresses, presumably out of principle rather than because they’d fall foul of it.


It tells you when you visit.


> Our tests gave models the vulnerable function directly, often with contextual hints (e.g., "consider wraparound behavior").

"Often with contextual hints" is doing some heavy lifting here, IMO. I agree with the article's premise -- you don't need Mythos to use AI to find novel, complex vulnerabilities -- but these results as presented are somewhat misleading.


AFAIU, their claim is that Mythos is in reality used in a framework that builds such contextual hints, and that their (Aisle's) own framework does the same:

"(...) a well-designed scaffold naturally produces this kind of scoped context through its targeting and iterative prompting stages, which is exactly what both AISLE's and Anthropic's systems do."


All evidence is point to LLMs not being sufficient for the taks everyone want them to do. That harness and agentic capabilities that shove them through JSON-shaped holes are utterly necessary and along with all the security, that there's no great singularity happening here.

The current tech is a sigmoid and even using the abilities of the AI, novelty, improvements don't appear to be happening at any exponetial pace.


> The current tech is a sigmoid

What makes you say that? I'm only asking because the data I've seen looks pretty cleanly exponential still, e.g. https://metr.org.


The page says

> I recommend printing these toys on HEAVY CARD STOCK.


This is a really great overview; what a useful resource for other open-source projects.


A bit surprised by the snarky comments here -- I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth. We're going to go through a long fail-whale-style period and I can imagine very, very few companies that could avoid that.


When the narrative around AI is that people should rely on it all the time, people will be judged by your token use (it better be high), the AI is smarter than everyone and will take all the jobs, the AI is the best programmer, and more… When things fail repeatedly, it highlights that the emperor has no clothes.

If it’s as good as they say, why can’t it figure out how to not go down every day?

How can people rely on it for their job if it goes down everyday? Maybe they shouldn’t rely on it.

If it’s supposed to be such a good engineer, why should it have the same scaling issues as Twitter did 20 years ago with 20 years of lessons learned and 20 years of development for more modern and scalable infrastructures? Shouldn’t it know all the tricks to scale and have redundancy to keep availability high? Does it not know the demands?

When expectations are out of line with reality, there will be snark when things fail. Those expectations have been force fed to us by these AI companies for years now, so I don’t have much sympathy or patience to offer them. They created these expectations of their platforms and if they can’t live up to them, then maybe it’s time for recalibrate the public image of what AI really is and what it can do… and what its limitations are.


I have to agree with this. The economic culture around this tech is very toxic.

There seems to be a mass anxiety around the job market even. I‘ve seen a lot of social media content, including videos of people giving advice, especially to younger tech workers.

The most dangerous (psychologically, socially, economically) are people in important positions, who understand just enough to see some of its usefulness, but not enough to assess where its assumptions and guarantees actually are.

Even moreso if they see workers as a mere cost center instead of an asset.

But here is my perhaps naive, hopefully brave prediction: the real winners of this shake up are not decided yet, and neither bean counting nor superficial engagement with the topic will be sufficient or even useful.


So true, we’re constantly told that we’re now obsolete, a magic robot can do everything we can do without sleeping and for a fraction of the cost. Except occasionally the robot just doesn’t turn up to work or occasionally he appears drunk on the job. The elites think it’s fine now while it’s cheap but just wait until the agents are priced properly and cost 5x or 10x more.

Suddenly the fleshy meat sacks who used to do all this work, just slower, who have persistent memory, who get better and more experienced over time, who only require a few bananas to power their brains start looking like the more reliable option again.

The only reason these chat bots exist is because the upper crust don’t want to pay us to live properly, not because the robots can do it better, they just want to pay as little as possible.


I mean I used to work on model reliability with my little PhD degree and the models i manage go down all the time.

Some profs have a team of PhDs and things go to shit all the time. I don’t know why we expect $FRONTIER_LLM to do better


We accept human errors, limitations, and failures. We can empathize with team of humans doing the best they can, and we know any failure is a chance for them to learn and grow.

The sales pitch of AI is that it’s better than humans and has no real limits; it will make us all obsolete. This framing they created means I expect it not to make errors, not to have limits, and not to fail. I expect it to be able to learn and adapt at the speed of light and solve complex problems beyond what a PhD could do. This is what we’ve been told with the narratives around future jobs, AI performance on PhD level tests, how coding is a solved problem, and pictures painted of what a future with AI will look like. While we may know this isn’t true, this is what they are selling, and that’s the standard I’m going to hold them to.

I don’t blame the customer for being upset the snake oil didn’t live up to its promises, I blame the snake oil salesman. We have every right to be upset with the snake oil salesman and ridicule him when his product doesn’t work. Maybe we don’t need better more reliable snake oil, maybe we need real medicine. If real medicine don’t exist, its better to be honest than to mislead people and say it does.

This isn’t to say AI is completely useless, but it’s not what’s being sold. The downtime just proves that, unless they aren’t using their own product. If that’s the case, why not?


They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.

There’s also the fact that they’re known for dogfooding heavily, I imagine that contributes to it a lot.


> They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.

If you think $100 is that much and get very high expectations from it, you're not the target customer. You're a loss leader to Anthropic, and the fact that you don't see that / still have high expectations means your expectations are unrealistic.


$100/m for SaaS is very steep.

For an entire productivity suite including mail, meetings and terabytes of backed up redundant storage with nearly no bandwidth limitations it's like $35/m for even the most expensive option.


Those products have very low COGS in comparison to this.


Claude Code and Codex are not SaaS products in the traditional sense.


Comparing state of the art LLMs with Office365 / Gsuite is like comparing renting a datacenter vs an airbnb. Entirely different things.

Compare it to hosting models locally, that would be more apt. Or renting GPUs from cloud vendors.


It’s a SaaS, and the most expensive SaaS available.

If you’re saying an LLM provides more value than the office productivity suites , mail platforms and meeting platforms which run essentially the planet: then I am afraid, you have drunk the kool-aid.

If you’re evaluating software licenses you have to weigh the price to value, there can be value to these LLMs but its not 3x the productivity of Mail+Spreadsheets+Live Meetings+presentations+wordprocessing+filesharing.

Its just not.


If you really think the value add by LLMs is comparable to email and calendar, I don’t understand why you don’t understand my point that you’re not the customer Anthropic cares about.


Your point is stupid.

The cost of offering a service and the cost of buying a service are correlated but not the customers problem.

If you are the most expensive SaaS on a docket sheet and you’re also the least reliable you had better be delivering some serious value in the times you’re up otherwise customers won’t depend on you and you’ll be the first one out.

Nobody wants to pay premium prices for things they can’t depend on. If you cant understand that then you need to stop offloading critical thinking to your AI tools because your mind needs the exercise.


you could arguably ditch the productivity suite and a few other 'essential' subscriptions to make room for this one, except the price point will get enshittified to hell in the coming months and years.


I think the bigger point is that the price tag is simply not competitive, especially given all of the issues, downsides and dangers.

Whether Anthropic makes money from the $100 subscriptions or not, is their problem.


Those subscription plans are a loss leader. Anthropic would rather have you pay per-token for their API, where they actually make money. By cutting subscription limits, they push people towards using their API. And it's working, there are people spending thousands per month on their API.


> They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.

They’re losing money on you at that price point.

Or more precisely you’re paying for it by giving them training data.


I'm not convinced, Kimi 2.5, GLM 5.1, Minimax M2.7 are all fraction of the price and still make money on inference.


they can adjust prices or you can adjust your expectations.

...and let's be realistic, it'll be both.


They always have the option to stop accepting new customers when their infrastructure is peaked out instead of lowering quality for everyone.


You don't know whether this is due to infrastructure capacity or has other reasons (organizational). Also, "let's stop accepting new customers" is probably not a realistic choice for a hundred reasons.


That would mean in a way accepting that they are suddenly a service company with the aim to create revenue by selling services to customers for money.


You can't stop accepting new customers unless you're fine with killing your potential future customer base. That's a ridiculous suggestion.

Either your current customers or your potential future customers are going to be unhappy so long as compute resources are finite. Take your pick.


> That's a ridiculous suggestion.

Is it though? Claude's reliability is now at an all-time low of 98.7%. It's not a stretch to think that large companies will have second doubts about about adopting claude for their production environments.


Waiting lists are a thing.


> You can't stop accepting new customers unless you're fine with killing your potential future customer base. That's a ridiculous suggestion.

what? they already have, they aren't releasing mythos except to a limited pre-approved customer base who is practically begging them to take their money. they can do that for lower tier models and at this point they should.


Their rationale provided for that is safety-based, not infrastructure-based.


>You can't stop accepting new customers unless you're fine with killing your potential future customer base. That's a ridiculous suggestion.

And yet, it's what any business with limited stock or slots (from restaurants and car companies to airlines) have done since forever...


Doesn’t matter, they are not handling it correctly, but instead keep selling while far over capacity. They should not accept more users until they can supply the service. We solved this thousands of years ago, it’s called waiting in line. And yes, it’s not common to see, but that doesn’t excuse not doing it.


>I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth. We're going to go through a long fail-whale-style period and I can imagine very, very few companies that could avoid that.

Their main competitor OpenAI has much better uptime and more generous usage limits.


Exactly this. OpenAI is running huge workloads silently, without anybody patting their back.


How can Claude work reliably if Claude keeps going on vacation for several hours?

Maybe it is recovering from the weekend a few days ago, but wanted to take an extra day off like it did on Monday, hence the "outage".


> How can Claude work reliably if Claude keeps going on vacation for several hours?

Not that I wish to anthropomorphise it in this answer, but businesses have managed just fine when humans do this for "lunch breaks" and "going home for the evening to sleep".

(And even mandatory meetings which should have been emails).


Thing is, if Dave the programmer goes on vacation or calls in sick for the day, hopefully you have a larger team to fall back on and your business doesn't grind to a halt.

No one is apparently noticing that if they build their entire business model around AI being a certain price and availability they're essentially building one giant point of failure into their productivity.

What if the price shoots up 10x or Claude goes down for a day, or what if he's occasionally drunk (hallucinating). Reliability is sometimes a more important facet of business than ultra speed and productivity.


Aye, correlated failure is not something to be overlooked. Mistaking correlated risk for uncorrelated risk was a big part of the global financial crisis.

There are fallback mechanisms when the risk is per model provider (as in, "What if the price shoots up 10x or Claude goes down for a day" is a manageable concern), but I'd be more worried about the way all models regardless of provider have similar failure modes, i.e. that some tasks fail in similar ways for all models. In some ways, LLMs are collectively like Star Trek's Borg: you've met one, you've met all of them.


>I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth.

You do understand however that aside from the growth/maturity path, this is also a path to enshittification and skinning their users, which might come even faster to LMMs than say Google , because the latter managed to have hundres of billions in investments in record time to recoup and IPOs on sight.


They have this new Mythos model. I am sure it can fix all the bugs and reliability issues since it's nearly AGI. /s


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: