Hacker Newsnew | past | comments | ask | show | jobs | submit | soiax's commentslogin

They should also check for hot dog vs not hot dog


Or, check there's no dog in the hot?


There it is... the usual ELON MUSK BAD.


Yeah that's false.

from: https://arcprize.org/blog/oai-o3-pub-breakthrough

"Note on "tuned": OpenAI shared they trained the o3 we tested on 75% of the Public Training set. They have not shared more details. We have not yet tested the ARC-untrained model to understand how much of the performance is due to ARC-AGI data."


This sound like you assume that the first thing someone thinks about is security, when building the next big thing.

They will just build something as fast as they can. Last thing you think about is "security".

There were prompt injections in all the big models, and still are. Why would it stop distruption?


The blog-poster is talking about long-term trends, so it doesn't matter if early-adopters skip on security, the time-horizon is long enough that the consequences will matter.

If we stop and carefully look at our world, security (safety against malicious peers) is an iceberg taken for granted. One might start by summing up the militaries of every country on earth. Add the budgets of most police departments, and a good chunk of the justice system. The energy, material, and labor poured into most weapons, fences, doors, and locks. The CPU cycles used in all encryption, and most of the hashing.

P.S.: "Investors, friends, I am pleased to announce that our bold and powerful new business-model which will completely disrupt the entire retail sector, worldwide, and change society forever. Behold! TTLMD: Take The Thing and Leave the Money in the Drawer! Existing industry dinosaurs will be unable to compete with our ultra-low-cost alternative which needs barely any staff."


You mentioned prompt injection, now when you talk about larger time horizons, that sounds like a AI alignment issue.

I'm sure there will be actors who don't care at all about "security", saying the positive outcomes outweight the negatives.


No, I'm still talking about prompt injection (and other more-normal reliability issues), because I do not believe LLMs are some inevitable stepping stone to an actual AI, one that has "alignment" to principles or goals beyond "what additional token completes this document the best." (Robot characters humans perceive when reading the document are not the author of the document.)

For any technology or product, there are issues which can be ignored or downplayed in the name of profit today, but they tend to pop up eventually. That's why it's very hard to buy leaded gasoline anymore, and the joke about how the "S" stands for "Security" in the term "IoT".


It's all renderer only RCE-s, no sandbox escape. So it doesn't work on your browser, only if you disable the sandbox.


> I then leverage this to achieve arbitrary memory read and write outside of the v8 heap sandbox, and in turn arbitrary code execution in the Chrome renderer process.

So the code is running in a process that runs as the same user running the browser. That's no longer much of a sandbox and you're now relying on the OS to protect your data, right?


No. There is a reason the author keeps repeating "arbitrary code execution in the Chrome renderer process." Because it's there, not in the browser process.


https://github.com/github/securitylab/tree/main/SecurityExpl...

> If successful, on Ubuntu 22.04, it should call launch xcalc when calc.html is opened in Chrome.

Then how does this work? It doesn't look like the provided build flags disable any sandbox that the distributed build doesn't.


You can disable it runtime, with --no-sandbox command line option.


No. You're relying on the OS's sandboxing features, which are much, much more granular than just "the same user running the browser."

https://chromium.googlesource.com/chromium/src/+/HEAD/docs/d...


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: