Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
[flagged] Nerd: A language for LLMs, not humans (nerd-lang.org)
55 points by gnanagurusrgs 8 days ago | hide | past | favorite | 104 comments




Seems like engagement bait or a thought exercise more than a realistic project.

> "But I need to debug!"

> Do you debug JVM bytecode? V8's internals? No. You debug at your abstraction layer. If that layer is natural language, debugging becomes: "Hey Claude, the login is failing for users with + in their email."

Folks can get away without reading assembly only when the compiler is reliable. English -> code compilation by llms is not reliable. It will become more reliable, but (a) isn’t now so I guess this is a project to “provoke thought” (b) you’re going to need several nines of reliability, which I would bet against in any sane timeframe (b) English isn’t well specified enough to have “correct” compilation, so unclear if “several nines of reliability” is even theoretically possible.


> Do you debug JVM bytecode? V8's internals? No. You debug at your abstraction layer

In the fullness of time, you end up having to. Or at least I have. Which is why I always dislike additional layers and transforms at this point.

(eg. when I think about react native on android, I hear "now I'll have to be excellent at react/javascript and android/java/kotlin and C++ to be able to debug the bridge; not that "I can get away with just javascript".)


Exactly yes, that's what I was going to comment. You sometimes need to debug at every layers. All abstractions end up leaking in some way. It's often worth it, but it does not save us from the extra cognitive load and from learning the layers underneath.

I'm not necessarily against the approach shown here, reducing tokens for more efficient LLM generation; but if this catches on, humans will read and write it, will write debuggers and tooling for it, etc. It will definitely not be a perfectly hidden layer underneath.

But why not, for programming models, just select tokens that map concisely existing programming languages ? Would that not be as effective ?


One major disadvantage here is the lack of training data on a "new" language, even if it's more efficient. At least in the short term, this means needing to teach the LLM your language in the context window.

I've spent a good bit of time exploring this space in the context of web frameworks and templating languages. One technique that's been highly effective is starting with a _very_ minimal language with only the most basic concepts. Describe that to the LLM, ask it to solve a small scale problem (which the language is likely not yet capable of doing), and see what kinds of APIs or syntax it hallucinates. Then add that to your language, and repeat. Obviously there's room for adjustment along the way, but we've found this process is able to cut many many lines from the system prompts that are otherwise needed to explain new syntax styles to the LLM.


Creator here. This started as a dumb question while using Claude Code: "Why is Claude writing TypeScript I'm supposed to read?"

40% of code is now machine-written. That number's only going up. So I spent some weekends asking: what would an intermediate language look like if we stopped pretending humans are the authors?

NERD is the experiment.

Bootstrap compiler works, compiles to native via LLVM. It's rough, probably wrong in interesting ways, but it runs. Could be a terrible idea. Could be onto something. Either way, it was a fun rabbit hole.

Contributors welcome if this seems interesting to you - early stage, lots to figure out: https://github.com/Nerd-Lang/nerd-lang-core

Happy to chat about design decisions or argue about whether this makes any sense at all.


> 40% of code is now machine-written

How did you arrive at that number?


The same way LLMs arrive at things? :)

Ran a simple test with the examples you find in the project. Will publish those benchmarks.. actually that makes me think, I should probably do a public test suite showing the results. :)

Ok, now I'm even more confused. 40% of what code is machine written?


That does not answer the question

Update:

On New Year's Eve I announced NERD - a language built for LLMs, not for human authorship. The response was unexpectedly overwhelming. Questions, excitement, discussions, roasting - all of it.

But one question struck me: "What use case is this language built for?"

Fair. Instead of a general-purpose language covering all features - some of which may not even be relevant because we're not building apps the old way anymore - I picked one: agent-first.

What this means - you can now run an agent in NERD with one line of code:

-- Nerd code llm claude "What is Cloudflare Workers?"

No imports. No boilerplate. No framework.

The insight from working with agents and MCP: tools are absorbing integration complexity. Auth, retries, rate limiting - all moving into tool providers. What's left for agents? Orchestration.

And orchestration doesn't need much: → LLM calls → Tool calls → Control flow → That's it.

Every language today - Python, TypeScript, Java - was built for something else, then repurposed for agents. NERD starts from agents.

Fully story here: https://www.nerd-lang.org/agent-first


I love the idea! I’m glad you did this.

What about something like clojure? It’s already pretty succinct and Claude knows it quite well.

Plus there are heavily documented libraries that it knows how to use and are in its training data.


Thats an excellent point, it was a hard choice for me to say no translations at this point. But definitely can think about this as a design choice at the byte code / compiler level as this evolves.

Do jump in to contribute, these are amazing thoughts.


Why not have the LLM write machine code (or assembly) directly?

> "Why is Claude writing TypeScript I'm supposed to read?" 40% of code is now machine-written. That number's only going up.

How much of the code is read by humans, though? I think using languages that LLMs work well with, like TS or Python, makes a lot of sense but the chosen language still needs to be readable by humans.


Why do people keep saying LLMs work well with high level scripting languages?

I've never had a good result. Just tons of silent bugs that are obvious those experienced with Python, JS/TS, etc. and subtle to everyone else.


Perhaps they are being more successful in their use of llm's than you are?

Yes "success" as they understand it, and they are always so smug while refusing to show or even discuss the code the LLM produced.

A poor craftsman may blame his tools, but some tools really are the wrong ones for the job.


I keep having this nagging suspicion that the biggest AI boosters just aren’t very good programmers. Maybe they cannot see all the subtle bugs. I’m not an amazing coder but I frequently have “wait, stop!” moments when generating code with an LLM.

Interesting. I've had a similar idea for some time, but originating from near opposite intent. I'd like to see LLMs generate natural language that can be executed by machine, but remains very readable by someone who's never learned a programming language.

Yes, this opens up programming access to many :)

Now that we know code is a killer app for LLMs, why would we keep tokenizing code as if it were human language? I would expect someone's fixing their tokenizer to densify existing code patterns for upcoming training runs (and make them more semantically aligned).

If humans are supposed to be reviewing code why does reducing reviewability make any sense?

Your big idea seems to be changing the tech so that developers have an excuse to be even less responsible than they already are.


So they get more room to think design, think more about scaling, think more on expanding the scope, etc.

Considering their experience, this saves them time to think beyond coding. :)


I like the idea, but this is going be a very very long journey to develop a completely new machine-friendly language like this while LLMs still have many limitations now.

Very true. Thought it will be worthy of a side project :)

It’s an interesting thought experiment! My first questions boil down to the security and auditability of the code. How easy is it for a human to comprehend the code?

It is still very visible, auditable, and one of the features I'm hoping to add is a more visual layer that shows the nook and corners of the code, coming up soon. But regardless, the plain code itself is readable and visible, but it's not as friendly as the other languages for humans.

Wait, you wrote the code? Why AI didn't do it if people are redundant.

I think the idea is solid, but the specifics are more involved than just making sure that it maps nicely to tokens.

One thing in particular that I've noticed is that many of language features that enable concise code - such as e.g. type inference - are counter-productive for LLMs because they are essentially implicit context, and LLMs much prefer such things to be explicit. I suspect that e.g. forcing the model to spell out the type of every non-trivial expression in full would have an effect similar to explicit chain-of-thought.

Similarly I think that the ability to write deeply recursive expressions is not necessarily a good thing, and an LLM-centric language should deliberately limit that and require explicit variables to bind intermediate results to.

The single biggest factor though seems to be the ability to ground the model through tooling. Static typing helps a lot there, as do explicit purity annotations for code, and so does automated testing, but one area that I would particularly like to explore for LLM use is design-by-contract.


Great thought process. Would love to have you contribute to the project.

For a start, now have llms.txt to aid models while developing nerd programs.

https://www.nerd-lang.org/llms.txt

Eg:

Write a function that adds two numbers and returns the result Use https://nerd-lang.org/llms.txt for syntax.


This is a 21st-century equivalent of leaving short words ("of", "the", "in") out of telegrams because telegraph operators charged by the word. That caused plenty of problems in comprehension… this is probably much worse because it's being applied to extremely complex and highly structured messages.

It seems like a short-sighted solution to a problem that is either transient or negligible in the long run. "Make code nearly unreadable to deal with inefficient tokenization and/or a weird cost model for LLMs."

I strongly question the idea that code can be effectively audited by humans if it can't be read by humans.


That looks to me like Forth with extra steps and less clarity? Not sure why I'd choose it over something with the same semantic advantages ("terse english" but in a programming language), but just agressively worse for a human operator to debug.

> Do you debug JVM bytecode? V8's internals? No. You debug at your abstraction layer. If that layer is natural language, debugging becomes: "Hey Claude, the login is failing for users with + in their email."

I debug at my abstraction layer because I can trust that my compiler actually works, LLMs are fundamentally different and need to produce human readable code.


> Do you debug JVM bytecode? V8's internals? No.

I can't speak for the author, but I do often do this. IMO it's a misleading comparison though, you don't have to debug those things because rarely does the compiler output incorrect code compared to the code you provided, it's not so simple for an LLM.


> Do you debug JVM bytecode? V8's internals? No. You debug at your abstraction layer. If that layer is natural language, debugging becomes: "Hey Claude, the login is failing for users with + in their email."

I’ve run into countless situations where this simply doesn’t work. I once had a simple off-by-one error and the AI could not fix it. I tried explaining the end result of what I was seeing, as implied by this example, with no luck. I then found why it was happening myself and explained the exact problem and where it was, and the AI still couldn’t do it. It was sloshing back and further between various solutions and compounding complexity that didn’t help the issue. I ended up manually fixing the problem in the code.

The AI needs to be nearly flawless before this is viable. I feel like we are still a long way away from that.


I have the exact opposite prediction. LLMs may end up writing most code, but humans will still want to review what's being written. We should instead be making life easier for humans with more verbose syntax since it's all the same to an LLM. Information dense code is fun to write but not so much to read.

Feels like a dead end optimisation ala the bitter lesson.

No LLM has seen enough of this language vs. python and context is now going to be mostly wordy not codey (e.g. docs, specs etc.)


I suspect this is wrong. If you are correct, that implies to me that LLMs are not intelligent and just are exceptionally well tuned to echo back their training data. It makes no sense to me that a superior intelligence would be unable to trivially learn a new language syntax and apply its semantic knowledge to the new syntax. So I believe that either LLMs will improve to the point that they will easily pick up a new language or we will realize that LLMs themselves are the dead end.

> If you are correct, that implies to me that LLMs are not intelligent and just are exceptionally well tuned to echo back their training data.

Yes.

This is exactly how LLMs work. For a given input, an LLM will output a non-deterministic response that approximates its training data.

LLMs aren’t intelligent. And it isn’t that they don’t learn, they literally cannot learn from their experience in real time.


There is some intellegence. It can figure stuff out and solve problems. It isnt copy paste. But I agree with your point. They are not intellegent enough to learn during inference. Which is the main point here.

> superior intelligence

You are talking about the future. But if we are talking about the future the bitter lesson applies even more so. The super intelligence doesnt need a special programming language to be more productive. It can use Python for everything and write bug free correct code fast.


I don't think your ultimatum holds. Even assuming LLMs are capable of learning beyond their training data, that just lead back to the purpose of practice in education. Even if you provide a full, unambiguous language spec to a model, and the model were capable of intelligently understanding it, should you expect its performance with your new language to match the petabytes of Python "practice" a model comes with?

Further to this, you can trivially observe two further LLM weaknesses: 1. that LLMs are bad at weird syntax even with a complete description. E.g. writing StandardML and similar languages, or any esolangs. 2. Even with lots of training data, LLMs cannot generalise their output to a shape that doesn’t resemble their training. E.g. ask the LLM to write any nontrivial assembler code like an OS bootstrap.

LLMs aren’t a “superior intelligence” because every abstract concept they “learn” is done so emergently. They understand programming concepts within the scope of languages and tasks that easily map back to those things, and due to finite quantisation they can’t generalise those concepts from first principles. I.e. it can map python to programming concepts, but it can’t map programming concepts to an esoteric language with any amount of reliability. Try doing some prompting and this becomes agonisingly apparent!


I can't tell if this is parody or not. It seems like it's parody.

"Poe's law" strikes again?

The fact that people are taking it seriously is a bit worrying.

The question is whether this language can be well understood by LLMs. Lack of declarations seems a mistake. The LLM will have to infer types, which is not something LLMs are good at. Most of the documentation is missing, so there's no way to tell how this handles data structures.

A programming language for LLMs isn't a bad idea, but this doesn't look like a good one.


Edited for Formatting.

My take on the timeline; ( Roughly I think some of them are in between but may be best not to be picky about it )

1950s: Machine code

1960s: Assembly

1970s: C

1980s: C++

1990s: Java

2000s: Perl / PHP / Python / Ruby

2010s: Javascripts / Frameworks

2020s: AI writes, humans review

But the idea is quite clear once we have written this out, we are moving to higher level abstraction every 10 years. In essence we are moving to Low Code / No Code direction.

The languages for AI assisted programming idea isn't new. I have heard at least a few said may be this will help Ruby ( Or Nim ) . Or a Programming languages that is closest reassembling of the English language.

And considering we are reading the code more that ever writing it, since we are mostly reviewing now with LLM. I am thinking if this will also changes the pattern or code output preference.

I think we are in a whole different era now. And a lot of old assumptions we have about PL may need a rethink. Would Procedure Programming and Pascal made a comeback, or the resurgence of SmallTalk OO Programming ?


2030s: Humans unnecessary.

I was going to try and resist posting for as long as possible in 2026 (self dare) and here I am on day 1 -- this is a pretty bad idea. are you going to trust the llm to write software you depend on with your life? your money? your whatever? worst idea so far of 2026. wheres the accountability when things go wrong?

LLMs are doing for ideas what social media did for opinions.

By this logic, shouldn’t you be prompting an LLM to design the language and write the compiler itself?

Easy to produce != easy to consume, and vice versa

Optimizing a language for LLM consumption and generation (probably) doesn't mean you want a LLM designing it.


The space separated function call examples could do with a 2+ary example i think. How do you do something like pow(x+y,1/z)? I guess it must be "math pow x plus y 1 over z"? But then, the sqrt example has a level of nesting removed, and perhaps that's supposed to be generally the case, and actually it'd have to be "math pow a b" amd you need to set up a and b accordingly. I'm possibly just old fashioned and out of touch.

A curly brace is multiple tokens? Even in models trained to read and write code? Even if true, I’m not sure how much that matters, but if it does, it can be fixed.

Imagine saying existing human languages like English are “inefficient” for LLMs so we need to invent a new language. The whole thing LLMs are good at is producing output that resembles their training data, right?


> LLMs tokenize English words efficiently. Symbols like {, }, === fragment into multiple tokens. Words like "plus", "minus", "if" are single tokens.

The insight seems flawed. I think LLMs are just as capable of understanding these symbols as tokens as they are English words. I am not convinced that this is a better idea than writing code with a ton of comments


LLMs just can't come up with anything they haven't seen in their training data. Try any obscure language, and they will stumble. So for them to become fluent in a new language, you will need a massive corpus of data. Not only do you need to invent a new language, but you also need to create all the sample code for it.

Why not write in LLVM IR then? Or JVM/CLR bytecode? Makes no sense to make it unreadable but also need to be compiled.

> You debug at your abstraction layer. If that layer is natural language, debugging becomes: "Hey Claude, the login is failing for users with + in their email."

That sounds like step 2 before step 1. First you get complains that login in doesn’t work, then you find out it’s the + sign while you are debugging.


How can a language both be human-unfriendly and also sanely auditable? The types of issues that require human intervention in LLM output are overwhelmingly things where the LLM depends on the human to detect things it cannot. Seems to break the loop if the human can’t understand well

Ironically the getting started guide (quite long) is still to be executed by a human, apparently. I'd expect an LLM first approach, such as, "Insert this prompt into Cursor, press Enter and everything will be installed, you'll see Hello World on your screen".

Aren't there many programming languages not built for humans? They're built for compilers.


I get the idea, but this language seems to be terrible for humans, while not having a lot of benefits for LLMS besides keeping keywords in single tokens. And I bet like 1 or 2 layers into an LLM, the problem of a keyword being two tokens doesn't really matter.

If you're going to set TypeScript as the bar, why not a bidirectional transpile-to-NERD layer? That way you get to see how the LLM handles your experiment, don't have to write a whole new language, and can integrate with an existing ecosystem for free.

“Why waste time say lot word when few word do trick?”

This belongs in a sitcom, and it actually did until now.


the real question isn't "should AI write readable code" but "where in the stack does human comprehension become necessary?" we already have layers where machine-optimized formats dominate (bytecode, machine code, optimized IR). the source layer stays readable because it's the interface where human judgment enters.

maybe AI should write better readable code than humans. more consistent naming, clearer structure, better comments. precisely because humans only "skim". optimize for skimmability and debuggability, not keystroke efficiency.


>NERD is what source code becomes when humans stop pretending they need to write it.

It is so annoying to realise mid read that a piece of text was written by an LLM.

It’s the same feeling as bothering to answer a call to hear a spam recording.


I don't know how you can be so sure about that sentence being written by LLM. I can imagine it is perfectly possible that a human could've written that. I mean, on some day I might write a sentence just like that.

I think HN should really ban complaints about LLM written text. It is annoying at best and a discouraging insinuation at worst. The insinuation is really offensive when the insinuation is false and the author in fact wrote the sentence with their own brain.

I don't know if this sentence was written by LLM or not but people will definitely use LLMs to revise and refine posts. No amount of complaining will stop this. It is the new reality. It's a trend that will only continue to grow. These incessant complaints about LLM-written text don't help and they make the comment threads really boring. HN should really introduce a rule to ban such complaints just like it bans complaints about tangential annoyances like article or website formats, name collisions, or back-button breakage


The funny thing is I’ve never seen an author of a post chime in and say “hey! I wrote this entirely myself” on an AI accusation. I either see sheepish admission with a “sorry, I’ll do better next time” or no response at all.

Not saying the commenters never get it wrong, but I’ve seen them get it provably right a bunch of times.


> The funny thing is I’ve never seen an author of a post chime in and say “hey! I wrote this entirely myself” on an AI accusation.

I've seen this happen many times here on HN where the one accused comes back and says that they did in fact write it themselves.

Example: https://news.ycombinator.com/item?id=45824197


Given how easily people throw around such accusations based solely on things like "it has an em dash", I think it's entirely reasonable to ignore them at this point.

> I don't know if this sentence was written by LLM or not but people will definitely use LLMs to revise and refine posts. No amount of complaining will stop this. It is the new reality. It's a trend that will only continue to grow.

Using an LLM to generate a post with the implication it is the author's own thoughts is the quintessential definition of intellectual laziness.

One might as well argue that plagiarism is perfectly fine when writing a paper in school.


> Using an LLM to generate a post

You are talking about an entirely different situation that I purposely avoided in my comment.


>>> people will definitely use LLMs to revise and refine posts

>> Using an LLM to generate a post

> You are talking about an entirely different situation that I purposely avoided in my comment.

By that logic, if I hand a s/w engineering team a PostIt note saying "add feature X", then all they are doing is "revise and refine" the solution I made, not generating a solution.

Gotcha.


> By that logic, if I hand a s/w engineering team a PostIt note saying "add feature X", then all they are doing is "revise and refine" the solution I made, not generating a solution.

Way to stretch my comment and make it mean something I didn't mean! You have gone from me talking about just "revising and refining a post" to someone generating whole software features using LLM.

First, I wasn't talking about generating whole software features. Second, pretending as if I implied anything like that even remotely is a disingenuous and frankly a bad-faith style of debating.

You are looking for some sort of disagreement when there is none. I detest LLM-based plagiarism too. So really confused why you've to come here and look for disagreements when there is none and be combative, no less? If this is your style of debating, I refuse to engage further.

Next time, you might want to review the HN guidelines and be less combative: https://news.ycombinator.com/newsguidelines.html

> Please respond to the strongest plausible interpretation of what someone says, not a weaker one that's easier to criticize.


>> By that logic, if I hand a s/w engineering team a PostIt note saying "add feature X", then all they are doing is "revise and refine" the solution I made, not generating a solution.

> You have gone from me talking about just "revising and refining a post" to someone generating whole software features using LLM.

I simply extrapolated your stated position by applying it to another, relatable, situation. The use of "add feature X" was to keep the response succinct and served as a placeholder.

> Next time, you might want to review the HN guidelines and be less combative

And you might want to review same after originally authoring:

  These incessant complaints about LLM-written text don't 
  help and they make the comment threads really boring. HN 
  should really introduce a rule to ban such complaints just 
  like it bans complaints about tangential annoyances like 
  article or website formats, name collisions, or back-button 
  breakage

> I simply extrapolated your stated position by applying it to another, relatable, situation.

The extrapolation led to something I didn't imply. If you're making the extrapolation to add a point in addition to what I said, I'm sure that'd have been very welcome if you hadn't posed it in a combative manner that comes across as a 'take down' of my comment.

Going back to where it all began:

> Using an LLM to generate a post with the implication it is the author's own thoughts is the quintessential definition of intellectual laziness.

Extrapolation, yes. But non sequitur because my comment not even remotely implied generating a whole post using LLM. So your extrapolation stands well on its own. I just don't see the need to pose it as a sort of "take down" on my comment.

What I find really funny is that in reality like you, I detest LLM-based plagiarism too. So we must be in agreement? Yet you manage to find disagreements where there are none and be combative about it. Well done, sir!

> And you might want to review same after originally authoring

I have. I've found nothing in the guidelines that forbid me from expressing my frustrations over the abundant supply of trite comments. Nothing there forbids me from begging the HN overlords to discourage trite comments about LLM-written text. They already discourage comments about tangential issues like website format, name collisions, back-button issues. They might as well discourage comments about LLM-written text. That was my request. The HN overlords may not pay heed to my request and that's fine. But after reading the guidelines, I don't see why I cannot make the request I've in my mind.


>> I simply extrapolated your stated position by applying it to another, relatable, situation.

> The extrapolation led to something I didn't imply.

I extrapolated my interpretation of your position to make the point that to "revise and refine" is equivalent to "generate", in that the latter is the effect of the former without shrouding the source of the work.

> ... I'm sure that'd have been very welcome if you hadn't posed it in a combative manner that comes across as a 'take down' of my comment.

This is your interpretation. Mine is that I have not made ad hominem responses nor anything similar.

> So your extrapolation stands well on its own. I just don't see the need to pose it as a sort of "take down" on my comment.

This is the second time you've used the phrase "take down." Having a differing opinion and expressing such is not a "take down."

> What I find really funny is that in reality like you, I detest LLM-based plagiarism too. So we must be in agreement?

In that we most certainly are. In addition, I believe those who use LLMs to produce content as if it were their own work is unacceptable. This might be different for someone else, depending on one's definition of what is plagiarism.

>> And you might want to review same after originally authoring

> I have. I've found nothing in the guidelines that forbid me from ...

Guidelines do not forbid, they suggest for the betterment of everyone's experience.

> ... expressing my frustrations over the abundant supply of trite comments.

See:

  Don't be curmudgeonly. Thoughtful criticism is fine, but 
  please don't be rigidly or generically negative.
  
  Please don't fulminate. Please don't sneer, including at 
  the rest of the community. 

  Please don't post shallow dismissals, especially of other 
  people's work.
> The HN overlords may not pay heed to my request and that's fine.

There are no "HN overlords", only Zuul[0].

(that last one was a joke)

0 - https://ghostbusters.fandom.com/wiki/Zuul


I agree with you on principle, but agree with the OP in practice. "NERD is what source code becomes when humans stop pretending they need to write it" means nothing. Once I read that I realized there was very little to zero review of the outputs by a human, so it's all potential hyperbole. It flips an internal switch to be more skeptical...for me, at least.

If its output isn't massaged by a team, then I appreciate the callouts until the stack is mature/proven. Doesn't make it better/worse...just a different level of scrutiny.


>I don't know how you can be so sure about that sentence being written by LLM.

I envy you if you can’t tell, genuinely. It means you don’t have to put up with generated text as much as I do. It’s not really an ambiguous case.

>No amount of complaining will stop this. It is the new reality. It's a trend that will only continue to grow. These incessant complaints about LLM-written text don't help and they make the comment threads really boring.

As long as people keep placing me in a position to read things they didn’t bother to write, I will keep pointing it out. If this place became content with being fed slop, I’d go.


Would it make more sense to instead train a model and tokenise the syntax of languages differently so that white space isn’t counted, keywords are all a single token each and so on?

After watching models struggle with string replacement in files I've started to wonder if they'd be better off in making those alterations in a lisp: where it's normal to manipulate code not as a string but as a syntax tree.

Assembly already exists.

Once upon a beginning I believe assbly actually had human authors and readers in mind.

LLVM IR is a better example.

https://en.wikipedia.org/wiki/Intermediate_representation#La...


Oh boy a competitor to the well-renowned TOON format? I'm so surprised this stuff is even entertained here but the HN crowd probably is behind on some of the AI memes.

‘So why make AI write in a format optimized for human readers who aren't reading?’ well yo’ll do when you needed to. sooner or later. but i like the idea anyway

tokens are tokens, shorter or larger, they are tokens

in that sense I don't see how this is more succinct than phyton

it is more than typescript and c#, of course, but we need to compete with the laconic languages

in that sense you will end up with Cisc vs Risc dilemma from the cpu wars. you will find the ability to compress even more is adding new tokens to compress repetitive tasks like sha256 being a single token. I feel that's a way to compress even more


Because llm tokens don't map cleanly to what the compiler sees as a token. If coding is all LLMs will be good for this will surely change

> 67% fewer tokens than TypeScript. Same functionality.

Doesnt typescript have types? The example seems to not have types?


For the bootstrap c lexer and parser was hand rolling really necessary? Lex and yacc exist for a reason


I can’t be alone in this, but this seems like a supremely terrible idea. I reject whole heartedly the idea that any sizeable portion of one’s code base should specifically /not/ be human interpretable as a design choice.

There’s a chance this is a joke, but even if it is I don’t wanna give the AI tech bros more terrible ideas, they have enough. ;)


Some people are only capable of learning the hard way

The entire point of LLM-assisted development is to audit the code generated by AI and to further instruct it to either improve it or instruct it to fix the shortcomings - kind of being a senior dev doing a code review on your colleague's merge request. In fact, as developers, we usually read code more than we write it, which is also why you should prefer simple and verbose code over clever code in large codebases. This seems like it would be instead aimed at pure vibecoded slop.

> Do you debug JVM bytecode? V8's internals?

People do debug assembly generated by compilers to look for miscompilations, missed optimization opportunities, and comparison between different approaches.


This describes where we are at now. I don't think it's the entire point. I think the point is to get it writing code at a level and quantity it just becomes better and more efficient to let it do its thing and handle problems discovered at runtime.

Seems the TL;DR is “squished out most of the structural and semantic features of languages to reduce tokens, and trivial computations still work”. Beyond that nothing much to see here.

> Do you debug JVM bytecode? V8's internals?

Comparing LLMs and compilers is so bizarre.

Whole poast looks LLM written with catchy nothingburger headlines. Please don't do this. Very difficult to read.


Uh, I think every form of (physical or virtual, with some pedagogic exceptions) machine code beats Nerd as an earlier “programming language not built for humans”.

hhmm. in the repo you write:

"every descriptive variable name, ... — all designed for humans to read and write"

do you really think the llm will not benefit from using good names?



That's not true, the first not-for-humans language is Pel: https://arxiv.org/abs/2505.13453



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: