Hacker Newsnew | past | comments | ask | show | jobs | submit | jey's commentslogin

(2014)


have you tried vim or neovim?

I don't think the point was to say "look, AI can just take care of writing a browser now". I think it was to show just how far the tools have come. It's not meant to be production quality, it's meant to be an impressive demo of the state of AI coding. Showing how far it can be taken without completely falling over.

EDIT: I retract my claim. I didn't realize this had servo as a dependency.


This is entirely too charitable. Basically all this proves is that the agent could run in a loop for a week or so, did anyone doubt that?

They marketed as if we were really close to having agents that could build a browser on their own. They rightly deserve the blowback.

This is an issue that is very important because of how much money is being thrown at it, and that effects everyone, not just the "stakeholders". At some point if it does become true that you can ask an agent to build a browser and it actually does, that is very significant.

At this point in time I personally can't predict whether that will happen or not, but the consequences of it happening seem pretty drastic.


> This is entirely too charitable. Basically all this proves is that the agent could run in a loop for a week or so, did anyone doubt that?

yes, every AI skeptic publicly doubted that right up until they started doing it.


I find it hard to believe after running agents fully autonomously for a week you'd end up with something that actually compiles and at least somewhat functions.

And I'm an optimist, not one of the AI skeptics heavily present on HN.

From the post it sounds like the author would also doubt this when he talks about "glorified autocomplete and refactoring assistants".


You don't run coding agents for a week and THEN compile their code. The best available models would have no chance of that working - you're effectively asking them to one-shot a million lines of code with not a single mistake.

You have the agents compile the code every single step of the way, which is what this project did.


With the agent running autonomously for a long time, I'd have feared it would break my build/verification tasks in an attempt to fix something.

My confidence in running an agent unsupervised for a long time is low, but to be fair that's not something I tried. I worked mostly with the agent in the foreground, at most I had two agents running at once in Antigravity.


It did not compile [1], so your belief was correct.

[1] https://news.ycombinator.com/item?id=46649046


It did compile - the coding agents were compiling it constantly.

It didn't have correctly configured GitHub Actions so the CI build was broken.


Then you should have no difficulty providing evidence for your claim. Since you have been engaging in language lawyering in this thread, it is only fair your evidence be held up to the same standard and must be incontrovertible evidence for your claims with zero wiggle room.

Even though I have no burden of proof to debunk your claims as you have provided no evidence for your claims, I will point out that another commenter [1] indicates there were build errors. And the developer agrees there were build errors [2] that they resolved.

[1] https://news.ycombinator.com/item?id=46627675

[2] https://news.ycombinator.com/item?id=46650998


I mean I interviewed the engineer for 47 minutes and asked him about this and many other things directly. I think I've done enough homework on this one.

I take back the implication I inadvertently made here that it compiled cleanly the whole time - I know that's not the case, we discussed that in our interview: https://simonwillison.net/2026/Jan/23/fastrender/#intermitte...

I'm frustrated at how many people are carrying around a mental model that the project "didn't even compile" implying the code had never successfully compiled, which clearly isn't true.


Okay, so the evidence you are presenting is that the entity pushing intentionally deceptive marketing with a direct conflict of interest said they were not lying.

I am frustrated at people loudly and proudly "releasing" a system they claim works when it does not. They could have pointed at a specific version that worked, but chose not to indicating they are either intentionally deceptive or clueless. Arguing they had no opportunity for nuance and thus had no choice but to make false statements for their own benefit is ethical bankruptcy. If they had no opportunity for nuance, then they could make a statement that errs against their benefit; that is ethical behavior.


See my comment here: https://news.ycombinator.com/context?id=46771405

I do not think Cursor's statements about this project were remotely misleading enough to justify this backlash.

Which of those things would you classify as "false statements"? The use of "from scratch"?


> Arguing they had no opportunity for nuance and thus had no choice but to make false statements for their own benefit is ethical bankruptcy.

absolutely

and clueless managers seeing these headlines will almost certainly lead to people losing their jobs


That is a good point. It is impressive. Llms from two years ago were impressive, llms a year ago were impressive, and from a month ago even more impressive.

Still, getting "something" to compile after a week of work is very different from getting the thing you wanted.

What is being sold, and invested in, is the promise that LLMs can accomplish "large things" unaided.

But they can't, as of yet, they cannot, unless something is happening in one of the SOTA labs that we don't know about.

They can however accomplish small things unaided. However there is an upper bound, at least functionally.

I just wish everyone was on the same page about their abilities and their limitations.

To me they understand conext well (e.g. the task, build a browser doesn't need some huge specification because specifications already exist).

They can write code competently (this is my experience anyway)

They can accomplish small tasks (my experience again, "small" is a really loose definition I know)

They cannot understand context that doesn't exist (they can't magically know what you mean, but they can bring to bear considerable knowledge of pre-existing work and conventions that helps them make good assumptions and the agentic loop prompts them to ask for clarification when needed)

They cannot accomplish large tasks (again my experience)

It seems to me there is something akin to the context window into which a task can fit. They have this compact feature which I suspect is where this limitation lies. Ie a person can't hold an entire browser codebase in their head, but they can create a general top level mapping of the whole thing so they can know where to reach, where areas of improvement are necessary, how things fit together and what has been and what hasn't been implemented. I suspect this compaction doesn't work super well for agents because it is a best effort tacked on feature.

I say all this speculatively, and I am genuinely interested in whether this next level of capability is possible. To me it could go either way.


Maybe so, but I don't think 3 million lines of code to ultimately call `servo.render()` is a great way to demonstrate how good AI coding is.


`servo.render()` does not appear to exist in the code base. Would you please point it out to us?


lmao okay, touché. I did not realize it had servo as a dependency.


Yeah, but starting with a codebase that is (at least approaching) production quality and then mangling it into something that's very far from production quality... isn't very impressive.


It didn't have Servo as a dependency.

Take a look in the Cargo.toml: https://github.com/wilsonzlin/fastrender/blob/19bf1036105d4e...


I haven't really looked at the fastrender project to say how much of a browser it implements itself, but it does depend on at least one servo crate: cssparser (https://github.com/servo/rust-cssparser).

Maybe there is a main servo crate as well out there, and fastrender doesn't depend on that crate, but at least in my mind fastrender depends on some servo browser functionality.

EDIT: fastrender also includes the servo HTML parser: html5ever (https://github.com/servo/html5ever).


Yes, it depends on cssparser and html5ever from Servo, and also uses Taffy which is a dependency shared with Servo.

I do not think that makes it a "Servo wrapper", because calling it that implies it has no rendering code of its own.

It has plenty of rendering code of its own, that's why the rendered pages are slow and have visual glitches you wouldn't get with Sero!


> I think it was to show just how far the tools have come.

In… terms of sheer volume of production of useless crap?


Dan Romik has a nice intro on the moving sofa problem: https://www.math.ucdavis.edu/~romik/movingsofa/


Yikes. When you’re taking the window frame out to get furniture in, I’m arguing for smaller furniture.


Yeah maybe time to send a sample to the “mass spec everything” guy



Here's a less editorialized article from Reuters: https://www.reuters.com/business/finance/banks-tap-record-li...


Thanks. Here's an alternate Reuters link: http://archive.today/GWDYr


Thank you. From that link: "Borrowing from the Fed at year-end is also tied to market forces, where an upward drift in money market rates can make it cheaper to borrow from the Fed compared to private sources. Most expect Wednesday's borrowing surge will dissipate over coming days as more normal trading conditions reassert themselves. The activity at the standing repo operation is highly unlikely to signal any sort of market trouble."


The breathlessly reported OP credits their man with scouring the halls of the Fed for a report that they post on their website every single day, which literally everyone on Wall Street knows how to find.

https://www.newyorkfed.org/markets/desk-operations/repo


That makes sense, but how do you efficiently evaluate the Gaussian kernel based approach (“operator-based data structures (OBDS)”)? Presumably you want to do it in a way that keeps a dynamically updating data structure instead of computing a low rank approximation to the kernel etc? In my understanding the upside of the kNN based approaches are fast querying and ability to dynamically insert additional vectors..?


Thank you for the thoughtful comment. Your questions are valid given the title, which I used to make the post more accessible to a general HN audience. To clarify: the core distinction here is not kernelization vs kNN, but field evaluation vs point selection (or selection vs superposition as retrieval semantics). The kernel is just a concrete example.

FAISS implements selection (argmax ⟨q,v⟩), so vectors are discrete atoms and deletion must be structural. The weighted formulation represents a field: vectors act as sources whose influence superposes into a potential. Retrieval evaluates that field (or follows its gradient), not a point identity. In this regime, deletion is algebraic (append -v for cancellation), evaluation is sparse/local, and no index rebuild is required.

The paper goes into this in more detail.


> When I ran this program, I expected the `CF_OEM­TEXT` string to have the byte 44, but it didn’t. It had the byte 90. We will start unraveling this mystery next time.

Whoa there exists something Raymond Chen didn’t know about Windows core APIs?


This seems to be incorporated into current LLM generations already -- when code execution is enabled both GPT-5.x and Claude 4.x automatically seem to execute Python code to help with reasoning steps.



I remember seeing that GPT-5 had two python tools defined in its leaked prompt one them would hide the output from user visible chain of thought UI.


Same with CoT prompting.

If you compare the outputs of a CoT input vs a control input, the outputs will have the reasoning step either way for the current generation of models.


Yeah, this is honestly one of the coolest developments of new models.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: