Hacker Newsnew | past | comments | ask | show | jobs | submit | flutas's commentslogin

While others will point to hardware or local LLMs or such IMO the biggest reason...

Because it's the easiest way to give "claw" iMessage access and that's the primary communication channel for a lot of the claw users I've seen.


Yup as context, in the same time Waymo had 101 collisions according to the same NHTSA dataset.


Waymo drives 4 million miles every week (500k+ miles each day). Vast majority of those collisions are when Waymos were stationary (they don’t redact narrative in crash reports like Tesla does, so you know what happened). That is an incredible safety record.


Is this the same time or the same miles driven? I think the former, and of course I get that's what you wrote, but I'm trying to understand what to take away from your comment.


For everyone's context, in the same time Waymo had 101 collisions according to the same dataset.


What dataset? Isn't the article clearly specified a different number?

Your context sucks, and it's good as a lie.

>Waymo reports 51 incidents in Austin alone in this same NHTSA database, but its fleet has driven orders of magnitude more miles in the city than Tesla’s supervised “


you are talking about 5 incidents, this is not statistics. Its just a fluctuation of random numbers, and random events like bus hits the taxi while idle. It's already 20% of your data is incorrect lol , since it's 1 out of 5.

So far , you can clearly tell : 1. tesla works decent in a limited environment, no crazy patterns 2. It's a limited env that means nothing. Scale is still not there. They ned to prove themself.


One of my earlier experiences with codex was actually reverse engineering, far before it was good at actual coding.

It was able to decompile a react native app (Tesla Android app), and fully trace from a "How does X UI display?" down to a network call with a payload for me to intercept.

Granted it did it by splitting the binary into a billion txt files with each one being a single function and then rging through it, but it worked.


I heard about this and tried quite a bit to reverse engineer a decompiled binary from a big game to find struct/schema information but could never get anything useful.


Working on reproducible test runs to catch quality issues from LLM providers.

My main goal is not just a "the model made code, yay!" setup, but verifiable outputs that can show degradation as percentages.

i.e. have the model make something like a connect 4 engine, and then run it through a lot of tests to see how "valid" it's solution is. Then score that solution as NN/100% accurate. Then do many runs of the same test at a fixed interval.

I have ~10 tests like this so far, working on more.


Sounds really interesting. What are you using for the tests/reports?


Nice. Sounds like will converge to QA as a Service


Reminds me of the "oopsie" by Reddit when they revealed Eglin Air Force Base as the "most addicted city."


It takes 20mins to fet from base houseing to the gate, lord k ows what traffic is like by the causeways, its an hour of driving before you're anywhere worth being and then its a coin flip if its exciting, so its either the ft. Walton beach strip clubs or onbase recreation.

No wonder eglin is addicted hahaha.

But in all seriousness, there are teams of people on the data crunch side of things that seems like a pedestrian insight


Waymo's remote drivers have literally caused accidents and we only know about it because journalists did digging. Waymo simply removed all details of the remote ops roles in it in the NHTSA reporting.


What journalists should I look for to see that


https://www.forbes.com/sites/bradtempleton/2024/03/26/waymo-...

The description there is

    In January, an incident took place where a Waymo robotaxi incorrectly went through a red light due to an incorrect command from a remote operator, as reported by Waymo. A moped started coming through the intersection. The moped driver, presumably reacting to the Waymo, lost control, fell and slid, but did not hit the Waymo and there are no reports of injuries. There may have been minor damage to the moped.
While the description in the official report to the NHTSA is (ID: 30270-6981)

    On January [XXX], 2024 at 10:52AM PT a rider of a moped lost control of the moped they were operating and fell and slid in front of a Waymo Autonomous Vehicle (Waymo AV) operating in San Francisco, California on [XXX] at [XXX] neither the moped nor its driver made contact with the Waymo AV.

    The Waymo AV was stopped on northbound [XXX] at the intersection with [XXX] when it started to proceed forward while facing a red traffic light. As the Waymo AV entered the intersection, it detected a moped traveling on eastbound [XXX] and braked. As the Waymo AV braked to a stop, the rider of the moped braked then fell on the wet roadway before sliding to a stop in front of the stationary Waymo AV. There was no contact between the moped or its rider and the Waymo AV. The Waymo AVs Level 4 ADS was engaged in autonomous mode.

    Waymo is reporting this crash under Request No. 1 of Standing General Order 2021-01 because a passenger of the Waymo AV reported that the moped may have been damaged. Waymo may supplement or correct its reporting with additional information as it may become available.


They've also been seen driving directly into flood waters, with one driving through the middle of a flooded parking lot.

https://www.reddit.com/r/SelfDrivingCars/comments/1pem9ep/hm...


curious what your take away from that is given the announcement.


It's a lot more iffy than that IME.

It's very happy to throw a lot into the memory, even if it doesn't make sense.


This is the core problem. The agent writes its own memory while working, so it has blind spots about what matters. I've had sessions where it carefully noted one thing but missed a bigger mistake in the same conversation — it can't see its own gaps.

A second pass over the transcript afterward catches what the agent missed. Doesn't need the agent to notice anything. Just reads the conversation cold.

The two approaches have completely different failure modes, which is why you need both. What nobody's built yet is the loop where the second pass feeds back into the memory for the next session.


even as far as we know they aren't

The Waymo blog post refused to say the word "child", instead using the phrase "young pedestrian" once.

The Waymo blog post switches to "the pedestrian" and "the individual" for the rest of the post.

The Waymo blog post also consistently uses the word "contact" instead of hit, struck, or collision.

The Waymo blog post makes no mention of the injuries the child sustained.

The Waymo blog post makes no mention of the school being in close proximity.

The Waymo blog post makes no mention of other children or the crossing guard.

The Waymo blog post makes no mention of the car going over the school zone speed limit (17 in 15).


The speed limit of a school zone in California is 25, not 15, which would explain why they didn't mention it.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: