Hacker Newsnew | past | comments | ask | show | jobs | submit | ryeguy_24's commentslogin

What expense app are you building? I really want an app that helps me categorize transactions for budgeting purposes. Any recommendations?


My advice is a little different. It’s make the life of your boss insanely easy. Similar in nature to post but slightly different optimization function. Don’t over communicate, communicate just the right amount. Anticipate questions. Don’t create any friction for them and be really helpful. Some of my people will anticipate things and be proactive. I love that and I constantly push to get them promoted.


Ive adopted this mindset recently and it really does work. That being said i feel it turns me into a bit of a “yes man”. I wish there was more room for more of my authentic personality


When I read the publications (the ACM magazine), I swear sometimes the content feels LLM generated. Does anyone else get that impression? In general, I'm not very impressed with the content (I'm used to WIRED, btw).


The way I think of it (might be wrong) but basically a model that has similar sensors to humans (eyes, ears) and has action-oriented outputs with some objective function (a goal to optimize against). I think autopilot is the closest to world models in that they have eyes, they have ability to interact with the world (go different directions) and see the response.


This is a very smart idea. I couldn't turn my Ring Alarm off and I was on the same Wifi connection as the system. In retrospect, it would be quite smart to switch over to local network.


Does anyone have this mystical report?



:) I mean, you always by a faster car. My interest is in the ability to enhance the voice quality of the uber-portable AirPods using AI.


Exactly. This was my point. Televisions can upconvert from 720p to 4k. In the same sense, the machine learning model would fill in the waveform and mimic a high powered mic. It can do this at the connection point (iPhone / computer).


Televisions have considerably more temporal data to work with than an audio stream does. It's very easy to hack together interpolated images, not so easy to predict/denoise/upres time-series audio information.

Past a certain point it's probably easier/more efficient to use the Airpods as a speech-to-text mic and then infer a "high quality" text-to-speech version on your connected device.


Isn’t there a whole bunch of dependency here related to prompting and methodology that would significantly impact overall performance? My gut instinct is that there are many many ways to architect this around the LLMs and each might yield different levels of accuracy. What do others think?

Edit: In reading more, I guess this is meant to be a dumb benchmark to monitor through time. Maybe that’s the aim here instead of viability as an auto close tool.


Agree. Also, with respect to training, what is the goal that we are maximizing? LLMs are easy, predicting the next word and we have lots of training data. But what are we training for in real world? Modeling the next spatial photograph to predict things that will happen next? It’s not intuitive to me what that objective function would be in spatial intelligence.


Why wouldn’t predicting the next frame in a video stream be as effective as predicting the next word?


Or that there is a sufficiently generalizable objective function for all “spatial intelligence.”


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: