Hacker Newsnew | past | comments | ask | show | jobs | submit | anupshinde's commentslogin

Just yesterday I had a moment

Claude's code in a conversation said - “Yes. I just looked at tag names and sorted them by gut feeling into buckets. No systematic reasoning behind it.”

It has gut feelings now? I confronted for a minute - but pulled out. I walked away from my desk for an hour to not get pulled into the AInsanity.


>It has gut feelings now?

I would say hard no. It doesn't. But it's been trained on humans saying that in explaining their behavior, so that is "reasonable" text to generate and spit out at you. It has no concept of the idea that a human-serving language model should not be saying it to a human because it's not a useful answer. It doesn't know that it's not a useful answer. It knows that based on the language its been trained on that's a "reasonable" (in terms of matrix math, not actual reasoning) response.

Way too many people think that it's really thinking and I don't think that most of them are. My abstract understanding is that they're basically still upjumped Markov chains.


It has a lot. I find by challenging it often, getting it to explain it's assumptions, it's usually guessing.

This can be overcome by continuously asking it to justify everything, but even then...


Trust shouldn't be inherent in our adoption of these models.

However, constant skepticism is an interesting habit to develop.

I agree, continually asking it to justify may seem tiresome, especially if there's a deadline. Though with less pressure, "slow is smooth...".

Just this evening, a model gave an example of 2 different things with a supposed syntax difference, with no discernible syntax difference to my eyes.

While prompting for a 'sanity check', the model relented: "oops, my bad; i copied the same line twice". smh


I don't find it tiresome at all. What I was getting at was, even with constant justifications you need to remain vigilant.

It's almost like an emergent feature of a tool that's literally built on best guesses is...guesswork. Not what you want out of a tool that's supposed to be replacing professionals!

Interesting perspective.

I guess I'm more interested in understanding what it can and can't do.


Even when used by humans, "gut feelings" is still a metaphor.

Glad to see this. I was tired of seeing posts that are on the extremes - "death of software by AI" vs "AI can't do this and that".

I took a break from software, and over the last few years, it just felt repetitive, like I was solving or attempting to solve the same kinds of problems in different ways every 6 months. The feeling of "not a for loop again", "not a tree search again", "not a singleton again". There's an exciting new framework or a language that solves a problem - you learn it - and then there are new problems with the language - and there is a new language to solve that language's problem. And it is necessary, and the engineer in me does understand the why of it, but over time, it just starts to feel insane and like an endless loop. Then you come to an agreement: "Just build something with what I know," but you know so much that you sometimes get stuck in analysis paralysis, and then a shiny new thing catches your engineer or programmer brain. And before you get maintainable traction, I would have spent a lot of time, sometimes quitting even before starting, because it was logistically too much.

Claude Code does make it feel like I am in my early twenties. (I am middle-aged, not in 60s)

I see a lot of comments wondering what is being built -

Think about it like this, and you can try it in a day.

Take an idea of yours, and better if it is yours - not somebody else's - and definitely not AI's. And scope it and ground it first. It should not be like "If I sway my wand, an apple should appear". If you have been in software for long, you would have heard those things. Don't be that vague. You have to have some clarity - "wand sway detection with computer vision", "auto order with X if you want a real apple", etc.. AI is a catalyst and an amplifier, not a cheat code. You can't tell it, "build me code where I have tariffs replacing taxes, and it generates prosperity". You can brainstorm, maybe find solutions, but you can't break math with AI without a rigorous theory. And if you force AI without your own reasoning, it will start throwing BS at you.

There is this idea in your mind, discuss it with ChatGPT, Gemini, or Claude. See the flaws in the idea - discover better ideas. Discuss suggestions for frameworks, accept or argue with AI. In a few minutes, you ask it to provide a Markdown spec. Give it to Claude Code. Start building - not perfect, just start. Focus on the output. Does it look good enough for now? Does it look usable? Does it make sense? Is the output (not code) something you wanted? That is the MVP to yourself. There's a saying - customers don't care about your code, but that doesn't mean you shouldn't. In this case, make yourself the customer first - care about the code later (which in an AI era is like maybe a 30min to an hour later)

And at this point, bring in your engineer brain. Typically, at this point, the initial friction is gone, you have code and something that is working for you in real - not just on a paper or whiteboard. Take a pause. Review, ask it to refactor - make it better or make it align with your way, ask why it made the decisions it made. I always ask AI to write unit tests extensively - most of which I do not even review. The unit tests are there just to keep it predictable when I get involved, or if I ask AI to fix something. Even if you want to remove a file from the project, don't do it yourself - acclimatize to prompting and being vague sometimes. And use git so that you can revert when AI breaks things. From idea to a working thing, within an hour, and maybe 3-4 more hours once you start reviews, refactors, and engineering stuff.

I also use it for iterative trading research. It is just an experiment for now, but it's quite interesting what it can do. I give it a custom backtesting engine to use, and then give it constraints and libraries like technical indicators and custom data indicators it can use (or you could call it skills) - I ask it to program a strategy (not just parameter optimize) - run, test, log, define the next iteration itself, repeat. And I also give it an exact time for when it should stop researching, so it does not eat up all my tokens. It just frees up so much time, where you can just watch the traffic from the window or think about a direction where you want AI to go.

I wanted to incorporate astrological features into some machine learning models. An old idea that I had, but I always got crapped out because of the mythological parts and sometimes mystical parts that didn't make sense. With AI, I could ask it to strip out those unwanted parts, explain them in a physics-first or logic-first way, and get deeper into the "why did they do this calculation", "why they reached this constant", and then AI obviously helps with the code and helps explain how it matches and how it works - helps me pin point the code and the theories. Just a few weeks ago, I implemented/ported an astronomy library in Go (github.com/anupshinde/goeph) to speed up my research - and what do I really know about astronomy! But the outputs are well verified and tested.

But, in my own examples, will I ever let AI unilaterally change the custom backtesting engine code? Never. A single mistake, a single oversight, can cost a lot of real money and wasted time in weeks or months. So the engine code is protected like a fortress. You should be very careful with AI modifying critical parts of your production systems - the bug double-counting in the ledger is not the same as a "notification not shown". I think managers who are blanket-forcing AI on their employees are soon going to realize the importance of the engineering aspect in software

Just like you don't trust just any car manufacturer or just any investment fund, you should not blindly trust the AI-generated code - otherwise, you are setting yourself up to get scammed.


The brainstorming, investigation and planning are so much fun, aren't they?

Having an infinitely patient, super smart colleague available all the time is amazing.


If you were forced to choose just one of all the competing players, which is "the one" you will use?

For me, the choice is ChatGPT, not for its Codex or other fancy tooling - just the chat. Not that Claude Code or Cowork is less important. Not that I like Codex over Claude Code.


Right now? Claude, so long as they don't fold to the Pentagon's demands. It's important to me that the company at least have a pretense of ethics. If they fold, I may just use open models via DDG – I don't find code assistants very useful for my workflow anyway.


Same. I wish it wasn’t the case, but even making a show of caring about ethics is about as much as we can hope to get from a company these days.


I don't think software engineering is going to die. Coding, as we know it, is going to change a lot. Short-term pain, but in the long term, we are likely to see an explosion of software. Having said that, AGI could change things - but then every profession would be dead.

Check this out: https://www.youtube.com/watch?v=OfMAtaocvJw


Exactly what I have experienced, I have tried with "thinking set" and saw varying results. An analogy - Each prompt in a conversation felt like I was talking to a different support representative, who missed one or the other part of the overall context, and I had to repeat myself many times. But it's also an inherent human part of me that forgets I'm talking to an AI.

However, I never explicitly told it to "think hard" - I will start doing that. I believe that is the key to making it work consistently.

Thanks!


What is the Apple hardware being used here? I see Apple Silicon but not the configuration.. what did I miss


Was looking for that too. Need to know whether I already own the hardware or can’t afford it.


I am just starting to feel that GPT-5 is more hype.

Just a day before GPT-5 launch, I made a video about making a tool with agents, Claude Sonnet 4 and GitHub Copilot.

There was so much hype on the launch day - on how good GPT-5 is, how it gets the code right the first time, and how little direction it needs.

So I was compelled to try it again with GPT-5 preview available in Copilot.

And for some reason, it struggled to align with directions.

a. It would make its own decisions, misaligned with what I mentioned. I had to explicitly say "DO NOT DO this....". ( explicit instructions were followed.)

b. It did not complete tasks and moved forward. This is the same style I used with both GPT-4o and Claude Sonnet 4.

Also, it might be good at coding, but it does feel geeky - need to try more.


An Investor is essentially a Trader operating on a very long timeframe, benefiting from a slow decision-making process, slow results, and slow emotional swings. And I know people will argue against this. But if you are a trader, identify a sweet spot for speed/frequency, and then your trading results will suddenly improve.


I am comfortable with both Python and Go. I prefer Go for performance; however, the earlier issue was verbosity.

It is easier to write things using a Python dict than to create a struct in Go or use the weird `map[string]interface{}` and then deal with the resulting typecast code.

After I started using GitHub Copilot (before the Agents), that pain went away. It would auto-create the field names, just by looking at the intent or a couple of fields. It was just a matter of TAB, TAB, TAB... and of course I had to read and verify - the typing headache was done with.

I could refactor the code easily. The autocomplete is very productive. Type conversion was just a TAB. The loops are just a TAB.

With Agents, things have become even better - but also riskier, because I can't keep up with the code review now - it's overwhelming.


I'm confused. My kid does this on my ChatGPT account all the time. What is new here?

I cannot emphasize how good a teacher ChatGPT is, until it misinforms (and human teachers also do). And it also stays open to questioning without making the student feel stupid for asking. The good part is that ChatGPT will accept a mistake, but a human teacher will get annoyed.

The only thing I keep reminding my kid is to keep the BS detector on and verify everything that ChatGPT says and never trust it blindly. (feels too similar to the "don't talk to strangers" advice)

Unrelated - check with kids and teenagers. Once, a teen told me, "AI is a dark evil force," and it's what their teachers told them.


Well said


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: