Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Fair point! I think my main idea was "prefer building with an API over training your own model" but that isn't as pithy.

The jury's still out on how much training and fine tuning are going to matter in the long run - my belief is that there are many great products that can exist without needing a new model architecture, or owning the model at all.



That advice makes sense if we're talking about 800B+ parameter models that require a gigantic investment of capital and time. For models that fit on a consumer GPU you're leaving chips on the table to not take advantage of training / fine-tuning. It's just too easy and powerful not to.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: