I really enjoyed this writeup. Having just finished a year in the field on the engineering side, and reading many papers without an academic background, it’s funny to recognize who are the “math” types and who are the “biology” types at work. Puts things in perspective.
I’ve been pushing for more ensembles and multi-label classifiers because I want to orchestrate the pieces with logic to fill gaps until stats methods outperform (given enough new data in nice annotation-friendly structures).
The “math” types seem to feel like there’s a neural net solution to every problem or that we can expand the multi class model to cover more domains despite being mostly saturated with high accuracy. Sounds awesome but I’m impatient!
The “biology” folks seem to be most attracted to neatness or parsimony. We had some great bikeshedding sessions around adjacency list vs materialized path (ltree) in postgres for label hierarchies. Abstractions can be useful too!
Any tips on being a better experimentalist and pushing academic colleagues towards better solutions in the field?
> I once had an occasion to ask a very prominent AI researcher for early career tips. His advice was simple: write!
I give the same advice to new PhDs. I wrote mini-overviews of stuff I was researching, doing and thinking for the first two years. Write it with LaTeX too. By the time you have to write your actual thesis you can merge papers, overviews, ideas, and will be done in a few months (instead of years for the people who didn't write along the way).
And after a week or so, "What important problems are you working on?"
And after some more time I came in one day and said, "If what you are doing is not important, and if you don't think it is going to lead to something important, why are you at Bell Labs working on it?"
there's a lot to unpack in that statement. A good problem is good for many different reasons: it's not enough that the problem be an important one, it should also be the case that you have the techniques and knowledge to solve it; otherwise you'll be banging your head against the wall without success.
As a student, you might know that a problem is important, but you might not yet have the tools to solve it; you might not even be aware of which tools can solve it.
> One of the most common and aggravating manifestations of hype in AI research is the renaming of old ideas with flashy new terms. Beware of these buzzwords -- judge a paper based primarily on its experiments and results.
I’ve been pushing for more ensembles and multi-label classifiers because I want to orchestrate the pieces with logic to fill gaps until stats methods outperform (given enough new data in nice annotation-friendly structures).
The “math” types seem to feel like there’s a neural net solution to every problem or that we can expand the multi class model to cover more domains despite being mostly saturated with high accuracy. Sounds awesome but I’m impatient!
The “biology” folks seem to be most attracted to neatness or parsimony. We had some great bikeshedding sessions around adjacency list vs materialized path (ltree) in postgres for label hierarchies. Abstractions can be useful too!
Any tips on being a better experimentalist and pushing academic colleagues towards better solutions in the field?