Takeaways

  • build tools just for yourself, across the entire stack. linus builds projects using his own programming language, his own css framework, his own backup service, etc
  • papers showing minor SOTA improvements on benchmarks are less interesting than papers that yield new insights on models. this is a good filter for the firehose of ML content
  • carefully curated dataset > billions of mediocre pieces of data for training
  • AI hardware products need to be sexy or have insane utility. iphone has both, google glass has neither
  • chatgpt is one of few first gen products that is easy, fast, useful out of the box
  • doing a research sabbatical is a good way to figure out your next career steps. start by reading for a couple months, then enter a prototyping phase where you build → get feedback from users → document failures in 2 week sprints. generate many artifacts and try to answer a research question each sprint

Rough notes

  • check out pair with google design stuff for neuron viz
  • anthropic paper on interpretable neurons
  • build software abstraction layers for yourself. makes iteration must faster and much easier since you know the stack
  • pile = automatically organized stack of info
  • chat interfaces solve a specific problem
  • cool interfaces: notion, adept, github, vision for gpt
  • how to avoid getting features scooped: lots of noise, linus only cares about new work that yields new understanding
  • lots of things are new SOTA with new benchmark, no new understanding or anything deeper. stuff like the anthropic paper are way more useful. only care about small improvements when building prod systems. scaling is interesting, more params/data, finetuning is harder than pretraining.
  • specifying a task closer to the end user’s specs is much harder than it looks but dataset quality is extremely important for training. quality > quantity, billions of text isn’t as useful as handpicked beautiful data. very carefully curated data > giant dataset for finetuning.
  • building on openai api is okay if you have a deep understanding of the underlying task, ie using wikipedia is extremely powerful since data quality is super high and fact checked
  • lots of room for failure for notion in terms of QA and hallucination
  • hard for gpt wrappers to succeed without customer lock in
  • building horizontally is definitely possible if you find a niche
  • LLMs haven’t hit an uncanny valley yet, disagree highly
  • for hardware you need to either be so sexy or have insane utility, ideally both like the iphone, google glass is neither for example. important to realize that a watch for example needs to do something a phone
  • chatgpt is super rare since it’s one of few first generation products that is easy, fast and useful right out of the box
  • tab/rewind thoughts: how sexy do people think it is? the people who built this are already obsessed with recording and searching everything, linus thinks he would be overwhelmed with transcribing everything he says. demonstrate for specific types of users that this is absolutely something they require, must do this to have a shot. transcribing/search alone isn’t particularly useful
  • tips for research sabbatical: just reading for a couple months (papers, links, etc). reading a lot gave a ton of ideas- implementation phase for a couple months. structure things into a two week cadence, two weeks to investigate each question. build, give to people to try, understand why it fails, document everything and move forward. structure in terms of projects: ie can we control how LMs generate text by controlling intermediate embedding layers. granularity and lots of artifacts is a useful medium of operation.