In twenty years, we cut the cost of sequencing a human genome by more than five orders of magnitude! And this feat is even more impressive if you look farther back in time. The first nucleic acid ever sequenced (by Robert Holley and others in 1965) was 76 nucleotides long and took 15 person-years to sequence. So if you had sequenced a full human genome in 1965 using Holley's methods, the labor cost alone would have been about 10 teradollars.[5]
Some authors have already tried to tell the story of how sequencing costs plummetted. For instance, Jay Shendure and others have a nice paper called "DNA sequencing at 40," where they recount the history of DNA sequencing technologies from the seventies to 2017. James Heather and Benjamin Chain's "The sequence of sequencers" covers roughly the same scope at a more comfortable pace. These papers are both pretty good, but I have three major complaints about them. (1) They're written for specialists. Unglossed biology jargon is everywhere, and when new sequencing methods are introduced, the authors don't do much to help readers picture what's happening physically. (2) They're out of date. Shendure & al.'s history of sequencing—the most recent high quality paper I could find—is eight years old, so it can't discuss important recent advances in paleogenomics, single-cell genomics, and probably other subfields of genomics I'm not aware of. (3) Neither paper presents a unified thesis on how sequencing got cheap. Instead, they give us a mosaic of incremental innovations and leave us to figure out how they all added up to ten OOMs of progress.
I see an opening here for someone to write the history of genomic progress. It should be sophisticated but accessible to non-experts with only college freshman level science knowledge. It should cover all the key innovations in nucleic acid sequencing technology up to 2025 and explain them on a physical level. And it should have one foot in economics so that it can answer the big question: How did sequencing get so cheap? Sarah Constantin has already taken an admirable swing at this one in her essay on "The Enchippening." She argues that sequencing got cheap by riding semiconductor manufacturing's coattails. The flow cells and photodetectors that make up the core of a modern DNA sequencer are made with the same methods we use to make computer chips, so as those methods have gotten exponentially better, sequencing has naturally gotten exponentially cheaper. I want to know how well this story holds up when you take a close look at the history of sequencing. When exactly did we start to enchipen sequencing, and was there really a dramatic change in the rate of progress at that point? What are the most important drivers of genomic progress besides the enchippening?
This book calls for a special author. They must be enough of a scientist to really understand how sequencing works (and has worked), but they also need some of the skills of an economist and a historian. The ideal author might be Saloni Dattani, James Somers, or some obscure guy living in Cambridge, Massachusetts who's seen the whole history of genomics firsthand and knows the field inside and out.
[1] | You can reach me at milesmkodama at Google's mail. |
[2] | Or so military historian Kamil Galeev claims in his excellent ChinaTalk interview. |
[3] | Cf The Art of War chapter 7.3: "When you surround an army, leave an outlet free." |
[4] | From Our World in Data using data from the National Human Genome Research Institute. |
[5] | I got this by assuming a Cornell professor made 20 kilodollars per year in 1965, so the labor cost per nucleotide would be 4 kilodollars. Multiplying by the 3 giganucleotides in the human genome gives us our ballpark answer. |