Neanderthal DNA and Modern Humans: Svante Pääbo Receives the 2022 Nobel Prize in Physiology or Medicine

What makes humans “human”?

Neanderthal DNA sequencing from ancient bone samples

On October 3, 2022, the Nobel Assembly at Karolinska Institutet announced the 2022 Nobel Prize in Physiology or Medicine had been awarded to Svante Pääbo, director of the Department of Genetics at the Max Planck Institute for Evolutionary Anthropology in Leipzig, Germany. The Assembly cited his “discoveries concerning the genomes of extinct hominins and human evolution”. They mentioned the highlight of his research: the seemingly impossible task, at the time, of sequencing the Neanderthal genome. The discoveries that followed from this sequencing project continue to redefine our understanding of modern human origins.

The award showcases the technological advancements made in the analysis of ancient DNA. However, Pääbo’s research had an inauspicious beginning. In 1985, he published the results of his early work, cloning and sequencing DNA fragments from a 2,400-year-old Egyptian mummy (1). Unfortunately, later analysis revealed that the samples could have been contaminated by the researchers’ own DNA (2).

Continue reading “Neanderthal DNA and Modern Humans: Svante Pääbo Receives the 2022 Nobel Prize in Physiology or Medicine”

High-Molecular Weight DNA for Long-Read Sequencing

Imagine that you’re putting together a large, complex jigsaw puzzle, comprising thousands of exceptionally small pieces. You lay them all out and attempt to make sense of them. It would be far easier to assemble this puzzle were the pieces larger, containing more of the image advertised on the box. The same can be said when sequencing a genome.

high-molecular weight DNA  Depiction of a DNA helix

Traditional short-read or next-generation sequencing relies on DNA spliced into small fragments (≤300 base pairs) and then amplified. While useful for detecting small genetic variants like single-base changes to the DNA, this type of sequencing can fail to illuminate larger variations (typically over 50 base pairs) in the genome. Long-read sequencing, or third generation sequencing, allows more accurate genome assemblies, facilitating better detection of structural variants like copy number variations, duplications, translocations and inversions that are too large to identify with short-read sequencing. Long-read sequencing has the capability to fill in “dark regions” of a genome that are unfinished and can be used to assemble larger, more complex genomes using longer fragments of DNA, or high-molecular weight (HMW) DNA.

Continue reading “High-Molecular Weight DNA for Long-Read Sequencing”

ProDye Brings Sanger Sequencing to Multiple Platforms

Researchers looking for new chemistry for Sanger sequencing need look no further than the ProDye™ Terminator Sequencing System, developed by Promega for use in capillary electrophoresis instruments. Sanger sequencing, or dye-terminator sequencing, has been the gold standard of DNA analysis for over 40 years and is a method commonly used in labs around the world. Even as new technologies emerge, Sanger sequencing remains the most cost-effective method for sequencing shorter pieces of DNA.

Sanger sequencing depicted as results on a musical cleft
Continue reading “ProDye Brings Sanger Sequencing to Multiple Platforms”

Harnessing the Power of Massively Parallel Sequencing in Forensic Analysis

The rapid advancement of next-generation sequencing technology, also known as massively parallel sequencing (MPS), has revolutionized many areas of applied research. One such area, the analysis of mitochondrial DNA (mtDNA) in forensic applications, has traditionally used another method—Sanger sequencing followed by capillary electrophoresis (CE).

Although MPS can provide a wealth of information, its initial adoption in forensic workflows continues to be slow. However, the barriers to adoption of the technology have been lowered in recent years, as exemplified by the number of abstracts discussing the use of MPS presented at the 29th International Symposium for Human Identification (ISHI 29), held in September 2018. Compared to Sanger sequencing, MPS can provide more data on minute variations in the human genome, particularly for the analysis of mtDNA and single-nucleotide polymorphisms (SNPs). It is especially powerful for analyzing mixture samples or those where the DNA is highly degraded, such as in human remains.  Continue reading “Harnessing the Power of Massively Parallel Sequencing in Forensic Analysis”

Better, Faster, Cheaper: Measuring the Speed of Science

Are we better off now than we were 10 years ago? Often times this question is answered subjectively and will vary from person to person. We can empirically show how life expectancy has increased over the centuries thanks to advances in the fields of agriculture and medicine, but what about quality of life? Science affects our lives every day, and the general notion is that better science will (eventually) translate into better lives. There is a burning curiosity shared by myself and others to quantify how we have progressed in science over the years:

publication-growth
Click for full article. Source: Bornmann, L. & Mutz, R. (2015). Growth rates of modern science: a bibliometric analysis based on the number of publications and cited references. Journal of the Association for Information Science and Technology, 66(11), 2215–2222.

Bornmann and Mutz demonstrate in the image shown above how we have been doubling scientific output every nine years since the 1940s. That is not to say that we have become twice as smart or efficient; this phenomenon could be partially fueled by a desire to gain prestige through a high number of publications. To better assess the topic of efficiency, we can measure how long it takes to perform specific procedures and how much they cost. This article compares the rate of improvement for DNA sequencing, PCR, GC-MS and general automation to the rate of improvement for supercomputers and video game consoles.

Continue reading “Better, Faster, Cheaper: Measuring the Speed of Science”

Remembering Frederick Sanger and Sanger Sequencing

It is with sadness that we recognize the passing of Dr. Frederick Sanger. Sanger is known to molecular biologists and biochemists worldwide for his DNA sequencing technique, which won for him the 1980 Nobel prize in Chemistry.

Also noteworthy, Sanger’s laboratory accomplished the first complete genome sequence, that of a viral DNA genome more than 5,000 base pairs in length.

The 1980 prize was Sanger’s second Nobel award, his first awarded in 1958 for determining the chemical structure of proteins. In this work, Sanger elucidated not only the amino acids that comprised insulin but also the order in which the amino acids occurred.

About Sanger Sequencing
Sanger DNA sequencing is also known as the chain-termination method of sequencing. The Sanger technique uses dideoxynucleotides or ddNTPs in addition to typical deoxynucleotides (dNTPs) in the reaction. ddNTPs result in termination of the DNA strand because ddNTPs lack the 3’-OH group required for phosphodiester bond formation between nucleotides. Without this bond, the chain of nucleotides being formed is terminated.

Sanger sequencing requires a single-stranded DNA, a DNA primer (either radiolabeled or with a fluorescent tag), DNA polymerase, dNTPs and ddNTPs. Four reactions are set up, one for each nucleotide, G, A, T and C. In each reaction all four dNTPs are included, but only one ddNTP (ddATP, ddCTP, ddGTP or ddTTP) is added. The sequencing reactions are performed and the products denatured and separated by size using polyacrylamide gel electrophoresis.

Diagram of Sanger dideoxy sequencing. (Courtesy  of Wikipedia and Estevez, J.)
Diagram of Sanger dideoxy sequencing. (Courtesy of Wikipedia and Estevez, J.)

This reaction mix results in various lengths of fragments representing, for instance, the location of each A nucleotide in the sequence, because while there is more dATP than ddATP in the reaction, there is enough ddATP that each ATP ultimately instead is replaced with a ddATP, resulting in chain termination. Separation by gel electrophoresis reveals the size of these ddATP-containing fragments, and thus the locations of all A nucleotide in the sequence. Similar information is provided for GTP, CTP and TTP.

The Maxam and Gilbert DNA sequencing method had the advantage at the time of being used with double-stranded DNA. However, this method required DNA strand separation or fractionation of the restriction enzyme fragments, resulting in a somewhat more time-consuming technique, compared to the 1977 method published by Sanger et al.

Dr. Sanger was born in Gloucestershire, U.K. in 1918, the son of a physician. Though he initially planned to follow his father into medicine, biochemistry became his life-long passion and area of research endeavor. Sanger retired at age 65, to spend more time at hobbies of gardening and boating.

References

Sanger, F. , Nicklen, S. and Coulson, A.R. (1977) DNA sequencing with chain-terminating inhibitors. Proc. Natl. Acad. Sci. USA 74, 5463-7.

Maxam, A.M. and Gilbert, W. (1977) A New Method for Sequencing DNA. Proc. Natl. Acad. Sci. USA

There is something special about seeing the original Sanger publication from 1977, available here as a scan.

Learning About the $1,000 Genome

Personalized MedicineAt the recent International Symposium on Human Identification, Kevin Davies, the keynote speaker and author of The $1,000 Genome, entertained attendees with a history of human genome sequencing efforts and discussed ways in which the resulting information has infiltrated our everyday lives. Obviously, there is enough material on the subject to fill a book, but I will describe just a few of the high points of his talk here.

Continue reading “Learning About the $1,000 Genome”

Dietary Analysis, DNA Style

dna testing of foodDNA testing methods are being used to solve problems in an ever-increasing number of fields. From crime scene analysis to tissue typing, from mammoths to Neanderthals, and from Thutmose I to Richard III, both modern mysteries and age-old secrets are being revealed. The availability of fast, accurate, and convenient DNA amplification and sequencing methods has made DNA analysis a viable option for many types of investigation. Now it is even being applied to solve such mundane mysteries as the precise ingredients used in a sausage recipe, and to answer that most difficult of questions “what exactly is in a doner kebab?” Continue reading “Dietary Analysis, DNA Style”

Sonnets in DNA

William ShakespeareFor sixty years now, scientists have studied the role of DNA as a vehicle for the storage and transmission of genetic information from generation to generation. We have marveled at the capacity of DNA to store all the information required to describe a human being using only a 4-letter code, and to pack that information into a space the size of the nucleus of a single cell. A letter published last week in Nature exploits this phenomenal storage capacity of DNA to archive a quite different kind of information. Forget CDs, hard drives and chips, the sum of human knowledge can now be stored in synthetic DNA strands. The Nature letter, authored by scientists from the European Bioinformatics Institute in Cambridge, UK, and Agilent Technologies in California, describes a proof-of-concept experiment where synthetic DNA was used to encode Shakespeare’s Sonnets, Martin Luther King’s “I Have a Dream” speech, a picture of the Bioinformatics Institute, and the original Crick and Watson paper on the double-helical nature of DNA. Continue reading “Sonnets in DNA”

DNA Sequencing from AutoRads and Gels to Nanopores

DNA SequencingLast week I read an article in Wired Science that described how an outbreak of antibiotic resistant Klebsiella pneuomiae was tracked in real-time at an NIH hospital using DNA sequencing technologies. The article described how whole genome sequencing of disease isolates and environmental samples from the hospital was used to track the source and spread of the outbreak.

The scientists monitoring the outbreak tracked spontaneous random mutations in the K. pneumoniea DNA sequence to determine that the outbreak was caused by a single source, and to track the spread of the organism within the hospital. The sequencing information helped investigators identify when and where infection occurred, and also to track transmission of the infection from person-to-person. It also revealed that the order of transmission was different from the order in which the cases presented with symptoms, and helped identify how the organism was spread between individuals.

The article describes how epidemiology, infection control and sequence identification were used together to influence outcome in this situation, but also shows the power of whole genome sequencing to find and track subtle differences between isolates that could not have been identified in any other way.

To me, this is a powerful illustration of just how far DNA sequencing has come over the last few years. Not so long ago, the idea of sequencing the entire genome of numerous disease isolates during an outbreak would have been almost laughable—an idea confined to episodes of the X-files or to science fiction stories. Now, thanks to advanced automated sequencing technologies and the computing power to analyze the results, it is doable within a reasonable timeframe for hospitals with access to the right facilities. Although this type of investigation is still beyond the capabilities of most hospitals, the costs and turnaround times for sequencing are coming down rapidly as new technologies capable of faster, cheaper analysis become available.

We have come a very long way since the days when DNA sequencing was a laborious process involving pouring a gel, running samples,and manually reading the resulting autoradiogram hoping to get a read of 50–100 bases. My reading of the wired article prompted me to find out more about the newer types of sequencing technology available today. Here’s what I learned about each: Continue reading “DNA Sequencing from AutoRads and Gels to Nanopores”