6. Techniques

Sequencing

Next-generation and long-read sequencing technologies, library prep, read mapping, and variant calling basics.

DNA Sequencing Technologies

Hey students! 🧬 Welcome to one of the most exciting frontiers in molecular biology - DNA sequencing! In this lesson, we're going to explore how scientists can read the genetic code that makes you uniquely you. By the end of this lesson, you'll understand how modern sequencing technologies work, how scientists prepare DNA samples for sequencing, and how they analyze the massive amounts of data these technologies generate. Think of this as learning to read the ultimate instruction manual - the one written in your DNA!

The Evolution of DNA Sequencing πŸ“ˆ

DNA sequencing has come a long way since the 1970s! To understand where we are today, let's start with some mind-blowing numbers. The Human Genome Project, completed in 2003, took 13 years and cost about 3 billion to sequence one complete human genome. Today, thanks to next-generation sequencing (NGS), we can sequence a human genome in just a few days for under $1,000!

Traditional sequencing methods, like Sanger sequencing, could only read about 800-1,000 base pairs at a time. While Sanger sequencing is still considered the gold standard for accuracy (99.9% accuracy), it's like reading a book one sentence at a time - very slow and expensive for large projects.

Next-generation sequencing changed everything by introducing massively parallel sequencing. Instead of reading one DNA fragment at a time, NGS can sequence millions of DNA fragments simultaneously. It's like having millions of people each reading a different page of a book at the same time! Popular NGS platforms include Illumina (which dominates about 80% of the market), Ion Torrent, and 454 sequencing (now discontinued).

The newest revolution comes from long-read sequencing technologies like Pacific Biosciences (PacBio) and Oxford Nanopore Technologies (ONT). These can read DNA fragments that are 10,000 to over 1,000,000 base pairs long - imagine reading entire chapters instead of just sentences!

Next-Generation Sequencing: The Workhorses of Modern Genomics πŸ”¬

NGS technologies work through a clever process called "sequencing by synthesis." Here's how it works in simple terms: scientists attach fluorescent tags to DNA building blocks (nucleotides A, T, G, C), each with a different color. As DNA polymerase adds these tagged nucleotides to growing DNA chains, cameras capture the flashing colors, revealing the sequence.

Illumina sequencing, the most popular NGS platform, typically produces reads of 150-300 base pairs with an error rate of about 0.1%. In a single run, an Illumina NovaSeq 6000 can generate up to 6 terabases of data - that's enough to sequence about 60 human genomes! The process takes 1-4 days depending on the specific machine and run parameters.

The strength of NGS lies in its incredible throughput and relatively low cost per base. However, the short read lengths can make it challenging to sequence repetitive regions of the genome or detect large structural variations. It's like trying to solve a jigsaw puzzle where many pieces look very similar - you need the bigger picture to put them in the right place.

Long-Read Sequencing: Seeing the Bigger Picture πŸ”

Long-read sequencing technologies have revolutionized our ability to study complex genomic regions. PacBio uses a technique called Single Molecule Real-Time (SMRT) sequencing, where DNA polymerase sits at the bottom of tiny wells called zero-mode waveguides. As the polymerase incorporates fluorescently labeled nucleotides, the light is detected in real-time.

Oxford Nanopore takes a completely different approach - it literally pulls DNA through tiny protein pores and measures changes in electrical current as different nucleotides pass through. This technology can sequence DNA fragments over 1 million base pairs long! The MinION, ONT's portable sequencer, is about the size of a smartphone and can even be used in remote locations or on the International Space Station.

While long-read sequencing has higher error rates (5-15% for raw reads), the long read lengths provide crucial advantages. They can span repetitive regions, detect large structural variations, and provide better genome assemblies. Recent improvements in accuracy, including consensus calling from multiple reads of the same molecule, have brought error rates down to less than 1%.

Library Preparation: Getting DNA Ready for Sequencing πŸ§ͺ

Before DNA can be sequenced, it needs to be prepared in a process called library preparation. Think of this as preparing ingredients before cooking - you need to cut, season, and organize everything properly.

The first step is DNA fragmentation. For NGS, DNA is broken into smaller pieces (typically 300-800 base pairs) using physical methods like sonication or enzymatic methods. For long-read sequencing, scientists try to keep DNA fragments as long as possible, sometimes using very gentle extraction methods to preserve molecules over 100,000 base pairs long.

Next comes adapter ligation. Adapters are short DNA sequences that act like barcodes and handles. They serve multiple purposes: they allow DNA to bind to the sequencing platform, provide primer binding sites for amplification, and can include unique molecular identifiers (UMIs) for tracking individual molecules.

Many NGS workflows include a PCR amplification step to generate enough DNA for sequencing. However, PCR can introduce bias - some sequences amplify better than others. Long-read sequencing often skips PCR amplification to avoid this bias, though this requires starting with more DNA.

Quality control is crucial throughout library preparation. Scientists use tools like the Agilent Bioanalyzer or Qubit fluorometer to check DNA concentration and fragment size distribution. A good library should have the right concentration (typically 2-20 nM for Illumina sequencing) and appropriate fragment sizes.

Read Mapping: Finding Where DNA Belongs πŸ—ΊοΈ

Once sequencing is complete, scientists face the challenge of analyzing millions or billions of short DNA sequences called "reads." The first major step is read mapping or alignment - figuring out where each read came from in the original genome.

Popular mapping algorithms include BWA (Burrows-Wheeler Aligner), Bowtie2, and STAR (for RNA sequencing). These tools use sophisticated algorithms to quickly compare each read against a reference genome. For a human genome with 3 billion base pairs, this is like finding the exact location of millions of puzzle pieces in a reference picture.

The mapping process considers that reads might contain errors or come from slightly different versions of the genome (genetic variants). Most aligners allow for a few mismatches and can handle small insertions or deletions. They assign quality scores to each alignment based on how confident they are about the mapping location.

Long-read mapping presents unique challenges because of the higher error rates and longer sequences. Specialized tools like minimap2 and NGMLR have been developed specifically for long-read data. These aligners are more tolerant of errors but can provide better resolution of complex genomic regions.

Variant Calling: Discovering Genetic Differences πŸ”

After mapping reads to a reference genome, the next step is variant calling - identifying differences between the sequenced DNA and the reference genome. These variants include single nucleotide polymorphisms (SNPs), small insertions and deletions (indels), and larger structural variations.

Popular variant calling tools include GATK (Genome Analysis Toolkit), FreeBayes, and SAMtools. These programs analyze pileups of reads at each genomic position and use statistical models to determine whether observed differences represent real variants or sequencing errors.

The process considers several factors: how many reads support the variant, the quality scores of those reads, and the expected error rate of the sequencing technology. For example, if 50 out of 100 reads at a position show a different nucleotide than the reference, and those reads have high quality scores, it's likely a real heterozygous variant.

Variant calling from long-read data can detect larger structural variations that short reads might miss. Tools like Sniffles and SVIM specialize in calling structural variants from long-read data, identifying insertions, deletions, inversions, and translocations that can be thousands of base pairs long.

Quality filtering is essential in variant calling. Scientists typically filter variants based on read depth (coverage), variant quality scores, and other metrics. A typical human genome contains about 4-5 million variants compared to the reference genome, but after quality filtering, researchers usually work with 3-4 million high-confidence variants.

Conclusion

DNA sequencing technologies have transformed from slow, expensive methods to high-throughput, cost-effective tools that are revolutionizing biology and medicine. NGS provides incredible throughput and accuracy for most applications, while long-read sequencing offers unique advantages for complex genomic regions and structural variation detection. The process from DNA sample to variant calls involves careful library preparation, sophisticated computational algorithms for read mapping, and statistical methods for variant calling. As these technologies continue to improve and costs decrease, DNA sequencing is becoming an increasingly powerful tool for understanding genetics, diagnosing diseases, and advancing personalized medicine.

Study Notes

β€’ Next-Generation Sequencing (NGS): Massively parallel sequencing technology that can sequence millions of DNA fragments simultaneously, typically producing reads of 150-300 base pairs with ~0.1% error rate

β€’ Long-Read Sequencing: Technologies like PacBio and Oxford Nanopore that can sequence DNA fragments 10,000+ base pairs long, useful for repetitive regions and structural variants

β€’ Library Preparation Steps: DNA fragmentation β†’ adapter ligation β†’ (optional PCR amplification) β†’ quality control

β€’ Read Mapping: Process of aligning sequencing reads to a reference genome using algorithms like BWA, Bowtie2, or minimap2 for long reads

β€’ Variant Calling: Identifying genetic differences from reference genome using tools like GATK, FreeBayes, or Sniffles for structural variants

β€’ Key Statistics: Human genome sequencing cost dropped from $3 billion (2003) to <$1,000 (2024); typical human genome has 4-5 million variants vs. reference

β€’ Illumina Dominance: Controls ~80% of sequencing market with high accuracy but short reads

β€’ Quality Metrics: Read depth (coverage), quality scores, and variant confidence scores are crucial for reliable results

β€’ Applications: Clinical genomics, cancer research, infectious disease monitoring, and personalized medicine

Practice Quiz

5 questions to test your understanding

Sequencing β€” Molecular Biology | A-Warded