Over the past two decades, the cost to comprehensively sequence a human genome has fallen from over $100 million in 2001 to under $200, with Ultima Genomics recently announcing the $100 genome. This drop is primarily due to technological and operational advancements. New chemistry and instrument designs from companies have driven down costs by increasing sample throughput and large batch runs, spreading fixed costs, and lowering per-base reagent expenses.
The long-read sequencing space has historically been more expensive, but providers are now offering more competitive pricing – while also improving accuracy and throughput. Because long reads can span complex genomic variants in a single run, they reduce the need for the downstream validation required with short-read techniques, making the overall sequencing process faster.
Improvements in lab automation and standardized workflows have also reduced manual labor and turnaround times, while falling costs for cloud computing and data storage have made large-scale sequencing more affordable. These advances are driven by growing demand from large-scale projects, such as national and pharmaceutical genomics programs and biobanks, which justifies investment in this high-capacity infrastructure.
Studying just one molecular layer, such as genomics alone, often missed the wider biological context. Diseases, drug responses, and health trajectories are rarely explained by DNA sequence in isolation; they require a holistic view across multiple omics datasets, including transcriptomics and epigenomics. Historically, scaling multi-omics was limited by three main challenges: cost, complexity, and data fragmentation. Early workflows were expensive and labor-intensive, which restricted studies to small cohorts, and the proliferation of different platforms created a mix of methods and data formats, making it difficult to integrate data on different “omes”. The sheer volume of data generated was also a major hurdle for storage and interpretation.
While recent advances in automation, robotics, and cloud computing have addressed many of these issues, true scalability still depends on the ability to orchestrate complex workflows and unify diverse data. As sequencing becomes more affordable and widespread, new regulatory challenges are emerging, with one key concern being the validation of results. As more genomes are produced, regulators must ensure that tests are both technically accurate and reliably predict disease risk.
Data privacy is another critical issue, as genomic data is inherently identifiable, making traditional anonymization nearly impossible, which challenges existing frameworks like HIPAA and GDPR, particularly around consent, data sharing, and secondary use.
The growing role of AI in interpreting genomic data introduces yet more new hurdles. Regulators need to determine how to assess AI algorithms for accuracy, transparency, and safety as these models continue to evolve. With the increased sequencing of infants and the ability to identify adult-onset diseases, complex policy questions also arise about how this information should be handled by insurers and health systems.
The trick is to coordinate multi-assay pipelines and integrate data across all omics layers into a single platform, and overcoming these challenges will come with benefits. Automation and AI have fundamentally changed the scale and speed of genomics. Automation has scaled genomics from small, specialized projects to industrial-level operations by reducing manual steps, increasing throughput, and ensuring consistent, high-quality data generation. New high-throughput sequencers can process thousands of samples per run, which is key for large-scale studies.
AI helps by managing and interpreting the massive data volumes that automation generates, supporting decision making by classifying variants and flagging potential risks in the process. Together, this synergy is powerful: automation generates terabytes of data with minimal manual input, while AI interprets that data in real time, extracting relevant findings and reducing turnaround times from weeks to hours.
Scalability in multi-omics is about more than just generating more data; it's about integrating various omics layers into a cohesive and reproducible system. Addressing key data and workflow bottlenecks generates a unified data backbone that integrates all modalities, standardizes protocols across different assays, and automates tracking to ensure a fully traceable, end-to-end solution that supports scientific progress at scale.
The future of genomics is not just about speed or volume. It's about building the infrastructure to make that scale meaningful. As sequencing expands, the primary challenge shifts from data generation to integration, insight, and action. This requires platforms that do more than just store information. They must be able to orchestrate workflows, manage compliance, and support a continuous cycle of discovery. By connecting automation, AI, and decision-making in a way that ensures traceability and reproducibility at every stage, genomics can move from raw potential to real-world impact – accelerating breakthroughs that are not only scientifically powerful, but clinically transformative.
