Flow cytometry is a core technique in modern pharmaceutical development pipelines. It functions as a flexible, high-throughput assay due to its capacity to process large quantities of cells in relatively short acquisition times, while simultaneously monitoring multiple cellular parameters. Flow cytometry is routinely used to assess cell death in cytotoxicity assays and observe changes in expression of one or more targets in response to stimuli. These measurements provide critical data for understanding drug safety, efficacy, and mechanism-of-action within early pipeline studies. Additionally, flow cytometry allows for the classification of heterogeneous cell populations in immunotyping assays, supporting the identification and characterization of immune cell subsets. The technique can be extended to other stages of cell and gene therapy R&D pipelines, including quality control assessments of genetically modified cells to ensure product consistency and safety. Despite its power, efficiency, and technological advancements, the modality is not without its challenges. Inefficiencies in experiment planning and execution, data management, and analysis highlight the need for ongoing innovation within and optimization of end-to-end workflows.
The complexity of flow cytometry begins long before data acquisition. Scientists often spend hours manually preparing plate-based assays, curating conditions for 96 or 384 wells by hand. Each well may require unique reagent combinations and concentrations, and without standardized plate maps, researchers must rely on their own conventions for format, terminology, and reagent naming. This lack of consistency not only slows down experimental setup but also complicates collaboration and reproducibility across teams and projects.
Capturing essential reagent information—such as lot numbers, expiration dates, and supplier details—is another persistent challenge. These details are typically recorded manually, if at all, and the process is so tedious that we often see it skipped. The result is incomplete metadata, which can hinder troubleshooting and compromise data integrity. Experimental design is further complicated by the need to juggle multiple disconnected systems for sample management, instrument scheduling, and documentation. The absence of standardized metadata capture and enrichment means that critical contextual information is often lost, making it difficult to reconstruct experiments or share insights with colleagues.
Once samples are acquired, scientists face a new set of hurdles in data analysis. Most rely on proprietary software to interpret results, manually adjusting gates for each sample to distinguish cell populations. This process is highly subjective and can vary significantly between users, introducing variability and making it difficult to compare results across experiments or operators.
Troubleshooting analysis is particularly challenging for less experienced users. When data does not look as expected, it can be difficult to pinpoint whether the issue stems from incorrect compensation, unexpected biological responses, or missing fluorophores. The lack of standardized troubleshooting workflows means that valuable time is often spent diagnosing issues that could be avoided with better documentation or automated checks.
To expedite routine analyses, scientists frequently build custom templates. While these templates can streamline workflows, they come at the cost of flexibility—experiments need to closely match the template structure, limiting the ability to modify markers or assay parameters. Capturing the rationale behind gating decisions or analysis choices is rarely systematic, and the lineage of analytical steps is often confined to proprietary software, making it difficult to audit or reproduce results outside these platforms.
Another challenge lies in managing the broad array of data and analysis files generated by flow cytometry. Data is saved in multiple formats—FCS, WSP, CSV, and others—some of which are proprietary and not easily accessible outside specific software environments. Files are often stored across a patchwork of locations, from instrument workstations to shared drives and individual laptops, with no standardized naming conventions to aid in tracking or retrieval.
This fragmentation makes it difficult to maintain a clear record of data provenance, track changes, or ensure that the correct version is used for downstream analysis. Retrieving files for reuse, quality control, or retrospective analysis is often a manual and error-prone process, limiting the ability to leverage historical data for new research or regulatory submissions. The lack of interoperability between systems and formats further compounds these challenges, creating barriers to efficient data management and collaboration.
To overcome these hurdles, there is a strong need for a solution that can automate and streamline the scientific workflow. Platforms that integrate experimental design, file capture, and data standardization/contextualization workflows can build a broad foundation for accelerating experiments and producing high-quality results. Solutions are open, modular, and cloud-native, leveraging industry standards to enhance flexibility, interoperability, and long-term value. The guiding principles we use to build these solutions for clients are detailed below and summarized in figure 1 for reference.
Figure 1: Core elements of the flow cytometry experimental workflows present opportunities to reduce burden on researchers and improve quality of scientific data.
Experimental planning needs to be supported by tools that standardize plate maps, field names, and formats, promoting consistency from the outset. Integration with reagent registration systems and streamlined intake of vendor or subcontractor order forms simplifies logistics and documentation. These capabilities help researchers gather complete, future-proof metadata, setting the stage for broad and repeatable downstream analysis.
Data capture systems are deployed to continuously monitor new instrument files (such as FCS) and automatically apply quality control checks as data is generated. Validated data sets are then securely transferred to a central cloud repository in near real time, reducing manual effort and minimizing data loss. Implementing automation at this stage enables data repositories to be current and ready for analysis.
Effective integration with leading Electronic Lab Notebooks (ELNs) allows experimental data, metadata, and additional annotations to flow automatically into a single, auditable source of truth. This unified approach simplifies compliance, traceability, and collaboration, making it easier for research teams to access and leverage relevant information.
The final goal of the above automation and aggregation workflows is to publish data in a form that enhances its value. Standardizing key attribute names using controlled ontologies and industry standards, such as the Allotrope Simplified Model (ASM) for instrument data, makes data easier and more reliable to search. Data, analysis results, and metadata are centralized in a documented and easily discoverable location, making them findable and reusable. By converting analyses into standard, open-source data models, the solution provides accessibility and interoperability, while cloud storage could enable programmatic access for future research.
An integrated, agentic research assistant can empower scientists to explore internal data sets, assess biological questions, assist with hypothesis generation, and strategize experiment design. Automated generation of key quality control and insight plots—such as histograms, scatter/density plots, gated panels, and cell population visualizations—can facilitate rapid interpretation. Previously time-consuming analyses, such as side-by-side sample comparison, will be facilitated by agent-powered capabilities. Comparisons of studies will be facilitated through interactive visuals and automatically generated statistical summaries to highlight meaningful differences.
Flow cytometry is an experimental modality of a dynamic and evolving cell and gene therapy R&D landscape. Organizations are optimistic about the opportunities that modern computational infrastructure, agentic AI, and strategic data reuse open.1 By designing solutions to fit expanding technical and automation needs and aligning these solutions to a future-looking strategy, research organizations can position themselves to conduct more efficient studies, spend more time and effort on insight generation (as opposed to data gathering), and increase their probability of success by better targeting questions within the pipeline.
Deloitte, in collaboration with Amazon Web Services (AWS), has developed a suite of cloud-based accelerators espousing the principles of the solution approach described in this article. The Deloitte and AWS Lab of the Future accelerators provide the ability to deliver value, focusing on many of the flow cytometry pain points mentioned above while being flexible enough to be fine-tuned to help address the specific needs of our clients.2 The accelerators enable automated instrument file management, data annotation and standardization, and integrated data reuse through data product creation.
Endnotes
1Jeff Morgan et al., “Pharma’s R&D lab of the future: Building a long-lasting innovation engine,” Deloitte, July 29, 2025.
2Deloitte, “The Lab of the Future: How AI and the cloud are transforming the science of discovery,” accessed January 2026.