This article provides a comprehensive guide for researchers and drug development professionals on applying user-centered design (UCD) to create impactful cancer quality improvement tools.
This article provides a comprehensive guide for researchers and drug development professionals on applying user-centered design (UCD) to create impactful cancer quality improvement tools. It explores the foundational need for UCD in oncology, details practical methodological approaches like co-design and iterative prototyping, addresses key implementation challenges including ethics and interoperability, and presents robust validation strategies. By synthesizing current evidence and real-world case studies, this resource aims to bridge the gap between technological innovation and clinical utility, ultimately fostering the development of digital tools that are both scientifically sound and readily adopted in cancer care.
Cancer remains a leading global health challenge, consistently ranking as a major cause of mortality worldwide [1]. Despite significant advancements, conventional diagnostic and therapeutic methods frequently lack the precision and adaptability required for complex cancer care environments [2]. The standard toolkit—comprising surgery, chemotherapy, radiation therapy, and imaging-based diagnostics—often falls short due to intrinsic limitations such as lack of personalization, collateral damage to healthy tissues, and inability to address dynamic tumor heterogeneity [1] [2].
Traditional diagnostics relying on symptoms, basic imaging, and biopsies often detect cancer at advanced stages, while treatments like chemotherapy and radiation struggle with toxicity, resistance, and imprecise targeting [1] [2]. These foundational limitations create critical gaps in patient care, prompting the oncology field to develop more sophisticated, user-centered tools that integrate molecular insights, artificial intelligence, and participatory design principles to overcome these challenges [3] [4] [5].
Table 1: Performance Comparison of Diagnostic Modalities in Oncology
| Diagnostic Method | Sensitivity | Specificity | Area Under Curve (AUC) | Key Limitations |
|---|---|---|---|---|
| Standard Imaging (CT, X-ray) | Not quantified | Not quantified | Not quantified | Fails to detect ~20% of breast cancers in dense tissue; high false positives for PSA tests [2] |
| Traditional Biopsy | Gold standard | Gold standard | Gold standard | Invasive, time-consuming, limited by tumor location/accessibility [2] [4] |
| AI-Based Lung Cancer Detection | 0.86 (0.84-0.87) | 0.86 (0.84-0.87) | 0.92 (0.90-0.94) | Requires large, high-quality datasets; integration challenges [4] |
| AI for EGFR Mutation Prediction | 0.78 (0.75-0.81) | 0.81 (0.77-0.84) | 0.86 (0.83-0.89) | Limited by imaging quality and algorithm transparency [4] |
Table 2: Limitations of Traditional Cancer Treatment Modalities
| Treatment Modality | Key Advancements | Persistent Challenges | Impact on Patient Outcomes |
|---|---|---|---|
| Surgery | Fluorescence-guided systems, robotic assistance, minimally invasive techniques [1] | Minimal residual disease (MRD), tumor heterogeneity, post-surgical metastatic progression [1] | Recurrence due to MRD; immunosuppression facilitating evasion [1] |
| Radiation Therapy | SBRT, IMRT, FLASH radiotherapy, radiation protectors/sensitizers [1] | Precision limitations, immune suppression, regional access issues [1] | Damage to healthy tissues; recurrence due to radioresistant cells [1] |
| Chemotherapy | Synthesized derivatives with amplified cytotoxicity [1] | Drug resistance, toxicity to healthy cells, limited efficacy [1] | Severe side effects; treatment failure due to resistance mechanisms [1] |
| Hormonal Therapy | Targeted approaches for hormone-dependent cancers [1] | Resistance development, quality of life impacts [1] | Limited long-term efficacy; adverse effects on wellbeing [1] |
Background: Digital health tools can potentially revolutionize supportive cancer care but often face implementation challenges due to limited stakeholder involvement in development [3]. This protocol outlines a participatory design methodology for creating patient-centered digital health applications.
Materials:
Methodology:
Expected Outcomes: Identification of critical user needs; development of a user-centered digital health application with improved adoption potential; comprehensive understanding of implementation facilitators and barriers [3].
Background: Artificial intelligence shows promise for addressing limitations in traditional cancer diagnosis but requires rigorous validation [4]. This protocol provides a framework for evaluating image-based AI tools in lung cancer management.
Materials:
Methodology:
Quality Assessment:
Expected Outcomes: Quantified performance metrics for AI algorithms in lung cancer diagnosis; identification of optimal AI approaches for specific clinical questions; framework for multicenter validation of AI tools [4].
Table 3: Essential Research Reagents and Platforms for Modern Oncology Investigations
| Research Reagent/Platform | Function | Application in Addressing Traditional Gaps |
|---|---|---|
| Next-Generation Sequencing (NGS) | Comprehensive genomic profiling to identify targetable mutations [2] [7] | Enables precision medicine by moving beyond one-size-fits-all treatment approaches [2] |
| Circulating Tumor DNA (ctDNA) Assays | Non-invasive liquid biopsy for monitoring treatment response and minimal residual disease [1] [5] | Overcomes limitations of traditional tissue biopsies; enables dynamic treatment adaptation [1] |
| DeepHRD | AI tool detecting homologous recombination deficiency from standard biopsy slides [7] | Identifies patients for PARP inhibitors; more accurate than current genomic tests with lower failure rates [7] |
| Prov-GigaPath, Owkin's Models | AI-powered diagnostic tools for cancer detection imaging [7] | Improves early detection accuracy beyond conventional imaging and human interpretation [4] [7] |
| Spatial Transcriptomics | High-resolution mapping of gene expression within tumor microenvironment [5] | Reveals tumor heterogeneity and therapy resistance mechanisms invisible to traditional histology [5] |
| Patient-Reported Outcome (PRO) Systems | Digital platforms for symptom monitoring and quality of life assessment [3] [6] | Addresses supportive care gaps by capturing patient experiences in real-world settings [3] |
| CAR T-Cells with Boolean Logic | Engineered cellular therapies with multiple receptors for enhanced cancer cell specificity [5] | Reduces off-target effects common in traditional chemotherapy; targets cancer stem cells [5] |
In the specialized field of cancer quality improvement research, the methodology employed to develop tools and interventions significantly influences their ultimate efficacy and adoption. While often used interchangeably, user-centered design (UCD), co-design, and participatory development represent distinct yet complementary approaches with unique philosophical and practical implications. For researchers, scientists, and drug development professionals working in oncology, understanding these nuances is critical for selecting the appropriate methodological framework. This application note delineates these core principles, provides structured protocols for their implementation, and contextualizes their application within cancer research, from clinical trial design to patient care tool development.
User-centered design is an iterative design process in which designers focus on the users and their needs in each phase of the design process [8]. UCD employs a mixture of investigative methods and tools (e.g., surveys, interviews) and generative ones (e.g., brainstorming) to develop an understanding of user needs [8]. The ultimate aim is to create highly usable and accessible products that address real user requirements [8].
Core Principles of UCD [9] [10] [11]:
Co-design represents a more collaborative approach, positioning itself as a structured process for involving patients throughout all stages of quality improvement [12]. In healthcare contexts, co-design captures the patient's lived experience, aims to understand these experiences, and implements improvements based on this understanding [12]. This approach has been characterized as involving six key phases: engage, plan, explore, develop, decide, and change [12].
In cancer care specifically, co-design has been successfully implemented to develop interventions such as information resource booklets and films [13], where patients and clinicians work collaboratively to identify improvement priorities and develop solutions.
Participatory development serves as an umbrella term encompassing various collaborative approaches. In cancer research, it often manifests as community-based participatory research (CBPR), which forms partnerships between researchers and communities to address disparities [14]. This approach is particularly valuable when developing interventions for populations experiencing cancer health disparities, as it ensures cultural relevance and community ownership [14].
Table 1: Comparative Analysis of Design Approaches in Cancer Research
| Dimension | User-Centered Design (UCD) | Co-Design | Participatory Development |
|---|---|---|---|
| Primary Focus | User needs and usability | Shared creation process | Community empowerment and capacity building |
| Typical Participant Role | Informant and tester | Active co-creator | Partner and decision-maker |
| Power Dynamic | Researcher-led | Shared ownership | Community-led |
| Key Strength | Optimizing usability and user experience | Leveraging lived experience for innovation | Ensuring cultural relevance and sustainability |
| Common Methods | Usability testing, interviews, personas | Joint workshops, prototyping | Community advisory boards, partnership development |
| Typical Output | Refined product or tool | Co-created intervention | Community-owned program |
This protocol outlines a systematic approach for developing digital health tools for cancer patients, such as symptom tracking applications or educational platforms.
Phase 1: Context Analysis and User Research [10]
Phase 2: Requirement Specification [10]
Phase 3: Iterative Prototyping and Evaluation [8] [10]
Figure 1: UCD Iterative Process for Digital Health Tool Development
This protocol details the experience-based co-design (EBCD) approach for developing cancer care interventions, adapted from successful implementations in oncology settings [12] [13].
Phase 1: Project Establishment and Ethical Approval
Phase 2: Experience Exploration [13]
Phase 3: Co-Design Workshops [12] [13]
Phase 4: Implementation and Reflection [13]
Table 2: Co-design Workshop Structure for Cancer Care Interventions
| Session | Duration | Participants | Key Activities | Materials |
|---|---|---|---|---|
| Introduction | 90 minutes | Patients, clinicians, facilitators | Project overview, confidentiality agreement, establishing group norms | Project information sheets, consent forms |
| Experience Sharing | 120 minutes | Patients, clinicians, facilitators | Watching touchpoint film, shared reflection, identifying key moments | Touchpoint film, audio recording equipment |
| Priority Setting | 90 minutes | Patients, clinicians, facilitators | Dot voting, discussion, consensus building on improvement areas | Voting materials, flip charts, colored markers |
| Ideation | 120 minutes | Patients, clinicians, facilitators | Brainstorming, storyboarding, concept development | Storyboard templates, sticky notes, prototyping materials |
| Refinement | 120 minutes | Patients, clinicians, facilitators | Prototype testing, feedback cycles, iteration | Prototypes, feedback forms, recording devices |
| Action Planning | 90 minutes | Patients, clinicians, facilitators | Implementation planning, responsibility assignment, evaluation planning | Action plan templates, implementation guides |
This protocol adapts the community-based participatory research (CBPR) framework for developing interventions to reduce cancer health disparities, drawing parallels to therapeutic drug development [14].
Stage 1: Community Engagement and Partnership Building [14]
Stage 2: Intervention Development and Adaptation [14]
Stage 3: Implementation and Evaluation [14]
Stage 4: Sustainability and Scaling [14]
Figure 2: Participatory Development Framework for Cancer Disparity Interventions
Table 3: Essential Methodological Reagents for Design Research in Cancer Quality Improvement
| Research Reagent | Function | Application Context | Implementation Considerations |
|---|---|---|---|
| Semi-structured Interview Guides | Elicit rich qualitative data on experiences and needs | UCD, Co-Design, Participatory Development | Must be adapted to cultural context and health literacy levels |
| Experience-Based Co-Design Toolkit | Facilitate collaborative design sessions | Co-Design | Requires trained facilitation; includes touchpoint films, journey mapping templates |
| HEART Framework Metrics | Align user experience goals with measurable outcomes | UCD | Must be customized for cancer-specific contexts (e.g., treatment adherence, symptom management) |
| Community Advisory Board | Ensure cultural relevance and community ownership | Participatory Development | Requires budget for stipends, transportation; must represent diversity of affected population |
| Usability Testing Protocol | Identify interface problems and usability issues | UCD | Should include cancer patients with varying levels of technological proficiency and health status |
| Co-Design Workshop Materials | Support creative collaboration and idea generation | Co-Design | Includes prototyping materials, journey mapping templates, voting materials |
| Cultural Adaptation Framework | Modify evidence-based interventions for specific cultural contexts | Participatory Development | Must address language, values, traditions, and historical trauma |
| Stakeholder Engagement Plan | Manage involvement of diverse stakeholders across project lifecycle | All approaches | Identifies key stakeholders, engagement frequency, methods, and communication channels |
The selection of UCD, co-design, or participatory development approaches in cancer quality improvement research should be guided by project goals, context, and desired outcomes. UCD excels when optimizing usability and user experience of existing tools or developing new digital health technologies. Co-design proves particularly valuable when leveraging lived experience to innovate cancer care services and interventions. Participatory development emerges as essential when addressing cancer health disparities and ensuring cultural relevance and community ownership.
Each approach demands distinct resources, timelines, and expertise, but all share the fundamental principle of engaging end-users in the development process. By applying these structured protocols and utilizing the provided toolkit, cancer researchers and drug development professionals can enhance the relevance, effectiveness, and adoption of quality improvement tools and interventions across the cancer care continuum.
The development and implementation of successful digital health tools in oncology depend on systematically identifying and engaging a diverse ecosystem of stakeholders. Human-centered design (HCD) methodologies provide a critical framework for ensuring that cancer quality improvement tools address the authentic needs of all end-users [15]. These approaches—including participatory design, co-design, and design thinking—prioritize the needs, desires, and behaviors of the people central to the problem being solved [16]. When applied to cancer care, this means actively involving patients, clinicians, and healthcare systems throughout the development process to create solutions that are not only technically robust but also clinically relevant, usable, and sustainable [3] [15].
The consequences of excluding key stakeholders are significant, leading to digital health tools with poor adoption, limited effectiveness, and ultimately, technological waste [15]. This application note provides a structured approach to stakeholder identification and engagement, framed within the context of cancer quality improvement research.
The stakeholder ecosystem for cancer quality improvement tools comprises three primary groups, each with distinct roles, needs, and influences. A comprehensive mapping is essential for targeted engagement strategies.
Table 1: Key Stakeholder Groups in Cancer Quality Improvement
| Stakeholder Group | Specific Roles | Primary Needs & Motivations | Influence on Implementation |
|---|---|---|---|
| Patients & Caregivers | - End-users of digital tools- Provide lived experience- Report outcomes and symptoms | - Access to clear information and support [17]- Streamlined communication with care team [3]- Self-efficacy and empowerment in care [3] [18] | - Ultimate determinants of adoption and engagement- Provide crucial feedback on usability and acceptability |
| Clinicians & Care Teams | - Primary facilitators of digital tool use- Interpret patient data and act on alerts- Integrate tools into clinical workflow | - Tools that save time and reduce administrative burden [18]- Seamless integration with existing EHR systems [19]- Clear clinical decision support [20] | - Gatekeepers to clinical integration- Crucial for championing the tool within the organization |
| Healthcare Systems & Leadership | - Provide infrastructure and resources- Establish governance and policies- Manage financial sustainability | - Improved patient outcomes and satisfaction [21]- Operational efficiency and cost-effectiveness [21]- Alignment with value-based care models and reimbursement [21] | - Create enabling environment (funding, IT, policy)- Drive organization-wide adoption and scaling |
Beyond these primary groups, other important stakeholders include payers who influence reimbursement models, regulators who ensure safety and efficacy, and health technology vendors who partner on development and integration [18] [22].
A phased, iterative approach to stakeholder engagement ensures that input is gathered meaningfully throughout the development lifecycle, from initial problem definition to post-implementation refinement.
The initial phase focuses on building empathy and deeply understanding the problem context from all stakeholder perspectives.
This phase translates insights into tangible solutions by collaboratively generating and refining ideas with stakeholders.
Stakeholders evaluate functional prototypes to identify usability issues and assess real-world fit before full-scale development.
The following diagram visualizes this iterative, multi-phase engagement process and its key outputs.
The following detailed protocol is adapted from successful implementations of Patient-Reported Outcome (PRO) systems in oncology, which demonstrate high-impact stakeholder engagement [20]. PROs are a critical cancer quality improvement tool, and their implementation exemplifies the principles discussed.
Table 2: Implementation Protocol for PRO Integration in Cancer Care
| Implementation Step | Stakeholder Engagement Activities | Key Outputs | Rationale & Evidence |
|---|---|---|---|
| 1. Needs Assessment & Planning | - Clinician input: Focus groups to identify key symptoms for monitoring (e.g., pain, fatigue) [20].- Patient input: Interviews to determine PRO acceptability, burden, and meaningful triggers for alerts [20]. | - List of target symptoms and PRO measures.- Defined thresholds for "concerning" PRO responses that trigger clinical review. | Ensures clinical relevance and patient acceptability, increasing long-term adoption [3] [20]. |
| 2. Development of Clinical Decision Support (CDS) | - Multidisciplinary team input: Oncology nurses, physicians, and pharmacists collaborate to develop evidence-based care pathways for concerning PROs [20]. | - PRO-based CDS tools (e.g., automated symptom management guidelines linked to specific PRO scores). | Standardizes care, supports nurses in managing alerts, and translates data into actionable clinical guidance [19] [20]. |
| 3. Training & Integration | - Role-specific training: Hands-on sessions for clinicians (interpreting PROs) and staff/CRAs (software use) [20].- Workflow integration: Map PRO review into existing clinical workflows (e.g., during pre-visit huddles) [3]. | - Trained clinical team.- Integrated workflow schematic.- Technical support plan. | Embeds the tool into routine practice, minimizing disruption and building clinician confidence [19] [20]. |
| 4. Pilot Testing & Iteration | - Stakeholder feedback: SUS surveys and interviews with patients and clinicians on usability and perceived value [23].- Workload assessment: Monitor alert frequency and nursing response time [20]. | - Refined PRO platform and CDS.- Understanding of resource needs for full implementation. | Identifies and resolves unforeseen technical and workflow issues in a controlled setting [3] [16]. |
| 5. Full Implementation & Sustainment | - Ongoing support: Dedicated coordinator for technical and logistical issues [19].- Feedback loops: Regular meetings with stakeholder champions to review metrics and address concerns. | - Fully operational PRO system.- Plan for continuous quality improvement. | Facilitates long-term sustainability and allows the system to adapt to evolving needs [18] [19]. |
Successful stakeholder engagement requires both methodological and practical tools. The following table details key "research reagents" for designing and executing this work.
Table 3: Essential Reagents for Stakeholder-Engaged Research
| Category & Item | Specification / Example | Primary Function in Research |
|---|---|---|
| Recruitment & Ethics | ||
| Purposive Sampling Framework | A predefined matrix to ensure diversity in stakeholder recruitment (e.g., by cancer type, treatment stage, clinical role) [17]. | Ensures a wide range of perspectives are captured, improving the validity and generalizability of findings. |
| Informed Consent Materials | Documents explaining study procedures, risks, benefits, and data handling in clear, accessible language. | Protects participant rights and meets ethical requirements for research. |
| Data Collection & Analysis | ||
| Semi-Structured Interview Guides | Question prompts tailored to each stakeholder group (patients, clinicians, etc.) [17] [23]. | Ensures consistent coverage of key topics while allowing flexibility to explore emergent themes. |
| Thematic Analysis Framework | A coding framework (e.g., based on Fitch's Supportive Care Framework) to analyze qualitative data [17]. | Provides a systematic method for identifying, analyzing, and reporting patterns (themes) across qualitative data sets. |
| Design & Evaluation | ||
| System Usability Scale (SUS) | A standardized 10-item questionnaire with a 5-point Likert scale [23]. | Provides a quick, reliable, and validated measure of a system's perceived usability from the user's perspective. |
| Low-Fidelity Prototyping Tools | Paper wireframes or digital mock-up tools (e.g., Figma, Adobe XD). | Allows for rapid, low-cost visualization of ideas for early-stage feedback and iteration before costly development. |
| Co-Design Workshop Kits | Physical (post-its, markers, printouts) or digital (Miro, Jamboard) tools for collaborative idea generation. | Facilitates creative collaboration and ensures all stakeholder voices are heard during the ideation phase [16]. |
Navigating the complex stakeholder landscape of cancer care is not a peripheral activity but a core component of developing effective, adoptable, and sustainable quality improvement tools. A structured, phased approach to engagement—Discover and Define, Ideation and Co-Design, and Prototyping and Testing—ensures that the resulting interventions are deeply rooted in the real-world needs of patients, clinicians, and healthcare systems. The provided protocols, visualization, and toolkit offer a practical starting point for researchers to design their own stakeholder engagement strategies, ultimately contributing to a more responsive and patient-centered oncology care ecosystem.
User-Centered Design (UCD) has emerged as a critical methodology for developing effective digital health tools in oncology. By prioritizing the needs, preferences, and workflows of end-users throughout the design process, UCD significantly enhances tool adoption, usability, and ultimately, clinical outcomes. This approach is particularly vital in cancer care, where digital tools must address complex patient needs and integrate seamlessly into clinical workflows. This application note synthesizes current evidence and provides structured protocols for implementing UCD in oncology quality improvement research.
The imperative for UCD in oncology is underscored by documented challenges with existing digital systems. Electronic Health Records (EHRs), for instance, often demonstrate significant usability failures, with physicians rating them in the bottom 9% of all software systems, a factor linked to increased burnout risk [24]. UCD directly addresses these shortcomings by ensuring digital tools are intuitive, efficient, and aligned with user expectations.
Research demonstrates that UCD methodologies lead to tangible improvements in usability metrics and implementation success. The following table summarizes key quantitative findings from recent studies.
Table 1: Quantitative Evidence for UCD in Oncology Digital Health Tools
| Digital Tool / Study | UCD Methodology | Key Usability & Outcome Metrics | Result |
|---|---|---|---|
| Step Proactive (AE Management Software) [25] | Usability testing with 6 patients & 6 HCPs; Iterative design per IEC 62366-1:2015 | System Usability Scale (SUS) Score; Scenario Completion Rate | SUS score in the 90th percentile (Grade A); 100% task completion rate |
| OncoSupport+ (Supportive Care App) [3] | Participatory co-design; Workshops, interviews, and focus groups with patients and HCPs | Identification of Critical Adoption Factors | Facilitators: Ease of use, workflow integration, professional introduction |
| Digital PRO Assessment (University Cancer Center) [26] | Interdisciplinary project group; Involvement of patient advisory board | Feasibility of Clinical Implementation | Successful development of a modular, clinically integrated PRO system |
| Young Adult Needs Assessment (NA-SB Intervention) [27] | Three-phase UCD: usability testing, contextual inquiry, prototyping | Optimization for Implementation | Intervention designed for implementation and scale-up across varied contexts |
UCD's impact extends beyond initial usability. The harmonization of evidence-based practices, implementation context, and implementation strategies through UCD methods potentially minimizes the need for elaborate, burdensome implementation strategies later, promoting sustainment [27].
This section provides detailed methodological protocols for key UCD experiments and processes cited in the evidence base.
Based on: Difrancesco et al. (2025), JMIR Human Factors [3]
Objective: To collaboratively design and develop a digital health application for supportive cancer care with patients and healthcare professionals.
Methodology Overview: A participatory study divided into three sequential phases: Predesign, Generative, and Prototyping.
Table 2: Phases of the Co-Design Protocol
| Phase | Primary Activities | Stakeholders Involved | Key Outcomes |
|---|---|---|---|
| Predesign | Understand context, challenges, and needs in supportive care. | Patients, survivors, HCPs, researchers. | Mapped challenges and needs at the clinical site. |
| Generative | Brainstorm app functionalities; Identify adoption facilitators/barriers. | Patients, survivors, HCPs. | Prioritized app features and implementation factors. |
| Prototyping | Iterative development of the app prototype and interface. | Patients, nurses. | A high-fidelity, user-validated app prototype (OncoSupport+). |
Detailed Procedures:
Based on: PMC Article 11924132 (2025) [25]
Objective: To assess the usability and user-friendliness of a medical device software for managing adverse events in oncology.
Methodology Overview: A multi-method usability test conforming to international standard IEC 62366-1:2015, involving both patients and healthcare professionals.
Detailed Procedures:
Based on: Hussaini et al. (2021), Implementation Science Communications [27]
Objective: To design a care coordination intervention and its implementation strategy to ensure fit with the clinical context.
Methodology Overview: A three-phase UCD process modeled as an iterative cycle of harmonization.
Diagram 1: The iterative UCD process for harmonization.
Detailed Procedures:
Table 3: Key Research Reagents and Solutions for UCD Experiments
| Item / Solution | Function / Description | Application in Protocol |
|---|---|---|
| System Usability Scale (SUS) | A reliable, 10-item Likert scale providing a global view of subjective usability assessments. | Quantitative usability assessment post-testing [25]. |
| Think-Aloud Protocol | A qualitative method where users verbalize their thoughts, feelings, and opinions while interacting with a prototype. | Identifying usability issues and understanding the user's mental model during prototyping [3]. |
| Interactive Prototyping Software | Tools (e.g., Figma, Adobe XD) to create high-fidelity, clickable mockups of digital applications. | Creating realistic prototypes for iterative testing in the prototyping phase without writing code [3] [27]. |
| IEC 62366-1:2015 Standard | International standard specifying a process for a manufacturer to analyze, design, develop and evaluate usability of a medical device. | Ensuring usability testing meets regulatory requirements for medical device software [25]. |
| Multidisciplinary Design Team | A core group including clinical experts (MDs, RNs), patients, designers, and software developers. | Ensuring all perspectives are integrated throughout the co-design and prototyping process [3] [27] [26]. |
The structured application of User-Centered Design is no longer optional but essential for developing successful digital health tools in oncology. The evidence demonstrates that UCD directly addresses the critical challenges of poor adoption and usability plaguing many healthcare technologies. By employing the detailed protocols and tools outlined in this document, researchers and drug development professionals can significantly enhance the implementation, effectiveness, and impact of their oncology quality improvement initiatives, ensuring that new technologies truly meet the needs of patients and clinicians.
The development of digital tools for cancer care has seen a significant shift towards methodologies that prioritize end-user needs through iterative design and evaluation. The following applications demonstrate the practical implementation of these approaches.
The Lion-App project exemplifies a rigorous, multi-stage iterative development process for a smartphone application that enables cancer patients to autonomously measure their quality of life (QoL). This research involved patients in a 3-stage process from conceptualization to deployment on private devices [28].
Key Quantitative Outcomes: The usability evaluation across development phases demonstrated consistent improvement, as captured by the User Experience Questionnaire (UEQ+) Key Performance Indicator (KPI), which ranges from -3 to +3 [28].
Table 1: Usability Evaluation Metrics Across Lion-App Development Cycles
| Development Phase | Participants (N) | Mean KPI (SD) | Key Findings | |------------------------||-------------------|------------------| | Usability Test 1 | 18 | 2.12 (0.64) | 94% response rate on UEQ+ | | Usability Test 2 | 14 | 2.28 (0.49) | Improvement from previous cycle | | Beta Test | 19 | 1.78 (0.84) | 74% UEQ+ response rate; age-dependent usage patterns |
The iterative refinements based on user feedback included restructuring the patient diary and integrating a shorter questionnaire for QoL assessment, demonstrating responsive adaptation to user needs [28]. The study found that age influenced engagement, with response rates decreasing with increasing age (P=.02), while sex demonstrated minimal influence on usability perceptions [28].
The development of OncoSupport+ employed a participatory study design with stakeholders at the University Hospital Zurich, integrating patients with cancer, survivors, and healthcare professionals throughout the development process [3].
Methodological Framework: The co-design process was structured into three distinct phases:
The resulting application featured two integrated components: (1) a patient dashboard for recording patient-reported outcome measures (PROMs) and accessing personalized supportive care information, and (2) a nurse dashboard for visualizing patient data during nursing consultations [3].
A German research team applied iterative development and early user testing to create a cancer prevention web application, demonstrating the value of formative evaluations during prototyping [29].
Usability Outcomes: The graphical user interface test yielded a System Usability Scale (SUS) score of 69.7/100 and a usefulness score of 75.8, indicating acceptable usability, though the learnability score was lower at 48.4, suggesting potential challenges in user understanding and satisfaction [29].
Table 2: Usability and Functional Assessment of Cancer Prevention Web Application
| Assessment Domain | Score/Outcome | Interpretation |
|---|---|---|
| Overall Usability (SUS) | 69.7/100 | Acceptable range |
| Usefulness | 75.8 | Above average |
| Learnability | 48.4 | Needs improvement |
| Identified Issues | 8 UX/UI categories | 1 severe, 3 moderate issues |
The qualitative feedback highlighted strengths in navigation, information presentation, and interactive features, particularly the risk simulation tool [29]. The identified usability issues primarily related to data input, user guidance, and risk visualization, providing clear direction for future iterations.
This protocol outlines the structured approach for iterative usability testing of digital health applications for cancer care, derived from the Lion-App development process [28].
2.1.1 Objectives
2.1.2 Materials and Equipment
2.1.3 Procedure
Phase 1: Focus Group Conduction
Phase 2: Initial Usability Testing
Phase 3: Iterative Refinement and Beta Testing
2.1.4 Data Analysis
This protocol details the participatory approach for engaging multiple stakeholders in the design of digital health applications for supportive cancer care [3].
2.2.1 Objectives
2.2.2 Participant Recruitment
Table 3: Stakeholder Inclusion Criteria for Co-Design Studies
| Stakeholder Group | Inclusion Criteria | Recruitment Source |
|---|---|---|
| Patients with Cancer | Current cancer treatment at participating clinic; Age ≥18 years; Language proficiency | Oncologic day clinic of Department of Oncology and Hematology |
| Patient Advocates | History of cancer; Age ≥18 years; Language proficiency | Swiss Group for Clinical Cancer Research (SAKK) |
| Cancer Nurses | Employment at participating institution; Direct patient care experience | Department of Oncology and Hematology |
| Supportive Care Specialists | Multidisciplinary expertise (nutrition, physiotherapy, psychology) | Comprehensive Cancer Center |
2.2.3 Procedure
Predesign Phase (Context Understanding)
Generative Phase (Functionality Brainstorming)
Prototyping Phase (Iterative Development)
2.2.4 Data Analysis
Diagram 1: Iterative UCD Process
Table 4: Essential Research Instruments and Tools for UCD Studies in Digital Health
| Research Reagent | Function/Application | Implementation Example |
|---|---|---|
| User Experience Questionnaire+ (UEQ+) | Modular assessment of usability metrics; calculates Key Performance Indicator (KPI) | Lion-App evaluation across development phases; KPI range -3 to +3 [28] |
| System Usability Scale (SUS) | Standardized 10-item scale for global usability assessment; scores 0-100 | Cancer prevention web app evaluation; overall score 69.7/100 [29] |
| Think-Aloud Protocols | Qualitative method where users verbalize thoughts while interacting with system | OncoSupport+ prototype testing; identification of UX/UI issues [3] |
| Focus Groups | Structured group discussions to gather perceptions and needs | Lion-App initial requirement gathering; 3 groups with 21 total participants [28] |
| Scoring Cards | Participatory prioritization technique for functionality ranking | OncoSupport+ co-design workshops; stakeholder-driven feature selection [3] |
| Patient-Reported Outcome Measures (PROMs) | Standardized instruments to capture patient health status | EORTC QLQ-C30 integration in Lion-App and OncoSupport+ for QoL assessment [28] [3] |
Within the paradigm of user-centered design, the development of effective cancer quality improvement tools hinges on a foundational principle: meaningful engagement with stakeholders. This includes patients, caregivers, healthcare professionals, and community partners. Moving beyond tokenistic involvement, structured engagement strategies are crucial for ensuring that tools and interventions are relevant, usable, and impactful. This application note details three core methodologies—charrettes, workshops, and qualitative interviews—providing protocols and data to guide their implementation in cancer research. These approaches are framed within the broader thesis that user-centered design is not merely an additive step but an integral component of rigorous, translational cancer research that ultimately enhances patient-centered care and tool efficacy.
This section provides a detailed exploration of three key stakeholder engagement methods, including experimental protocols and quantitative outcomes.
The CBPR Charrette model is an intensive, facilitated workshop designed to rapidly develop and strengthen collaborative research partnerships between community, academic, and medical stakeholders [30]. Its structured process fosters transparency and collective negotiation.
Experimental Protocol: Implementing the CBPR Charrette
The following diagram illustrates the logical workflow and participant interactions in a CBPR Charrette:
Application in Cancer Research: The CHAMPS (Cancer Health Accountability for Managing Pain and Symptoms) Study leveraged the CBPR Charrette to develop its partnership, which led to greater transparency, accountability, and trust among community, academic, and medical partners [30]. The process served as a catalyst for capacity building and allowed for the exploration of challenges with expert support.
Stakeholder workshops, such as those conducted by the EMA Cancer Medicines Forum (CMF) and the Extension for Community Healthcare Outcomes (ECHO) model, are organized meetings for collaborative discussion, education, and problem-solving around specific topics in cancer care [31] [32].
Experimental Protocol: Conducting a Virtual Training Workshop (e.g., ACS ECHO Model)
Quantitative Outcomes from ACS ECHO Programs: The table below summarizes aggregated quantitative data from four distinct ACS ECHO programs, demonstrating the model's effectiveness in engaging professionals and improving self-reported outcomes [32].
Table 1: Quantitative Outcomes from ACS ECHO Cancer Care Programs (2023-2024)
| Metric | Program A (Tobacco Cessation) | Program B (Colorectal Cancer Screening) | Program C (Prostate Cancer Screening) | Program D (Caregiving) | Aggregated Average |
|---|---|---|---|---|---|
| Unique Participants | 195 | 45 | 59 | 132 | 108 |
| Session Count | 4 | 7 | 9 | 7 | 6.75 |
| Participants Likely to Use Information | 59% | 59% | 59% | 59% | 59% |
| Mean Knowledge Increase | +0.84 | +0.84 | +0.84 | +0.84 | +0.84 |
| Mean Confidence Increase | +0.77 | +0.77 | +0.77 | +0.77 | +0.77 |
Data sourced from a quantitative evaluation of four ACS ECHO programs [32]. Likelihood to use, knowledge, and confidence data are reported as averages across all programs.
Qualitative interviews provide an in-depth understanding of patients' lived experiences, unmet needs, and perspectives on interventions, which is critical for developing truly patient-centered tools.
Experimental Protocol: A Two-Stage Qualitative Interview Study
The workflow for this two-stage qualitative interview process is as follows:
Key Findings from an ePROM Study: A Norwegian study employing this method revealed two central themes: 1) Symptom management in the shadow of disease-centered care, where patients felt personally responsible for bringing symptoms to clinicians' attention, and 2) ePROMs: bridging holistic care and disease management, where patients viewed ePROMs as a promising tool to amplify their voice and enable more holistic, responsive follow-up [33] [34].
Successful execution of these engagement strategies requires a suite of conceptual and practical "research reagents." The following table details key tools and their functions.
Table 2: Key Research Reagent Solutions for Stakeholder Engagement
| Item | Function/Application in Engagement Research |
|---|---|
| Semi-Structured Interview Guide | Ensures consistent coverage of key topics (e.g., patient experiences, unmet needs) while allowing flexibility to explore novel participant-led insights [34]. |
| System Usability Scale (SUS) | A ten-item, Likert-scale questionnaire used to quickly and reliably assess the perceived usability of a tool or system prototype [23]. |
| Conceptual Prototype (e.g., PowerPoint, wireframe) | A low-fidelity visualization of a proposed tool used in interviews or workshops to elicit concrete, actionable feedback from stakeholders before significant resources are invested in development [34]. |
| CBPR Charrette Facilitator's Guide | A structured protocol for guiding the intensive partnership workshop, ensuring all critical elements (strengths, challenges, goals, roles) are addressed collaboratively [30]. |
| Pre-/Post-Program Survey (5-point Likert scale) | A quantitative instrument for measuring changes in stakeholders' (e.g., clinicians') self-reported knowledge and confidence before and after an educational workshop or intervention [32]. |
Charrettes, workshops, and qualitative interviews are not merely data collection techniques but are foundational processes for embedding user-centered design into cancer quality improvement research. The structured protocols and supporting data presented here provide a roadmap for researchers to implement these methods effectively. By rigorously engaging stakeholders through these tailored approaches, the field can ensure that the resulting tools—from clinical databases and ePROM systems to professional education programs—are grounded in real-world needs, thereby enhancing their relevance, adoption, and ultimate impact on patient care.
In the development of digital tools for cancer quality improvement (QI), the transition from abstract requirements to a functional prototype is a critical phase. For researchers, scientists, and drug development professionals, this process must be rigorous, evidence-based, and efficient. Unvalidated, feature-heavy software can lead to clinician burnout, data integrity issues, and failed implementation. This document outlines application notes and protocols for wireframing, creating low-fidelity mockups, and employing structured feature prioritization, all framed within a user-centered design (UCD) methodology essential for creating effective cancer QI tools.
Objective: To quickly visualize and validate the core layout and navigation of a cancer QI tool (e.g., an Adverse Event Management System) with clinical stakeholders.
Materials:
Methodology:
Table 1: Quantitative Feedback Summary from Wireframe Walkthroughs (n=5 Clinical Stakeholders)
| Feedback Metric | Pre-Iteration (Average Score 1-5) | Post-Iteration 1 (Average Score 1-5) | Post-Iteration 2 (Average Score 1-5) |
|---|---|---|---|
| Ease of Navigation | 2.8 | 3.6 | 4.4 |
| Clarity of Layout | 3.0 | 3.8 | 4.6 |
| Alignment with Clinical Workflow | 2.6 | 3.8 | 4.4 |
| Overall Usability Perception | 2.8 | 3.7 | 4.5 |
Objective: To assess the usability and functional logic of an interactive, low-fidelity prototype before any code is written.
Materials:
Methodology:
Table 2: Usability Test Results for Low-Fidelity Prototype (n=6 Participants)
| Task Description | Success Rate (%) | Avg. SEQ (1-7) | Critical Errors Encountered |
|---|---|---|---|
| Log a new patient symptom | 100% | 6.5 | 0 |
| Report a CTCAE-gradeable adverse event | 83% | 5.2 | 1 (Difficulty finding grading scale) |
| Generate a standard therapy response report | 67% | 4.3 | 2 (Confusion over report parameters) |
Given the constrained resources in clinical research, a systematic approach to feature prioritization is paramount. The RICE scoring model provides a quantitative framework.
RICE Score = (Reach × Impact × Confidence) / Effort
Table 3: RICE Prioritization for a Hypothetical Cancer QI Tool
| Feature Idea | Reach (users/quarter) | Impact (1-3 scale) | Confidence (%) | Effort (person-months) | RICE Score |
|---|---|---|---|---|---|
| Automated CTCAE v6.0 grading | 500 | 3.0 | 100% | 4 | 375.0 |
| EHR Bi-directional Integration | 500 | 2.0 | 80% | 12 | 66.7 |
| Patient-Reported Outcome (PRO) Portal | 1000 | 2.0 | 100% | 8 | 250.0 |
| Customizable Dashboard Widgets | 200 | 1.0 | 50% | 3 | 33.3 |
Title: UCD Workflow: Wireframing to Prototype
Title: RICE Scoring Model Components
Title: MoSCoW Prioritization Framework
Table 4: Key Research Reagent Solutions for Prototyping Cancer QI Tools
| Item / Solution | Function / Explanation |
|---|---|
| Figma / FigJam | A collaborative, web-based platform for creating wireframes, low-fidelity mockups, and interactive prototypes. Essential for distributed team collaboration. |
| User Story Map | A visual artifact that organizes user stories into a logical workflow model, ensuring feature development aligns with the complete user journey. |
| RICE Scoring Sheet | A quantitative model (Spreadsheet) for prioritizing features based on Reach, Impact, Confidence, and Effort, reducing subjective bias. |
| MoSCoW Method | A prioritization framework for categorizing features into Must-haves, Should-haves, Could-haves, and Won't-haves, crucial for managing scope. |
| Think-Aloud Protocol | A usability testing methodology where participants verbalize their thought process, providing direct insight into user cognition and interface problems. |
| System Usability Scale (SUS) | A reliable, 10-item questionnaire for measuring the perceived usability of a system. Provides a quick, standardized usability score. |
This document presents a synthesis of successful applications of user-centered design principles in cancer quality improvement (QI) tools. Framed within a broader thesis on user-centered design, these case studies demonstrate how quantitative evaluation, structured implementation, and accessible design are critical for developing effective cancer care tools for researchers, scientists, and drug development professionals. The integration of rigorous data collection and adherence to usability standards ensures that these tools meet the complex needs of both clinicians and patients.
The Extension for Community Healthcare Outcomes (ECHO) model, developed by the University of New Mexico, was utilized by the American Cancer Society (ACS) to address cancer-related knowledge gaps among healthcare professionals in underserved communities. This virtual telementoring program creates a collaborative "all-teach, all-learn" environment that connects community providers with specialist mentors, thereby increasing local expertise and improving patient care without requiring patient travel [32].
Program Structure: Four distinct ACS ECHO programs were conducted between 2023 and 2024, focusing on various cancer care topics including tobacco cessation, colorectal cancer screening, prostate cancer screening, and caregiver needs [32].
Data Collection: A quantitative approach was employed using:
Analysis Methods: Descriptive statistics summarized quantitative survey data. Mean differences in knowledge and confidence were calculated by subtracting pre-program scores from post-program scores. Percentage data was derived by dividing participants in each category by total sample size. All analyses were performed using Excel and GraphPad Prism software [32].
Table 1: ACS ECHO Program Participation and Outcomes (2023-2024)
| Program Characteristic | Program A | Program B | Program C | Program D | Aggregate |
|---|---|---|---|---|---|
| Cancer Focus | Lung | Colorectal | Prostate | All | - |
| Topic | Prevention | Screening | Screening | Caregiving | - |
| Program Length (months) | 4 | 7 | 9 | 7 | - |
| Number of Sessions | 4 | 7 | 9 | 7 | 27 |
| Unique Participants | 195 | 45 | 59 | 132 | 431 |
| Average Participants/Session | - | - | - | - | 20.15 |
| Participants Planning to Use Information Within 1 Month | - | - | - | - | 59% |
| Mean Knowledge Increase (5-point scale) | - | - | - | - | +0.84 |
| Mean Confidence Increase (5-point scale) | - | - | - | - | +0.77 |
The quantitative evaluation demonstrated statistically significant improvements in knowledge and confidence among participants. The "all-teach, all-learn" approach successfully created a collaborative environment that fostered professional development. The study highlighted that quantitative methods provide an objective approach to evaluating model implementation and program impact, addressing a gap in previous predominantly qualitative evaluations of ECHO programs [32].
Future Health Today (FHT) is a quality improvement and clinical decision support (CDS) tool implemented in general practice to assist with appropriate follow-up of patients at risk of undiagnosed cancer. This complex intervention addresses the challenge of timely cancer detection in primary care, where initial presentations and routine blood tests are critical for determining whether patients require further investigation [35].
Tool Development and Integration: FHT was integrated within the general practice electronic medical record (EMR) and consisted of:
Cancer Module Algorithms: The FHT cancer module employed three central algorithms to flag patients with abnormal blood test results associated with increased risk of undiagnosed cancer:
Implementation Strategy: A multifaceted implementation approach included:
Evaluation Framework: A process evaluation was conducted alongside a pragmatic cluster-randomized controlled trial to understand implementation gaps, explore differences between general practices, and contextualize effectiveness outcomes. Data collection included semistructured interviews, usability and educational session surveys, engagement metrics, and technical logs [35].
The process evaluation revealed critical insights for user-centered design:
These findings underscore the importance of designing flexible tools that accommodate varying practice contexts and resource constraints while minimizing workflow disruption.
Driven by technological advancements and emerging high-throughput molecular data, cancer biology has evolved into a more quantitative discipline. This framework supports the translation of laboratory findings into clinically relevant applications and therapeutics through standardized quantitative approaches [36].
Modeling Drug Dose Response: Quantitative chemical biology research utilizes mathematical models to quantify biological processes and chemical effects on biological systems. Key methodologies include:
Michaelis-Menten Enzyme Kinetics:
IC50 Determination for Inhibitors:
Table 2: Essential Research Materials and Their Applications in Quantitative Cancer Biology
| Research Reagent/Tool | Function/Application | Experimental Context |
|---|---|---|
| Cell Titer Glo (CTG) | Measures viable cell ATP levels to quantify cellular viability | Phenotypic/cell-based inhibitor response assays [36] |
| Patient-Derived Cell Lines | Provides tractable in vitro system for drug screening | High-throughput screening of compound libraries [36] |
| Patient-Derived Xenografts (PDXs) | Maintains tumor microenvironment for drug response modeling | Preclinical drug efficacy studies [36] |
| Purified Protein-Ligand Binding Assays | Measures direct chemical-protein interactions | Target-based drug screening [36] |
| Parametric Proportional Hazard Survival Model | Analyzes time course of overall survival | Model-based meta-analysis of clinical outcomes [37] |
CDS Workflow for Cancer Risk Identification
Quantitative Drug Screening Pipeline
These case studies demonstrate that successful cancer quality improvement tools share common characteristics: rigorous quantitative evaluation, thoughtful implementation strategies that address workflow integration, and user-centered design that accommodates varying contexts and resources. The integration of quantitative frameworks with practical clinical tools creates a powerful paradigm for advancing cancer care from prevention through survivorship.
Future development should focus on creating more adaptable tools that can be tailored to specific practice environments while maintaining rigorous evaluation standards. Additionally, further research is needed to optimize the balance between comprehensive functionality and usability within the constraints of busy clinical settings.
The development of the OncoSupport+ digital health application exemplifies a successful user-centered design approach for supportive cancer care. Researchers conducted a participatory study involving patients with cancer, survivors, healthcare professionals, and patient advocates to co-design an application that directly addresses workflow integration challenges in clinical oncology settings [3].
The co-design process was structured into three distinct phases:
This structured participatory approach resulted in an application consisting of two integrated dashboards: (1) a patient dashboard for recording patient-reported outcome measures (PROMs) and accessing personalized supportive care information, and (2) a nurse dashboard for visualizing patient data during nursing consultations [3].
Table: Healthcare Workflow Automation Challenges and Impact (2025)
| Challenge Area | Current Status | Projected Impact | Automation Solution |
|---|---|---|---|
| Clinical Staff Shortages | 47.8% of hospitals report >10% vacancy rates [38] | Projected 10% RN shortage by 2026 (350,540 positions) [38] | Automated staffing adjustments based on patient acuity |
| Administrative Burden | 15-30% of U.S. healthcare spending is administrative ($285-570 billion) [38] | Physicians spend >50% of workdays on EHR [38] | AI-powered smart documentation and automated clinical notes |
| Revenue Cycle Inefficiency | Over 35% of healthcare organizations use RPA for revenue cycle [38] | Predictive claim denial prevention and automated appeals [38] | Real-time prior authorization and billing error detection |
Table: Global Data Protection Regulations Impacting Cancer Research
| Regulation | Key Provisions | Research Impact | Compliance Strategies |
|---|---|---|---|
| EU GDPR | Strict consent requirements, data anonymization mandates [39] | 39% decline in pharma R&D spending 4 years post-implementation [40] | Differential privacy, federated learning, dynamic consent platforms [39] [41] |
| U.S. DOJ Rule (2025) | Restrictions on bulk sensitive data access by "covered persons" [42] | Prohibits access to human genomic data for >100 US persons by entities in countries of concern [42] | Regulatory approval exemptions, de-identification, contractual restrictions on onward transfers [42] |
| HIPAA | De-identification standards, minimum necessary disclosure [40] | Creates hurdles for large-scale data collection and sharing [40] | Model data use agreements, single institutional review boards for multi-site studies [40] |
Purpose: To enable collaborative AI model training across multiple institutions without sharing raw patient data, addressing both privacy concerns and data siloing challenges in cancer research.
Materials:
| Reagent/Solution | Function | Application Context |
|---|---|---|
| Differential Privacy | Adds calibrated noise to query responses to prevent re-identification [40] | Protecting genetic data in rare disease studies with small population sizes [41] |
| Homomorphic Encryption | Enables computation on encrypted data without decryption [40] | Secure analysis of genomic data across institutions without exposing raw data |
| Secure Enclaves | Isolated processing environments that protect code and data [40] | Clinical trial data analysis while maintaining confidentiality |
| Federated Learning Framework | Distributed machine learning approach where models move to data rather than data to models [39] | Multi-institutional cancer studies without transferring sensitive patient data |
| Dynamic Consent Platforms | Enables granular patient control over data usage preferences [41] | Ongoing consent management for longitudinal cancer studies and AI-driven analysis |
Procedure:
Model Initialization:
Federated Training Cycle:
Validation and Testing:
Purpose: To facilitate global cancer clinical trials while complying with the 2025 DOJ data rules and international data protection regulations.
Materials:
Procedure:
Data Preparation:
Secure Transfer Implementation:
Compliance Documentation:
Purpose: To responsibly integrate AI tools into clinical cancer workflows while addressing ethical concerns and maintaining clinical autonomy.
Materials:
Procedure:
AI Model Validation:
Workflow Integration:
Continuous Monitoring and Evaluation:
Purpose: To establish a robust technical infrastructure supporting AI-driven cancer research while maintaining data privacy and workflow efficiency.
Materials:
Implementation Guidelines:
Computational Infrastructure:
Security and Compliance:
Interoperability Standards:
The protocols and application notes presented demonstrate that overcoming hurdles in cancer quality improvement tools requires an integrated approach addressing workflow, privacy, and technical challenges simultaneously. By adopting user-centered design principles, implementing privacy-enhancing technologies, and establishing robust technical infrastructure, researchers can develop transformative cancer tools that both advance scientific discovery and maintain ethical responsibility.
Designing digital tools for cancer quality improvement demands a rigorous, user-centered approach that prioritizes equity and accessibility. Researchers and drug development professionals must ensure that these critical tools are usable and effective for all populations, including older adults, individuals from diverse cultural backgrounds, and users with disabilities. Disparities in cancer care remain pervasive, often driven by socioeconomic, racial, and insurance-related inequities [45]. A failure to address accessibility can exclude users, create frustration, and lead to legal issues under laws like the Americans with Disabilities Act (ADA) [46]. This document provides detailed application notes and experimental protocols to embed equity into the design process, ensuring that digital cancer research tools are inclusive by design.
Adherence to established technical standards is the baseline for accessible design. The Web Content Accessibility Guidelines (WCAG) serve as the foundational framework.
Color contrast is a critical factor in making content readable for users with visual impairments or color blindness. The following table summarizes the key WCAG 2.1 Level AA requirements [46] [47]:
Table 1: WCAG 2.1 Level AA Color Contrast Requirements
| Element Type | Definition | Minimum Contrast Ratio | Examples |
|---|---|---|---|
| Normal Text | Text smaller than 18 point or 14 point bold. | 4.5:1 | Body text, labels. |
| Large Text | Text that is 18 point or larger, or 14 point and bold. | 3:1 | Headings, large call-to-action text. |
| User Interface (UI) Components | Visual information required to identify components and states. | 3:1 | Button borders, icons, focus indicators, form field outlines [48]. |
| Graphical Objects | Parts of graphics required to understand content. | 3:1 | Charts, graphs, diagrams, infographics [48]. |
Beyond contrast, a holistic approach is necessary. The following table outlines key considerations for the target user groups:
Table 2: Key User Group Considerations & Design Responses
| User Group | Key Considerations | Design Responses & Requirements |
|---|---|---|
| Older Adults | May have declining vision, fine motor skills, and cognitive abilities [49]. Higher prevalence of multimorbidity and frailty. | Prioritize clear information architecture; use larger, legible fonts; simplify complex navigation; capture outcomes like quality of life and independence [49]. |
| Users with Visual Impairments | Includes low vision, color blindness (affecting ~300M people globally), and contrast sensitivity [47]. | Strictly adhere to WCAG contrast ratios; never use color as the sole means of conveying information; provide text alternatives for graphics. |
| Culturally Diverse Users | Color symbolism and associations vary significantly across cultures [50] [51]. | Research cultural connotations of colors and imagery; support multiple languages; involve diverse users in co-design processes. |
This protocol provides a methodology for testing the color contrast of user interface components, a requirement under WCAG 1.4.11 [48].
Objective: To verify that all interactive UI components (buttons, form fields, icons) and their states (default, focus, hover) meet a minimum 3:1 contrast ratio against adjacent colors.
Materials:
Procedure:
Logical Workflow: The following diagram outlines the sequential and iterative process for validating color accessibility.
This protocol outlines a methodology for moving beyond tokenistic involvement to genuine partnership with older adults, ensuring research tools and endpoints reflect their priorities [49].
Objective: To engage older adults and their caregivers as co-investigators in the design of cancer quality improvement tools, ensuring outcomes and usability align with what matters most to this population.
Materials:
Procedure:
Logical Workflow: The following diagram illustrates the cyclical, integrated process of co-design.
This section details essential tools and materials for implementing the protocols and ensuring equitable, accessible design.
Table 3: Essential Research Reagents & Tools for Accessible Design
| Tool / Solution Name | Function | Application in Protocol |
|---|---|---|
| Colour Contrast Analyser (CCA) | Desktop application that measures contrast ratios between two colors for any on-screen element against WCAG 2.1 standards. | Protocol 1: Primary tool for validating 3:1 contrast for UI components and graphical objects [46]. |
| Stark Plugin (Figma) | Integrated plugin for design software that checks contrast, simulates color blindness, and suggests accessible color palettes in real-time. | Protocol 1: Used during the design phase to prevent contrast issues before implementation [46]. |
| Color Blindness Simulators (e.g., Colour Oracle, Coblis) | Software tools that simulate how designs appear to users with various forms of color vision deficiency (e.g., deuteranopia, protanopia). | Protocol 1: Used to validate that color is not the sole means of conveying information and that charts/graphs remain decipherable [52]. |
| WAVE (Web Accessibility Evaluation Tool) | A browser extension that scans web pages for accessibility violations, including low-contrast text, and provides visual feedback. | Protocol 1: Automated testing of live web tools to identify low-contrast text and other common errors [46]. |
| Patient Partner Compensation Framework | A budgeted, institutional policy for financially compensating patient partners for their time and expertise as co-investigators. | Protocol 2: Critical for establishing equitable partnerships and acknowledging the value of patient contributions [49]. |
| Accessible Consent Form Templates | Pre-designed consent forms that use plain language, high contrast, large fonts, and clear navigation to ensure comprehension. | Protocol 2: Used to ensure the research process itself is accessible to participants with diverse abilities and health literacy levels [49]. |
The principles and protocols outlined above have direct and critical implications for the development of digital tools in cancer research.
By integrating these application notes and protocols, researchers and drug development professionals can create cancer quality improvement tools that are not only compliant with standards but are truly equitable, accessible, and effective for the diverse populations they serve.
The integration of artificial intelligence (AI) and data-driven tools into cancer care presents unprecedented opportunities for improving quality and personalizing treatment. However, these technologies introduce significant ethical challenges, including algorithmic bias, data privacy concerns, and potential threats to patient autonomy. The embedded ethics approach has emerged as a transformative methodology for proactively addressing these challenges by integrating ethicists directly into research and development teams. This application note provides detailed protocols for implementing embedded ethics within cancer quality improvement initiatives, emphasizing practical strategies for bias mitigation, fairness preservation, and psychological safety. Drawing from real-world case studies in oncology research, we demonstrate how this collaborative framework enables the development of ethically responsible digital tools that maintain scientific rigor while protecting patient welfare and promoting equitable care delivery.
Embedded ethics represents a fundamental shift in how ethical considerations are integrated into technology development, moving from retrospective review to proactive, continuous collaboration. In the context of cancer quality improvement tools, this approach addresses the critical need to anticipate and resolve ethical challenges throughout the development lifecycle. Where traditional ethics frameworks operate as external checkpoints, embedded ethics positions ethicists as core team members who work iteratively with developers, clinicians, and researchers from project inception through implementation [53]. This integrated approach is particularly crucial for AI-driven cancer tools, where algorithmic decisions can directly impact patient diagnosis, treatment selection, and survivorship care.
The transformative potential of embedded ethics lies in its ability to bridge the distinctive cultures of ethics and technology development. Where ethics emphasizes thorough critique and theoretical analysis, technology development often prioritizes concrete results and efficiency. Embedded ethics creates a structured collaboration that respects both perspectives, enabling the identification of ethical issues that might otherwise remain hidden until clinical deployment [53]. For cancer researchers and drug development professionals, this approach provides a practical methodology for navigating the complex ethical landscape of predictive algorithms, patient-reported outcome systems, and clinical decision support tools while maintaining focus on improved patient outcomes.
Embedded ethics is characterized by three core principles that distinguish it from traditional ethical review processes:
Continuous Integration: Ethical consideration becomes an ongoing process throughout development rather than occurring at discrete review points. This continuous engagement allows for early identification of potential concerns and more effective mitigation strategies [53].
Interdisciplinary Collaboration: Ethicists, developers, clinicians, and patients collaborate as equal partners, each contributing unique expertise to address complex challenges that transcend individual disciplines [54].
Proactive Anticipation: The approach focuses on anticipating ethical issues before they manifest in deployed systems, enabling preventative design rather than retrospective correction [53].
The theoretical foundation of embedded ethics aligns with and extends several established frameworks, including Value Sensitive Design (VSD), which systematically considers human values throughout the design process [55]. VSD employs conceptual, empirical, and technical investigations to identify stakeholder values and translate them into design requirements, providing a structured methodology for implementing embedded ethics principles [55].
Table 1: Critical Ethical Dimensions in Cancer Quality Improvement Tools
| Ethical Dimension | Definition | Relevance to Cancer Tools |
|---|---|---|
| Algorithmic Fairness | Absence of systematic discrimination in algorithmic outputs | Ensures equitable performance across diverse patient demographics [56] [57] |
| Transparency | Understandability of system logic and operations | Enables clinical validation and appropriate trust in decision support [53] |
| Data Privacy | Protection of sensitive patient information | Maintains confidentiality of oncology health records and genetic data [54] |
| Accountability | Clear assignment of responsibility for system outcomes | Establishes protocols for addressing errors in diagnostic or prognostic tools [53] |
| Psychological Safety | Environment where team members feel safe expressing concerns | Facilitates open discussion of ethical concerns within development teams [54] |
Successful implementation of embedded ethics begins with thoughtful team construction and role definition:
Protocol 3.1.1: Ethics Team Integration
Protocol 3.1.2: Stakeholder Mapping
Table 2: Bias Taxonomy in Cancer AI Systems
| Bias Category | Definition | Detection Methods | Mitigation Strategies |
|---|---|---|---|
| Data Bias | Systematic skew in training data representation | Statistical analysis of dataset demographics [56] | Data augmentation, stratified sampling [57] |
| Algorithmic Bias | Fairness issues introduced by model architecture | Fairness metrics calculation across subgroups [57] | Adversarial debiasing, regularization techniques [57] |
| Interaction Bias | Bias emerging from human-system interaction | Usability testing with diverse user groups [56] | Interface redesign, user education protocols [55] |
| Temporal Bias | Performance degradation due to practice evolution | Monitoring model drift over time [56] | Continuous learning frameworks, scheduled retraining [56] |
Protocol 3.2.1: Comprehensive Bias Assessment
Protocol 3.3.1: Team Psychological Safety Framework
The following diagram illustrates the continuous, iterative process of embedded ethics implementation:
Table 3: Essential Research Reagents for Embedded Ethics Implementation
| Tool Category | Specific Tools | Application Context | Implementation Considerations |
|---|---|---|---|
| Fairness Assessment | AI Fairness 360 Toolkit, Fairlearn, Aequitas | Bias detection in classification models [57] | Requires predefined protected attributes and fairness definitions |
| Stakeholder Engagement | Persona development templates, Journey mapping worksheets, Co-design workshop guides | Identifying values and needs of diverse stakeholders [17] [3] | Must include vulnerable and marginalized patient groups |
| Ethical Analysis | Ethical matrix, Value hierarchy worksheets, Case analysis templates | Structured analysis of ethical dilemmas [53] [54] | Benefits from facilitation by trained ethicist |
| Psychological Safety | Team safety assessment surveys, Anonymous feedback systems, Conflict resolution protocols | Creating environment for open ethical discussion [54] | Requires leadership commitment and regular reinforcement |
The 4D PICTURE project provides a compelling case study of embedded ethics in oncology research. This European consortium aims to develop data-driven decision support tools (DSTs) for breast cancer, prostate cancer, and melanoma care path redesign [54]. The project integrated embedded ethics through several key protocols:
Implementation Approach:
Key Outcomes:
The project demonstrated that embedded ethics requires approximately 15-20% time commitment from core ethics personnel but results in more robust ethical integration and potentially reduces costly redesign cycles post-development.
The embedded ethics approach represents a paradigm shift in how cancer quality improvement tools are developed, moving ethical consideration from an external review process to an integrated, continuous collaboration. By implementing the protocols outlined in this application note, research teams can more effectively address complex challenges around algorithmic bias, fairness, and psychological safety while developing tools that are both scientifically rigorous and ethically sound.
Future development in embedded ethics will likely focus on standardized metrics for evaluating ethical integration success, specialized training programs for ethicists working in cancer domains, and automated tools for continuous bias monitoring in production systems. As AI technologies become increasingly sophisticated and pervasive in cancer care, the embedded ethics approach provides a essential framework for ensuring these powerful tools developed responsibly and focused squarely on improving patient outcomes.
Implementing a successful cancer data ecosystem requires balancing technical excellence with human-centered design principles. Systems must be Complete (avoiding missing data), Consistent (ensuring semantic and scope uniformity throughout study timelines), and Correct (identifying outliers and duplicates while considering their potential significance) [58]. This foundation enables researchers to aggregate data efficiently while supporting accurate filtering, selection, and calculation operations essential for cancer research.
The integration of User-Centered Design (UCD) creates an iterative process where designers focus on users and their needs throughout each design phase [59]. This approach involves understanding the context in which users may use a system, specifying their requirements, developing solutions, and evaluating outcomes against users' context and requirements [59]. For cancer researchers, this translates to systems that align with actual workflow patterns rather than imposing artificial technical constraints.
Table 1: Data Quality Metrics for Cancer Research Systems
| Quality Dimension | Target Threshold | Measurement Method | Validation Protocol |
|---|---|---|---|
| Data Completeness | >95% for core fields | Statistical analysis of missing values | Comparison against minimum dataset requirements |
| Record Duplication | <0.1% duplicate rate | Algorithmic matching across identifiers | Manual review of potential duplicates |
| Temporal Consistency | 100% format uniformity | Time-series analysis of data entry patterns | Audit of semantic consistency across study periods |
| Code Standardization | >98% adherence to standards | Terminology service validation | Cross-check with NCI Thesaurus or mCODE specifications |
Purpose: To establish a reproducible workflow for identifying and resolving data quality issues in cancer research datasets, with particular attention to the specialized requirements of oncology data management [60].
Materials and Reagents:
Procedure:
Data Quality Audit
Anomaly Detection and Resolution
Terminology Standardization
Quality Verification
Troubleshooting: For datasets with >10% missing core elements, implement multiple imputation techniques rather than deletion. For terminology mapping conflicts >15%, convene clinical review panel to establish consensus mappings.
Modern cancer data interoperability requires implementing layered standards that support both general healthcare data exchange and oncology-specific requirements. The HL7 FHIR (Fast Healthcare Interoperability Resources) standard provides the foundational framework for health data exchange, with specific implementation guides for cancer research [61]. The Minimal Common Oncology Data Elements (mCODE) initiative defines approximately 30 FHIR profiles covering patient characteristics, disease details, genomics, cancer treatments, and outcomes [61].
The USCDI+ Cancer extension addresses specialized use cases for cancer registry data and research applications, while the Central Cancer Registry Reporting Content Implementation Guide specifies how to use the MedMorph reporting infrastructure for automated, standardized exchange of cancer surveillance data [61]. This standards ecosystem enables seamless data flow from electronic health records to research databases while maintaining semantic consistency.
Purpose: To optimize and digitize the workflow of multidisciplinary team (MDT) meetings in cancer care through implementation of an integrated information platform using the FHIR standard [62].
Materials:
Procedure:
Process Analysis
FHIR Resource Mapping
System Integration
Workflow Implementation
Validation Metrics: A successful implementation demonstrated a 60% reduction in process steps (from 83 to 33 steps) and decreased coordination time from 30 to 5 minutes per case [62].
Purpose: To establish a systematic approach for continuous user engagement and feedback incorporation that maintains long-term viability of cancer data systems [59] [63].
Materials:
Procedure:
Stakeholder Identification
Feedback Collection Framework
Feedback Analysis and Prioritization
Iterative Implementation
Validation: Successful implementation demonstrates increased user adoption rates, decreased support requests, and improved task completion times measured through standardized usability metrics.
Table 2: Essential Tools for Cancer Data System Implementation
| Tool Category | Specific Examples | Primary Function | Implementation Considerations |
|---|---|---|---|
| Terminology Services | NCI Thesaurus, SNOMED-CT, LOINC | Standardized coding of clinical concepts | Version control, mapping maintenance |
| FHIR Infrastructure | HAPI FHIR Server, IBM FHIR Server | FHIR resource storage and retrieval | Profile customization, API management |
| Data Quality Tools | OpenRefine, Talend, custom scripts | Anomaly detection and cleaning | Integration with research workflows |
| User Feedback Platforms | UXtweak, UsabilityHub, custom surveys | Collection and analysis of user input | Participant management, bias mitigation |
| Interoperability Testing | Inferno FHIR Validator, Touchstone | Conformance testing with standards | Automated testing integration |
Cancer data systems must implement rigorous privacy protections including de-identification of personal identifiers, compliance with HIPAA requirements, and ethical data handling practices [58]. The integration of AI technologies demands near-zero tolerance for errors while maintaining explainability and human oversight [64]. Systems should implement data governance policies that assign specific responsibilities for data quality and establish clear standards for data entry and handling [65].
Long-term engagement requires designing systems capable of evolving with changing research requirements, standards updates, and technological advancements. This includes implementing scalable architectures that can handle growing data volumes, establishing governance processes for standards updates, and creating knowledge preservation systems that document design rationales and evolution paths [66]. The investment in user-centered design returns value through increased adoption, reduced error rates, and more efficient research processes [59].
Within user-centered design for cancer quality improvement tools, robust validation frameworks are essential for ensuring that digital solutions are effective, efficient, and satisfactory for their intended users. The System Usability Scale (SUS) and Mobile App Rating Scale (MARS) represent two established methodologies for quantifying usability and quality. The SUS provides a "quick and dirty" assessment of perceived usability through a standardized ten-item questionnaire [67] [68]. Developed by John Brooke in 1986, it has become a industry standard for evaluating everything from hardware to mobile applications [69] [68]. In parallel, the MARS was specifically developed to address the need for a reliable, multidimensional tool to classify and assess the quality of mobile health (mHealth) applications [70]. For cancer-focused tools, which often handle critical patient data and complex clinical workflows, employing these structured validation frameworks ensures that applications meet high standards of usability, functionality, and information quality before deployment in clinical or patient-facing contexts.
The SUS is a ten-item questionnaire using a five-point Likert scale from "Strongly Disagree" to "Strongly Agree" [69] [68]. It is designed to be administered immediately after a user has interacted with the system being evaluated. The instrument includes both positive and negative statements to prevent response bias [67]. To calculate the SUS score, follow this protocol:
Table 1: System Usability Scale (SUS) Questionnaire Items
| Item Number | Statement | Scale Direction |
|---|---|---|
| 1 | I think that I would like to use this system frequently. | Positive |
| 2 | I found the system unnecessarily complex. | Negative |
| 3 | I thought the system was easy to use. | Positive |
| 4 | I think that I would need the support of a technical person to be able to use this system. | Negative |
| 5 | I found the various functions in this system were well integrated. | Positive |
| 6 | I thought there was too much inconsistency in this system. | Negative |
| 7 | I would imagine that most people would learn to use this system very quickly. | Positive |
| 8 | I found the system very cumbersome to use. | Negative |
| 9 | I felt very confident using the system. | Positive |
| 10 | I needed to learn a lot of things before I could get going with this system. | Negative |
When implementing SUS for cancer quality improvement tools, researchers should adhere to the following experimental protocol:
Participant Recruitment: Recruit a minimum of 15-20 participants representing the target user groups (e.g., oncology clinicians, cancer patients, researchers) to achieve reliable results [69]. For quantitative studies with statistical power, larger sample sizes of at least 50-60 responses are recommended [69].
Contextual Administration: Administer the SUS immediately after participants have completed a standardized set of tasks with the cancer tool or after sufficient exposure to simulate real-world usage [69] [71]. The testing environment should reflect the actual context of use where possible.
Benchmarking: Compare obtained scores against established benchmarks. The widely accepted average SUS score is 68 (SD 12.5) [67] [71]. Scores above 68 are considered above average, with scores over 85 considered excellent [69]. A recent meta-analysis of digital health apps found a mean SUS score of 76.64, though this was skewed by particularly high-performing physical activity apps [67].
The MARS provides a comprehensive framework for evaluating the quality of mHealth applications across multiple dimensions. The original MARS contains 23 items across four objective quality subscales (Engagement, Functionality, Aesthetics, Information) and a subjective quality scale [70]. Each item is rated on a 5-point scale from 1 (Inadequate) to 5 (Excellent). The uMARS (user version) is a simplified 20-item end-user version with similar subscales but rewritten in plain English for easier comprehension by non-experts [72].
Table 2: Mobile App Rating Scale (MARS) Evaluation Dimensions
| Domain | Subscale Components | Sample Assessment Criteria |
|---|---|---|
| Engagement | Entertainment, Interest, Customization, Interactivity, Target Group | Fun, interesting, customizable content, interactive feedback, well-targeted |
| Functionality | Performance, Ease of Use, Navigation, Gestural Design | Accurate function, easy to learn, logical navigation, effortless touch interface |
| Aesthetics | Layout, Graphics, Visual Appeal | Clean layout, visually appealing, high-quality graphics |
| Information | Accuracy, Goals, Quality, Quantity, Visual Information, Credibility | High-quality information, from credible source, logically structured |
| Subjective Quality | Overall rating, Recommendation frequency, Would pay for app | Overall star rating, willingness to recommend, perceived impact |
When implementing MARS for cancer-related mobile applications, follow this experimental protocol:
Rater Training: For the professional MARS, train 2-3 independent raters with expertise in mHealth and the relevant health domain. Training should include practice ratings with sample applications to establish consistency [70]. The uMARS requires minimal training as it is designed for end-users [72].
Application Testing: Raters should interact with the cancer application for a minimum of 10-15 minutes to adequately explore all features and content [70]. For complex cancer tools with multiple modules, this may require extended interaction time.
Independent Rating: Each rater completes the MARS assessment independently, rating each item on the 5-point scale. The uMARS can be administered to multiple end-users (patients or clinicians) after sufficient app usage [72].
Score Calculation: Calculate mean scores for each subscale and an overall mean app quality score. The MARS has demonstrated excellent internal consistency (Cronbach alpha = 0.90) and interrater reliability (ICC = 0.79) [70].
The choice between SUS and MARS depends on the specific research objectives and context. The following table provides guidance for selecting the appropriate framework for cancer quality improvement tool validation:
Table 3: SUS vs. MARS Framework Selection Guide
| Evaluation Aspect | System Usability Scale (SUS) | Mobile App Rating Scale (MARS) |
|---|---|---|
| Primary Purpose | Overall perceived usability assessment | Comprehensive app quality evaluation |
| Administration Time | Quick (3-5 minutes) | Extended (15-20 minutes) |
| Output | Single usability score (0-100) | Multiple dimension scores + overall quality |
| Expertise Required | Minimal (end-users) | Moderate (trained raters for professional version) |
| Cancer-Specific Applications | Ideal for rapid comparison of multiple interface alternatives | Suitable for comprehensive pre-release quality assessment |
| Established Benchmarks | Mean: 68 (SD 12.5); Above 70 = Good; Above 85 = Excellent | Score >3.0 considered acceptable; >4.0 considered high quality |
Recent applications in cancer research demonstrate the utility of these frameworks. A 2022 study evaluating digital health apps found that while the overall mean SUS score was 76.64, this distribution exhibited asymmetrical skewness (-0.52) and was not normally distributed according to Shapiro-Wilk test (P=.002) [67]. Physical activity apps specifically drove this skewness with a mean SUS score of 83.28 (SD 12.39), while other health apps aligned closely with the standard SUS distribution (mean 68.05, SD 14.05) [67].
In a 2024 study focusing specifically on cancer mobile applications, researchers developed and validated a content quality evaluation tool consisting of 8 main themes (prevention, diagnosis, treatment, follow-up, education, communication, requests/order and other) with 43 question items [73]. The tool demonstrated high reliability with a Cronbach's alpha score of 0.967 [73].
A performance and usability evaluation of a mobile health data capture application in clinical cancer trials follow-up reported a mean SUS score of 87 points, indicating excellent usability in this specialized context [74].
Table 4: Essential Research Reagents for Usability Validation
| Research Reagent | Primary Function | Implementation Notes |
|---|---|---|
| System Usability Scale (SUS) | Quantifies perceived usability through 10-item questionnaire | Use for rapid assessment; ideal for A/B testing of interface alternatives |
| Mobile App Rating Scale (MARS) | Comprehensive evaluation of mHealth app quality across multiple domains | Employ for thorough pre-release assessment; requires trained raters |
| User Version MARS (uMARS) | End-user assessment of app quality | Simplified version for patient or clinician feedback; requires no special training |
| mHealth App Usability Questionnaire (MAUQ) | Specialized usability assessment for mHealth apps | Available in versions for standalone (18 items) and interactive (21 items) apps |
| NASA-Task Load Index (TLX) | Measures perceived workload across 6 dimensions | Suitable for complex cancer tools where cognitive load is a concern |
For comprehensive validation of cancer quality improvement tools, we recommend an integrated approach:
Formative Evaluation Phase: Employ SUS during iterative development cycles to compare design alternatives and identify usability issues early. The quick administration and scoring enables rapid iteration.
Summative Evaluation Phase: Implement MARS for comprehensive quality assessment before deployment. The multidimensional assessment ensures all aspects of app quality are addressed.
Comparative Benchmarking: Utilize established SUS benchmarks (mean 68, SD 12.5) [67] and MARS thresholds (score >3.0 acceptable, >4.0 high quality) [75] to contextualize results within the broader landscape of digital health tools.
Cancer-Specific Validation: Supplement with domain-specific instruments when available, such as the content quality evaluation tool for cancer applications [73], to address unique requirements of oncology contexts.
This integrated validation protocol ensures that cancer quality improvement tools meet both general usability standards and the specific needs of oncology applications, ultimately supporting the development of more effective digital interventions in cancer care and research.
Mixed-methods evaluation systematically integrates quantitative and qualitative data collection and analysis to address complex research questions that cannot be fully understood through a single methodological approach [76] [77]. In the context of user-centered design for cancer quality improvement tools, this approach provides both the statistical power of measurable outcomes and the rich, contextual understanding of user experiences, needs, and barriers [78] [3].
The fundamental value of mixed-methods research lies in its capacity to answer not only what is happening (through quantitative metrics) but also why it is happening and how it occurs in specific contexts (through qualitative feedback) [79] [80]. For cancer quality improvement tools, this means researchers can simultaneously measure tool effectiveness and understand the contextual factors influencing implementation success, such as workflow integration, user acceptance, and organizational barriers [78] [81].
Triangulation, where findings from different methods are compared and contrasted, strengthens research validity [76] [77]. When quantitative results showing improved patient outcomes align with qualitative reports of positive user experiences, confidence in the tool's value increases significantly. Conversely, when methods yield conflicting findings—such as when usage metrics are high but qualitative feedback reveals significant usability problems—researchers are prompted to investigate more deeply, often revealing important insights about user behavior and adaptation strategies [79].
Mixed-methods research employs specific designs that determine the sequencing, priority, and integration of quantitative and qualitative components. The choice of design should align with the research questions and practical constraints of the study context [77] [79].
Table 1: Mixed-Methods Research Designs and Applications in Cancer Care
| Design Type | Sequence | Rationale | Cancer Tool Application Example |
|---|---|---|---|
| Sequential Explanatory [79] | Quant → Qual | Explains quantitative results with qualitative insights | Analyze usage statistics, then conduct interviews to understand reasons for low adherence |
| Sequential Exploratory [79] | Qual → Quant | Explores concepts qualitatively before testing quantitatively | Identify user needs through focus groups, then develop and test prototypes with larger samples |
| Convergent Parallel [79] | Quant + Qual simultaneously | Obtains complementary data on same phenomenon | Collect survey data and conduct usability testing concurrently, then merge findings |
| Embedded Design [79] | One method nested within another | Answers different questions within single study | Main trial of tool effectiveness with nested qualitative study of implementation context |
Objective: To systematically measure usage patterns, clinical outcomes, and user engagement metrics for cancer quality improvement tools.
Materials: Digital analytics platforms, electronic health record systems, validated survey instruments (e.g., System Usability Scale [23], Mobile App Rating Scale [23]), structured data collection forms.
Procedure:
Table 2: Core Quantitative Metrics for Cancer Quality Improvement Tool Evaluation
| Metric Category | Specific Measures | Collection Method | Interpretation |
|---|---|---|---|
| Usage Metrics | Frequency of use, session duration, feature adoption | Digital analytics | Engagement level with tool |
| Usability Metrics | System Usability Scale (SUS) [23], task success rates, error counts | Surveys, performance testing | Perceived and actual ease of use |
| Clinical Outcomes | Symptom tracking accuracy, adherence rates, clinical guideline compliance | EHR extraction, self-report | Potential impact on care quality |
| Behavior Change Potential | App Behavior Change Scale (ABACUS) [23] | Structured evaluation | Capacity to modify health behaviors |
Objective: To understand user experiences, perceived benefits, implementation barriers, and contextual factors influencing cancer tool adoption and effectiveness.
Materials: Semi-structured interview guides, focus group protocols, audio recording equipment, transcription services, field note templates.
Procedure:
Mixed-Methods Evaluation Workflow: This diagram illustrates the sequential and parallel pathways for integrating quantitative and qualitative approaches in cancer quality improvement tool evaluation.
Table 3: Essential Research Materials for Mixed-Methods Evaluation of Cancer Quality Improvement Tools
| Tool/Resource | Primary Function | Application Context | Examples from Literature |
|---|---|---|---|
| System Usability Scale (SUS) [23] | Standardized usability assessment | Quantitative measurement of perceived usability | Used in mobile app evaluation for physical activity in cancer care [23] |
| Mobile App Rating Scale (MARS) [23] | Comprehensive app quality assessment | Quantitative evaluation of engagement, functionality, aesthetics, information | Applied in rating cancer physical activity apps [23] |
| App Behavior Change Scale (ABACUS) [23] | Measurement of behavior change potential | Quantitative assessment of goal-setting, planning, self-monitoring features | Used to evaluate behavior change potential in cancer care apps [23] |
| Semi-Structured Interview Guides | Elicitation of user experiences and perspectives | Qualitative understanding of user needs, barriers, and facilitators | Employed in user needs assessment for chemotherapy toxicity management tool [78] |
| Think-Aloud Protocols [81] | Real-time usability problem identification | Qualitative identification of interface and workflow issues | Utilized in personal health record evaluation for childhood cancer survivors [81] |
| Focus Group Protocols | Capture of group perspectives and dynamics | Qualitative exploration of shared experiences and consensus views | Applied in co-design of digital health app for supportive cancer care [3] |
| Affinity Diagramming Methods [78] | Thematic analysis and pattern identification | Qualitative data synthesis and theme development | Used to analyze field observations and interviews in toxicity management tool design [78] |
| Unique Participant Identifiers | Data integration and longitudinal tracking | Connecting quantitative and qualitative data at individual level | Enables person-level mixed methods analysis across multiple data sources [82] |
Objective: To systematically combine and interpret quantitative and qualitative findings to generate comprehensive insights about cancer quality improvement tools.
Materials: Statistical analysis software (e.g., R, SPSS), qualitative analysis tools (e.g., NVivo, Dedoose), data integration frameworks, visualization tools.
Procedure:
The successful application of these mixed-methods protocols in cancer quality improvement research requires careful planning, methodological flexibility, and attention to both technical and human factors influencing tool implementation and effectiveness [78] [3] [81]. By systematically integrating quantitative metrics with qualitative user feedback, researchers can develop a comprehensive understanding of how cancer quality improvement tools function in real-world settings and generate evidence-based recommendations for optimization and scaling.
This document provides application notes and protocols for assessing digital health tools beyond usability, focusing on their impact on behavior change, clinical outcomes, and quality of life (QoL) within cancer care. The framework integrates human-centered design (HCD) principles with rigorous implementation science methodologies to evaluate how digital interventions improve patient-centered outcomes in real-world settings [83] [84]. As evidence confirms that cancer patients' quality of life measures can predict survival, comprehensive assessment frameworks become increasingly critical for research and development [85].
Digital health tools for cancer quality improvement should be evaluated across multiple, interconnected domains. The table below summarizes core assessment areas, associated metrics, and example instruments derived from current research.
Table 1: Core Domains for Assessing Digital Health Tools in Cancer Care
| Assessment Domain | Primary Metrics | Example Instruments & Methods | Exemplary Findings from Literature |
|---|---|---|---|
| Usability & Design Quality | Effectiveness, efficiency, satisfaction, learnability [83] [84] | Intervention Usability Scale (IUS) [84], heuristic evaluation [83] | IUS ratings of 80.5/100 for an exposure therapy protocol indicated "good" usability with room for improvement [83]. |
| Behavior Change | Self-efficacy, behavioral intentions, engagement, motivation (COM-B) [86] [87] | Theory-driven surveys (e.g., self-efficacy), engagement analytics, the COM-B model [86] | The PREVENT tool led to greater increases in self-efficacy and vigorous activity among AYA cancer survivors [87]. |
| Clinical & Biomedical Outcomes | Cardiovascular health indicators, symptom burden, survival [87] [85] | American Heart Association's Life's Simple 7 [87], analysis of EHR data | A major study confirmed that baseline physical functioning, pain, and appetite loss are predictive of survival [85]. |
| Health-Related Quality of Life (HRQoL) | Physical, emotional, social functioning; overall health status [88] [85] | EORTC QLQ-C30 [88] [85], EQ-5D-5L [88] | HRQoL remains stable for most of a cancer patient's journey but deteriorates considerably in the last three months of life [88]. |
The following diagram illustrates a sequential, multi-method workflow for evaluating digital health tools, from initial user-centered design to the assessment of long-term outcomes.
This protocol adapts the Usability Evaluation for Evidence-Based Psychosocial Interventions (USE-EBPI) methodology for digital health tools [83].
Objective: To identify and prioritize usability issues that may impede adoption and effectiveness.
Procedure:
This protocol outlines a pilot RCT, as demonstrated by the PREVENT digital intervention, to assess impact on behavior and clinical outcomes [87].
Objective: To determine the feasibility and preliminary effectiveness of a digital health tool on behavior change mediators and clinical outcomes.
Procedure:
This protocol details methods for tracking HRQoL over time, a critical outcome in cancer care [88] [85].
Objective: To describe the trajectory of HRQoL in patients with cancer and understand its determinants, particularly near the end of life.
Procedure:
The following table catalogs key instruments and methodologies for developing and evaluating user-centered digital health tools in cancer research.
Table 2: Essential Research Instruments and Materials
| Item Name | Type/Format | Primary Function and Application |
|---|---|---|
| Intervention Usability Scale (IUS) [84] | Psychometric Scale (10-item questionnaire) | Quantifies the usability of complex interventions. Provides a total score (0-100) and subscales for "Usable" and "Learnable." |
| EORTC QLQ-C30 [88] [85] | Patient-Reported Outcome Measure (30-item questionnaire) | Assesses health-related quality of life in cancer patients across multiple domains, including physical functioning, symptoms, and global health status. |
| EQ-5D-5L [88] | Generic Preference-Based HRQoL Measure (5 dimensions + VAS) | Provides a utility score for health states used in cost-effectiveness analyses and allows comparison of burden across different diseases. |
| COM-B Model [86] | Theoretical Framework | Informs the design and evaluation of interventions by diagnosing barriers and enablers of behavior change through the lens of Capability, Opportunity, and Motivation. |
| USE-EBPI Methodology [83] | Qualitative & Quantitative Evaluation Framework | A structured 4-step method for identifying, organizing, and prioritizing usability issues in complex interventions via user testing and heuristic evaluation. |
| AHA Life's Simple 7 Algorithm [87] | Clinical Algorithm | Generates a cardiovascular health (CVH) profile and score from clinical and behavioral data to educate and motivate patients during clinical encounters. |
This application note provides a structured analysis of major oncology data platforms and digital health initiatives, framing their capabilities within a user-centered design framework for cancer quality improvement tools. We synthesize quantitative data and experimental protocols to offer researchers, scientists, and drug development professionals actionable methodologies for developing effective cancer care technologies. The analysis emphasizes participatory design principles, data integration architectures, and implementation strategies critical for creating tools that address real-world clinical and patient needs.
The integration of digital health tools into oncology represents a paradigm shift from generalized cancer treatment to precision care. Oncology data analytics—the systematic collection, processing, and analysis of cancer-related data—is revolutionizing patient outcomes, drug discovery, and treatment decisions [89]. However, a significant implementation gap persists; while technological capabilities advance, adoption remains limited by inadequate attention to user-centered design principles [3] [78]. This analysis examines leading platforms and initiatives through the lens of user-centered design to extract transferable methodologies for developing effective cancer quality improvement tools that are both technologically sophisticated and clinically implementable.
The oncology data platform landscape encompasses diverse solutions specializing in real-world evidence, clinical workflow integration, and patient-centered applications. The table below summarizes key platforms, their primary applications, and distinctive technological approaches.
Table 1: Comparative Analysis of Major Oncology Data Platforms and Digital Health Initiatives
| Platform/Initiative | Primary Application | Core Technology | Data Sources | User-Centered Features |
|---|---|---|---|---|
| Komodo Health [90] | Healthcare intelligence, care gap analysis | AI-powered Healthcare Map, Marmot analytics engine | 330M+ de-identified patient journeys, claims data | Transparent, verifiable insights; healthcare-native analytics |
| Flatiron Health [91] | Real-world evidence, research | EHR-integrated platform, analytics tools | Electronic health records, oncology-specific data | Workflow integration for academic cancer centers |
| Tempus [91] | Precision medicine, clinical decision support | AI, machine learning, molecular sequencing | Genomic data, clinical data, imaging data | Therapeutic matching, clinical trial identification |
| OncoSupport+ [3] | Supportive cancer care, symptom management | Patient and nurse dashboards, PROM collection | Patient-reported outcomes, clinical assessments | Co-designed with patients and clinicians; workflow integration |
| Lifebit [89] | Biomedical data analysis, precision medicine | Federated learning, AI-powered analytics | Genomic, transcriptomic, proteomic, clinical data | Privacy-preserving data analysis; multi-omics integration |
| bridges [78] | Chemotherapy toxicity management | Web-based tool, self-management advice | Patient-reported symptoms, clinical data | Just-in-time self-management advice; HCP communication |
| MSK iHub [92] | AI-driven drug discovery | AI algorithms, clinical data access | De-identified clinical datasets, research data | Mentoring from clinical experts; clinical workflow validation |
Platforms demonstrate divergent but complementary approaches to oncology data challenges. Komodo Health and Lifebit employ comprehensive data aggregation strategies, creating expansive datasets (Komodo's 330 million patient journeys [90]) and sophisticated analytics infrastructures (Lifebit's federated learning for multi-omics data [89]). In contrast, patient-focused tools like OncoSupport+ and bridges prioritize targeted functionality through intensive user involvement in development [3] [78].
A critical differentiator is workflow integration depth. EHR-integrated platforms (Flatiron, Cerner Oncology [91]) embed directly into clinical workflows, while standalone tools (bridges [78]) face implementation barriers despite robust functionality. This underscores the importance of considering integration capabilities early in the design process for cancer quality improvement tools.
Understanding the scale and impact of oncology data initiatives requires examination of quantitative performance indicators across commercial, clinical, and technical dimensions.
Table 2: Quantitative Performance Metrics of Oncology Data Technologies
| Metric Category | Specific Metric | Representative Values | Source/Example |
|---|---|---|---|
| Commercial Impact | Global oncology market value | Projected USD 903.81B by 2034 (CAGR 10.9%) | [93] |
| AI in oncology market | Projected growth from $2.4B (2025) to $9.1B (2035) | [93] | |
| Healthcare analytics market | Projected $85.9B by 2027 (CAGR 25.7%) | [94] | |
| Platform Scale | Patient data coverage | 330M+ de-identified patient journeys | Komodo Health [90] |
| Healthcare organization reach | 9 in 10 U.S. hospitals | symplr [90] | |
| Clinical Volume | Testing volume | 850,000+ tests processed (Q2 2025) | Natera [90] |
| Revenue impact | $546.6M revenue (Q2 2025, 32% YoY growth) | Natera [90] | |
| Data Scale | Genomic data volume | 2.5+ petabytes from 20,000 cancer samples | The Cancer Genome Atlas [89] |
| Future data projections | 1 zettabase of sequence data annually by 2025 | [89] |
This section provides detailed methodologies for implementing user-centered design approaches in cancer quality improvement tool development, derived from successful initiatives documented in the literature.
Based on the development of OncoSupport+ [3], this protocol outlines a structured approach to engaging stakeholders in digital health design.
4.1.1 Research Objectives and Ethical Considerations
4.1.2 Participant Recruitment and Sampling
4.1.3 Study Design and Phased Implementation The protocol implements a three-phase approach adapted from the framework of Noorbergen et al. [3]:
4.1.4 Data Collection Methods
4.1.5 Analysis Framework
Derived from the bridges toxicity management tool [78], this protocol focuses on integrating digital tools into existing clinical workflows.
4.2.1 Implementation Context Assessment
4.2.2 Prototype Development and Refinement Cycle
4.2.3 Implementation Readiness Assessment
This section details critical methodological components and their functions in developing user-centered cancer quality improvement tools, presented as a toolkit for researchers.
Table 3: Research Reagent Solutions for User-Centered Digital Health Development
| Research Component | Function | Implementation Example | Considerations |
|---|---|---|---|
| Affinity Diagramming [3] [78] | Thematic analysis of qualitative data through visual clustering | Grouping coded data from focus groups into natural themes | Requires multidisciplinary team; enables consensus-based theme development |
| Cognitive Walk-throughs [78] | Usability evaluation method using task scenarios | Participants complete realistic tasks while thinking aloud | Identifies usability issues; reveals mental models and expectations |
| Event-Focused Ethnography [78] | Contextual understanding of clinical workflows and challenges | Field observation of clinical interactions with follow-up interviews | Provides insights into real-world constraints and information flows |
| Co-Design Workshops [3] | Collaborative generation of design ideas with stakeholders | Scoring cards for feature prioritization; design thinking exercises | Engages diverse perspectives; builds stakeholder investment |
| Federated Learning Models [89] | Privacy-preserving analysis across distributed datasets | Training AI algorithms on data across institutions without transferring raw data | Addresses data privacy regulations; enables multi-institutional collaboration |
| Patient-Reported Outcome Measures [3] | Direct capture of patient symptoms and quality of life | Integration of EORTC QLQ-C30, NCCN Distress Thermometer | Requires validation for context; enables patient-centered assessment |
| Low-Fidelity Prototyping [78] | Rapid concept testing without development investment | Screen shots without interactive functionality | Facilitates early feedback; reduces development costs |
The following diagram maps the critical pathway from tool development to successful implementation, synthesizing factors identified across successful initiatives [3] [89] [78].
This comparative analysis yields critical lessons for developing user-centered cancer quality improvement tools. First, participatory design is non-negotiable for implementation success; tools developed without deep stakeholder engagement face fundamental adoption barriers regardless of technical sophistication [3] [78]. Second, workflow integration precedes value realization; tools must align with clinical workflows and information systems to avoid creating additional burden [90] [78]. Third, data interoperability and privacy preservation must be foundational design requirements rather than afterthoughts [89].
The protocols and frameworks presented provide actionable methodologies for developing the next generation of cancer quality improvement tools. By applying these structured approaches, researchers and drug development professionals can create digital health solutions that not only demonstrate technical innovation but also achieve meaningful adoption and impact in clinical practice, ultimately advancing cancer care quality and patient outcomes.
The integration of rigorous user-centered design is paramount for developing cancer quality improvement tools that are effective, equitable, and sustainable. The evidence synthesized from foundational principles to validation studies consistently demonstrates that engaging patients and clinicians throughout the development process leads to higher usability, better adoption, and improved health outcomes. Key takeaways include the necessity of iterative co-design, the importance of addressing ethical challenges proactively, and the value of robust, mixed-methods evaluation. Future efforts must focus on scaling these methodologies, improving data interoperability across platforms like CancerLinQ and genomic commons, and exploring the role of AI in personalized tool development. For biomedical and clinical research, this signifies a paradigm shift towards creating digitally-enabled, patient-centered cancer care ecosystems that are firmly grounded in the real-world needs of all stakeholders.