Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Developing an evaluation framework for public health environmental surveillance: Protocol for an international, multidisciplinary Delphi consensus study

  • Douglas G. Manuel ,

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – original draft, Writing – review & editing

    dmanuel@ohri.ca

    Affiliations Methodological and Implementation Research Program, Ottawa Hospital Research Institute, Ottawa, Ontario, Canada, Department of Family Medicine, University of Ottawa, Ottawa, Ontario, Canada, School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario, Canada, C.T. Lamont Primary Health Care Research Centre Program, Bruyère Research Institute, Ottawa, Ontario, Canada

  • Carol Bennett,

    Roles Conceptualization, Methodology, Writing – original draft, Writing – review & editing

    Affiliations Methodological and Implementation Research Program, Ottawa Hospital Research Institute, Ottawa, Ontario, Canada, Department of Family Medicine, University of Ottawa, Ottawa, Ontario, Canada

  • Emma Brown,

    Roles Methodology, Writing – original draft, Writing – review & editing

    Affiliation Methodological and Implementation Research Program, Ottawa Hospital Research Institute, Ottawa, Ontario, Canada

  • David L. Buckeridge,

    Roles Writing – review & editing

    Affiliation School of Population and Global Health, McGill University, Montreal, Quebec, Canada

  • Yoni Freedhoff,

    Roles Writing – review & editing

    Affiliation Department of Family Medicine, University of Ottawa, Ottawa, Ontario, Canada

  • Sarah Funnell,

    Roles Writing – review & editing

    Affiliation Department of Family Medicine, Faculty of Health Sciences, Queen’s University, Kingston, Ontario, Canada

  • Farah Ishtiaq,

    Roles Writing – review & editing

    Affiliation Tata Institute for Genetics and Society, National Centre for Biological Sciences, Bengaluru, Karnataka, India

  • Matthew J. Wade,

    Roles Writing – review & editing

    Affiliations Data Analytics and Surveillance Group, UK Health Security Agency, London, United Kingdom, School of Engineering, Newcastle University, Newcastle-upon-Tyne, United Kingdom

  • David Moher,

    Roles Methodology, Writing – review & editing

    Affiliations School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario, Canada, Centre for Journalology, Methodological and Implementation Research Program, Ottawa Hospital Research Institute, Ottawa, Ontario, Canada

  • on behalf of the PHES-EF Executive Group

    Membership of the PHES-EF Executive Group is provided in the Acknowledgments.

Abstract

Introduction

Public health environmental surveillance has evolved, especially during the coronavirus pandemic, with wastewater-based surveillance being a prominent example. As surveillance methods expand, it is important to have a robust evaluation of surveillance systems. This consensus study will develop an evaluation framework for public health environmental surveillance, informed by the expanding practice of wastewater-based surveillance during the pandemic.

Methods

The public health environmental surveillance evaluation framework will be developed in five steps. In Step 1, a multinational and multidisciplinary executive group will be formed to guide the framework development process. In Step 2, candidate items will be generated by conducting relevant scoping reviews and consultation with the study executive group. In Step 3, an international electronic Delphi will be conducted over two rounds to develop consensus on items for the framework. In Step 4, the executive group will reconvene to finalize the evaluation framework, discuss standout items, and determine the dissemination strategies. Lastly, Step 5 will focus on disseminating the evaluation framework to all parties involved with or affected by wastewater-based surveillance using traditional and public-oriented methods.

Discussion

The Delphi consensus study will provide multidisciplinary and multinational consensus for the evaluation framework, by providing a set of minimum criteria required for the evaluation of public health environmental surveillance systems. The evaluation framework is intended to support the sustainability of environmental surveillance and improve its implementation, reliability, credibility, and value.

Introduction

Public health relies on effective surveillance systems to support informed decision-making and timely interventions [1]. Wastewater-based surveillance, a form of environmental-based surveillance, has historically informed public health responses to threats like typhoid, polio, and illicit drug use [26]. More recently, the coronavirus pandemic (Covid-19) marked a turning point, expanding wastewater-based monitoring and surveillance to an unprecedented scale and underscoring its potential for wider public health applications [7]. Since then, wastewater-based surveillance, has supported public health practice for controlling monkeypox [8], influenza [9,10], antimicrobial resistance [11,12], and other pathogens and health risks beyond infectious diseases [13,14]. It is now used globally across diverse settings [15], from metropolitan to rural areas [16,17]. Cited advantages include its ability to track disease transmission independently of clinical testing availability or health-seeking behaviour, offering a non-invasive, resource-efficient surveillance method [18,19].

Current frameworks for evaluating surveillance performance require reassessment and updating. Established guidelines, such as the U.S. Centers for Disease Control and Prevention (CDC) 2001 framework, were developed before widespread environmental-based surveillance and do not adequately address unique challenges like environmental measurement variability or the need for broad, multidisciplinary collaboration [5,20,21]. Meanwhile, the pandemic prompted shifts in surveillance practices—such as increased public engagement and open-access data—and encouraged consensus-driven guidance through international, multidisciplinary collaboration. Yet no systematic review has examined how these developments influence surveillance evaluation, and there is no rigorous consensus on measuring the performance of wastewater-based surveillance and other forms of environmental-based surveillance.

This study aims to develop a comprehensive public health evaluation framework. Wastewater-based surveillance will be an example; however, it will apply to broader public health environmental surveillance systems, including air and surface monitoring. Surveillance targets considered include infectious agents and health risks beyond SARS-CoV-2.

Study objectives

This project will use a multidisciplinary and multinational consensus approach to develop an evaluation framework for public health environmental surveillance (Public Health Environmental Surveillance Evaluation Framework, PHES-EF). See https://PHES-EF.org. An evaluation framework provides structured guidance to facilitate a systematic approach to program evaluation [2224]. See below for the study’s definition of an evaluation framework.

Specific objectives for the framework development process are as follows: (1) engage a multidisciplinary executive group of international experts to inform the development of the evaluation framework; (2) review the literature to identify current guidance on surveillance evaluation items; (3) prioritize items for an evaluation framework for public health environmental surveillance using an international electronic-Delphi (e-Delphi) process; (4) conduct a consensus meeting, which engages experts across disciplines, to create an evaluation framework for public health environmental surveillance based on feedback from the international e-Delphi process; and, (5) develop a comprehensive dissemination plan for the evaluation framework.

Definitions

We define key terms as follows:

  • Environmental-based surveillance (environmental surveillance): Surveillance of substances or organisms in air, water, soil, and living organisms to detect changes or threats that may affect human health, ecosystems, or both [25].
  • Evaluation framework: A structured system or schema to assess a program systematically [1,22,24]. An evaluation framework presents a clear list of items or elements that should appear in an evaluation, such as constructs or concepts. It provides an organisation or structure for the list of items, defines how the list and organisation were developed, and potentially includes details about measuring the evaluation items. Importantly, an evaluation framework facilitates informed decision-making by providing a comprehensive understanding of program effectiveness and areas for improvement.
  • Public health surveillance: The ongoing systematic collection, analysis, interpretation and dissemination of health data for the planning, implementation and evaluation of public health interventions and strategies to protect and improve the health of populations [26].
  • Wastewater: Water that has been in contact with people (e.g., for washing) or used for cleansing and sanitation (e.g., for flushing away faecal matter), and is discharged via sewers or other sanitation systems [6].
  • Wastewater-based surveillance (wastewater surveillance): A public health approach that examines specific substances or organisms in wastewater (sewage). Wastewater surveillance can be referred to as wastewater-based epidemiology; however, the latter can be specific to using wastewater to study and analyse the distribution, patterns, and determinants of health and disease conditions in defined populations [26].

Materials and methods

Study design

This protocol was developed a priori. The development of PHES-EF is guided by recommendations for Delphi techniques in the health sciences [27]. Fig 1 provides an overview of the study flow.

Step 1: Establish the executive group

A multinational and multidisciplinary executive group will oversee the development of the evaluation framework. The group will bring expertise from various fields, including public health, infectious disease, epidemiology, environmental and physical sciences, mathematical sciences, social sciences, communication, knowledge translation and exchange, knowledge users, and an engaged public.

The composition of the study executive group, which includes content expert and knowledge user members, will be identified by examining relevant author lists, internet-based searching, and the study working group’s professional networks. In contrast, the engaged public member of the executive group will be identified through an open call for expressions of interest. The S1 Table details the composition of the study executive group.

The executive group will be responsible for providing feedback on the framework development process (e.g., framework organization, completeness, additions, redundancy), supporting and advising working groups in their area of expertise, connecting the study to other related projects, recruiting e-Delphi panellists, participating in the consensus meeting, disseminating the evaluation framework, advising and/or participating in post-publication activities, and promoting the adoption of the evaluation framework among their professional networks. Co-chairs will also be responsible for decision-making when the executive group cannot reach a consensus, leading executive group activities, and the overall framework development process.

Step 2: Generate candidate evaluation items

We will systematically compile a list of e-Delphi candidate items through scoping reviews, content analysis and extraction, and executive group consultation.

Scoping reviews

Two scoping reviews will be performed to provide a comprehensive view of public health environmental surveillance system evaluations. The first review will identify frameworks (e.g., sources which guide the planning, management, or conduct of processes) and criteria (e.g., principles on which a judgment may be based) for evaluating public health or environmental surveillance systems. The second review will explore insights from the Covid-19 pandemic, particularly focusing on the principles of equity, trust, open science, community engagement, and One Health in the context of wastewater surveillance [2830]. Detailed protocols for the scoping reviews are available on OSF (https://doi.org/10.17605/OSF.IO/GSKF6; https://doi.org/10.17605/OSF.IO/C9A3Q). The completed scoping reviews will be reported following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for scoping reviews (PRISMA-ScR) [31]. See S1 and S2 Appendices for associated PRISMA-ScR checklists.

Content analysis and extraction

We will conduct a content analysis on items identified during the data extraction phase for both scoping reviews. The content analyses will identify and group criteria for evaluating public health and environmental surveillance systems. In the first step, one team member will group all the extracted items into initial domains and themes. In the second step, two researchers will independently compare the category system of the first researcher to agree on the final category system consensually. In the third step, all criteria will be reviewed again to merge or modify duplicate concepts and create a parsimonious framework for the initial list of e-Delphi candidate items.

Executive group consultation

The executive group will review the initial list of e-Delphi candidate items and discuss the initial structure and presentation of Round 1 items. Their review will focus on creating an evidence-informed public health environmental surveillance framework centred around SARS-CoV-2 wastewater surveillance, and incorporating lessons learned from the pandemic (e.g., equity, open science, community engagement, and One Health). Additional items will be added if the executive group determines that key evaluation areas have not been addressed.

Step 3: International e-Delphi survey

A Delphi technique will be used to develop an internationally accepted evaluation framework for public health environmental surveillance. The Delphi method is an iterative multi-round approach that uses a series of sequential surveys, interspersed by controlled feedback, to elicit consensus among a group of individuals while maintaining anonymity [32,33]. An e-Delphi method will be used to overcome geographic barriers and allow us to engage panellists internationally across various time zones. Delphi surveys relating to health sciences have typically been conducted for 2–3 rounds [27,34,35]. Therefore, a two-round e-Delphi survey and a subsequent consensus meeting will be conducted to balance feasibility with rigour. Each round will be active for approximately 4–6 weeks to allow panellists sufficient time to participate in the survey while balancing study timeline constraints [36].

Recruitment of panellists

We will recruit a multinational, multidisciplinary panel of wastewater surveillance experts, knowledge users, and engaged members of the public to complete the e-Delphi survey. Content experts are adult individuals with knowledge in their respective subdiscipline groups (see S2 Table). Panellists will be selected to capture the multiple perspectives of those that influence the design, implementation, evaluation, use, and reporting of wastewater surveillance activities, including the following subdisciplines groups: public health, infectious disease, epidemiology; environmental and physical sciences; mathematical sciences; social sciences; and communication, knowledge translation and exchange. Targeted recruitment will also be conducted to identify knowledge users and engaged members of the public. We will use purposive sampling to identify potentially eligible panellists by examining relevant author lists from the scoping reviews, professional networks, and internet-based searching, among other potential methods. Potential panellists may also be recruited using social media, a study website or calls for expressions of interest during relevant meetings, presentations, or correspondence. Snowball sampling may also be employed by executive group members and potential panellists to maximize recruitment.

Potentially eligible panellists will receive a recruitment email from the study working group, executive group members, or other potential panellists. A call for panellists to share the e-Delphi with other potentially eligible panellists may be included in recruitment materials and the e-Delphi survey. Inclusion criteria for e-Delphi panellists are outlined in Table 1.

thumbnail
Table 1. Eligibility criteria for e-Delphi survey panellists.

https://doi.org/10.1371/journal.pone.0310342.t001

We will conduct an all-rounds e-Delphi invitation approach, meaning panellists will be invited to every round regardless of whether they participated in the previous round [37]. Previous Delphi studies have commonly ranged from 20 to 30 panellists; however, there is no agreed upon standard [35,38]. Studies have also shown that the number of panellists and their diversity can impact the quality of outcomes [38]. Therefore, assuming only half of registered panellists will participate in at least one round, we will aim to recruit at least 50 panellists. Preferably, with at least 8–10 people per discipline subgroup. The study working group will monitor the distribution of registered panellists based on their demographic information and will try to distribute appropriately across discipline subgroups and other demographic markers.

Registration to complete the e-Delphi will open approximately four weeks before the first round, and e-Delphi round surveys will be active for roughly 4–6 weeks. The tentative start and end dates for recruitment are 08/09/2025 and 19/12/2025, respectively. Recruitment or retention efforts will occur throughout each e-Delphi round and the four weeks of targeted recruitment before Round 1. Email reminders to complete the survey will be sent twice per e-Delphi round. Panellists may also receive up to two reminders to complete the screening questionnaire during the targeted recruitment period. The time required to complete each e-Delphi round survey will be approximately one hour. The estimated time commitment will be provided to all potential panellists during recruitment and before starting the consensus process.

Panellist withdrawal or termination

Panellists will not be able to withdraw from the study due to the safeguards in place to maintain anonymity; however, they may choose to stop actively participating in the study at any time.

Panellist compensation

Content experts and knowledge users who complete all rounds of the e-Delphi process will receive a $100 CAD pre-paid Visa card in compensation for the time required to complete the e-Delphi process. Engaged members of the public who complete all rounds will be entered to win one of five $100 CAD pre-paid Visa cards. Remuneration has been shown to encourage participation and retention from diverse groups, thereby minimizing bias, and motivating thoughtful, high-quality responses, thereby improving data quality [39].

Procedure

We will conduct a two-round e-Delphi survey to generate consensus on evaluation criteria (see Fig 2). The full survey will be pre-tested and validated before administration. A pilot test of the survey will be conducted with up to five test panellists. Test panellists may include members of the executive group or other potential panellists matching our recruitment criteria. The pilot test will focus on assessing the clarity and relevance of background information provided for candidate items. The survey will be revised as necessary following feedback from pilot testing. Summaries of Round 1 will be compiled for the subsequent round.

thumbnail
Fig 2. E-Delphi and consensus meeting methodology, methods to achieve consensus on core items for PHES-EF.

Consensus: if ≥ 70% of panellist votes fall within the same category (irrelevant, equivocal, relevant); no consensus if <70% agreement. Item relevancy: core item: highly or moderately relevant; further discussion: slightly relevant, neutral, slightly irrelevant; excluded item: moderately or highly irrelevant.

https://doi.org/10.1371/journal.pone.0310342.g002

We will use the Surveylet Delphi platform to administer the survey [40]. The platform facilitates interactive dialogue between panellists by allowing them to justify their responses and comment on the responses of others. Panellists can also save their progress and answer survey questions across multiple sessions. Custom survey pathways will be generated for each panellist discipline subgroup (i.e., panellists from different discipline subgroups will be shown a different collection of candidate items). Within each custom discipline subgroup survey pathway, panellists can skip items or self-declare that they are not qualified to assess certain candidate items. To ensure the independent evaluation of items, the order of candidate items may be randomized for each panellist.

Individuals will complete the screening questionnaire separate from the e-Delphi platform using LimeSurvey [41]. The study working group will review eligibility, and if eligible to participate in the study, individuals will receive an individualized link for anonymous participation in the e-Delphi survey. Implied consent will be obtained electronically for all panellists using the Surveylet Delphi platform before participation in the survey. Panellists who are eligible for compensation or wish to receive acknowledgement for their contributions will be able to do so by clicking a link at the end of each survey round. The link will bring them to a separate LimeSurvey form that will not be linked to their survey responses, allowing for e-Delphi responses to remain anonymous.

Defining consensus

Candidate items will be assessed on the following four rating properties: (1) relevance and practical utility; (2) scientific rigor, validity and reliability; (3) feasibility, adaptability and resource implications; and (4) equity, inclusiveness, and mitigation of bias.

A 7-point Likert scale (1 = highly irrelevant to 7 = highly relevant) will be used to rate each property for all candidate items. A summary score for each candidate item will be created by calculating the median of the four property ratings. Panellist summary scores will then be categorized as excluded item (irrelevant: 1–2), further discussion (equivocal: 3–5), or core item (relevant: 6–7). Consensus for each item is defined as ≥ 70% of the panellist votes falling within the same category (1–2, 3–5, or 6–7).

The approach to use a 7-point fully-labeled Likert scale – a higher number of categories compared to many studies – is informed by the consideration that panellists are professionals and engaged members of the public with good cognitive skills; therefore, more categories and labels will be more discriminating and reproducible [4245]. The ≥ 70% consensus cut-off – a lower cut-off compared to many studies – is informed by the rapidly evolving nature of environmental surveillance for public health and the wide range of disciplines invited to the panel; therefore, a high level of agreement may not occur [4245].

The rating properties were synthesized from previous performance evaluation frameworks from public health and health systems [1,24,4651]. Equity is a property that was not previously identified but was added following an expert group discussion of potential properties and the study’s goals and objectives. See S3 Appendix for further details on the rating properties for e-Delphi candidate item scoring.

Stopping rules

When consensus is reached for a given item, no subsequent ranking rounds for that item will be performed. Consensus items that are considered equivocal, in addition to remaining non-consensual items, will be further deliberated during the consensus meeting.

Round 1

Panellists will be invited to rate their level of agreement with candidate items generated from scoping review results and consultation with the study executive group. Free-text boxes will be included for panellists to provide feedback or identify additional candidate items to be included in the next e-Delphi survey round.

Round 2

Regardless of whether they participated in the previous round, panellists will be invited to participate in Round 2 of the e-Delphi survey. All panellists will be invited to rate new items and re-rate previous items that did not reach consensus. When re-rating their level of agreement, panellists will be presented with their previous round scores alongside the aggregate group results. Anonymous feedback from Round 1 will also be compiled and presented during Round 2. Any newly suggested items during Round 2 will be deliberated during the consensus meeting.

Summary panellist characteristics

At the start of the e-Delphi survey, panellists will provide basic demographic details to ascertain their discipline subgroup, ensuring they receive the correct list of candidate items. To maintain anonymity, this demographic data will not contain identifiable information, and only aggregate responses of 5 or more panellists will be analyzed. We will report aggregate panel characteristics based on gender, discipline group or subgroup (depending on the number of panellists), level and type of experience, country income level, and World Bank work regions.

Quantitative analysis

We will report response rates, agreement levels, consensus, and stability across rounds using median, range, and IQR for each criterion item. The proportion of panellists who skipped items or selected “not qualified to respond” will be reported. Fisher’s exact test will be used to evaluate differences in item ratings by panellist characteristics: primary panellist disciplines (see S2 Table), country income level (high-income versus low-and-middle income), and World Bank world regions (see S3 Table). Fisher’s exact test was chosen for its suitability in analyzing small sample sizes and categorical data, as it does not rely on the assumption of large sample sizes required by other tests, such as the chi-squared test. This makes it a good choice for comparing item ratings across subgroups, particularly given the potential variability in panellist numbers across disciplines, income levels, and regions.

Qualitative analysis

Content analyses will be conducted on panellists’ feedback to extract common themes and insights. A summary will describe the evolution of each criterion across rounds, including reasons for item modification or elimination.

Grouping of subgroups

Subgroups with fewer than 5 panellists will be combined with similar groups based on panellist numbers and group likeness. Decisions on combined subgroups will be made prior to analyses of the items.

Exclusion of data

All panellists that participate in at least one round will have their responses included in analyses unless the study working group determines that a panellist’s response warrants being excluded from analyses for reasons that are newly developed or not previously recognized.

Step 4: Consensus meeting

In preparation for the consensus meeting, the study working group will categorize the remaining candidate items for consideration into a preliminary evaluation framework. This will include merging and/or modifying the remaining items based on gathered feedback from the e-Delphi rounds.

We will invite all members of the executive group to participate in the consensus meeting. An in-person or hybrid (in-person and online) consensus meeting may be held using the Surveylet Delphi survey platform after the results of the e-Delphi have been compiled and analyzed. If in a hybrid format, Microsoft Teams will be used for messaging and videoconferencing. Video recording and transcription services may also be used to aid notetaking. The primary objective of the meeting will be to achieve expert consensus on the final list of items for PHES-EF, through review and discussion of salient items. This process will be guided by the empirical evidence that was identified during the scoping reviews, and the opinions gathered during the e-Delphi process. The secondary objective will be to discuss publication and dissemination strategies for the final evaluation framework.

Steps to produce the final list of items are as follows: (i) present the results of the e-Delphi exercise (name, rationale, and score of each item); (ii) discuss the rationale and relevance for including the items in the framework; and (iii) vote on equivocal and non-consensual items. PHES-EF will be developed based on the final list of items that receive consensus during the executive group consensus meeting.

Step 5: Dissemination of evaluation framework

The dissemination of this evaluation framework will start with the publication of the PHES-EF protocol. The development of the evaluation framework will be reported in a statement document that will include the rationale and a brief description of the meeting, and the panellists involved. While the publication strategy will be finalized based on the consensus meeting discussion, the preliminary approach will be to post to OSF as a preprint, before publication in an open-access peer-reviewed journal. During this time, we will actively seek feedback from experts, knowledge users, and the public within our networks and via social media. PHES-EF will be made available using an open license (CC-BY-SA-4.0 license).

Post-publication, we will aim to keep the framework dynamic and up to date using a website, Discourse server, or other social media platforms to elicit ongoing feedback. Post-publication strategies may also include the creation of guides, toolkits, or ontologies to enhance the practical applicability of the evaluation framework and facilitate its adoption for policy and decision-making. Active dissemination approaches will include presenting at relevant scientific conferences, holding webinars, and conducting workshops. Additional dissemination practices will be determined during the consensus meeting.

Equity, diversity, and inclusion (EDI)

Incorporating EDI into scientific research enhances creativity and innovation, and improves the quality of results by embracing diverse perspectives and minimizing biases. It also helps promote fairness and equity by aiming to ensure that research outcomes serve various communities and global needs.

“Equity is defined as the removal of systemic barriers and biases enabling all individuals to have equal opportunity to access and benefit from the program” [52]. To achieve this, members of the research team are committed to developing a strong understanding of the systemic barriers faced by individuals from underrepresented groups. “Diversity is defined as differences in race, colour, place of origin, religion, immigrant and newcomer status, ethnic origin, ability, sex, sexual orientation, gender identity, gender expression and age” [52]. Panellists will be asked to provide characteristic information (e.g., country of residence, gender, primary discipline) to assess and help ensure (through more targeted recruitment) a large diversity in perspectives. “Inclusion is defined as the practice of ensuring that all individuals are valued and respected for their contributions and are equally supported” [52]. To help ensure that all research team members are integrated and supported, we will have a code of conduct statement in the executive group Terms of Reference.

Where feasible, and with sufficient statistical power, e-Delphi results will be further analyzed by panellist demographic information, such as by distinguishing respondents from high-, middle-, and low-income settings, as well as different world regions. This will allow for a nuanced understanding of how these contexts may have influenced the development of the framework, and which candidate items may be most valued in these areas.

All study working group members will complete EDI-related Canadian Institutes of Health Research (CIHR) training [53], and the First Nations Principles of OCAP© (Ownership, Control, Access, and Possession) training for research [54]. The executive group will receive an overview of EDI and OCAP© and will be encouraged and supported to complete such training. Furthermore, staff trained in EDI and OCAP© principles will review the framework from an EDI perspective to ensure appropriate language.

Knowledge user and public involvement

The international e-Delphi survey will collect responses from content experts, knowledge users, and engaged members of the public to ensure that all parties involved with or affected by wastewater surveillance are represented in the consensus-building approach. Apart from what has already been described, more detailed recruitment or public engagement methods may be informed by the executive group's engaged member of the public. Additionally, to ensure accurate and transparent reporting of knowledge user and public involvement throughout the study, we will refer to the Guidance for Reporting Involvement of Patients and the Public (GRIPP2) checklist [55]. We will document the methods used to engage knowledge users and members of the public, report the impacts and outcomes of their engagement, and report on lessons learned from the experience.

Acknowledgments

E-Delphi participation will not satisfy authorship criteria; however, panellists will have the option to receive individual acknowledgement for their contributions to the framework, given they provide consent to be named. Individuals who participate in the consensus meeting will have the opportunity to be named as authors.

Research data management

The project will follow an open-science approach and make available the data, code, and materials for all project stages, including the search strategy, search findings, initial curation of evaluation measures, summary of executive group discussions, e-Delphi processes, and agreement procedures. The management of research data will follow a data management and sharing plan.

All personal data will be managed using LimeSurvey. LimeSurvey is a free, open-source, online survey tool. Personal data will only be collected from individuals during the screening questionnaire, and from panellists who wish to receive compensation and/or individual acknowledgement for their contributions to the evaluation framework. LimeSurvey data management practices are overseen by The Ottawa Hospital (TOH) Research Institute Methods Centre that protects privacy and confidentiality while assisting authorized users in collecting and managing survey data. It has been installed on a TOH server in the TOH Demilitarized Zone and all data collected is stored in a TOH database within the TOH firewall. Because LimeSurvey is installed outside of the TOH firewall in the TOH Demilitarized Zone, surveys can be responded to from anywhere using an internet connection.

Calibrum’s Surveylet Delphi platform will be used to manage e-Delphi responses. It may also be used for the consensus meeting. The e-Delphi will only collect anonymous data using a modified Delphi approach; meanwhile, the consensus meeting may collect identifiable data using a real-time Delphi approach. Surveylet complies with the EU-U.S. Privacy Shield Framework and Swiss-U.S. Privacy Shield Framework regarding the collection, use, and retention of personal information. The Surveylet platform also allows its users to be General Data Protection Regulation (GDPR) compliant. Calibrum’s servers are protected by high-end firewall systems and vulnerability scans are conducted regularly. Calibrum uses Transport Layer Security (TLS) encryption (also known as HTTPS) for all transmitted data. Surveys are also protected with passwords and HTTP referrer checking. All data are encrypted at rest and hosted by SSAE-16 SOC II certified third-party data centres. Data on deprecated hard drives are destroyed by U.S. DOD methods and delivered to a third-party data destruction service.

Study record retention

Any data exported from either survey company will be stored securely in a password protected location within TOH’s SharePoint. All study records will be retained securely for 10 years following study completion, then destroyed in accordance with Ottawa Hospital Research Institute (OHRI) requirements.

Ethical considerations

This study has received ethical approval (20230428-01H) from the Ottawa Health Science Network Research Ethics Board (OHSN-REB). Implied consent will be obtained from all panellists prior to participating in the e-Delphi survey. Information about the survey and consent practices will be provided on the first page of the survey. Panellists will be informed that by providing their demographic information and proceeding to the next page, they are providing their implied consent to participate in the study.

Only anonymous panellist responses will be used for analyses. All panellist data and responses will be password protected and only REB-approved study working group members from the Ottawa Hospital will have access to the data.

Discussion

This paper outlines the protocol for a two-round e-Delphi study that will be used to develop a multinational, multidisciplinary evaluation framework for public health environmental surveillance. Developing an evaluation framework will contribute to a better understanding of the ideal scope and use cases for wastewater surveillance and improve the reliability and credibility of surveillance data for public health officials. An evaluation framework will also help promote the sustainability of wastewater surveillance systems by providing an effective performance measurement tool.

We strove to design the PHES-EF protocol using a rigorous consensus approach. The development of the PHES-EF study protocol was informed by ACCORD (ACcurate COnsensus Reporting Document): A reporting guideline for consensus methods in biomedicine developed via a modified Delphi [38]. Surveillance evaluation frameworks are uncommonly developed using a consensus approach. PHES-EF will be developed for environmental surveillance and informed by wastewater surveillance during the pandemic. However, associated scoping reviews and prior surveillance evaluation frameworks are related largely to any public health surveillance system [56]. Therefore, PHES-EF may be applicable for evaluating public health surveillance beyond environmental systems.

Limitations

Although the Delphi technique uses a systematic process to generate consensus among a panel of experts, the method is not without its limitations. One of the primary concerns when using this technique is the ability to implement a truly representative expert panel. Despite our plan to use various recruitment methods to enroll panellists from all discipline subgroups with varying demographic characteristics, consensus results may ultimately be influenced by the final panel composition [57].

Another potential limitation in using a modified Delphi approach is the potential influence of controlled feedback on the convergence of panellist opinions. Although providing controlled feedback between survey rounds aids in generating consensus, this approach may bias consensus results by influencing individual panellists’ opinions [58].

Furthermore, due to feasibility and resource limitations, only sources that are available in English will be included in associated scoping reviews. Additionally, only individuals who are proficient in English will be eligible to participate in the e-Delphi survey and consensus meeting. Therefore, the findings of this study may not reflect relevant knowledge that is not available in English.

Supporting information

S2 Table. Executive group and e-Delphi panellist disciplines.

The list of specializations or related concepts is not extensive.

https://doi.org/10.1371/journal.pone.0310342.s002

(PDF)

S3 Table. World region classifications.

Note: The regional classifications are used by the World Bank for organizing economic data, development projects, and policy analysis. These classifications are not strictly geographical but also consider economic and developmental similarities and relationships.

https://doi.org/10.1371/journal.pone.0310342.s003

(PDF)

S1 Appendix. PRISMA-ScR checklist for scoping review #1.

Current Evaluation Practices for Public Health and Environmental Surveillance Systems: A Scoping Review Protocol.

https://doi.org/10.1371/journal.pone.0310342.s004

(PDF)

S2 Appendix. PRISMA-ScR checklist for scoping review #2.

Lessons Learned from Wastewater-Based Surveillance During the COVID-19 Pandemic: A Scoping Review Protocol.

https://doi.org/10.1371/journal.pone.0310342.s005

(PDF)

S3 Appendix. Rating properties for e-Delphi candidate item scoring.

https://doi.org/10.1371/journal.pone.0310342.s006

(PDF)

S4 Appendix. Draft e-Delphi panellist screening questionnaire.

https://doi.org/10.1371/journal.pone.0310342.s007

(PDF)

Acknowledgments

We thank Mary Jessome (PhD candidate, University of British Columbia) for her contributions to the study as a past executive group member with expertise in social sciences. We would also like to thank Peter Farrell (Research Librarian, University of Ottawa) for developing the associated scoping review search strategies for this study.

PHES-EF Executive Group (ordered alphabetically according to surname except for the first two authors):

Douglas G. Manuel* (0000-0003-0912-0845), David L. Buckeridge* (0000-0003-1817-5047), Yoni Freedhoff (0000-0001-9391-3455), Sarah Funnell, Bernd M. Gawlik (0000-0002-4406-5362), Farah Ishtiaq (0000-0002-6762-7014), Amy E. Kirby, Kerrigan McCarthy (0000-0001-8958-9795), David Moher (0000-0003-2434-4206), Beate Sander (0000-0003-2128-9133), Jeremy Veillard, Matthew J. Wade (0000-0001-9824-7121)

* Co-chairs of PHES-EF Executive Group

References

  1. 1. Woodward G, Manuel DG, Goel V. Developing a balanced scorecard for public health. Toronto: Institute for Clinical Evaluative Sciences; 2004.
  2. 2. McClary-Gutierrez JS, Mattioli MC, Marcenac P, Silverman AI, Boehm AB, Bibby K, et al. SARS-CoV-2 Wastewater Surveillance for Public Health Action. Emerg Infect Dis. 2021;27(9):1–8. pmid:34424162
  3. 3. Diamond MB, Keshaviah A, Bento AI, Conroy-Ben O, Driver EM, Ensor KB, et al. Wastewater surveillance of pathogens can inform public health responses. Nat Med. 2022;28(10):1992–5. pmid:36076085
  4. 4. Zuccato E, Chiabrando C, Castiglioni S, Bagnati R, Fanelli R. Estimating community drug abuse by wastewater analysis. Environ Health Perspect. 2008;116(8):1027–32. pmid:18709161
  5. 5. Hrudey SE, Bischel HN, Charrois J, Chik AHS, Conant B, Delatolla R, et al. Wastewater Surveillance for SARS-CoV-2 RNA in Canada. FACETS. 2022;7:1493–597.
  6. 6. World Health Organization. Environmental surveillance for SARS-COV-2 to complement public health surveillance. World Health Organization; 2022 Apr.
  7. 7. Manuel DG, Delatolla R, Fisman DN, Fuzzen M, Graber T, Katz GM, et al. The Role of Wastewater Testing for SARS-CoV-2 Surveillance. Ontario COVID-19 Science Advisory Table; 2021 Aug. https://doi.org/10.47326/ocsat.2021.02.40.1.0
  8. 8. Tiwari A, Adhikari S, Kaya D, Islam MA, Malla B, Sherchan SP, et al. Monkeypox outbreak: Wastewater and environmental surveillance perspective. Sci Total Environ. 2023;856(Pt 2):159166. pmid:36202364
  9. 9. Wolfe MK, Duong D, Bakker KM, Ammerman M, Mortenson L, Hughes B, et al. Wastewater-Based Detection of Two Influenza Outbreaks. Environ Sci Technol Lett. 2022;9(8):687–92.
  10. 10. Sims N, Kasprzyk-Hordern B. Future perspectives of wastewater-based epidemiology: Monitoring infectious disease spread and resistance to the community level. Environ Int. 2020;139:105689. pmid:32283358
  11. 11. Huijbers PMC, Flach C-F, Larsson DGJ. A conceptual framework for the environmental surveillance of antibiotics and antibiotic resistance. Environ Int. 2019;130:104880. pmid:31220750
  12. 12. Tiwari A, Kurittu P, Al-Mustapha AI, Heljanko V, Johansson V, Thakali O, et al. Wastewater surveillance of antibiotic-resistant bacterial pathogens: A systematic review. Front Microbiol. 2022;13:977106. pmid:36590429
  13. 13. Yang D, Zheng Q, Thai PK, Ahmed F, O’Brien JW, Mueller JF, et al. A nationwide wastewater-based assessment of metformin consumption across Australia. Environ Int. 2022;165:107282. pmid:35597112
  14. 14. Escolà Casas M, Schröter NS, Zammit I, Castaño-Trias M, Rodriguez-Mozaz S, Gago-Ferrero P, et al. Showcasing the potential of wastewater-based epidemiology to track pharmaceuticals consumption in cities: Comparison against prescription data collected at fine spatial resolution. Environ Int. 2021;150:106404. pmid:33578067
  15. 15. Woytus A. What’s in the wastewater? In: St. Louis [Internet]. 22 Sep 2021 [cited 1 Jun 2023. ]. Available from: https://www.stlmag.com/api/content/3f60d498-1b26-11ec-a8ea-1244d5f7c7c6/
  16. 16. D’Aoust PM, Towhid ST, Mercier É, Hegazy N, Tian X, Bhatnagar K, et al. COVID-19 wastewater surveillance in rural communities: Comparison of lagoon and pumping station samples. Sci Total Environ. 2021;801:149618. pmid:34418622
  17. 17. Shah S, Gwee SXW, Ng JQX, Lau N, Koh J, Pang J. Wastewater surveillance to infer COVID-19 transmission: A systematic review. Sci Total Environ. 2022;804:150060. pmid:34798721
  18. 18. Liu P, Ibaraki M, VanTassell J, Geith K, Cavallo M, Kann R, et al. A sensitive, simple, and low-cost method for COVID-19 wastewater surveillance at an institutional level. Sci Total Environ. 2022;807(Pt 3):151047. pmid:34673061
  19. 19. Larsen DA, Wigginton KR. Tracking COVID-19 with wastewater. Nat Biotechnol. 2020;38(10):1151–3. pmid:32958959
  20. 20. Gretchen Vogel. Signals from the sewer—Measuring virus levels in wastewater can help track the pandemic. But how useful is that? Science. 9 Mar. 2022375. Available from: https://www.science.org/content/article/pandemic-signals-sewer-what-virus-levels-wastewater-tell-us
  21. 21. Manuel D, Amadei CA, Campbell JR, Brault J, Zierler A, Veillard J. Strengthening Public Health Surveillance Through Wastewater Testing: An Essential Investment for the COVID-19 Pandemic and Future Health Threats. World Bank Group. The World Bank; 2022. p. 1–67.
  22. 22. Fynn JF, Hardeman W, Milton K, Jones AP. A scoping review of evaluation frameworks and their applicability to real-world physical activity and dietary change programme evaluation. BMC Public Health. 2020;20(1):1000. pmid:32586301
  23. 23. Nilsen P. Making sense of implementation theories, models, and frameworks. Implement Sci. 2020;30:53–79.
  24. 24. Calba C, Goutard FL, Hoinville L, Hendrikx P, Lindberg A, Saegerman C, et al. Surveillance systems evaluation: a systematic review of the existing approaches. BMC Public Health. 2015;15:448. pmid:25928645
  25. 25. Andrews JR, Yu AT, Saha S, Shakya J, Aiemjoy K, Horng L, et al. Environmental Surveillance as a Tool for Identifying High-risk Settings for Typhoid Transmission. Clin Infect Dis. 2020;71(Suppl 2):S71–8. pmid:32725227
  26. 26. Porta M. A Dictionary of Epidemiology. A Dictionary of Epidemiology. Oxford University Press; 2016. Available from: https://www.oxfordreference.com/display/10.1093/acref/9780199976720.001.0001/acref-9780199976720
  27. 27. Spranger J, Homberg A, Sonnberger M, Niederberger M. Reporting guidelines for delphi techniques in health sciences: a methodological review. Z Evid Fortbild Qual Gesundh Wesen. 2023;172:1–11.
  28. 28. Editorial. How epidemiology has shaped the COVID pandemic. Nature. 2021;589(7843):491–2. pmid:33505041
  29. 29. Devine D, Gaskell J, Jennings W, Stoker G. Trust and the Coronavirus Pandemic: What are the Consequences of and for Trust? An Early Review of the Literature. Polit Stud Rev. 2021;19(2):274–85. pmid:35082554
  30. 30. World Health Organization. Defining collaborative surveillance: a core concept for strengthening the global architecture for health emergency preparedness, response, and resilience (HEPR). Geneva: World Health Organization; 2023. Available from: https://iris.who.int/handle/10665/367927
  31. 31. Tricco AC, Lillie E, Zarin W, O’Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and Explanation. Ann Intern Med. 2018;169(7):467–73. pmid:30178033
  32. 32. Nasa P, Jain R, Juneja D. Delphi methodology in healthcare research: How to decide its appropriateness. World J Methodol. 2021;11(4):116–29. pmid:34322364
  33. 33. Shang Z. Use of Delphi in health sciences research: A narrative review. Medicine (Baltimore). 2023;102(7):e32829. pmid:36800594
  34. 34. Jünger S, Payne SA, Brine J, Radbruch L, Brearley SG. Guidance on Conducting and REporting DElphi Studies (CREDES) in palliative care: Recommendations based on a methodological systematic review. Palliat Med. 2017;31(8):684–706. pmid:28190381
  35. 35. Niederberger M, Spranger J. Delphi Technique in Health Sciences: A Map. Front Public Health. 2020;8:457. pmid:33072683
  36. 36. Schmalz U, Spinler S, Ringbeck J. Lessons Learned from a Two-Round Delphi-based Scenario Study. MethodsX. 2020;8:101179. pmid:33365260
  37. 37. Boel A, Navarro-Compán V, Landewé R, van der Heijde D. Two different invitation approaches for consecutive rounds of a Delphi survey led to comparable final outcome. J Clin Epidemiol. 2021;129:31–9. pmid:32991995
  38. 38. Gattrell WT, Logullo P, van Zuuren EJ, Price A, Hughes EL, Blazey P, et al. ACCORD (ACcurate COnsensus Reporting Document): A reporting guideline for consensus methods in biomedicine developed via a modified Delphi. PLoS Med. 2024;21(1):e1004326. pmid:38261576
  39. 39. Permuth-Wey J, Borenstein AR. Financial remuneration for clinical and behavioral research participation: ethical and practical considerations. Ann Epidemiol. 2009;19(4):280–5. pmid:19230712
  40. 40. Surveylet. United States: Calibrum. Available from: https://calibrum.com
  41. 41. LimeSurvey: An open source survey tool. Hamburg, Germany: LimeSurvey GmbH; Available from: https://www.limesurvey.org
  42. 42. Diamond IR, Grant RC, Feldman BM, Pencharz PB, Ling SC, Moore AM, et al. Defining consensus: a systematic review recommends methodologic criteria for reporting of Delphi studies. J Clin Epidemiol. 2014;67(4):401–9. pmid:24581294
  43. 43. Belton I, MacDonald A, Wright G, Hamlin I. Improving the practical application of the Delphi method in group-based judgment: A six-step prescription for a well-founded and defensible process. Technological Forecasting and Social Change. 2019;147:72–82.
  44. 44. Weijters B, Cabooter E, Schillewaert N. The effect of rating scale format on response styles: The number of response categories and response category labels. International Journal of Research in Marketing. 2010;27(3):236–47.
  45. 45. Lange T, Kopkow C, Lützner J, Günther K-P, Gravius S, Scharf H-P, et al. Comparison of different rating scales for the use in Delphi studies: different scales lead to different consensus and show different test-retest reliability. BMC Med Res Methodol. 2020;20(1):28. pmid:32041541
  46. 46. Locke EA, Latham GP. A theory of goal setting & task performance. Prentice-Hall, Inc. 1990.
  47. 47. Veillard J, Champagne F, Klazinga N, Kazandjian V, Arah OA, Guisset A-L. A performance assessment framework for hospitals: the WHO regional office for Europe PATH project. Int J Qual Health Care. 2005;17(6):487–96. pmid:16155049
  48. 48. Kaplan RS, Norton DP. Balanced scorecard. Springer; 2007.
  49. 49. Mitchell RJ, Williamson AM, O’Connor R. The development of an evaluation framework for injury surveillance systems. BMC Public Health. 2009;9:260. pmid:19627617
  50. 50. Nothacker M, Stokes T, Shaw B, Lindsay P, Sipilä R, Follmann M, et al. Reporting standards for guideline-based performance measures. Implement Sci. 2016;11:6. pmid:26772173
  51. 51. Groseclose SL, German RR, Nsubuga P. Evaluating Public Health Surveillance. Principles & Practice of Public Health Surveillance. Oxford University Press. 2010. p. 166–97. https://doi.org/10.1093/acprof:oso/9780195372922.003.0008
  52. 52. Government of Canada. Best Practices in Equity, Diversity and Inclusion in Research. In: New Frontiers in Research Fund [Internet]. 22 Jun 2021. Available from: https://www.sshrc-crsh.gc.ca/funding-financement/nfrf-fnfr/edi-eng.aspx
  53. 53. Government of Canada. Introduction to GBA Plus. 31 Mar 2021 [cited 1 Jun 2023. ]. Available from: https://women-gender-equality.canada.ca/gbaplus-course-cours-acsplus/eng/mod00/mod00_01_01.html
  54. 54. The First Nations Information Governance Centre. The First Nations Principles of OCAP®. In: The First Nations Information Governance Centre [Internet]. 1 Jun 2023 [cited 1 Jun 2023. ]. Available from: https://fnigc.ca/ocap-training/
  55. 55. Staniszewska S, Brett J, Simera I, Seers K, Mockford C, Goodlad S, et al. GRIPP2 reporting checklists: tools to improve reporting of patient and public involvement in research. BMJ. 2017;358:j3453. pmid:28768629
  56. 56. Yang L, Weston C, Cude C, Kincl L. Evaluating Oregon’s occupational public health surveillance system based on the CDC updated guidelines. Am J Ind Med. 2020;63(8):713–25. pmid:32483871
  57. 57. Winkler J, Moser R. Biases in future-oriented Delphi studies: A cognitive perspective. Technological Forecasting and Social Change. 2016;105:63–76.
  58. 58. Goodman CM. The Delphi technique: a critique. J Adv Nurs. 1987;12(6):729–34. pmid:3320139