Journal:Management and disclosure of quality issues in forensic science: A survey of current practice in Australia and New Zealand

From LIMSWiki
Jump to navigationJump to search
Full article title Management and disclosure of quality issues in forensic science: A survey of current practice in Australia and New Zealand
Journal Forensic Science International: Synergy
Author(s) Heavey, Anna L.; Turbett, Gavin R.; Houck, Max M.; Lewis, Simon W.
Author affiliation(s) Curtin University, PathWest Laboratory Medicine, Florida International University
Primary contact s dot lewis at curtin dot edu dot au
Year published 2023
Volume and issue 7
Article # 100339
DOI 10.1016/j.fsisyn.2023.100339
ISSN 2589-871X
Distribution license Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International
Website https://www.sciencedirect.com/science/article/pii/S2589871X23000268
Download https://www.sciencedirect.com/science/article/pii/S2589871X23000268/pdfft (PDF)

Abstract

The investigation of quality issues detected within the forensic process is a critical feature in robust quality management systems (QMSs) to provide assurance of the validity of reported laboratory results and inform strategies for continuous improvement and innovation. A survey was conducted to gain insight into the current state of practice in the management and handling of quality issues amongst the government service provider agencies of Australia and New Zealand. The results demonstrate the value of standardized quality system structures for the recording and management of quality issues, but also areas where inconsistent reporting increases the risk of overlooking important data to inform continuous improvement. With new international changes requiring mandatory reporting of quality issues, this highlights compliance challenges that agencies will face. This study reinforces the need for further research into the standardization of systems underpinning the management of quality issues in forensic science to support transparent and reliable justice outcomes.

Keywords: forensic science, quality issues, error, accreditation, disclosure

Introduction

The origins of modern quality management in forensic science can be traced back to the mid-twentieth century with the introduction of drunk driving legislation compelling testing laboratories to develop protocols for result validation, chain of custody, and sample storage conditions.[1] Since then, the field of forensic quality management has flourished through the development of international networks and professional organizations, reviews and inquiries into best practices, and the standardization of testing systems and methodologies. At the same time, accreditation of forensic science service provider agencies has become the expected norm and increased worldwide.[1] Whether operating as an accredited service or not, the importance of robust and fit-for-purpose quality management systems (QMSs) in forensic science cannot be understated in their ability to provide assurance that results being produced by forensic service providers are accurate, consistent, and on-time.

In Australia and New Zealand, the major providers of forensic science services are government agencies, and the structure of these providers varies between the countries and states/territories. Services in each jurisdiction are provided by one or more government sectors, with some jurisdictions’ service provided solely by the police, and others split between multiple government sectors, including health, justice, and science.[2] The directors of the government forensic agencies in Australia and New Zealand form the Australia New Zealand Forensic Executive Committee (ANZFEC), which sits as a governing body of the Australia New Zealand Policing Advisory Agency – National Institute of Forensic Science (ANZPAA-NIFS). ANZPAA-NIFS was established in 1992 with a strategic intent to promote and facilitate excellence in forensic science in the region.[2] The cross-jurisdictional oversight of ANZFEC and ANZPAA-NIFS is an important component in the overarching quality management of forensic science service provision in Australia and New Zealand, facilitating collaborative relationships between agencies and the wider forensic community to champion innovation, address priority needs, and promote the ongoing development and quality of the field.[3]

Crucial to the continuous improvement of any QMS, including forensic QMSs, is the identification and prevention of risks that may adversely affect the quality of the product or result. These risks may be identified proactively (preventative) or after an issue has been identified and addressed (corrective).[4] Forensic QMSs, such as those designed to comply with the requirements of ISO/IEC 17025[4], are required to have processes in place for the management of these laboratory quality issues. In the ISO/IEC 17025 standard, this requirement is detailed under section 7.10 “Nonconforming work,” and the wording used is: “The laboratory shall have a procedure that shall be implemented when any aspect of its laboratory activities or results of this work do not conform to its own procedures or the agreed requirements of the customer.” The instructions for how the nonconformity is to be actioned are provided later in the standard under section 8.7 “Corrective actions” and can be summarized as follows:

  • Take action to control the nonconformity and correct it;
  • Take action to address the consequences of the nonconformity;
  • Analyze the nonconformity to determine the root cause and, where necessary,;
  • Take action to address the root cause to prevent reoccurrence of the nonconformity, and subsequently;
  • Assess the effectiveness of the new control measures.

With many forensic service agencies operating QMSs compliant with standards such as ISO/IEC 17025, a wealth of information on the types of issues detected in forensic processes is available, with the potential to be shared and collated for purposes including detecting trends, identifying opportunities for research and development, and facilitating interagency comparison and benchmarking.[5]

However, there is limited published data on quality issues detected within forensic service provider agencies internationally. A fundamental concern for agencies with sharing this data publicly may be the risk of potential misuse or misunderstanding of the information. This concern may be further compounded by the lack of standardization in how that information is collected between agencies and the terminology used, making the sharing of this information difficult.[5] The implications of a lack of a standardized approach to quality issue investigation and management were highlighted in late 2022 by The Commission of Inquiry into Forensic DNA Testing in Queensland, particularly with regards to the perceived risk to transparency where a consistent approach is not apparent.[6]

Recent changes to new statutory guidance documents in the UK and international accreditation requirements highlights that this lack of standardization presents a current and urgent challenge to the field. The UK Forensic Science Regulator Code of Practice (released March 2023 and coming into effect in October 2023) requires the forensic unit (defined within the Code as “a legal entity or a defined part of a legal entity that performs any part of a forensic science activity”) to inform the Regulator of non-conforming work or “quality failures” that have the potential to attract adverse public comment, be against the public interest, or lead to a miscarriage of justice.[7] The American National Standards Institute - National Accreditation Board (ANAB) published AR 3125:2023 Accreditation Requirements for Forensic Testing and Calibration Laboratories in February 2023, with a new clause for proficiency testing requiring the laboratory to notify ANAB within 30 days when an expected result is not attained.[8] Both of these documents go further than the ISO/IEC 17025 standard in providing examples of the types of “non-conforming work” or “unexpected results” that would compel reporting. However, the examples are not exhaustive in what might be seen in an operational forensic agency and thus are open to interpretation. The subjective application of requirements such as these invites risks, for example, the over-reporting of unwarranted issues may choke the resources of both the reporting agency and the regulator or accreditation body being reported to, preventing timely and effective carriage of justice and action on critical findings. Under-reporting of critical issues, even if done with no malicious intent, risks not only the perceived integrity of the forensic agency, but also the non-identification of systemic issues potentially threatening justice outcomes and public confidence in forensic science. The matter of mandatory reporting is complicated further by the fact that some quality issues only become apparent a considerable time later, whether being undetected at the time or not categorized as a quality issue at the time as technology or reporting thresholds evolve. For example, in the case of reanalysis of cold case DNA evidence collected many years prior, when testing methods were not sensitive enough to detect underlying contaminants[9], or where retrospective review of method validation studies identifies deficiencies in the analysis used to set analytical thresholds that have been in operation for a considerable time.[6] Therefore, the sense of security that a system of mandatory reporting provides to current investigations or court matters may be false, as undetected and unreported issues will inevitably still exist even in the most willing and compliant of environments.

As the foundation to a program of research investigating quality issues in forensic science, a survey was designed and conducted on the current state of practice in forensic quality management and the management of quality issues amongst the government forensic service provider agencies of Australia and New Zealand. The objective of the survey was to gain perspective on the diversity of forensic services offered by these agencies and the quality structures surrounding them, along with insights into the current categorization, management, and disclosure of quality issues within those agencies. With the focus on transparent reporting of quality issues in forensic science now gaining momentum in international jurisdictions, the issue of effective communication on these issues is a clear and present concern to the field. This paper presents the results of this survey within the context of current practice in forensic quality management. Further we discuss these results and their significance in the environment of operational forensic science service provider agencies.

Survey methodology and participant demographics

The survey was designed and conducted using the online tool Qualtrics for ease of use and submission by participants.

The target participants for this survey were forensic quality practitioners or forensic practitioners involved in the management of quality issues working in the government forensic science service provider agencies of Australia and New Zealand. A request to approach suitable participants was made to ANZFEC via ANZPAA-NIFS. The agencies with representatives on the Quality Specialist Advisory Group (QSAG) of ANZPAA-NIFS were targeted. Permission was granted by all of the 15 agencies represented on QSAG to allow the research team to approach their staff. An invitation to potential participants was circulated to the QSAG members by ANZPAA-NIFS on behalf of the research team, and the invitation stated that it could also be forwarded to other suitable practitioners within the agency. Invitees who responded positively were forwarded a participant information sheet and participation consent form for return. Upon confirmation of consent, a unique alphanumeric identifier was assigned and forwarded to the participant along with a link to the online survey tool. The unique identifier was generated using online random number and letter generators and was to be used to ensure that each survey response was de-identified to anyone outside of the research team.

The participants were required to enter their unique identifier on the front page of the survey and to confirm consent before they were able to progress into the survey questions. Other than the unique identifier, no identifying information on the participants or the agency they represented was collected in the survey tool. The survey consisted of multiple sections, with certain questions only being displayed based on previous answers. The survey was designed to allow participants to return to previous questions and amend responses or pause and restart the survey at any time prior to final submission. The section designations in the survey are reflected in the headings presented in the results below.

Survey questions were designed as either requiring selection from a set of provided options (nominal scale), or as free text fields for unstructured data collection. All questions with options provided also included a space for additional free text comment to be made. All questions were set as forced response.

The full survey logic and functionality was pretested prior to deployment by volunteers outside of the research team with familiarity of the subject matter.

Uptake for the survey was successful with 16 participants from 13 of the 15 QSAG agencies responding. The participants represent all of the states and territories of Australia (including the federal jurisdiction), and New Zealand. All participants in the survey self-declared as being employed in a quality manager or quality practitioner role, with more than 85% of respondents indicating that their quality management role was responsible for multiple forensic science disciplines.

Survey results and discussion

Accreditation of disciplines

With accreditation of ANZFEC agencies being a mandated aim for the region[1], it is not surprising that all agencies represented by the survey participants are accredited to ISO/IEC 17025. However, the responses presented in Fig. 1 show that, of the various forensic disciplines that are represented across the agencies surveyed, some are not accredited across all the agencies where that discipline is offered and, in a few cases, none of the agencies where the discipline is available are accredited for that service.


Fig1 Heavey ForSciIntSyn2023 7.jpg

Figure 1. Accreditation status of forensic disciplines represented in survey participant agencies.

The survey responses with regards to the non-accreditation or partial accreditation of some disciplines may be indicative of the particular challenges associated with these areas. For example, in the case of anthropology and odontology, very few of the agencies surveyed were responsible for the provision of these services (three agencies and two agencies, respectively), and all are part of multidisciplinary forensic service agencies. The highly specialized nature of these fields, coupled with accreditation for these disciplines being under a different international standard (ISO 15189) to the other disciplines within the same agencies (ISO/IEC 17025) may be a consideration made for business decisions not to pursue accreditation. Entomology is a similarly rare field amongst the surveyed agencies, only offered by one. In the case of disciplines where there is a sole subject matter expert, the logistics of requirements such as technical peer review of results can be a barrier to accreditation in cases of limited staffing resources and may lend a preference to testing validity of the science and results as part of the court evidence process.[10]

The field of electronic evidence deals in an environment of rapidly changing technology and evidence types. This presents other unique challenges from an accreditation perspective, for example, the swift adoption of new and novel analysis techniques may preclude lengthy validation studies or the ability to source appropriate and relevant proficiency testing programs.[11] The field of facial identification (or facial image comparison) faces similar challenges to electronic evidence management, with the rapidly changing nature of image samples being presented and technology available to perform analysis. Further, the contrasting techniques of manual examiner comparison and automated facial recognition, along with the diversity of applications the techniques may be used for in different jurisdictions can make the development of standardized practice difficult, although expert scientific working groups are progressing this space.[12][13]

Quality issue management

What is considered a “quality issue”?

Evidence of how the forensic facility identifies and mitigates risks to the quality of results and how it identifies and manages nonconforming work is an accreditation requirement, meaning that records must be kept and available. Whilst the standards are explicit in the necessity for these quality issues to be noted, managed, and recorded, the nature of what constitutes a “quality issue” worthy of recording is less clear.[4][14]

All participants indicated that their agency has a documented process for the management of quality issues (including one which is a new document currently being drafted). Other than the one in draft, all participants also noted that the documented process included instructions for the logging/recording of quality issues.

Participants were asked to indicate which types of issues would generally be considered as a “quality issue” that needed to be managed/logged against a list of provided examples (Table 1).

Table 1. Survey responses to “Which of the following types of issues would generally be considered as a 'quality issue' to be managed/logged?”
Description of issue Number of affirmative responses from 16 survey participants (%)
Customer/client complaints 16 (100%)
A documented procedure was not followed 15 (94%)
A reported/disclosed result requires correction 15 (94%)
Non-conformance identified at audit 15 (94%)
Results could have been invalidated/delayed 14 (88%)
Failure/investigation of an external QAP 14 (88%)
Results were invalidated/unable to be reported 13 (81%)
Critical equipment failure 13 (81%)
The content of a disclosed report, other than the result, requires correction 12 (75%)
Customer/client compliments 12 (75%)
Court testimony feedback 12 (75%)
Failure/investigation of an internal QAP 11 (69%)
Recommendations identified at audit 11 (69%)
Customer/client feedback (non-complaint or compliment) 9 (56%)
Results were delayed 8 (50%)
Non-critical equipment failure 7 (44%)

The responses across the participants demonstrate high agreement in some categories, with all participants managing/logging customer complaints and 15 out of the 16 participants agreeing that adherence to documented procedures, correction of issued results, and non-conformances identified at audit were all considered “quality issues” requiring management/logging. The single negative response for each of these categories was received from participants representing the same agency but responsible for different disciplines. One of these disciplines within the agency is not accredited to ISO/IEC 17025, which may influence the response given.

Participant responses were split on whether delayed results (exceeding agreed reporting time to client) would be managed/logged as “quality issues.” The issue of timely provision of results has been noted as an area of concern for the industry for some time, both in the risk of delays to the justice system[15][16] and as a risk to the provision of quality scientific services.[17] Reasons for this divide in answers may be due to the nuances of individual agency client service level agreements (or equivalent arrangements) where there is either no agreed reporting times specified, or an accepted percentage of delayed reports allowed before it is considered a nonconformity. To analyze the root cause of this discrepancy between participants, this category will be explored further in future research stages.

Participants were provided with an opportunity to list any other types of issues, in their own words, that would generally be considered as quality issues for management/logging. Comments were received from three participants, and included:


  • “Risk identified”;
  • “Planned deviation from process”; and
  • “Issues relating to chain of custody, training, procedural errors or deficiencies, fieldwork or environmental conditions, facilities and software (upgrades, limitations, outages).”

Those responses where limited detail was provided will be explored further in later phases of the research.

How are “quality issues” recorded?

Much like the category of issues which require management/logging, the specifics of how these issues are to be recorded is also undefined in the standard requirements. The predominant tool (or form) associated with QMSs prescribed by ISO/IEC 17025 and ISO 9001[14], (along with other management systems, such as ISO 45001: Occupational health and safety management systems[18]) is the corrective action request (CAR). A CAR is a mechanism of documenting a problem or potential problem identified in the management system which requires that the root cause of the problem be mitigated or removed to prevent recurrence.[19] Generally, the CAR process will also include documentation of the details of the investigation conducted to determine the root cause of the problem, actions taken to address the root cause, and evidence to demonstrate that the effectiveness of those actions has been reviewed and additional actions taken as required.

A root cause is the underlying issue that resulted in the non-conformance or issue. Root cause analysis (RCA) is the process used to identify the root cause of any issue.[20] There are several approaches to conducting an RCA involving a variety of methods and tools.[21] RCA can be considered a component of problem-solving, with the outcome providing information on the fundamental cause of an issue to be removed or mitigated to facilitate continuous improvement of the QMS.

CARs may be recognized by other names depending on the organization, for example, an “opportunity for improvement request” or a “continuous improvement request”. However named, the process of having a mechanism for the documentation of issues detected (actual or potential) in the QMS and the actions taken to address these is a critical component of these systems and one of the most valuable tools in the continuous improvement cycle, if utilized effectively.[21]

In the context of forensic science service provision, maintaining records of issues identified is not just a necessary feature of the QMS, but also crucial in demonstrating the robustness of the analysis process and subsequent results being provided.[9] Transparency with regards to the analytical and interpretive process involves not only demonstrating the scientific validity of the methodology being used but also the clear and accurate communication of any limitations to be associated with the results provided, which includes the clear disclosure of issues that were identified during the process and how these may influence the validity of the results obtained.

As noted above, the specifics of what types of issues are recorded and how these issues are recorded is not defined in the guiding management system standards. All forensic service providers operating to one of the ISO standards will have some version of a CAR in their systems; however, this may not be used to record all issues detected, and alternative methods of documentation will be used either instead of, or in addition to, the CAR.

To better understand the various means used to record the types of quality issues identified in Table 1, survey participants who indicated that a particular issue would be logged were asked to indicate how each type of issue would be recorded within their agency. To accommodate types of issues which may be recorded in different ways dependent on the particular situation, participants were able to select multiple methods of recording for each issue type, as applicable. The responses were grouped into the following categories for the purposes of review and discussion: Reporting and delivery of results; Quality assurance programs; Equipment issues; Audits and documented procedures, and; Customer feedback and monitoring of court testimony.

Reporting and delivery of results

The provision of accurate, timely, and unambiguous results is a fundamental measure of the successful provision of forensic science services, and therefore issues that are identified which affect the delivery of those results require close attention.[22] Issues related to the reporting and delivery of results were among the highest logged by participants in the survey (Table 1), with the highest proportion of participants indicating that issues where a reported or disclosed result required correction (15 out of 16) would be recorded. The next two highest recorded issues included scenarios where results were invalidated or unable to be reported (13 out of 16), and where results could have been invalidated or delayed (14 out of 16).

However, whilst all respondents logging these issues indicated that a root cause analysis would be performed in the case of invalid or unreportable results, only 72% indicated the same would be performed for results that could have been delayed (Fig. 2).


Fig2 Heavey ForSciIntSyn2023 7.jpg

Figure 2. Methods of recording issues related to reporting and delivery of results that participants indicated may be used within their agency.

Although only half of participants indicated that actual delayed reporting of results was deemed as a quality issues requiring recording (Table 1), all of those participants who do record actual delays would document these as a CAR (or equivalent) and perform root cause analysis (Fig. 2).

Quality assurance programs

Quality assurance programs (QAPs) are used by forensic science agencies to test the systems and processes in place.[22] The nature of QAPs can vary greatly and will be dependent on the nature of the services being provided. QAPs can be both internal (programs developed in-house to monitor system capability and competency such as testing of samples where ground truth is known by the program administrator or monitoring of contamination minimization processes through environmental samples) or external (programs sourced externally such as commercial proficiency testing programs or inter-laboratory comparison exercises). When used effectively, QAPs assist the agency to identify vulnerabilities in the quality system and are a valuable tool in the proactive identification of risks to the quality of results. Just as importantly, QAPs can highlight successes in the quality system and provide tangible evidence of the robustness of the procedures in place to protect the agency's quality of output.

Where the results of QAPs indicate a possible failure or risk in the system this may be deemed a quality issue for investigation, and in the case of external QAPs, 14 of the 16 respondents indicated this would be recorded. By comparison, only 11 respondents would log the same for an internal QAP (Table 1). All of those who would log these issues indicated that they may involve root cause analysis in the management of the issue (Fig. 3).


Fig3 Heavey ForSciIntSyn2023 7.jpg

Figure 3. Methods of recording issues related to Quality Assurance Programs (QAPs) that participants indicated may be used within their agency.

Audits and documented procedures

Audits are a flagship component of any QMS as a means of monitoring the compliance of the agency's system against the standard it has been designed to meet, as well as how that system aligns with the actual processes and procedures being performed within the agency.[22] A robust forensic science audit program will include review of both the competence of the QMS and processes, as well as technical (scientific) competence. Where the actual practice does not conform with the required (or documented) practice, a non-conformance will be noted by the auditor to be investigated and addressed by the auditee.

Supporting the QMS is the requirement for the agency to document the system, including the processes and procedures used, as part evidence of its compliance with the guiding standard.[4] It is therefore the expectation that actual practice will be performed in accordance with the documented practice. Misalignment of these two factors can be indicative of risks to quality outputs, introducing the potential for inconsistent application of processes, or procedures which are outside the scope of required standards. Fifteen out of 16 survey respondents indicated that they would log both non-conformances identified at audit and where a documented procedure was not followed as quality issues (Table 1), and over 90% of those indicated that the investigation may include a root cause analysis (Fig. 4). This high level of recording and investigative analysis of issues related to audit findings and documented procedures suggests that surveyed agencies place great significance of these types of issues as a symptom of risk to quality.

Fig4 Heavey ForSciIntSyn2023 7.jpg

Figure 4. Methods of recording issues related to audits and documented procedures that participants indicated may be used within their agency.

Equipment issues

Equipment involved in the provision of forensic science services can be loosely grouped into the categories of "critical" and "non-critical." The distinction between the two being the degree to which the failure of the equipment impacts on the validity of results. For example, a genetic analyzer used to perform DNA analysis is critical to the DNA testing process, whereas the flask used to measure a general cleaning reagent for dilution is non-critical.

Equipment which has an effect on the validity of results must be considered with regards to its selection, implementation, performance monitoring, and ongoing maintenance.[4] However, dependent on the criticality of the piece of equipment, the procedures required to assure this may vary. In the survey, participants were asked to consider whether equipment failure would be considered a quality issue and whether the distinction of critical versus non-critical became a factor. Thirteen out of 16 participants would record a critical equipment failure, whilst only seven would record failure of non-critical equipment (Table 1). Of those, 100% would consider logging the critical failures as a CAR with a root cause analysis, whist non-critical equipment failures are more likely to be managed without root cause analysis (Fig. 5).


Fig5 Heavey ForSciIntSyn2023 7.jpg

Figure 5. Methods of recording issues related to equipment that participants indicated may be used within their agency.

This demonstrates the perceived distinction between critical and non-critical with regards to equipment used in the forensic process and the impact it has on the provision of a quality service. In this survey, the terms "critical" and "non-critical" were not defined for the participants, therefore the responses received will be subjective dependent on the understanding of these terms. How these terms are defined and designated for the purposes of managing quality issues will be explored in future studies.

Customer feedback and monitoring of court testimony

For any service-based business, a strong knowledge of customer needs and expectations is crucial to designing fit-for-purpose service delivery strategies. Seeking regular and varied customer feedback to measure how the service aligns with those expectations is a standard feature of such models.[14] In that respect, forensic science service provision is a complex beast. In Australia and New Zealand, the majority of forensic science services are operated by government agencies to provide for the jurisdictions they represent, so when it comes to “seeking customer feedback,” who do we define as the customer? For the purposes of this standard requirement, the baseline used by most agencies will be those individuals or institutions who have a direct liaison with the outputs of the agency, for example the investigators who submit the exhibits for testing or the legal counsel who receive the reports for trial and call the analysts as expert witnesses. Other sources of external feedback may also be incorporated, such as independent reviews of the service provider or feedback related to the forensic service in the form of legal judgement transcripts.

“Customer feedback” for the purposes of this survey was categorized as either complaints or compliments/feedback (non-complaint or compliment). “Customer complaints” was the only type of quality issue surveyed which was noted as being logged by 100% of participants (Table 1), with almost all respondents indicating that root cause analysis would be considered in the event of a complaint (Fig. 6). The use of CARs as a logging tool across all types of customer feedback was much lower than amongst other types of issues, and a higher proportion of respondents indicated that other methods of logging customer feedback (inclusive of complaints, compliments and feedback) would be used, generally being some form of customer feedback database.


Fig6 Heavey ForSciIntSyn2023 7.jpg

Figure 6. Methods of recording issues related to customer feedback that participants indicated may be used within their agency.

The role of the forensic scientist as expert witness is often the ultimate conclusion in the forensic service process, and one which is responsible for translating analytical methodologies and the significance of the results into a public forum to be evaluated in the context of an adversarial legal system (in Australia and New Zealand). Thus, the monitoring of forensic expert court testimony plays an important part in quality assurance of the forensic process. Although court testimony monitoring is a requirement for forensic facilities accredited to ISO/IEC 17025, how this is to be managed is not specified.[8][23] Individual agency procedures may account for the few participants who did not indicate that “Court testimony feedback” would be logged as a quality issue. How this feedback is captured and managed will be investigated further in subsequent stages of this project.

Recording of "near miss" events

A number of the issues participants were asked to respond to may be considered as indicative of a potential issue in the quality system that requires action, but not a scenario where an actual non-conformance has occurred. These types of events may be referred to as “near miss” events. The results of the survey indicated that the proportion of participants who would consider logging “near miss” events as quality issues is notably lower than for events where a non-conformance has occurred (Fig. 7).


Fig7 Heavey ForSciIntSyn2023 7.jpg

Figure 7. Type of quality issue grouped by whether non-conformance has occurred, or is indicative of potential issue (“near miss”), and percentage of participants who would log this as a “quality issue.”

The lists in Fig. 7 were also compared with the corresponding data for each issue on whether root cause analysis may be used in the investigation from Fig. 2, Fig. 3, Fig. 4, Fig. 5, and Fig. 6. For the events where a non-conformance has occurred, the combined average of responses indicating that root cause analysis may be used across all events listed was 92%, whereas the average for the listed “near miss” events was only 74%.

The differences in the rates of recording and level of investigation used for “near miss” events compared with other quality issues highlights an important area for consideration with regards to quality issue management in forensics. Whilst there can be limited arguments against the importance of the identification, analysis, and correction of actual non-conformances within the QMS, a core feature of robust quality management is the design of systems to identify risks to quality so that they can be proactively strengthened to prevent non-conformances from even occurring in the first place.[24] This is where the identification of “near miss” events becomes critical to inform continuous, preventive improvement of the quality system, and consideration should be given to applying root cause analysis at this stage in order to determine the most effective means of preventing issues from occurring. Failing to log and action near-miss events will likely lead to an increase in actual non-conformances, some of which could have been prevented from occurring.

Terminology

Inconsistent definitions of terminology related to “error” and associated issues in forensic science have been noted as a key challenge to the transparent communication on this topic not just with end users of forensic information and the public, but also between forensic agencies themselves.[25] Only two participants in this survey indicated that their agency has a documented glossary of terms related to quality issues, with a further seven respondents indicating that “some” terms are defined within relevant standard operating procedures as appropriate.

All respondents were asked to define (based on their own knowledge and expertise) a number of terms commonly associated with quality issues, with selected results for some of the terms surveyed presented in Table 2.

Table 2. Selected survey participant definitions of terms commonly associated with quality issues in forensic science.
Term Selected participant definitions provided
Non-conformance “Departure from an approved procedure, standard or regulatory requirement. Can also include non-conforming results where a quality check (blank, known standard) has not passed acceptance criteria.” - Participant 10
 
“A failure to meet or not adhering to a requirement in the standard against which the facility is accredited - e.g. not performing a procedure as per the facility's documented standard operating procedure.” - Participant 13
 
“Non-compliance with a standard, unexpected result, issue that has or potential to have impact on quality of work.” - Participant 8
 
“Something deviating from process or procedures in the quality management system.” - Participant 5
 
“A failure to meet requirements set out in the Quality Manual, procedures/methods, AS ISO/IEC 17025 or other regulating bodies.” - Participant 15
Contamination “The presence of an artefact that was not part of the original item.” - Participant 1
 
“When another variable has been introduced to a sample or test, reducing the confidence of the test as being a 'true' result.'” - Participant 10
 
“As a more general definition, contamination would be the unintentional or undesired introduction of material that is exogenous to the exhibit/subsample after the point it is contained/bought into control/custody.” - Participant 13
 
“Where an artefact has impacted on a result.” - Participant 16
 
“The presence of an unexpected or unintended agent (e.g., biological, chemical), foreign material (e.g., trace) or substance (e.g., drug) that has the potential to negatively impact results.” - Participant 14
Error “Data that has been reported on incorrectly.” - Participant 6
 
“A mistake in the work being carried out. Mostly referring to human as the cause - but could be instrument related.” - Participant 5
 
“When a mistake has been made due to not following procedures or factors that could influence results have not been detected.” - Participant 10
 
“Something demonstrably inaccurate e.g., wrong number, units, result, name, date, time or documented procedure not followed.” - Participant 2
 
“Difference between the result obtained with the correct or expected result.” - Participant 9
 
“A mistake or accident - related to accuracy.” - Participant 15
 
“There is no consistency to the term error. It could mean where there is an incorrect result, a problem with documentation or an issue with an instrument or computer.” - Participant 16

Consistency in the definitions provided for some terms was high between participants (see “non-conformance”), whereas others indicated a range of rationales behind understandings of the terms. In particular, the definitions of “error” provided by participants demonstrate a breadth of potential meanings for the term, with one participant even noting that “this term is not used by our organization or categorized under our notification (non-conformance) system.”

Differences in key quality management terminology, including commonly used terms such as “error,” are apparent in the survey results, highlighting the difficulties that may arise in being able to compare quality issue data between agencies to identify trends at an inter-jurisdictional level. In the context of recently released mandatory reporting requirements on quality issues from the UK Forensic Regulator and ANAB[7][8], these differences in language used have the potential to influence inconsistent reporting, even inadvertently, which may have implications for the agency with regards to conclusions made from data based on the rate of reported quality issue types.

Disclosure of quality issues

All participants indicated that records of quality issues are disclosed to external recipients; however, less than half indicated that these would be disclosed at the same time as case results. The remaining responses indicated that such records are disclosed upon request. Some participants noted that contributing factors as to whether such records are disclosed include whether the quality issue affects the item being reported, whether it is being requested under the Freedom of Information Act or court order, or where the decision is at the discretion of the reporting analyst.

Inconsistent approaches to disclosure may be overcome by mandatory reporting frameworks, such as the Code of Practice issued by the UK Forensic Science Regulator.[7] Frameworks such as these are designed with the aim of enhancing transparency and oversight of risks to quality, the value of which cannot be argued. However, unless the underlying systems to ensure that the integrity and consistency of the reporting are sufficiently standardized, there will always be a risk of unnecessary over-reporting or inadvertent under-reporting of issues, both having outcomes which undermine the just intent of the measures. In addition, it is inevitable that certain issues may not be detected and reported until after legal proceedings are complete or may only become apparent as an issue as the sensitivity of testing methodologies increases over time. Mandatory reporting frameworks will have associated expectations and, potentially, public expectations as to their effectiveness in assuring safety and transparency. Delayed reporting of issues may be viewed adversely against those expectations if not managed and communicated effectively, and this should be considered by the forensic community.

Conclusions

Within the context of the Australian and New Zealand forensic community, the results of this survey demonstrate how the commitment to accreditation has resulted in considerable consistency in the QMSs in place within agencies. This consistency also translates to how several types of quality issues detected within those systems would be recorded, investigated, and managed, with RCA being applied for the majority of non-conformances identified. However, the survey results also demonstrate that this level of analysis is less consistently applied for events that could be indicators of an underlying quality issue, or a “near miss.” ISO/IEC 17025 requirements provide limited definition of the criteria by which a laboratory determines the significance of a nonconformity, which requires action and subsequent analysis to determine the root cause and, therefore, for compliance the laboratory must make its own determinations on the significance of issues. This can lead to “near miss” events not being recorded in the same manner as observed non-conformances and, more significantly, not being investigated with the same level of scrutiny to determine the root cause. This appears to be at odds with the risk-based and proactive intent of modern QMSs and highlights an opportunity for all forensic agencies and networks to advocate practice beyond just compliance to enhance their continuous improvement programs through the thorough and rigorous identification and analysis of “near miss” events.

Further, it is important to note that the management of quality issues for the improvement of systems is not just about the detection of adverse, negative events. “Near miss” events—such as where results may have been invalidated or where an internal QAP picked up an issue in the system before it affected results, although they may be considered “quality issues”—are positive indicators that the QMS is working effectively by proactive identification of risks, which can be investigated and actioned to prevent reoccurrence. Additional positive system indicators may come in the form of customer compliments or positive feedback on court testimony. Combining the data from the analysis of positive indicators alongside the cause analysis of adverse events provides the agency with a deeper understanding of the root causes identified and, therefore, has the potential to inform more meaningful and effective corrective action to eliminate risks.

This foundational study supports the need for further research into the development of standardized systems of critical issue classification, management and disclosure in forensic science, as highlighted by the increased international demand for mandatory reporting and transparency.[6][7][8] Although the Australian and New Zealand forensic service provision community may be small when considered against international counterparts, the size and isolation of this community has enabled an enviable level of control and oversight across all its jurisdictions for an extended period. Since the creation of the National Institute of Forensic Science in 1992 and subsequent development of governance structures from both policing and scientific branches of forensic investigation, the progression of standardized practice in forensic science across Australia and New Zealand has moved largely by consensus, led by the government agencies that provide the majority of forensic services. This highly effective and active governance network also places Australia and New Zealand in a unique position to make holistic change to forensic science service provision policy across the entire region to assure equitable access to high-quality, contemporary forensic services.

Through further investigations involving the Australian and New Zealand forensic science community, along with examination of international jurisdictions and approaches to critical issue management and analysis from other high-risk fields, this ongoing research aims to develop an evidence-based standard terminology for quality issues in forensic science to support data sharing and reporting, enhance understanding of quality issues, and promote transparency in forensic science. A common language forms the basis for a taxonomic system for the categorization of critical issues to support the consistent and effective communication of quality issues, whether through mandatory reporting obligations, effective delivery of forensic information in investigations or as evidence, or to enhance public understandings of forensic science and the significance of “error” in the pursuit of open and reliable forensic science services to support justice outcomes.

Acknowledgements

The authors wish to thank the survey participants for their responses and generosity of time. We would also like to acknowledge the involvement and support of ANZFEC for the approval to conduct this survey, and ANZPAA-NIFS for their assistance in approaching ANZFEC and disseminating the invitation to participate. Thanks also to Dr. Matthew Thompson and the members of the Perceptual Expertise Advanced Research Lab at Murdoch University for their advice on survey research and design. Portions of this manuscript were presented at the 25th International Symposium of the Australia New Zealand Forensic Science Society held in Brisbane, Australia, 11 - 15 September 2022. This study has been approved by the Curtin University Human Research Ethics Committee (approval number HRE2021-0252).

Author contributions

  • Anna Heavey – Conceptualization, Methodology, Formal analysis, Visualization, Investigation, Writing – Original Draft, Writing – Review & Editing.
  • Gavin Turbett - Conceptualization, Methodology, Writing – Review & Editing, Supervision.
  • Max Houck – Conceptualization, Methodology, Writing – Review & Editing, Supervision.
  • Simon Lewis - Conceptualization, Methodology, Writing – Review & Editing, Supervision, Project administration.

Competing interests

The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: At the time of the survey being conducted, Anna Heavey was Chair of the Quality Specialist Advisory Group,3. and Dr. Gavin Turbett was Chair of the Australia New Zealand Forensic Executive Committee. As an Editorial Board Member of Forensic Science International: Synergy, Professor Simon Lewis had no involvement in the peer review of this manuscript. Dr. Max Houck, the Editor in Chief of Forensic Science International: Synergy, is a member of the first author's doctoral committee. He has had input on early drafts of this manuscript prior to selecting a journal for publication. A Senior Editor has handled the administration and peer review process for the manuscript, and Dr. Houck has had no input on the processing of the manuscript except as co-author.

References

  1. 1.0 1.1 1.2 Ross, Alastair; Neuteboom, Wim (4 May 2021). "Implementation of quality management from a historical perspective:the forensic science odyssey" (in en). Australian Journal of Forensic Sciences 53 (3): 359–371. doi:10.1080/00450618.2019.1704058. ISSN 0045-0618. https://www.tandfonline.com/doi/full/10.1080/00450618.2019.1704058. 
  2. 2.0 2.1 Wilson-Wilde, L. (2021). "The State of Forensic Science in Australia and New Zealand". Forensic Science Review 33 (1): 2–6. ISSN 1042-7201. http://forensicsciencereview.com/content.htm#v33. 
  3. Wilson-Wilde, Linzi (2 January 2017). "The future of the National Institute of Forensic Science – implications for Australia and New Zealand" (in en). Australian Journal of Forensic Sciences 49 (1): 1–8. doi:10.1080/00450618.2016.1253773. ISSN 0045-0618. https://www.tandfonline.com/doi/full/10.1080/00450618.2016.1253773. 
  4. 4.0 4.1 4.2 4.3 4.4 "ISO/IEC 17025:2017 General requirements for the competence of testing and calibration laboratories". International Organization for Standardization. March 2018. https://www.iso.org/standard/66912.html. 
  5. 5.0 5.1 Heavey, Anna L.; Turbett, Gavin R.; Houck, Max M.; Lewis, Simon W. (1 July 2022). "Toward a common language for quality issues in forensic science" (in en). WIREs Forensic Science 4 (4): e1452. doi:10.1002/wfs2.1452. ISSN 2573-9468. https://wires.onlinelibrary.wiley.com/doi/10.1002/wfs2.1452. 
  6. 6.0 6.1 6.2 Sofronoff, W. (13 December 2022). "Final Report: Commission of Inquiry into Forensic DNA testing in Queensland" (PDF). https://www.health.qld.gov.au/__data/assets/pdf_file/0036/1196685/final-report-coi-dna-testing-qld-dec-2022.pdf. 
  7. 7.0 7.1 7.2 7.3 Forensic Science Regulator (March 2023). "Code of Practice" (PDF). https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/1178250/FINAL_2023.1.18_Code_of_Practice.pdf. 
  8. 8.0 8.1 8.2 8.3 "Accreditation Requirements for Forensic Testing and Calibration (2023)" (PDF). ANSI National Accreditation Board. 1 February 2023. https://anab.qualtraxcloud.com/ShowDocument.aspx?ID=12371. 
  9. 9.0 9.1 "The State of Western Australia vs Edwards [No 7 [2020] WASC 339"] (PDF). State of Western Australia. https://ecourts.justice.wa.gov.au/eCourtsPortal/Decisions/DownloadDecision?id=9582e4ba-91ec-4119-998b-269447363aed. 
  10. McCartney, Carole; Nsiah Amoako, Emmanuel (1 July 2019). "Accreditation of forensic science service providers" (in en). Journal of Forensic and Legal Medicine 65: 143–145. doi:10.1016/j.jflm.2019.04.004. https://linkinghub.elsevier.com/retrieve/pii/S1752928X19300253. 
  11. Sommer, Peter (1 June 2018). "Accrediting digital forensics: What are the choices?" (in en). Digital Investigation 25: 116–120. doi:10.1016/j.diin.2018.04.004. https://linkinghub.elsevier.com/retrieve/pii/S1742287618301701. 
  12. "Best Practice Manual for Facial Image Comparison ENFSI-BPM-DI-01" (PDF). European Network of Forensic Science Institutes. January 2018. https://enfsi.eu/wp-content/uploads/2017/06/ENFSI-BPM-DI-01.pdf. 
  13. "Facial Identification Scientific Working Group". Facial Identification Scientific Working Group. 2022. http://www.fiswg.org/index.html. 
  14. 14.0 14.1 14.2 "ISO 9001:2015 Quality management systems — Requirements". International Organization for Standardization. September 2015. https://www.iso.org/standard/62085.html. 
  15. Casey, Eoghan; Ferraro, Monique; Nguyen, Lam (1 November 2009). "Investigation Delayed Is Justice Denied: Proposals for Expediting Forensic Examinations of Digital Evidence*" (in en). Journal of Forensic Sciences 54 (6): 1353–1364. doi:10.1111/j.1556-4029.2009.01150.x. ISSN 0022-1198. https://onlinelibrary.wiley.com/doi/10.1111/j.1556-4029.2009.01150.x. 
  16. Charan, J.L.; Manikyam, K.S. (2023). "Forensic Science and its Limitations in Rape and Murder Cases in India". Journal of Forensic Science and Medicine 9 (1): 91–97. doi:10.4103/jfsm.jfsm_98_21. 
  17. Airlie, Melissa; Robertson, James; Krosch, Matt N.; Brooks, Elizabeth (1 March 2021). "Contemporary issues in forensic science—Worldwide survey results" (in en). Forensic Science International 320: 110704. doi:10.1016/j.forsciint.2021.110704. https://linkinghub.elsevier.com/retrieve/pii/S0379073821000244. 
  18. "ISO 45001:2018 Occupational health and safety management systems — Requirements with guidance for use". International Organization for Standardization. March 2018. https://www.iso.org/standard/63787.html. 
  19. Robitaille, D. (2011). The Corrective Action Handbook (2nd ed.). Paton Professional. ISBN 9781932828405. 
  20. "What Is Root Cause Analysis (RCA)?". American Society for Quality. 2023. https://asq.org/quality-resources/root-cause-analysis. 
  21. 21.0 21.1 Houck, Max M. (6 December 2016). "Risk, Reward, and Redemption: Root Cause Analysis in Forensic Organizations" (in en). Forensic Science Policy & Management: An International Journal 7 (3-4): 106–112. doi:10.1080/19409044.2016.1224278. ISSN 1940-9044. https://www.tandfonline.com/doi/full/10.1080/19409044.2016.1224278. 
  22. 22.0 22.1 22.2 Heavey, Anna L. (2023), "Principles of Quality Assurance" (in en), Encyclopedia of Forensic Sciences, Third Edition (Elsevier): 290–297, doi:10.1016/b978-0-12-823677-2.00173-2, ISBN 978-0-12-823678-9, https://linkinghub.elsevier.com/retrieve/pii/B9780128236772001732. Retrieved 2023-10-17 
  23. "Specific Accreditation Criteria, ISO/IEC 17025 Application Document, Legal (including Forensic Science) - Appendix" (PDF). National Association of Testing Authorities. September 2023. https://nata.com.au/files/2021/05/Forensic-Science-ISO-IEC-17025-Appendix-effective-feb-2020.pdf. 
  24. Sheikhtaheri, Abbas (1 June 2014). "Near Misses and Their Importance for Improving Patient Safety". Iranian Journal of Public Health 43 (6): 853–854. ISSN 2251-6085. PMC 4475608. PMID 26110160. https://pubmed.ncbi.nlm.nih.gov/26110160. 
  25. Christensen, Angi M.; Crowder, Christian M.; Ousley, Stephen D.; Houck, Max M. (1 January 2014). "Error and its Meaning in Forensic Science" (in en). Journal of Forensic Sciences 59 (1): 123–126. doi:10.1111/1556-4029.12275. ISSN 0022-1198. https://onlinelibrary.wiley.com/doi/10.1111/1556-4029.12275. 

Notes

This presentation is faithful to the original, with changes to presentation, spelling, and grammar as needed. The PMCID and DOI were added when they were missing from the original reference. Everything else remains true to the original article, per the "NoDerivatives" portion of the distribution license.