Keywords
Administrative healthcare data, Patient safety, Adverse events, Validation, Chart review
Patient safety is a key issue for health systems and a growing global public health challenge. Administrative healthcare data provide a coded summary of a patient and their encounter with the healthcare system. These aggregated datasets are often used to inform research and decisions relating to health service planning and therefore it is vital that they are accurate and reliable. Given the reported inaccuracy of these datasets for detecting and recording adverse events, there have been calls for validation studies to explore their reliability and investigate further their potential to inform research and health policy. Researchers have since carried out validation studies on the rates of adverse events in administrative data through chart reviews therefore, it seems appropriate to identify and chart the evidence and results of these studies within a scoping review.
The scoping review will be conducted in accordance with the Joanna Briggs Institute (JBI) methodology for scoping reviews. A search of databases such as PubMed, CINAHL, ScienceDirect and Scopus will be conducted in addition to a search of the reference lists of sourced publications and a search for grey literature. Following this, Covidence will be used to screen the sourced publications and subsequently extract data from the included sources. A numerical summary of the literature will be presented in addition to a charting based on the qualitative content analysis of the studies included.
This protocol provides the structure for the conduct of a review to identify and chart the evidence on validation studies on rates of adverse events in administrative healthcare data. This review will aim to identify research gaps, chart the evidence of and highlight any flaws within administrative datasets to improve extraction and coding practices and enable researchers and policy makers to use these data to their full potential.
Administrative healthcare data, Patient safety, Adverse events, Validation, Chart review
The manuscript has been amended to provide further clarity based on the reviewers’ feedback and comments. The manuscript has also been reduced to remove redundant text as per the feedback received. The manuscript has also amended to place further emphasis on the prevention of adverse events and the improvement of quality care as the ultimate goal of validating administrative datasets and ensuring their accuracy. Furthermore, the potential benefit of this work to highlight the capacity of linking accurate administrative data with staffing information to demonstrate the impact of safe/unsafe staffing has also been addressed in the revised manuscript.
See the authors' detailed response to the review by Keith Marsolo
See the authors' detailed response to the review by Angela Flynn
Patient safety is a key issue for health systems and a growing global public health challenge. The human cost of patient safety incidents is high with adverse events causing injury, disability or death and resulting in suffering for patients and family members. The financial and economic burdens of patient safety incidents are also of concern with further healthcare expenditure required for increased length of stays in hospital and additional tests, treatments and healthcare to be provided to patients (WHO, 2021b). The Harvard Medical Practice Study (Brennan et al., 1991) and the Institute of Medicine (2000) identified high rates of adverse events in hospital admissions and marked the beginning of the patient safety movement. This led to rates of adverse events becoming an important quality indicator across health systems. Adverse events have since been reported and measured however, challenges associated with the accuracy of their recording in administrative data has been widely acknowledged (Kim et al., 2022; Rodrigo-Rincon et al., 2015). Evidently, patient safety is a key healthcare quality concern and with adverse events as main contributors to patient harm, their monitoring and measurement is a major priority.
Routinely collected healthcare data are regularly collected for purposes other than research (Welk & Kwong, 2017). These data, frequently called administrative data, capture the patient’s characteristics, their diagnoses, treatments and procedures and, essentially provide a coded summary of the patient and their encounter with the healthcare system (Healthcare Pricing Office, 2020).
The key events of a patient’s stay in hospital are documented in a discharge summary within their clinical chart or in their electronic health record (EHR). Once a patient has been discharged from hospital, their chart is translated into alphanumerical data by professional coding specialists (Tang et al., 2017). In addition to a patient’s discharge summary, clinical coders may also code information in documents such as nursing notes, consultation reports, progress notes, operative reports, pre- and post-operative reports and pathology reports (Healthcare Pricing Office, 2020). Nursing notes have previously been identified as important sources of data for clinical coders (Alonso et al., 2020; Doktorchik et al., 2020; Lucyk et al., 2017) and are particularly insightful in relation to identifying adverse events such as pressure injuries, which are a well-recognised quality indicator (Weller et al., 2022). The data captured by the alphanumerical codes include patient diagnoses, procedures, hospital services used and any complications that arose during a patient’s stay.
Various classification systems such as the International Classification of Diseases (ICD) are used to code this data. The ICD translates diagnoses and other health-related issues from words into alphanumerical data, allowing for an efficient method of storing, retrieving, and analysing data (WHO, 2016). Coding systems such as ICD allow for a consistent and standard way of recording, reporting and monitoring diseases. ICD is the is used by all members states of WHO, therefore data can be shared, understood and compared across hospitals, regions and countries (WHO, 2023a). Various derived classifications or extensions of the fundamental ICD classifications have been developed and are used to monitor diagnoses in specific areas or settings (WHO, 2023b). Clinical coders follow clearly defined ICD coding guidelines to assign alphanumeric codes to each diagnosis or event recorded in a patient’s chart (O’Malley et al., 2005). As the most widely used classification of diseases, ICD has had 10 iterations since 1900 (O’Malley et al., 2005). The 10th revision of the ICD (ICD-10), which contains more than 155,000 codes (DiSantostefano, 2009), has been widely used since its development about 30 years ago (WHO, 2021a). An 11th revision of the ICD (ICD-11 was adopted by World Health Assembly in 2019 and came into effect in 2022 (WHO, 2023c). Various other coding systems such as the Current Procedural Terminology (CPT), which was developed by the American Medical Association, are used to capture and assign codes to aspects of healthcare data (Dotson, 2013) however, ICD is recognised globally with all WHO member states committed to using this classification system.
Administrative data are often used at national and international levels in health service planning and to inform policy and epidemiological, clinical and biomedical research (Hemkens et al., 2016). They have also been utilised for carrying out audits, developing financial strategies and determining the distribution of health resources (Burns et al., 2012). Their use in the evaluation of the quality of health care delivery has also been well-recognised (Clarke et al., 2019). Given that these datasets are often used for purposes other than those for which they were initially collected, their use in research and evaluation has been labelled as a secondary use (Jorm, 2015). Administrative data can contain rich and valuable information and their use in research has been increasingly recognised. These databases are advantageous for research as they allow research questions that require large sample sizes to be answered (Harron et al., 2017). Their potential to inform research to enhance the efficiency of health systems and improve population health is well-documented and has led to their increased use in research (Jorm, 2015; Mitchell & Braithwaite, 2021; Moorthie et al., 2022). Patient safety researchers in particular have been increasingly recognising the potential of using administrative data for research purposes and for the monitoring and reporting of patient safety events (Raleigh et al., 2008). Given the association between lower registered nurse staffing levels, increased mortality rates and other adverse outcomes (Griffiths et al., 2016), validating administrative datasets and ensuring their accuracy may also be beneficial in encouraging the potential for these datasets to be linked to nurse staffing data to demonstrate the impact of nurse staffing on patient outcomes and highlight areas for improvement.
Given that these data are used to inform research that may help to improve patients’ health and well-being and inform decisions that will impact on the quality of care that patients receive, it is vital that these administrative datasets are of high-quality and that precise and accurate diagnostic codes are used to generate them (Gologorsky et al., 2014; Ibrahim et al., 2021). Accurate measurement of adverse events is central to improving patient safety and healthcare quality by allowing priority areas to be identified, improvement strategies to be designed and implemented interventions to be evaluated (Classen et al., 2011). Accurate data on adverse events may contribute to the development, implementation and evaluation of more targeted and evidence-based patient safety interventions that can prevent adverse events and improve quality of care (Dillner et al., 2023). Although these datasets can provide rich and detailed information, they are subject to limitations. Previous studies indicate that administrative datasets are not always accurate and therefore, may contribute to inaccuracies in research and lead to poorly informed decisions (McGuckin et al., 2022; Nicholls et al., 2017). In particular, the inaccuracy of rates of adverse events in administrative data has been previously identified (Kim et al., 2022). Coding of primary diagnoses and procedures in administrative data is generally accurate however, coding of secondary diagnoses is less complete and therefore, adverse events rates are often underestimated (Raleigh et al., 2008).
Various challenges associated with using administrative data in research have been reported. Such challenges include not all data being abstracted from the patient chart, coded and represented in the database, incomplete data, difficulty determining the number of out-patient visits due to imprecise coding and difficulty capturing all relevant tests due to various diagnostic codes being used for the same test (McGuckin et al., 2022). The inconsistency in the definitions and terminology used in practice have previously been identified by Naessens et al. (2009) as another challenge to the accurate reporting of patient safety issues. Given that these administrative datasets are generated for purposes other than research, their accuracy and appropriateness for detecting adverse events is varied and their poor representation in some coding schemes has also previously been reported (Bates et al., 2003). The limited accuracy and variability of rates of hospital acquired infections in administrative data has also been identified as a challenge associated with using these datasets for research or benchmarking purposes. This has also resulted in calls for validation of these datasets and improvement of the algorithms used to generate them (van Mourik et al., 2015). Although the inaccuracy of administrative data for recording rates of adverse events has been previously reported by researchers such as Walther et al. (2021), who identified poor recording of postpartum haemorrhage within administrative data, the potential of these datasets to provide accurate information is also being recognised. Ackroyd-Stolarz et al. (2014) found a relatively high degree of accuracy in such datasets for identifying adverse events in older patients, therefore demonstrating the potential value of these datasets to inform research and improve patient safety.
Given the lack of consistency in the accuracy of these datasets for recording adverse events, there have been calls for validation studies to explore their reliability and investigate further their potential to inform research and health policy (Ehrenstein et al., 2016; Jorm, 2015; Nicholls et al., 2017). Again, the potential of these datasets to inform research and policy in relation to patient safety and the human and economic cost of adverse events is evident, however specific calls for validating rates of adverse events within these datasets have been made (Raleigh et al., 2008; van Mourik et al., 2015). Many researchers have since carried out validation studies on various administrative datasets therefore, it seems necessary to chart the evidence and results of these studies within this scoping review.
Validating administrative datasets is vital in determining the reliability and credibility of research based on this data and a common way of validating such data by manually comparing the coded data to the data recorded in a patient’s healthcare record (Nissen et al., 2019). Patient charts have been referred to as the ‘gold standard’ reference point in many validation studies (Ehrenstein et al., 2016), therefore this scoping review aims to present an overview of how this method of validating administrative data has been used by previous researchers. Various methodological approaches such as the Harvard Medical Practice Study (HMPS) and the Global Trigger Tool (GTT) can be used to conduct chart reviews. The HMPS method is a two-stage approach involving a nurse review for the triggering of charts, followed by a physician review of triggered charts to confirm rates of patient safety incidents using set definitions (Brennan et al., 1991) The GTT was developed by the Institute for Healthcare Improvement in response to demands for a less labour-intensive approach to chart reviewing. This two-stage approach involves a 20-minute screening phase for a comprehensive list of triggers, followed a physician review of the triggers using broader definitions to confirm the occurrence of an adverse event (Rafter et al., 2015). The HMPS method is a two-stage approach involving a nurse review for the triggering of charts, followed by a physician review of triggered charts to confirm rates of patient safety incidents using set definitions (Brennan et al., 1991) The GTT was developed by the Institute for Healthcare Improvement in response to demands for a less labour-intensive approach to chart reviewing. This two-stage approach involves a 20-minute screening phase for a comprehensive list of triggers, followed a physician review of the triggers using broader definitions to confirm the occurrence of an adverse event (Rafter et al., 2015).
In line with the JBI methodological framework for carrying out scoping reviews (Peters et al., 2020), the inclusion and exclusion criteria for studies that may potentially be included in the review will be defined by the population, concept and context screening criteria (Table 1).
As this is a protocol for a scoping review exploring the use of chart reviews to validate rates of adverse events in administrative data, the population of interest is hospital-based patients. These are Patients who have been discharged from hospital and have had the information within their medical chart extracted, assigned ICD codes, and recorded in administrative data Hospital-based patients’ data will have been used to inform the publications that will be included in this scoping review, therefore making them the population of interest in this review. This research will seek to focus on identifying and charting the evidence of reviewing hospital-based patient’s charts as a method of carrying out validation studies on rates of adverse events in administrative datasets.
Validating the accuracy of the coding of administrative data is key for ensuring that the policy, practice and research informed by such datasets is credible and robust (Nicholls et al., 2017). A review of patient charts is a common and effective way of validating the accuracy of administrative data as it allows the coded data recorded in the administrative dataset to be compared to the data maintained within hospital notes or discharge summaries within patient charts (Burns et al., 2012). The chart review methodology is an efficient way of determining whether the administrative data is an accurate reflection of the patient’s physical medical chart (Nissen et al., 2019). The concept explored in this scoping review will be the validation of rates of adverse events in administrative data through chart review.
Validation studies that have been conducted in hospital-based settings will be considered for inclusion in this review. Chart reviews that have been conducted as part of a validation of rates of adverse events in administrative datasets that have been carried out in any country will be included in this review. There will be no limitation on the context of the administrative data in order to allow a comprehensive review to be carried out.
The types of sources of evidence to be included in this review will be left open to allow a comprehensive review of all available literature to be carried out. This will allow for the inclusion of various types of publications and therefore, should provide an in-depth presentation of the previous literature published in relation to using chart reviews to validate administrative data.
This scoping review will be conducted in accordance with the Joanna Briggs Institute (JBI) methodology for scoping review (Peters et al., 2020). The JBI methodological framework will be used to guide this scoping review and ensure that the review is carried out appropriately.
The sources of evidence that will be included in this scoping review will be identified and collated using a three-step process as outlined by the JBI (Peters et al., 2020). The first step of this process will involve an initial search of two databases. The databases included in this initial search will be PubMed (MEDLINE) and CINAHL. Following this preliminary search, an analysis of the text words in the titles and abstracts of identified papers will be carried out. The indexing terms used to describe the retrieved papers will also be analysed. The second step of this process will involve using the identified keywords and index terms to perform a second search of all included databases. The databases included in this step will include PubMed (MEDLINE), CINAHL, Web of Science and Scopus. The third and final step of this process will involve a search of the reference lists of identified publications for additional sources of materials to be included in this scoping review. The reference lists of all publications selected from full text or those included in the review will be searched in order to ensure a comprehensive and in-depth search for existing literature. A search for grey literature will also be conducted to ensure that the review is as thorough as possible (Arksey & O’ Malley, 2005). Grey literature refers to literature that is not commercially published and not usually available within standard databases. Examples of such literature include government reports, conference proceedings, theses and dissertations. A search for this type of literature is beneficial as many researchers publish through other means, such as those mentioned above, and therefore, this allows for the inclusion of a wider scope of material. Also, although such materials may not be included in the final scoping review, they can aid in identifying data sources, research currently being undertaken in the area or complete study reports (Pawliuk et al., 2021).
Sources included in this scoping review will be limited to publications written in English as it will not be feasible for the research team to translate publications published in other languages into English. Literature published over the past 30 years, between 1991 and 2023, will be eligible for inclusion in this study. This time period has been chosen as the patient safety movement was instigated by Brennan et al. (1991) about 30 years ago following their study that investigated the incidence of adverse events and negligence in hospitalised patients through the use of chart reviews.
A librarian from Dublin City University has been consulted and their advice was sought in relation to developing a sample search strategy (Table 2). Advice from the librarian will continue to be sought in relation to the search strategy to ensure that a comprehensive search of the available literature is completed. The research team will continue to work closely with the librarian throughout the data collection phase of the scoping review to make certain that repetitions of the search strategy result in all key and relevant literature being gathered.
After the search for literature to be potentially included in this scoping review, all identified publications will be gathered and their citations will be stored in the bibliographic software Zotero.
The review management tool Covidence will be used to screen the collected publications and subsequently extract data from included sources. The references of the sourced publications will be imported to Covidence from Zotero and any duplicates will be automatically identified and removed. Using Covidence, two members of the research team will independently screen each of the imported studies for inclusion based on title and abstract examination. The researchers will individually assess the eligibility of each publication for inclusion in the review by making reference to the inclusion criteria. The researchers will each have the option of selecting yes, no or maybe at this stage of the selection process. Any conflicts that arise between the two researchers at this initial screening stage will be resolved by discussion and by consulting a third member of the research team to reach a final decision if necessary. Sources of data that received a yes vote from both researchers and are deemed potentially relevant based on the initial assessments made by researchers after examining the titles and abstracts will be retrieved in their full-text version. The full-text versions of potential studies for inclusion will be read and again, be assessed against the inclusion criteria by two researchers independently. The researchers will have the option in Covidence to include or exclude publications at this stage. If the researchers vote to exclude any publication, they will be prompted to select a reason for exclusion. This decision-making process will be documented and reported in the scoping review and also displayed in a PRISMA flowchart. The researchers will then be guided to the consensus process. The consensus screen in Covidence highlights the number of conflicts between the two researchers and allows a final decision to be reached in relation to the inclusion of publications. In the event that any disagreement between the two members of the research team assessing the studies for inclusion in the scoping review arises, a third member of the research team will be consulted, and their input will be used to resolve such conflicts of opinions. In line with the JBI guidelines (Peters et al., 2020), a PRISMA flowchart of the review process will be developed and included in the review. This flowchart will detail the full flow of the review process from the initial search to the presentation of included sources.
The data extraction method for this scoping review will firstly involve the development of a data extraction template. As previously mentioned, the data will be extracted from the publications selected for inclusion in the scoping review within Covidence. The key information that will be extracted will include the following: name of author(s), year of publication, geographical location of study, aim of the study, methodology and key findings that relate to the research question. The data extraction method will be piloted and trialled on two or three sources to ensure that all relevant data is extracted. After this step, the extraction template will be updated to ensure that it is an appropriate method of collecting all necessary data. The piloting of the template will be carried out by two members of the research team as recommended by Peters et al. (2020). Once these data have been extracted, they will be exported to Microsoft Excel.
In line with the JBI methodological framework for scoping reviews, the extracted data will be arranged into a charting table in order to provide the reviewers and readers with a clear and descriptive summary of the publications included in the scoping review (Peters et al., 2020). According to Peters et al. (2020), the data charting can be an iterative process and therefore, the charting tool will be piloted, and the table will be continuously updated. The results of this scoping review will be mapped descriptively in order to answer the research question. A narrative account of the included literature will be presented. A numerical summary of the literature will be presented in addition to a qualitative content analysis of the studies included in the scoping review. Both the numerical summary and qualitative content analysis will be of a descriptive nature as the aim of a scoping review is to present an overview of the literature reviewed rather than an assessment of the quality of the included materials (Arksey & O’Malley, 2005; Peters et al., 2020). The descriptive numerical summary will include information such as the overall number of publications that have been included, the years of publication, the geographical location where the studies were conducted, and the types of study designs used. The qualitative content analysis will answer the research question by providing a description of the results of each of the included studies. The data extracted from the included studies will be presented in a tabular and descriptive format.
Considering the implications of the findings produced by a scoping review has been highlighted as an important part of the scoping review methodological framework (Levac et al., 2010). Given that this is viewed as an essential element of conducting a scoping review, the implications of the results of this scoping review will be considered in relation to the identification of research gaps or calls for further research in the area.
The scoping review is currently in the preliminary stages of searching the databases. Step 1 of the three-step search strategy, as outlined by Peters et al. (2020) for identifying and collating sources of evidence to inform the scoping review, is currently underway. An initial search of databases for potentially relevant sources of evidence is being performed.
This protocol provides the structure for the conduction of a review to identify and chart the evidence of validation studies of rates of adverse events in administrative data. In line with the JBI guidelines (Peters et al., 2020), any deviations of the scoping review from the protocol will be acknowledged and discussed in the scoping review. It is widely acknowledged that administrative data have the potential to enhance patient safety and be used as a quality indicator however, given that the rates of adverse events in these datasets have been frequently reported as inaccurate, there is a strong rationale to conduct this scoping review. This review will aim to present an overview of the chart review approaches and tools that have been used to validate rates of AEs in administrative datasets and ultimately highlighting the need to develop a systematic approach to measuring rates of AEs in administrative data. This will allow the accuracy of these datasets to be improved, which will enable researchers and policy makers to use them to their full potential to improve patient safety and healthcare quality.
Is the rationale for, and objectives of, the study clearly described?
Yes
Is the study design appropriate for the research question?
Yes
Are sufficient details of the methods provided to allow replication by others?
Yes
Are the datasets clearly presented in a useable and accessible format?
Not applicable
Competing Interests: No competing interests were disclosed.
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: informatics, real-world data research.
Is the rationale for, and objectives of, the study clearly described?
Yes
Is the study design appropriate for the research question?
Yes
Are sufficient details of the methods provided to allow replication by others?
Yes
Are the datasets clearly presented in a useable and accessible format?
Not applicable
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: Nursing, Health inequities, Inclusion Health
Is the rationale for, and objectives of, the study clearly described?
Partly
Is the study design appropriate for the research question?
Partly
Are sufficient details of the methods provided to allow replication by others?
Yes
Are the datasets clearly presented in a useable and accessible format?
Not applicable
Competing Interests: No competing interests were disclosed.
Reviewer Expertise: informatics, real-world data research.
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | |||
---|---|---|---|
1 | 2 | 3 | |
Version 2 (revision) 12 Dec 24 |
read | read | |
Version 1 23 Mar 23 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Register with HRB Open Research
Already registered? Sign in
Submission to HRB Open Research is open to all HRB grantholders or people working on a HRB-funded/co-funded grant on or since 1 January 2017. Sign up for information about developments, publishing and publications from HRB Open Research.
We'll keep you updated on any major new updates to HRB Open Research
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Comments on this article Comments (0)