Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

The OMOP common data model in Australian primary care data: Building a quality research ready harmonised dataset

  • Roger Ward,

    Roles Conceptualization, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Health & Biomedical Research Information Technology Unit (HaBIC R2), Department of General Practice and Primary Care, Faculty of Medicine, Dentistry & Health Sciences, The University of Melbourne, Parkville, Victoria, Australia

  • Christine Mary Hallinan ,

    Roles Conceptualization, Methodology, Validation, Visualization, Writing – review & editing

    hallinan@unimelb.edu.au

    Affiliation Health & Biomedical Research Information Technology Unit (HaBIC R2), Department of General Practice and Primary Care, Faculty of Medicine, Dentistry & Health Sciences, The University of Melbourne, Parkville, Victoria, Australia

  • David Ormiston-Smith,

    Roles Data curation, Formal analysis, Methodology, Resources, Writing – review & editing

    Affiliation Health & Biomedical Research Information Technology Unit (HaBIC R2), Department of General Practice and Primary Care, Faculty of Medicine, Dentistry & Health Sciences, The University of Melbourne, Parkville, Victoria, Australia

  • Christine Chidgey,

    Roles Data curation, Formal analysis, Methodology, Writing – review & editing

    Affiliation Health & Biomedical Research Information Technology Unit (HaBIC R2), Department of General Practice and Primary Care, Faculty of Medicine, Dentistry & Health Sciences, The University of Melbourne, Parkville, Victoria, Australia

  • Dougie Boyle

    Roles Conceptualization, Writing – review & editing

    Affiliation Health & Biomedical Research Information Technology Unit (HaBIC R2), Department of General Practice and Primary Care, Faculty of Medicine, Dentistry & Health Sciences, The University of Melbourne, Parkville, Victoria, Australia

Abstract

Background

The use of routinely collected health data for secondary research purposes is increasingly recognised as a methodology that advances medical research, improves patient outcomes, and guides policy. This secondary data, as found in electronic medical records (EMRs), can be optimised through conversion into a uniform data structure to enable analysis alongside other comparable health metric datasets. This can be achieved with the Observational Medical Outcomes Partnership Common Data Model (OMOP-CDM), which employs a standardised vocabulary to facilitate systematic analysis across various observational databases. The concept behind the OMOP-CDM is the conversion of data into a common format through the harmonisation of terminologies, vocabularies, and coding schemes within a unique repository. The OMOP model enhances research capacity through the development of shared analytic and prediction techniques; pharmacovigilance for the active surveillance of drug safety; and ‘validation’ analyses across multiple institutions across Australia, the United States, Europe, and the Asia Pacific. In this research, we aim to investigate the use of the open-source OMOP-CDM in the PATRON primary care data repository.

Methods

We used standard structured query language (SQL) to construct, extract, transform, and load scripts to convert the data to the OMOP-CDM. The process of mapping distinct free-text terms extracted from various EMRs presented a substantial challenge, as many terms could not be automatically matched to standard vocabularies through direct text comparison. This resulted in a number of terms that required manual assignment. To address this issue, we implemented a strategy where our clinical mappers were instructed to focus only on terms that appeared with sufficient frequency. We established a specific threshold value for each domain, ensuring that more than 95% of all records were linked to an approved vocabulary like SNOMED once appropriate mapping was completed. To assess the data quality of the resultant OMOP dataset we utilised the OHDSI Data Quality Dashboard (DQD) to evaluate the plausibility, conformity, and comprehensiveness of the data in the PATRON repository according to the Kahn framework.

Results

Across three primary care EMR systems we converted data on 2.03 million active patients to version 5.4 of the OMOP common data model. The DQD assessment involved a total of 3,570 individual evaluations. Each evaluation compared the outcome against a predefined threshold. A ’FAIL’ occurred when the percentage of non-compliant rows exceeded the specified threshold value. In this assessment of the primary care OMOP database described here, we achieved an overall pass rate of 97%.

Conclusion

The OMOP CDM’s widespread international use, support, and training provides a well-established pathway for data standardisation in collaborative research. Its compatibility allows the sharing of analysis packages across local and international research groups, which facilitates rapid and reproducible data comparisons. A suite of open-source tools, including the OHDSI Data Quality Dashboard (Version 1.4.1), supports the model. Its simplicity and standards-based approach facilitates adoption and integration into existing data processes.

Introduction

The use of routinely collected health data for secondary research purposes is increasingly recognised as a methodology that advances medical research, improves patient outcomes, and guides policy [14]. This secondary data, as found in electronic medical records (EMRs), can be optimised through conversion into a uniform data structure to enable analysis alongside other comparable health metric datasets [46]. This can be achieved using the Observational Medical Outcomes Partnership Common Data Model (OMOP-CDM), which employs a standardised vocabulary to facilitate systematic analysis across various observational databases. The OMOP-CDM is a data schema that uses a standardised vocabulary for the systematic analysis of multiple distinct observational repositories (databases) [7]. The concept behind the OMOP-CDM, is the conversion of data into a common format through the harmonisation of terminologies, vocabularies, and coding schemes within a unique repository (Fig 1) [7].

The primary purpose of an EMR is to record information related to patient care as it naturally occurs in the clinical setting [8]. The use of this data in medical research is a secondary, albeit useful function, as it provides the opportunity to establish ‘real world’ evidence on patient outcomes, healthcare quality, comparative effectiveness, and health system policy [9]. Yet, the quality of data recorded in an EMR varies in its completeness, conformance, plausibility, and currency, hence it is imperative that a measure of its quality is ascertained to determine if is suitable for research purposes. Kahn et. al. [10] established a comprehensive set of quality checks that have since become the widely adopted ’de facto’ standard across the globe. We used these standards to ascertain data quality in this research.

The OMOP model enhances independent institutional research capacity through the development of: shared advanced analytic and prediction techniques; pharmacovigilance for the active surveillance of drug safety; and ‘validation’ analyses across multiple institutions across Australia, the United States, Europe and the Asia Pacific [7]. In this research we aim to investigate the use of the open-source OMOP-CDM in the PATRON primary care data repository [11].

Methods

Data warehouse

The dataset was sourced from a data warehouse Primary Care Audit, Teaching and Research Open Network Program (PATRON) curated by the University of Melbourne [11]. The database collects de-identified EMR data from 129 Australian general practices, chiefly in Victoria. The repository comprises over 700 consenting general practitioners (GPs) who work in Australian general practices that use Best Practice™, Medical Director™, and ZedMed™ proprietary EMR systems (Table 1).

EMR data are extracted from these systems using the data extraction tool GRHANITE™ [13], and the data is then sent via encrypted transmission into the repository. The GRHANITE™ tool de-identifies each patient by replacing the patients name with a unique patient identifier that links the patient to the individual visit data in each patient table [13]. Identifiers including patient address, date of birth, Medicare number (i.e., health insurance number/healthcare identifier), general practitioner, and staff member details are either removed or deidentified prior to extraction to the data repository.

Each EMR system held in the PATRON repository has a data structure that is unique. Hence, to facilitate use of the whole database, the data from each system are harmonised to provide consistency. For instance, to provide a standardised version to the database, all data pertaining to ‘patient history’ from each EMR are merged into a single table, and likewise information relating to ‘medications prescribed’ are also merged into a table. Whilst data standardisation provides a single unified view to simplify researcher use, no data is lost in this harmonisation process (Fig 2).

Mapping process

Primary care EMR systems incorporate both free text terms and proprietary coding, which is uniquely specified by each individual EMR vendor. Therefore, one of the challenging aspects of extracting meaningful data is mapping the free text data to numerical codes in vocabularies such as the Systematized Nomenclature of Medicine Clinical Terms (SNOMED) and RxNorm (United States, medication terminology). It is important to note SNOMED is considered the ‘standard terminology’ for conditions in the OMOP CDM, and likewise, RxNorm is the ‘standard’ for medications. An additional advantage of employing these vocabularies for the preparation of data for OMOP compatibility, is that the mapping process results in a repository that conforms to international standards.

The mapping process was facilitated through the utilisation of a tool called USAGI, developed by the multi-stakeholder international collaborative body, Observational Health Data Sciences and Informatics (OHDSI) [7]. The USAGI tool is crafted with the purpose of transforming text terms extracted from the data warehouse into standardised SNOMED and RxNorm terms. Despite its utility, USAGI still relies on manual input from data mappers who possess specialised domain knowledge. This expertise is essential for ensuring the accuracy of the mapping process, as it enables a deep understanding of the clinical expertise within the field of medicine, guaranteeing precise representation of medical conditions and treatments. In our approach, we enlisted the expertise of three final-year medical students, these students independently conducted data mappings, where pairwise comparisons of their mappings were assessed for concordance using Excel. There was a ninety percent agreement in the student mappings, clinical insights from a physician were sought to resolve discrepancies.

The volume of distinct free text terms from each unique EMR within the PATRON data repository presented challenges. For example, there were 96,000 distinct medication terms, consisting of some combination of a drug’s brand name, generic name, form, strength, and packet size. Only a small fraction, less than five percent, could be automatically linked to the standard RxNorm concept through direct text matching. This left a substantial number of terms requiring manual assignment. To manage this, our mappers were instructed to map all terms that occurred with sufficient frequency.

For each domain, we set specific threshold values to ensure that over ninety-five percent of all records were associated with an appropriately mapped SNOMED or RxNorm concept. As a result, terms occurring 200 times, or more were mapped. A similar approach was taken for fields other than medications. For example, in the ’reason for visit’ table, only conditions that occurred 50 or more times were mapped.

There was no direct match of a ‘conditions’ table in the EMR to the ‘conditions occurrence’ table in the OMOP-CDM. Unlike the OMOP-CDM, which integrates pertinent SNOMED terms into a unified ’condition occurrence’ table to depict a patient’s condition, the EMR uses separate ’reason for visit’ and ’history’ tables to store information about a patient’s current condition, past medical history, and clinical observations.

If an entry in the EMR was accompanied with a date, we considered it a current condition and recorded it as a ‘condition occurrence’ in the OMOP table. Conversely, when a condition did not have an associated date, we considered it as a past observation and therefore recorded it as an ‘observation’ in the OMOP-CDM.

Extract Transform and Load (ETL) Structured Query Language (SQL) scripts

We used mapping tables, constructed from the mappings generated by the mapping team, to facilitate the transformation of EMR data into the OMOP-CDM during the Extract, Transform, and Load (ETL) process.

ETL incorporates the process of retrieving data from a source system, converting it into the format specified by the OMOP model, and then inserting it into the OMOP-CDM database. This transformation is achieved through the use of SQL scripts, which are sets of instructions designed to manage and modify data within the database.

Removal of inactive patients

The Royal Australian College of General Practitioners (RACGP) employs ‘active patients’ as the specific group or denominator for the purpose of reporting clinical indicator measures [14]. This ensures clinical indicator assessments and reports refer only to patients who care classified as ‘active’ or ‘current’ within their practice. As per RACGP definition of an ‘active patient’, an inactive patient’ is defined a someone who has not attended a practice at least three times in the past two years [14]. Initially, this ‘inactive’ definition was applied to the dataset and the inactive patient records were excluded from the analysis. However, we found adherence to this definition resulted in the exclusion of new patients, with only 1 visit, so we also included patients with at least one visit over the last 2 years into the data set.

Data analysis

Once the primary care data was converted into an OMOP compliant format it was securely connected to the ATLAS application for data visualisation and analysis purposes.

Data governance

For research using the OMOP-CDM, a data governance framework has been developed building on the existing comprehensive framework implemented for primary care repository (Fig 3). This framework encompasses various topics including consent, privacy, and risk management [12, 15]. All governance procedures employed in our analysis were underpinned by the principal of beneficence ‘to do no harm’ [16].

Consent

Medical practices provided consent for their practice data to be accessed for research purposes via the primary care repository. Practices are also informed they can change their consent options or withdraw at any time, without prejudice. Regarding individual patient consent for the secondary use of their EMR data, a waiver of consent is applied. A ’waiver of consent’ is often granted under the premise of ’easy rescue,’ where the perceived benefits of data access are deemed substantial, and the potential harm from the risk of privacy loss is considered minimal [17]. Practices inform patients their data is used for research using various communication strategies (i.e., on practice websites and practice posters), they are also informed they can withdraw consent at any point.

Patient privacy

The GRHANITE™ data extraction tool de-identifies patient data in the practice before it is sent to our primary care repository. At no point is identifying data present within PATRON.

Risk management

We conducted a structured risk assessment of the entire process considering privacy, organisational, and technical risks (Table 2).

Results

Data

The PATRON repository contains data for circa 5.6 million patients, which we harmonised. The results were harmonised and converted to version 5.4 of the OMOP Common Data Model (Table 3).

thumbnail
Table 3. Data in the data repository and the resultant OMOP CDM after conversion.

https://doi.org/10.1371/journal.pone.0301557.t003

Mappings

As described above, we mapped a term from the EMR if it exceeded a specified frequency threshold. This threshold level was established individually for each table (Table 4).

Medications occurrence

To illustrate the frequency-based approach for mapping, Table 5 shows the numbers that appear in the medications table in the EMR’s.

Data quality

This data quality DQD assessment comprised a total of 3,570 individual evaluations. Each evaluation compared the outcome against a predefined threshold. A ’FAIL’ occurred when the percentage of non-compliant rows exceeded the specified threshold value. In this assessment of the primary care OMOP database described here, we achieved an overall pass rate of 97 percent (Table 6).

Discussion

The project

In our study, we conducted a comprehensive investigation into the implementation of the open-source OMOP-CDM in a primary care data repository. Utilising a substantial primary care dataset, we underscored the potential and wide-ranging implications of integrating the OMOP-CDM into real-world primary care settings. Additionally, we ascertained that adhering to the OHDSI-recommended conversion processes proved to be both practical and adaptable for achieving our research objectives.

The OMOP common data model proved to be a logical extension of our existing data warehouse. Once the tables and ETL scripts were established it proved to be simple to re-run the process every time we ran a new data extraction. As the CDM is primarily based around SQL it integrated well into our existing processes. SQL is a commonly used query language so resourcing the conversion work was achieved with existing staff with some additional on-line training from OHDSI and the European Health Data Evidence Network (EHDEN) Academy [18].

The OMOP clinical tables and the derived tables once connected to ATLAS were straightforward to work with. The pre-configured dashboards in ATLAS facilitated streamlined data visualisation. However, the real strength in ATLAS lies in the ability to rapidly construct concept sets, clinical cohorts, and design studies.

Another key advantage of the CDM, is its ability to facilitate network studies across multiple centres, both locally and across the globe. With ATLAS, research groups can design analysis packages and easily share them with other groups. These packages can then be imported and executed on local OMOP datasets. Such standardisation enables direct comparison of de-identified patient data between regions and countries without the need to transfer data from the source repository. This approach yields benefits in governance and security, allowing organisations to maintain control over data access while only allowing aggregated data to leave their systems. Additionally, from a security perspective, existing secure data repositories can be utilised for storing and analysing the data, eliminating the need to secure the data elsewhere.

One challenge we encountered, was the large amount of free text terms found in local EMR systems. We noted that the ‘source of truth’ was with the text–where the most reliable and accurate information is found in the textual descriptions or notes that a GP reads, rather than in the numerical code values, such as SNOMED codes. In other words, when it comes to understanding patient conditions or information, the narrative or textual information is considered more authoritative and informative than the numerical codes. In some cases, textual descriptions did not match allocated codes and the codes were clearly erroneous. This made the task of mapping the terms time consuming with the potential to introduce errors. The ETL scripts to convert terms to the OMOP CDM can also be complex so we found it was important to document the scripts to ensure future maintainability.

A good understanding of data quality is central to the use of any dataset. It is especially important for OMOP datasets as the data is derived from source data and it has the potential to mask data quality issues, which only become apparent when comparing data to other datasets. During our quality assessment, we identified minor issues that required correction, such as discovering some values that were implausible. For instance, in some cases, missing dates of birth were recorded as 1900 in the EMR. This practice stems from the necessity for EMR systems to handle missing or incomplete data, including dates of birth, to uphold the accuracy and reliability of patient records. We also found whole tables that contained incomplete or no data, but this is normal with this type of primary care data. Two of the CDM tables were not populated in our data, the ‘PAYER_PLAN_PERIOD’ table. which refers to insurance data that is not applicable in the Australian context, and the ‘NOTE’ table as the narrative clinical note data is not extracted from the EMR source.

The data quality tool has a default threshold for incomplete data set at zero percent. This means any table or field with less than one-hundred percent data completion will be classified as a ’Fail.’ However, this default threshold was not suitable for our dataset. For instance, the ’CONDITION_OCCURRENCE’ table had only sixty-two percent of patients with an entry, which commonly occurs in this type of data due to reasons, such as GP documentation practices, where the condition occurrence maybe recorded as free text in the narrative notes. The recording of condition occurrence can also be influenced by factors related to clinical workflow, patient presentation, and limitations of the EMR system itself. Providentially, the data quality dashboard allows for customisation of the pass and fail thresholds to account for local conditions and prior knowledge, recognising that a one-size-fits-all approach may not be applicable in all cases.

The ETL proved to be a complex and time-consuming activity. But in future studies this could be made more efficient by modifying existing practise. The first and most obvious modification would be to have more coding in the source EMR systems rather than free text terms. This could be implemented as predicative text to reduce manual data entry for ‘time poor’ health care professionals. This would have the added benefit of allowing for validation of data input reducing typographical errors. However, changing proprietary systems is not a simple endeavour and will require cooperation from vendors and regulators.

In order to make the mapping process more manageable we employed a frequency-based approach to mapping terms. This relied on the fact that mapping frequently occurring terms converted a high percentage of required concepts. As an example, we covered 97% of medications by mapping only terms that appeared more than 200 times in our data. This is a pragmatic way of reducing the effort required to map terms. It is also important to note if rarely occurring concepts falling below the frequency threshold are to be studied, they can be mapped simply on a case-by-case basis. Hence, the frequency-based approach allows efficiency and flexibility in the mapping process.

To streamline the mapping process, we adopted a frequency-based strategy for term mapping, leveraging the observation that mapping commonly encountered terms effectively covered a significant portion of the necessary concepts. For instance, we achieved coverage for ninety-seven percent of medications by mapping terms that appeared in our data more than 200 times. This pragmatic approach serves to reduce the overall effort required for mapping while maintaining flexibility. It is worth noting that for rarely occurring concepts falling below the frequency threshold, they can still be mapped on a case-by-case basis when needed. In this manner, the frequency-based approach offers both efficiency and adaptability within the mapping process. The medication fields included, product names (generic and trade), strength, and pack size, we removed pack size to optimise RxNorm text matching in USAGI. Regarding medication vocabulary, Australia uses the Australian Medicines Terminology (AMT), however we chose to use RxNorm, because this is the standard drug vocabulary in the Observational Medical Outcomes Partnership-Common Data Model (OMOP-CDM). When dealing with situations where a one-to-one mapping of a RxNorm code to a medication is not possible, particularly with multi-ingredient medications (e.g., medications containing both pseudoephedrine and codeine), we employed RxNorm ingredient mapping. This approach allows for the application of multiple RxNorm codes to a single medication, ensuring accurate representation and comprehensive coverage of its constituent ingredients.

Another modification to the process that would help with subsequent primary care studies is to share the mappings produced for the ETL process. Indeed, we shared mappings with the University of New South Wales who had already mapped a number of general practice terms. However, in future shared mappings could potentially be shared programmatically via an Application Program Interface (API) with other research teams using Fast Healthcare Interoperability Resources (FHIR) servers, such as the Commonwealth Scientific and Industrial Research Organisation (CSIRO) Ontoserver [19]. This would streamline the process and cut down on manual tasks.

For the information contained in OMOP data sets to be used in research it is essential that good data governance is in place [12]. This ensures the rights of individual patients are respected and that the data is managed responsibly. Transparency in the governance process underpins trust in the data and is fundamental to successful research. We have developed an extensive data governance process that is being adopted by our data governance committee and we believe this provides a good model for other groups managing OMOP data to adopt and learn from.

In this research, we have made significant contributions to the ongoing research on the OMOP-CDM by validating it using real-world primary care data. Additionally, we have leveraged the data quality dashboard and discussed its utility within this paper. Our research also offers a comprehensive examination of implementing the open-source OMOP-CDM within a primary care data repository. By harnessing the PATRON primary care dataset, we illustrate the potential and wide-ranging implications of integrating the OMOP-CDM into real-world primary care settings. In doing so, we demonstrate the practicality and adaptability of adhering to the OHDSI-recommended conversion processes to achieve our research objectives.

Future options for data linkage

As part of this program of work, we are also investigating patient record linkage between primary care and hospital datasets. Whilst this paper describes the conversion of a primary care data repository to the OMOP-CDM, we are also in the process of converting a hospital dataset to OMOP using the same linkage keys. This provides opportunity, as when linkage keys exist in both primary care and hospital data, records can be linked to create a more comprehensive patient record. This linkage method has previously been demonstrated by other groups [2022]. In these studies, data from EMR’s have been enhanced with more detailed health data from other healthcare sources, such as cancer registries, to increase completeness of the datasets.

Conclusion

The OMOP CDM’s widespread international use, support, and training provides a well-established pathway for data standardisation in collaborative research. Its compatibility allows the sharing of analysis packages across local and international research groups, which facilitates rapid and reproducible data comparisons. A suite of open-source tools, including the OHDSI Data Quality Dashboard, supports the model. Its simplicity and standards-based approach facilitates adoption and integration into existing data processes.

Acknowledgments

This research utilised de-identified patient data sourced from the PATRON primary care data repository, which is derived from consenting general practices. The repository is established and maintained by the Department of General Practice and Primary Care, The University of Melbourne: www.gp.unimelb.edu.au/datafordecisions. We acknowledge the contributions of the three final year medical students, Irwan Jaeson, Jun Yong, and Ronnen Leizerovitz, who conducted a significant proportion of the data mappings.

References

  1. 1. Jensen PB, Jensen LJ, Brunak S. Mining electronic health records: towards better research applications and clinical care. Nature Reviews Genetics. 2012;13(6):395–405. pmid:22549152
  2. 2. Hersh WR. Adding value to the electronic health record through secondary use of data for quality assurance, research, and surveillance. Clin Pharmacol Ther. 2007;81:126–8.
  3. 3. Overhage JM, Ryan PB, Reich CG, Hartzema AG, Stang PE. Validation of a common data model for active safety surveillance research. Journal of the American Medical Informatics Association. 2012;19(1):54–60. pmid:22037893
  4. 4. Biedermann P, Ong R, Davydov A, Orlova A, Solovyev P, Sun H, et al. Standardizing registry data to the OMOP Common Data Model: experience from three pulmonary hypertension databases. BMC medical research methodology. 2021;21(1):238. pmid:34727871
  5. 5. Kent S, Burn E, Dawoud D, Jonsson P, Østby JT, Hughes N, et al. Common Problems, Common Data Model Solutions: Evidence Generation for Health Technology Assessment. PharmacoEconomics. 2021;39(3):275–85. pmid:33336320
  6. 6. Tan HX, Teo DCH, Lee D, Kim C, Neo JW, Sung C, et al. Applying the OMOP Common Data Model to Facilitate Benefit-Risk Assessments of Medicinal Products Using Real-World Data from Singapore and South Korea. Healthc Inform Res. 2022;28(2):112–22. pmid:35576979
  7. 7. OHDSI. Observational Health Data Sciences and Informatics Data Standardization 2021 [Available from: https://www.ohdsi.org/data-standardization/the-common-data-model/.
  8. 8. Ludwick DA, Doucette J. Adopting electronic medical records in primary care: Lessons learned from health information systems implementation experience in seven countries. International Journal of Medical Informatics. 2009;78(1):22–31. pmid:18644745
  9. 9. Reimer AP, Milinovich A, Madigan EA. Data quality assessment framework to assess electronic medical record data for use in research. International Journal of Medical Informatics. 2016;90:40–7. pmid:27103196
  10. 10. Kahn MG, Callahan TJ, Barnard J, Bauck AE, Brown J, Davidson BN, et al. A Harmonized Data Quality Assessment Terminology and Framework for the Secondary Use of Electronic Health Record Data. EGEMS (Wash DC). 2016;4(1):1244. pmid:27713905
  11. 11. Boyle D, Sanci L, Emery J, Gunn J, Hocking JS, Manski-Nankervis J-A, et al. PATRON Primary Care Research Data Repository. University of Melbourne. Dataset. 2019 [Available from: https://medicine.unimelb.edu.au/school-structure/general-practice/engagement/data-for-decisions.
  12. 12. Hallinan CM, Ward R, Hart GK, Sullivan C, Pratt N, Ng AP, et al. Integrated Data Governance, Digital Health, and the Common Data Model (OMOP-CDM). BMJ Health & Care Informatics. 2024.
  13. 13. Boyle DI. Middleware Supporting Next Generation Data Analytics in Australia. Stud Health Technol Inform. 2015;216:1019. pmid:26262320
  14. 14. RACGP. RACGP Standards for general practices 4thEdition ed. East Melbourne, Australia: Royal Australian College of General Practitioners; 2015.
  15. 15. Boyle D, Sanci L, Emery J, Gunn J, Hocking J. S, Manski-Nankervis J.-A, et al. Patron Data Governance Framework Department of General Practice Faculty of Medicine, Dentistry & Health Sciences2019 [Available from: https://medicine.unimelb.edu.au/__data/assets/pdf_file/0016/3020272/Governance-Framework.pdf.
  16. 16. Varkey B. Principles of Clinical Ethics and Their Application to Practice. Med Princ Pract. 2021;30(1):17–28. pmid:32498071
  17. 17. NHMRC. National Statement on Ethical Conduct in Human Research (2007)—Updated 2018. National Health and Medical Research Council; 2018.
  18. 18. Hughes N, John H, Miani M, Soininen I, Tischler R, Ryan P, et al. EHDEN Academy a global training collaboration with OHDSI. 2022 OHDSI European Symposium; Rotterdam2022.
  19. 19. Metke-Jimenez A, Steel J, Hansen D, Lawley M. Ontoserver: a syndicated terminology server. J Biomed Semantics. 2018;9(1):24. pmid:30223897
  20. 20. Angela W, Rachel D, Sam H, Jennifer C, Samantha I, Venexia W, et al. Linked electronic health records for research on a nationwide cohort of more than 54 million people in England: data resource. Bmj. 2021;373:n826. pmid:33827854
  21. 21. Snijders BEP, van der Hoek W, Stirbu I, van der Sande MAB, van Gageldonk-Lafeber AB. General practitioners’ contribution to the management of community-acquired pneumonia in the Netherlands: a retrospective analysis of primary care, hospital, and national mortality databases with individual data linkage. Primary Care Respiratory Journal. 2013;22(4):400–5. pmid:24042173
  22. 22. Bouras G, Markar SR, Burns EM, Mackenzie HA, Bottle A, Athanasiou T, et al. Linked Hospital and Primary Care Database Analysis of the Incidence and Impact of Psychiatric Morbidity Following Gastrointestinal Cancer Surgery in England. Ann Surg. 2016;264(1):93–9. pmid:26649592