Figures
Abstract
The objective of this project was to evaluate the accuracy of owner-reported health diagnoses in dogs compared to veterinary electronic medical records (VEMRs) using data from the Dog Aging Project (DAP), a longitudinal study of companion dogs in the United States. We hypothesized that owner-reported diagnoses would align more closely with VEMRs for acute or emergent conditions and less so for benign or self-limiting conditions. A subsample of 350 dogs was selected from the 2020 curated DAP dataset (n = 27,541). Dogs were included if they had VEMRs covering ≥ 85% of their life up to the time of survey completion. Forty-one dogs were excluded due to incomplete VEMRs, resulting in a final sample of 309 dogs included in this observational study comparing owner responses in the Health and Life Experience Survey (HLES) (https://github.com/dogagingproject/dataRelease/tree/master/SurveyInstruments/HLES) to VEMR data abstracted by masked reviewers. VEMR reviewers completed a survey mirroring HLES and identified supporting evidence for each diagnosis. Agreement between owner-reported and VEMR-verified diagnoses was assessed across 20 disease categories, with a focus on five target disease categories (TDCs). Agreement between owner and VEMR data was ≥ 90% in 10 of 20 disease categories. Agreement was highest in endocrine, immune, and “other” categories (99%) and lowest in dental/oral (44%). Among TDCs, orthopedic and traumatic conditions had higher agreement (>78%), while dermatologic and dental/oral categories showed the least. Owner-reported dog health data showed substantial concordance with VEMRs for many disease categories, supporting their utility in research. Thus, owner-reported diagnoses might provide a reliable and scalable supplement or alternative to VEMRs in veterinary research and epidemiology.
Citation: Schmid SM, Sexton CL, Yoerger A, Kauffman M, McClelland RL, D.A.P. Consortium, et al. (2026) Accuracy of owner-reported diagnoses for dogs enrolled in the Dog Aging Project as compared to veterinary electronic medical records. PLoS One 21(3): e0342427. https://doi.org/10.1371/journal.pone.0342427
Editor: Joshua Kamani, National Veterinary Research Institute (NVRI), NIGERIA
Received: June 25, 2025; Accepted: January 22, 2026; Published: March 4, 2026
Copyright: © 2026 Schmid et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: These data are housed at the Broad Institute of MIT and Harvard and it can be accessed by the public through the Terra platform: [https://dogagingproject.org/opendataaccess/].
Funding: This research is based on publicly available data collected by the Dog Aging Project, which was supported by U19 grant AG057377 from the National Institute on Aging, a part of the National Institutes of Health (KEC, AR). There was no additional external funding received for this study. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Introduction
Conducting veterinary research that endeavors to improve real-world health outcomes for animals depends on access to large quantities of diverse and, ideally, longitudinal data. Veterinary electronic medical records (VEMRs) are generally considered the gold standard source for collecting a crucial segment of these data in companion animals, specifically clinical phenotypes [1–6]. While there are indeed many advantages to using VEMRs to extract health-related phenotypes, there remain persistent challenges when trying to extract consistent and compatible phenotype data for many patients due to differences in interface operability and standardization [7]. Differing identification requirements across systems and lower accessibility to systems (e.g., due to cost) compared to human EMRs [8], along with the lack of a unifying system for standardizing diagnoses across providers and institutions, such the International Classification of Diseases (ICD) codes used in human medicine [9], present additional complications when using VEMRs as a source for data extraction.
Furthermore, compared to hospital-based veterinary practices and those involved in medical professional training, independent veterinary practices are less likely to utilize a VEMR system [10], therefore potentially omitting a substantial source of data from large studies. Still, despite these current obstacles, VEMRs demonstrate the potential for expanding on veterinary informatics and are among the most comprehensive sources of medical information for pets, especially when it comes to data that can advance the efficacy of the research-to-real world practice pipeline.
When compared to VEMRs, the potential benefits of using multi-source surveys to collect pet health-related data include the ease and speed of data acquisition, data standardization, and customization potential. Survey data have been successfully utilized in numerous studies aiming to find associations between phenotypes of interest (e.g., physical traits, lifestyle factors) and other paired data (e.g., genetic data, health outcomes). For example, similar to human health survey data being successfully used to identify genetic associations for hair color, eye color, and freckling in humans [11], Embark Veterinary, Inc. used owner-reported phenotype data combined with canine genotype data from high-density single nucleotide polymorphism arrays to detect a locus associated with blue eye color in dogs [12].
However, survey data can also present their own challenges and limitations. For example, these data can be prone to error, including measurement variability, as when respondent answers do not correspond directly with the intended question from the researcher due to factors such as poor survey design, decreasing motivation of respondents during long surveys, or inability of a respondent to remember the correct answer [13]. In humans, accuracy of self-reported data compared to EMR data has also been shown to vary with the perceived social desirability of one answer over another [13,14]. For example, studies assessing human cancer screening procedures indicate that self-reporting tends to overestimate the population that has been screened [1,5,6]. Despite these limitations, self-administered health surveys have shown satisfactory concordance with EMR data in several settings and populations including self-reports of injuries and various chronic conditions [2,4,15,16].
Such concordance among human data sources could be encouraging from a veterinary perspective if the same holds true for VEMRs and owner-reported information about their pets, as there is a high dependency on such reports among clinicians and researchers alike. Furthermore, integrating available VEMR data with novel owner-reported survey responses can enhance both the quantity and diversity of information available to support a wide range of clinical and research needs across veterinary fields.
Recognizing that access to both types of data is essential for advancing our understanding of disease in dogs, the Dog Aging Project (DAP), a longitudinal study of aging and age-related disease in US companion dogs, gathers VEMR data alongside survey data [17]. Consequently, the DAP provides a unique opportunity to compare owner-reported data to VEMR data for the same individuals and health-related phenotypes. To assess the accuracy and validity of owner-reported pet health data, the aim of this study was to compare the concordance of owner-reported health diagnoses with VEMR data among a cohort of dogs enrolled in the DAP. We hypothesized that dog-owners would be in highest concordance with VEMRs when reporting diagnoses that were acute or emergent (e.g., traumatic bone fracture) or chronic (e.g., hypothyroidism requiring daily medication) and less concordant when describing more innocuous diagnoses (e.g., flank alopecia, grade 1 dental disease, or self-limiting vomiting).
Methods
Data collection
The DAP is a community science project in which owners enroll their companion dogs through a series of online surveys [17]. Owners who opt to enroll their dogs in the study are first led through an extensive written informed consent process. Once consent is confirmed, all owners complete a Health and Life Experience Survey (HLES) which collects information on the dogs’ signalment, behavior, environment, physical activity, diet, and health. As part of the HLES, each participant is asked to identify their primary care veterinarian and indicate whether they are willing to share their dogs’ medical records with the DAP for research purposes. Participants who express willingness to share their dogs’ medical records are invited to obtain VEMRs from their veterinarian and submit them through electronic file transfer. Once uploaded, the VEMR is evaluated by DAP personnel to ensure it meets the standard for eligibility and to capture the dates for which the VEMR entries are recorded [18]. The study data from DAP are collected and managed using REDCap (Research Electronic Data Capture) tools hosted at the University of Washington [19,20].
As an open data project, DAP survey instruments are made publicly available on GitHub: https://github.com/dogagingproject/dataRelease/tree/master/SurveyInstruments. The University of Washington Institutional Review Board (IRB) deemed that recruitment of dog owners for the DAP, and the administration and content of each AP questionnaire, are human subjects research that qualifies for category 2 exempt status (IRB ID no. 5988, effective 10/30/2018). No interactions between researchers and privately owned dogs occurred; therefore, IACUC oversight was not required.
Study population
The curated 2020 release of the HLES data contained 27,541 survey records collected between January 1, 2020 and December 31, 2020. In the health status section of HLES, dog owners were asked if their dogs have ever been diagnosed with various medical conditions within a set of pathophysiologic process or organ system categories. Pathophysiologic processes that often affect more than one organ system (e.g., cancer, trauma, infection, etc.) were presented at the beginning of the survey to encourage owners to record such conditions in those categories. The top five diagnostic categories reported by owners within the DAP pack in the 2020 curated dataset included dermatologic, traumatic, oral, infectious, and orthopedic disease. We identified these frequently reported disease categories as target disease categories (TDC) to be evaluated in this study.
Dogs were eligible for inclusion if they had veterinary electronic medical records (VEMRs) covering at least 85% of their life up to the time of HLES available for review. To determine an effective sample size for evaluation from among the eligible population, we conducted power analysis using Cohen’s kappa. Our analysis suggested that a sample of 300 dogs would yield 90% power to detect whether HLES reports of dogs with no diseases in the TDCs matched VEMR records with no diseases in the TDCs better than expected by chance. Thus, to adequately assess agreement between VEMR data and the owner-reported HLES for the five TDCs, a total study sample of 350 dogs was randomly selected from among those eligible – 250 dogs with a minimum of one TDC diagnosis and 100 dogs without a TDC diagnosis – with the additional 50 above the 300-dog target included to account for an anticipated 16% dropout rate due to incomplete records.
EMR data extraction
For each dog included in this study, one of two clinical veterinarian authors (SS, AY) independently filled out a dog Health Status questionnaire identical to that in the HLES (referred to as VEMR Health Status in this report) using the dogs’ VEMR. Investigators completing the VEMR Health Status questionnaires were unaware of the owners’ original responses in the HLES Health Status questionnaire. For a condition to be coded as a diagnosis in the VEMR Health Status questionnaire, the VEMR had to contain sufficient information to support either a definitive diagnosis or a presumptive diagnosis for which treatment was initiated. For example, a diagnosis of neoplasia was made if cytology or histopathology confirmed the diagnosis, whereas a diagnosis of parvovirus was based on a positive viral PCR in the face of supportive clinical signs. Many of the diagnoses could be identified on a physical examination (e.g., heart murmur, dental calculus, medial patellar luxation) and thus were listed as a diagnosis if they were documented on at least one physical examination. If a condition was only listed as a differential in the VEMR it was not recorded as a diagnosis in the survey. As HLES includes clinical-sign based diagnoses (e.g., pruritus, lameness, alopecia) as well as more broad diagnoses in which other diagnoses might fall (e.g., granuloma, gastrointestinal parasites), diagnoses based on VEMR review were recorded to the most specific level of understanding. For example, if a dog was noted to have gastrointestinal parasites and the fecal floatation results showed roundworms, the diagnosis was recorded as “roundworms” instead of “gastrointestinal parasites” in the VEMR Health Status Questionnaire. In contrast, if a dog was noted to have pruritic skin disease but there was no evidence in the VEMR that fleas, other ectoparasites, or food allergies had been ruled out to provide a diagnosis of atopic dermatitis, the dog was classified as having “pruritus”. All HLES diagnoses identified in the VEMR, regardless of if they fell within or outside of a TDC, were recorded. To ensure that data extraction was replicable, reviewer concordance was tested on a subset of 25 randomly selected dog VEMRs. Both reviewers independently completed the VEMR Health Status questionnaires based on VEMR data for each individual in the subset, after which the completed questionnaires were compared. Upon confirming appropriate agreement [21], the remaining 325 records were reviewed and VEMR Health Status questionnaires completed. The original owner-completed HLES Health Status questionnaires (referred to as HLES Health Status in this report) and the new reviewer-completed VEMR Health Status questionnaires (referred to as VEMR Health Status in this report) were then compared.
Statistical analyses
For each TDC we include an overall percentage agreement, numbers of agreements and disagreements, the Kappa statistic and the Gwets AC1 statistic [22]. The Kappa statistic quantifies how much the observed agreement is over-and-above what would be expected based on chance alone. If the prevalence of a condition is low, then expected agreement based on chance alone is high (0/0 agreement is high by chance), so Kappa tends to be very low. Gwets AC1 is a measure of chance-corrected agreement that is less impacted by the prevalence of the condition than Kappa. Levels of agreement statistics over 0.60 are considered to reflect substantial agreement [23]. Agreements included conditions that were reported to be present in both the HLES Health Status and VEMR Health Status as well as conditions that were reported to be not present in both the HLES Health Status and VEMR Health Status. For these analyses at the TDC level (Table 2), if any sub-condition was reported for the TDC in both HLES Health Status and the VEMR Health Status it is counted as an agreement, regardless of whether the specific condition was a match. Subsequent tables describe agreement/disagreement at the level of the specific conditions within categories.
Results
The study population included 250 dogs for which the owners identified at least one diagnosis within five target disease categories and 100 dogs for which no diagnoses were reported in any of the target categories per HLES Health Status. Forty-one dogs were determined to have an incomplete VEMR available for review and were excluded, resulting in a total population of 309 dogs, Table 1. The study sample included 54% single-breed and 46% mixed-breed dogs. Fifty-five percent were male and 45% were female, with the vast majority of dogs in both sexes spayed or neutered. Most dogs (68%) were mature adults [24] and lived in suburban areas (65%). The majority of the owners of these dogs self-identified as White (96%), and were primarily 55 or older (63%) and educated (86% with a Bachelor’s degree or higher). Median annual household income varied, with the plurality reporting $60K-$120K (30%). Dogs and owners lived in all regions of the U.S., with the highest number (39%) in the West.
Overall agreement between owner-reported health data (HLES) and VEMR-collected health data among target disease categories (TDCs)
Three of the five TDCs – Orthopedic, Traumatic, and Infectious – had overall agreement scores above 70%, and two – Dermatologic and Dental/Oral – were below 70% agreement (Fig 1). Overall agreement was highest in the Orthopedic category (85%; Gwet’s AC1 0.77), and agreement as reflected by Gwet’s AC1 indicates that only Orthopedic and Traumatic TDCs have a substantial level of agreement. Overall agreement scores were lowest for conditions classified as Dental/Oral disease (44%), followed by those classified as Dermatologic (65%), Table 2. Gwet’s AC1 likewise indicates very low agreement in these TDCs (−0.12 and 0.33, respectively). In both of these TDCs there were frequently conditions detected in VEMR Health Status that were not reported in HLES Health Status. In all but one disease category (Dental/Oral), agreement on the absence of disease was higher than agreement on the same disease being present.
Agreement within individual TDCs
Twenty-one individual diseases comprised the Orthopedic category, Table 3. Overall agreement was above 88% for all 21 conditions with 13 conditions (62%) having an overall agreement equal to or greater than 99%. Across the Orthopedic conditions, there was lowest agreement regarding the presence of osteoarthritis (88.3%) and degenerative joint disease (89%) (Fig 2). There was a lack of agreement regarding the presence of osteoarthritis in 36 dogs, with 30 dogs having osteoarthritis recorded only in VEMR Health Status and six dogs having osteoarthritis recorded only in HLES Health Status. Similarly, 34 dogs were recorded to have degenerative joint disease according to VEMR Health Status, with no corresponding diagnosis reported in HLES Health. Across all orthopedic disease categories, a specific orthopedic disease diagnosis was noted to be present in HLES Health Status, VEMR Health Status, or both in 187 instances. Agreement on the presence of a specific orthopedic disease was noted in 49 instances, while HLES and VEMR Health Status disagreed on the presence of a disease in 138 instances. Specifically, the VEMR Health Status indicated disease presence in 108 instances, while the HLES Health Status indicated disease presence in 30 instances.
Across all conditions classified as Traumatic, HLES and VEMR Health Status agreed that a specific trauma had occurred in 48 instances, Table 4. Disagreements were noted in 136 cases with the VEMR Health Status reporting a specific trauma in 71 instances, while the HLES Health Status indicated a specific trauma in 65 instances. The injuries with the highest potential (i.e., indicated in HLES Health Status, VEMR Health Status, or both) number of dogs affected were torn/broken toenails (n = 43) and bite wounds from other dogs (n = 42) (Fig 2). For these common injuries, agreement between HLES Health Status and VEMR Health Status for incidents of broken toenails was 90% and when disagreements were noted, HLES Health Status indicated presence (14 dogs) almost as frequently as VEMR Health Status (16 dogs). For bite wounds caused by another dog, while overall agreement was 93% (including absence), VEMR Health Status showed half the number of bite wounds (7) compared to HLES Health Status (14). The condition impacting the third highest potential number of dogs was laceration (n = 37). While overall agreement (including absence) was high (89%) for lacerations, a substantial number of owners did not report lacerations (9 reports in HLES Health Status vs. 24 reports in VEMR Health Status).
Nearly half (19/40) of the individual diseases in the Infectious disease category were not identified in either HLES Health Status nor VEMR Health Status and were therefore reported as 100% agreement based on the absence of these diseases, Table 5. Giardia had the highest representation in the Infectious category, reportedly affecting up to 59 (19%) of all dogs with infectious diseases (Fig 2). Agreement between HLES Health Status and VEMR Health Status for Giardia was 88%; however, many owners did not report Giardia (13 dogs in HLES Health Status vs. 25 dogs in VEMR Health Status). The next most prevalent conditions were Coccidia, Bordetella, and roundworm, potentially affecting up to 26, 25, and 20 dogs, respectively. There were five reports of Coccidia in HLES Health Status, versus 19 reports in VEMR Health Status six HLES Health Status versus 13 VEMR Health Status reports indicated presence of Bordetella; and seven HLES Health Status compared to 12 VEMR Health Status reports indicated presence of roundworm.
HLES Health Status and VEMR Health Status agreed that disease was present in 35 instances of Dermatologic diseases, Table 6. When disagreements were noted, the VEMR Health Status indicated presence of a diagnosis in 268 instances, while the HLES Health Status indicated a diagnosis in 95 instances. Among skin-related disorders, the conditions that potentially affected the highest number of dogs were pruritus (n = 88; 29%), pyoderma or bacterial dermatitis (n = 48; 16%), seasonal allergies (n = 34; 11%), sebaceous cysts (n = 31; 10%), and chronic skin infections (n = 31; 10%). In each of these conditions there was substantial disagreement between HLES Health Status and VEMR Health Status reports regarding disease presence. Four HLES Health Status reports indicated disease presence for pruritis compared to 73 VEMR Health Status reports; two HLES Health Status compared to 46 VEMR Health Status reports indicated pyoderma or bacterial dermatitis; 27 HLES Health Status reports indicated seasonal allergies, compared to six VEMR Health Status reports; and six HLES Health Status reports indicated chronic/recurrent skin infections compared to 22 VEMR Health Status reports.
Overall agreement was the lowest (44%) in the Dental/Oral disease category compared to other TDCs, Table 2. For all diseases in this category, the VEMR and HLES Status agreed on a specific diagnosis in 84 instances. When disagreements were noted, the VEMR Health Status indicated presence of a diagnosis in 362 instances, while the HLES Health Status indicated 24 disease diagnoses, Table 7. Among all oral diseases, the conditions that potentially impacted the highest number of dogs were dental calculus (n = 228; 74%) and gingivitis (n = 100; 32%) (Fig 2). These two conditions also accounted for the highest disagreement between HLES and VEMR Health Status in the category, with dental calculus at 37% agreement and gingivitis at 69% agreement, Table 6. Notably, dental calculus was not reported in HLES Health Status when it was absent in VEMR Health Status, while VEMR Health Status reported 195 diagnoses that were not noted in HLES Health Status. Similarly, gingivitis was reported in only two HLES Health Status reports, while 93 VEMR Health Status reports indicated presence of gingivitis.
Discussion
In comparing owner-provided dog health data (HLES Health Status) with data extracted in a standardized manner from VEMRs (VEMR Health Status) for the same dogs, we find that dog owners who are asked to report veterinary diagnoses are generally reliable in their report of the absence of health conditions across five target disease categories, but are less reliable reporting the presence of specific diseases within categories. Existing discrepancies between owner-reported health outcomes and reviews of the VEMR data— especially among the most frequently reported disease categories (those designated here as target disease categories)— warrant consideration. To that end, we identified potential sources of such discrepancies in three areas, including the design of the survey instrument, survey end-user (owner/reporter) error, and the nature of the disease or condition. Regarding survey design, inconsistent reporting could arise in cases where the HLES instrument does not offer discrete diagnoses that are recognizable by name or description to be selected by dog owners, such as “kennel cough” not being recognized as Bordetella, or a “cut” as a laceration. Indeed, while VEMR Health Status indicated 24 instances of laceration, owners selected this diagnosis in HLES Health Status only nine times.
Lack of clarity in the HLES survey instrument on how to locate and select the correct disease category for reporting individual diagnoses could also impact owner response, leading to an inaccurate or null selection, or reporting of disease present in a category other than intended. To help owners locate and report their dogs’ medical conditions, HLES was designed with categories of medical conditions presented first, and discrete conditions provided within those categories. Some conditions that could logically fit in more than one category (e.g., discoid lupus erythematosus as both Dermatologic and Immune-Mediated) were presented in more than one category to assist the respondent in locating conditions of interest. Owners might have chosen to report the condition in either or both categories. The analysis presented here focused on categorical agreement as a first assessment and some owner-reported conditions might not have been included if they were reported in an alternate category. Future analyses of the accuracy of reporting specific disease conditions could evaluate discrete diagnoses, independent of category.
Owner self-diagnoses of dogs’ conditions and/or misinterpretation or poor recall of diagnoses determined by their veterinarian present additional room for reporting error and thus disagreement between VEMR and HLES data. For example, pruritus, a non-specific term meaning itchy skin, had the lowest agreement among listed dermatologic conditions, with the condition noted far more frequently in VEMR Health Status than HLES Health Status. In veterinary practice, the degree to which owners pursue definitive diagnosis for their dogs’ conditions varies widely. For this reason, medical conditions offered as response variables in owner-facing surveys within the DAP intentionally include both definitive diagnoses and clinical signs. A clinical sign such as pruritus could be recorded in the medical record even without an etiologic explanation but might not be recognized by an owner as a medical condition warranting reporting in a health history. A related challenge involves conditions that owners might frequently perceive as “normal”. For example, while veterinarians could be recording the presence of plaques and other dental calculus in clinical notes, owners might not recognize these conditions as true “diagnoses” or might not consider them to warrant reporting.
Conversely, when it is clear and obvious that a condition is present or absent, such as an infectious disease for which blood or stool might be tested, we see high agreement in that condition, which consequently impacts the overall category agreement. For example, Traumatic disease is a category within which most individual diagnoses have high agreement (one of two TDCs having substantial agreement per Gwet’s AC1), because it is often quite clear whether trauma occurred (e.g., getting hit by a car). Where we do see disagreement in Traumatic disease, it does not seem to be a matter of the trauma not having occurred, but rather whether it was reported. For example, bite wounds were found in HLES Health Status more often than in VEMR Health Status, which suggests owners might recognize the trauma has occurred but might not have sought care, in which case the incident would not be recorded in the VEMR. The opposite seems to be true for lacerations, wherein there were a higher number of instances in VEMR Health Status than in HLES Health Status, which implies the dog was seen by a veterinarian at the time of trauma, but owners did not report the laceration, perhaps because of non-recognition of the term, resolution of the problem, not considering it a medical condition, or other reasons.
High agreement among most individual conditions but low overall categorical agreement in the categories evaluated here were most frequently due in part to (1) one condition with low agreement representing the bulk of the number of dogs affected in the category (i.e., Giardia in the Infectious disease category), (2) fewer instances of the more definitive conditions described above, such as getting kicked by a horse (clear and specific) vs. laceration (requires interpretation) in the Trauma category, or a combination of these two.
This study was subject to potential limitations. While we have a very diverse sample of dogs distributed across the entire United States, dog owners choose to nominate their dog and therefore the data are subject to self-selection bias. Owners who tend to participate in surveys are more likely to nominate their dogs and complete the survey, possibly leading to a biased sample. Also in terms of sampling biases, the dog owner population is skewed toward being highly educated and financially stable, which may potentially impact both the degree of survey agreement and decisions around seeking veterinary care, which would ultimately affect what is reported on VEMR and thus also agreement. From a record standpoint, the owner-provided VEMR utilized in this study might not have included every veterinary visit that dog has had. To decrease gaps in medical record data, we selected a study sample of dogs that had VEMR available covering greater than 85% of their lifetime to date. In addition, most VEMRs come from primary care veterinary clinics, which might have excluded notes from emergency veterinary visits, resulting in under-reporting of traumatic and other urgent conditions in the VEMR Health Status. Furthermore, information is not always transcribed into the medical record, with one previous study revealing that only 64% of observed problems discussed during a first opinion veterinary practice consultation are recorded in the VEMR [25]. Finally, the prevalence of most diagnoses included in this study was low. Consequently, while our findings suggest that dog owners are generally reliable when reporting the absence of veterinary diagnoses, this reliability could differ in study samples with higher disease prevalence.
Compared to use of VEMR-derived dog health data, use of owner-reported dog health data presents important opportunities for large-scale longitudinal research that includes both populations that might infrequently access veterinary care, and medical conditions for which veterinary care might not be needed. The study reported here compared owner-reported and medical record dog health data using the DAP’s HLES and identified some notable areas of disagreement that might be related to term recognition, perceived importance of a medical condition, or survey design factors. Quality control is vital to getting the most out of this longitudinal dataset. Work is ongoing to modify the presentation of survey items in the Health Section instrument – such as through reviewing and editing language, clarifying descriptions of conditions, and incorporating easy reporting help functions – to improve its ability to accurately capture health information from dog owners in order to take full advantage of the data they uniquely provide.
Conclusions
Owner-reported dog health data have the potential to be a reliable source of information for use in research. Owner-reported data showed good agreement with medical records data at the categorical level for some of the most commonly reported disease categories. Owner-reported data showed poorer agreement at the level of specific conditions within disease categories. Further adaptation of reporting mechanisms is crucial to improve the accuracy and utility of owner-reported dog health data.
Supporting information
S1 File. Results of VEMR analyses indicating presence or absence of evidence of diagnosis with specific conditions included.
https://doi.org/10.1371/journal.pone.0342427.s001
(XLSX)
S2 File. Tabular data comparing owner-reported vs VEMR abstracted diagnoses.
https://doi.org/10.1371/journal.pone.0342427.s002
(XLSX)
Acknowledgments
The authors thank Dog Aging Project participants, their dogs, and community veterinarians for their important contributions. These data are housed on the Terra platform at the Broad Institute of MIT and Harvard. Members of the Dog Aging Project Consortium: Drs. Creevy and Ruple and the following authors of this report: Joshua M. Akey6, Brooke Benton7, Elhanan Borenstein8, Marta G. Castelhano9, Amanda E. Coleman10, Matthew D. Dunbar4, Virginia R. Fajt11, Annette L. Fitzpatrick12, Jessica M. Hoffman13, Erica C. Jonlin7, Matt Kaeberlein7, Elinor K. Karlsson14, Kathleen F. Kerr5, Jing Ma15, Evan L. MacLean16, Daniel E. L. Promislow17, Sandi Shrager5, Noah Snyder-Mackler18, and M. Katherine Tolbert1.
6Lewis-Sigler Institute for Integrative Genomics, Princeton University, Princeton, NJ, USA.
7Department of Laboratory Medicine and Pathology, University of Washington School of Medicine, Seattle, WA, USA.
8Department of Clinical Microbiology and Immunology, Sackler Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel.
9Cornell Veterinary Biobank, College of Veterinary Medicine, Cornell University, Ithaca, NY, USA.
10Department of Small Animal Medicine and Surgery, College of Veterinary Medicine, University of Georgia, Athens, GA, USA.
11Department of Veterinary Physiology and Pharmacology, Texas A&M University School of Veterinary Medicine & Biomedical Sciences, College Station, TX, USA.
12Department of Family Medicine, University of Washington, Seattle, WA, USA.
13Department of Biological Sciences, Augusta University, Augusta, GA, USA.
14Broad Institute of MIT and Harvard, Cambridge, MA, USA.
15Division of Public Health Sciences, Fred Hutchinson Cancer Research Center, Seattle, WA, USA.
16College of Veterinary Medicine, University of Arizona, Tucson, AZ, USA
17Jean Mayer USDA Human Nutrition Research Center on Aging at Tufts University, Boston, MA, USA.
18School of Life Sciences, Arizona State University, Tempe, AZ, USA.
References
- 1. Rauscher GH, Johnson TP, Cho YI, Walk JA. Accuracy of self-reported cancer-screening histories: A meta-analysis. Cancer Epidemiol Biomarkers Prev. 2008;17(4):748–57. pmid:18381468
- 2. Schuh-Renner A, Canham-Chervak M, Grier TL, Jones BH. Accuracy of self-reported injuries compared to medical record data. Musculoskelet Sci Pract. 2019;39:39–44. pmid:30472439
- 3. Carsley S, Birken CS, Parkin P, Pullenayegum E, Tu K. Completeness and accuracy of anthropometric measurements in electronic medical records for children attending primary care. J Innov Health Inform. 2018;25(1):963. pmid:29717951
- 4. Skinner KM, Miller DR, Lincoln E, Lee A, Kazis LE. Concordance between respondent self-reports and medical records for chronic conditions: Experience from the Veterans Health Study. J Ambul Care Manage. 2005;28(2):102–10. pmid:15923944
- 5. Gupta V, Gu K, Chen Z, Lu W, Shu XO, Zheng Y. Concordance of self-reported and medical chart information on cancer diagnosis and treatment. BMC Med Res Methodol. 2011;11:72. pmid:21592352
- 6. Gordon NP, Hiatt RA, Lampert DI. Concordance of self-reported data and medical record audit for six cancer screening procedures. J Natl Cancer Inst. 1993;85(7):566–70.
- 7. Janett RS, Yeracaris PP. Electronic medical records in the American Health System: Challenges and lessons learned. Cien Saude Colet. 2020;25(4):1293–304. pmid:32267432
- 8. Krone LM, Brown CM, Lindenmayer JM. Survey of electronic veterinary medical record adoption and use by independent small animal veterinary medical practices in Massachusetts. J Am Vet Med Assoc. 2014;245(3):324–32.
- 9. Lustgarten JL, Zehnder A, Shipman W, Gancher E, Webb TL. Veterinary informatics: Forging the future between veterinary medicine, human medicine, and One Health initiatives-a joint paper by the Association for Veterinary Informatics (AVI) and the CTSA One Health Alliance (COHA). JAMIA Open. 2020;3(2):306–17. pmid:32734172
- 10. Simon SR, Kaushal R, Cleary PD, Jenter CA, Volk LA, Poon EG, et al. Correlates of electronic health record adoption in office practices: A statewide survey. J Am Med Inform Assoc. 2007;14(1):110–7. pmid:17068351
- 11. Eriksson N, Macpherson JM, Tung JY, Hon LS, Naughton B, Saxonov S, et al. Web-based, participant-driven studies yield novel genetic associations for common traits. PLoS Genet. 2010;6(6):e1000993. pmid:20585627
- 12. Deane-Coe PE, Chu ET, Slavney A, Boyko AR, Sams AJ. Direct-to-consumer DNA testing of 6,000 dogs reveals 98.6-kb duplication associated with blue eyes and heterochromia in Siberian Huskies. PLoS Genet. 2018;14(10):e1007648. pmid:30286082
- 13. Stratton SJ. Assessing the accuracy of survey research. Prehosp Disaster Med. 2015;30(3):225–6. pmid:25997500
- 14. Nederhof AJ. Methods of coping with social desirability bias: A review. Eur J Soc Psychol. 1985;15(3):263–80.
- 15. Wagaw F, Okoro CA, Kim S, Park J, Rachman F. Linking Data From Health Surveys And Electronic Health Records: A demonstration project in Two Chicago Health Center Clinics. Prev Chronic Dis. 2018;15:E09. pmid:29346063
- 16. Ockander M, Hjerppe M, Timpka T. Patient-doctor concordance in elderly women’s self-reported health and medical records. Methods Inf Med. 2002;41(2):119–24. pmid:12061118
- 17. Creevy KE, Akey JM, Kaeberlein M, Promislow DEL, Dog Aging Project Consortium. An open science study of ageing in companion dogs. Nature. 2022;602(7895):51–7. pmid:35110758
- 18. Praczko D, Tinkle AK, Arkenberg CR, McClelland RL, Creevy KE, Tolbert MK. Development and evaluation of a survey instrument to assess veterinary medical record suitability for multi-center research studies. Front Vet Sci. 2022;9:941036.
- 19. Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. 2009;42(2):377–81. pmid:18929686
- 20. Harris PA, Taylor R, Minor BL, Elliott V, Fernandez M, O’Neal L, et al. The REDCap consortium: Building an international community of software platform partners. J Biomed Inform. 2019;95:103208. pmid:31078660
- 21. McHugh ML. Interrater reliability: The kappa statistic. Biochem Med (Zagreb). 2012;22(3):276–82. pmid:23092060
- 22. Gwet KL. Computing inter-rater reliability and its variance in the presence of high agreement. Br J Math Stat Psychol. 2008;61(Pt 1):29–48. pmid:18482474
- 23. Wongpakaran N, Wongpakaran T, Wedding D, Gwet KL. A comparison of Cohen’s Kappa and Gwet’s AC1 when calculating inter-rater reliability coefficients: A study conducted with personality disorder samples. BMC Med Res Methodol. 2013;13:61. pmid:23627889
- 24. Creevy KE, Grady J, Little SE, Moore GE, Strickler BG, Thompson S, et al. 2019 AAHA Canine Life Stage Guidelines. J Am Anim Hosp Assoc. 2019;55(6):267–90.
- 25. Jones-Diette J, Robinson NJ, Cobb M, Brennan ML, Dean RS. Accuracy of the electronic patient record in a first opinion veterinary practice. Prev Vet Med. 2017;148.