Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

An open-source probabilistic record linkage process for records with family-level information: Simulation study and applied analysis

  • John Prindle ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    jprindle@usc.edu

    Affiliation Suzanne Dworak-Peck School of Social Work, University of Southern California, Los Angeles, Los Angeles, California, United States America

  • Himal Suthar,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Validation, Writing – original draft, Writing – review & editing

    Affiliation Keck School of Medicine, University of Southern California, Los Angeles, Los Angeles, California, United States America

  • Emily Putnam-Hornstein

    Roles Conceptualization, Funding acquisition, Investigation, Supervision, Writing – original draft, Writing – review & editing

    Affiliation School of Social Work, University of North Carolina at Chapel Hill, Chapel Hill, North Carolina, United States America

Abstract

Research with administrative records involves the challenge of limited information in any single data source to answer policy-related questions. Record linkage provides researchers with a tool to supplement administrative datasets with other information about the same people when identified in separate sources as matched pairs. Several solutions are available for undertaking record linkage, producing linkage keys for merging data sources for positively matched pairs of records. In the current manuscript, we demonstrate a new application of the Python RecordLinkage package to family-based record linkages with machine learning algorithms for probability scoring, which we call probabilistic record linkage for families (PRLF). First, a simulation of administrative records identifies PRLF accuracy with variations in match and data degradation percentages. Accuracy is largely influenced by degradation (e.g., missing data fields, mismatched values) compared to the percentage of simulated matches. Second, an application of data linkage is presented to compare regression model estimate performance across three record linkage solutions (PRLF, ChoiceMaker, and Link Plus). Our findings indicate that all three solutions, when optimized, provide similar results for researchers. Strengths of our process, such as the use of ensemble methods, to improve match accuracy are discussed. We then identify caveats of record linkage in the context of administrative data.

Introduction

Record linkage definition and motivation

Record linkage, also known as entity resolution, is the process of identifying records across unique datasets that capture information for the same entity. In the context of research, record linkage has been applied to merge administrative data to address research questions related to social sciences and public policy [13]. These studies leverage fields of interest (i.e., variables related to specific administrative processes) collected by public-sector and nonprofit agencies to assemble secondary datasets with a broader set of information than any single agency can provide.

Record linkage can be applied using deterministic (i.e., strictly identical matches and nonmatches) or probabilistic (i.e., calculating match probabilities between potential pairs) methods. Given the nature of administrative data, which may include missing, incomplete, or inaccurate values for fields (e.g., misspelled names or incorrect birth dates), there is an advantage to implementing probabilistic linkage methods, and studies have shown data simulated to mimic real linkage parameters are more accurately linked with probabilistic methods than deterministic methods [4]. Additionally, applying probabilistic linkage methods to large datasets, for which it is often not feasible to conduct human reviews, can output quality linkage results without a cumbersome and resource-intensive manual review process [5].

Probabilistic record linkage process

The key steps in probabilistic record linkage are: (1) standardizing and cleaning data; (2) blocking; (3) generating a comparison feature matrix; and (4) assigning match probability. The standardization and cleaning step ensures that all columns are comparable between the two datasets in the linkage by performing transformations such as removing special characters from fields, capitalizing letters in name and address fields, and setting invalid data to blanks (e.g., Social Security number values of 00-00-0000 or first names values of “Baby Boy”). The blocking step reduces all possible record pairs between the two datasets in the linkage to only the most likely pairs, which can be implemented via hard rules (e.g., only pairs that match on first letter of last name), fuzzy matching (e.g., only pairs whose birth years are within 2 years of each other), or both. The next step performs pairwise comparisons between the blocked pairs, exporting the results in the form of a Boolean matrix, where a row represents a pair to be compared and a column represents the comparison (e.g., first names are an exact match). Based on the results of this matrix, a match probability is assigned to each row via human review or machine algorithm and a final decision is made on whether the pair is a match. Further reading on probabilistic record linkage can be found in Christen [6], Sayers et al. [7], and Harron et al. [8].

Current landscape of probabilistic record linkage

Link Plus, developed by the Centers for Disease Control and Prevention [9], is a manual review solution designed to detect duplicates in cancer registries and link these registries to external sources. It assigns a match score to potential match pairs based on similarity of personally identifiable information values, presents these values of potential pairs in descending order of score, and allows end users to determine a match threshold after clerically reviewing the scored pairs. This solution does not apply a standard cleaning process to the datasets to be linked (i.e., linkage results may differ across users if different cleaning mechanisms are implemented), demands a high time investment (due to the manual review process), has an upper limit of 4.5 million to 4.8 million rows for the first dataset in the linkage (dependent on the hardware specifications of machine; the second dataset does not have a row limit), only runs on Windows, and outputs results that are highly subjective to the end user determining the match threshold score. When done correctly, however, it produces highly accurate results because most positive matches have been reviewed manually [2]. It is also relatively easy to implement for users with less technical knowledge.

ChoiceMaker [10] is a fully automated solution that only requires users to specify the datasets they wish to link and the training data from which to generate scoring algorithms. Due to its low computing and user input requirements, it is scalable to large datasets and can handle a high throughput of linkages with little manual involvement. Further, due to the automated nature of the scoring process (i.e., performed by a model), results of linkages are consistent across runs, with accuracy in line with Link Plus. Still, this solution does not include a standard cleaning process or guidelines, is compatible only with specific computing environments, requires a significant time commitment in specifying a training dataset, and offers only limited user customization options outside of training data procurement and determining a match probability cutoff.

Other solutions such as Link King [11], Dedupe [12], Match*Pro [13], and splink [14], although not benchmarked in the current paper, have different combinations of the strengths and shortcomings relative to the previously mentioned solutions. The RecordLinkage package in R was considered due to its open-source nature but ultimately not reviewed due to its inability to be easily translated to run parallel processing [15]. No solutions reviewed adequately address the shortcomings of Link Plus and ChoiceMaker while also duplicating their strengths.

Probabilistic record linkage for families (PRLF) and how it differs

Definition.

PRLF was developed to allow end users more control over the linkage task while automating as many steps as possible. The solution is an automated, Python-based, end-to-end linkage pipeline built on the open-source RecordLinkage package (see https://recordlinkage.readthedocs.io/). PRLF performs data cleaning, blocking, feature extraction, model training, and match probability assignment while also generating optimized models for scoring based on various algorithms of interest and training data supplied. All components of PRLF are accessible via a single install, and although it has default settings that allow users to perform linkages “out of the box,” users easily can control parameters at each step of the linkage process to fit their needs with minimal editing of Python code. Further, PRLF is agnostic to system software and hardware and can be run on Windows, Mac OS, and Linux machines of varying hardware capabilities. A detailed process flow and available user customizations can be found in S1 Text and S1 Fig.

Key features.

PRLF is automated by design, and the Python architecture allows for throughput and scalability comparable to ChoiceMaker with a low level of human intervention during the linkage process. The high degree of customizability over each step provides the potential for a high degree of match accuracy in line with Link Plus. The Python architecture obviates the need for third-party support, instead empowering the user to edit the open-source Python code to tune the tool to a specific linkage. The user arguments pertaining to hardware performance allow PRLF to be portable enough to run on even the most basic hardware specifications. Finally, PRLF is expandable to allow for numerous machine learning algoirthms (e.g., regularized regression, random forests, XGBoost) to estimate match probabilities, which allows for the creation of ensembles of probability scores for further scrutiny to identify matched pairs. Thus, PRLF builds on the combined strengths of current linkage solutions while addressing their gaps. A comparison of PRLF with other linkage solutions is displayed in Table 1.

thumbnail
Table 1. Features of probabilistic linkage solutions available for record linkage applications.

https://doi.org/10.1371/journal.pone.0291581.t001

Study aims

In the current study, we compared several known probabilistic record linkage solutions and PRLF, a novel record linkage solution designed to provide researchers with greater control over characteristics dictating linkage performance. We conducted comparisons using two methods: simulation and applied analysis.

Simulation analysis.

First, we ran PRLF to link sets of simulated data with known parameters to provide insight into the performance of the linkage solution. Our simulation analysis was guided by several questions: (1) How well does blocking correctly document true matches? (2) Are blocking accuracy and block size associated with dataset size, percentage of overlap, and percentage of rows with typos introduced? (3) What is the proportion of matches identified in the blocked pairs and how does it compare to the total match pairs? (4) What is the proportion of false positives?

Applied analysis.

Next, we compared the linkage results of PRLF to other linkage solutions using real-world administrative data. For our applied analysis, we sought to address the following questions: (1) Does PRLF, with trained algorithms, perform similarly to other linkage solutions? (2) Does varying the PRLF algorithm threshold for identifying a match pair affect analyses based on linkages? (3) Is it possible to establish recommended thresholds that researchers can use as a starting point and modify based on desired accuracy?

Methods

Simulation analysis

Data.

The data created through simulation were designed to mimic data observed in records from California (i.e., administrative data for health and human services agencies serving residents of the state) [16]. Simulated datasets were created in pairs with the purpose of treating each pair as two distinct datasets to be linked (see S2 Text for details on the data creation process, along with the Python script used to generate the simulated datasets). Paired datasets had the same three parameters: (a) number of unique individuals in the dataset (i.e., row count); (b) percentage of overlap between the two datasets (i.e., percentage of identical rows); and (c) percentage of rows in each dataset with some form of information error introduced. Each dataset contained the following fields: source ID, first name, middle name, last name, birthdate, street address, city, state, and ZIP code. The source ID field served as a unique identifier for each row in a dataset and was created using a random number generator, ensuring no duplicates in a dataset. The fields of first name, middle name, last name, street address, and city were generated using the Python package Faker version 13.15.1 (see https://faker.readthedocs.io/en/master/index.html). The birthdate field was created using a random number generator, ensuring all birthdays were in 2016. The state field was set to “California” for all individuals, and the ZIP code field was populated using random selection (with replacement) from all valid California ZIP codes (performed once per row). The dataset sizes included 10,000, 50,000, and 100,000 rows. The row percentage overlap parameter values were 1%, 3%, 5%, 10%, 20%, 40%, and 50%. The percentage error parameter values were 0%, 2%, 5%, 10%, 15%, 20%, 30%, and 40%. In total, 168 pairs of datasets were created, one pair for each parameter.

Linkage.

All simulated pairs underwent the “out of the box” PRLF processes (S1 Fig) for data cleaning and standardization, blocking, and generation of a comparison feature matrix. The first step identified which fields were to be used for linkage, standardized variable names used in the matching process, and allowable values in records to be compared between sources. Potential match pairs were then selected through a preliminary blocking stage that identified similar first or last names through a term frequency–inverse document frequency (TF-IDF) algorithm [17] with parameters of 4 n-gram length, top 3 matches of names matching at least 80%, and an additional constraint of matching birth year and month between records. Features were then extracted from this filtered set of potential match pairs using comparison functions across fields of unique and nonunique identifiers (see S3 Text and S1 Table for a sample of comparison functions). Using the resulting Boolean matrix from the previous step, record pairs were scored using an extreme gradient boosting (XGBoost) algorithm optimized through a grid search on prototypical training data [18]. XGBoost was selected for examining simulation and applied data linkages to provide a simplified interpretation of implementation of probabilistic data linkage in the context of PRLF. Further comparisons of alternative algorithms and ensemble methods may be produced. The grid search hyperparameters included learning rate (0.01, 0.10, 0.25), max depth (5, 10, 15), and number of estimators (250, 500, 1,000, 2,000). All potential match pairs after blocking in each dataset pair were assigned match probabilities using the same model.

Simulation analysis.

Comparison of model performance was done with two main simulation parameters, overlap between dataset pairs and percentage of rows with administrative errors. Several metrics for exploring the resulting linkages from simulated data were presented from a typical confusion matrix. That is, pairs identified as matches (positives) and nonmatches (negatives) through model scoring were identified as true or false based on the data simulation procedure. We calculated the sensitivity, specificity, precision, and F1 score for each condition in the simulation grid. We additionally calculated metrics of blocking performance (proportion of true match pairs in blocked data and proportion of true matches in all blocked data).

Applied data analysis

Data.

Data from the California Department of Public Health Vital Statistics was obtained for the birth cohort year 2016 (N = 487,853). These data were linked to child protection service (CPS) records from the California Department of Social Services from 2016 (N = 5.8 million unique individuals). Identifiers for both datasets included child names (first, middle, last); date of birth; parent or guardian names (first, middle, last) and dates of birth; and complete addresses (street, city, state, ZIP code). These identifiers were used in three linkage solutions: Link Plus, ChoiceMaker, and PRLF. Link Plus and ChoiceMaker have been used for previous analyses with these data [19, 20], and PRLF was compared to these accepted linkage solutions. Data available to the research team were stripped of direct identifiers and anonymized prior to analysis when accessed in January 2023. Analyses were approved by the California Health and Human Services Agency’s Committee for the Protection of Human Subjects and met criteria for a waiver of informed consent. Work with these data was also reviewed by the Vital Statistics Advisory Committee and data security protocols were approved by a university institutional review board.

Linkage parameters.

Link Plus allows users to identify blocking variables and subsequent to the blocking procedure, use various field comparisons to identify potential matched pairs between data sources. ChoiceMaker similarly has a blocking step, with more sophisticated field comparison functions that are used as features for a machine learning model to output a match probability per potential record pair. Although Link Plus requires manual review of potential record pairs without prior model training, ChoiceMaker allows the user to set a threshold for which a pair of records will be called a match (set at 80% match probability for the model used on the current data). For PRLF, we used the combination of these approaches with birth year, birth month, first name, middle name, and last name meeting a similarity threshold based on TF-IDF to establish a set of blocked record pairs for which to generate comparison features. Pairs were placed into a block if their first names or last names were in the top 3 matches as determined by TF-IDF, with a match probability cutoff of 80% using n-grams of length 4 (excluding leading and trailing spaces). Of these pairs, only those that matched on exact birth month and birth year continued to feature extraction. The extracted features of these blocked pairs were then used to determine a match probability using the XGBoost algorithm. The parameterization for this model is based on a grid search outlined in the simulation data linkage solution. We compared results of the PRLF linkage with varied thresholds for establishing a positive match between data sources to those of the established solutions.

CPS outcome variable.

Referral for abuse or neglect to CPS (the first stage of contact with CPS) was identified by a positive linkage between the birth cohort and a CPS maltreatment record with a referral within 3 years of the birthdate of birth (yes or no). A 3-year window of observation was used based on data available at the time of data linkage (i.e., maltreatment referrals through 2019).

Birth cohort characteristic variables.

Birth records offered covariates as a measure of model accuracy using various linkage solutions. Child sex (male or female), low birthweight (yes or no) and variables related to the mother: age (< 20, 20–25, 26–30, >30), race and ethnicity (White, Black, Hispanic, Asian or Pacific Islander, other), non-U.S.-born (yes or no), and education (high school graduate; yes or no). Variables related to the household were also coded: public insurance (yes or no), first trimester prenatal care (yes or no), established paternity (yes or no), and WIC (supplemental services provided for low-income women, infants, and children with nutrition risk; yes or no). Later sections show the distribution of these variables in the subset of children for each linkage solution.

Applied analysis.

An initial bivariate comparison of linkage tools tested the equality of group distributions on the outcome (CPS referral) and model covariates from the birth records using chi-square tests. Then, a generalized linear model with a Poisson distribution with a log link function and robust standard errors was used to show the multivariable relationship of the covariates with CPS referral likelihood [21]. To test the effect of match threshold selection, a range of thresholds above and below a relatively conservative 80% threshold was established (40%–95%) and the tests previously outlined were performed (chi-square tests and generalized linear modeling analysis). Proportions for varied thresholds were tested for trends to provide a starting point for establishing a threshold with linkages performed with the PRLF solution.

Results

Simulation analysis

Simulations were run for the grid of 168 linkages specified in the methodology. Table 2 shows an iteration of the results for a simulation run with 50,000 rows, with varying row overlap between datasets and varying errors introduced in a subset of rows. The results of the XGBoost algorithm are presented with a prediction probability threshold for match set at 80%. The table shows seven metrics of model performance: (a) accuracy; (b) sensitivity; (c) specificity; (d) precision; (e) F1 score; (f) proportion of true matches in blocks; and (g) sensitivity adjusted to include true matches not in block. As depicted in Panel A, two trends occurred: an increased proportion of true matches between datasets was indicative of increased accuracy, and these gains were seen across increasing levels of jitter or noise added to records. Accuracy improved as errors in records increased for low levels of record overlap (0.732 to 0.820 at 0.01 overlap), but the opposite was true for higher levels of record overlap (0.940 to 0.910 at 0.50 overlap). Panel B shows that as errors increased, sensitivity decreased, but there was no interaction with the amount of overlap between datasets. Panel C exhibits how specificity increased with the proportion of rows with introduced error (0.727 to 0.818 at 0.01 overlap), with no interaction with percentage of overlap. On the other hand, in Panel D, we see precision increasing with percentage of overlap (0.063 to 0.928 at 0.00 row errors) and no interaction with proportion of rows with introduced error. Panel E documents the F1 score decreasing with proportion of introduced error and increasing with data overlap. These effects are the harmonic mean of precision and sensitivity, which researchers may value over the combined metric when determining model appropriateness for specific use cases. In Panel F, we observe that the proportion of true matches identified by blocking decreased with an increase in row errors, in line with the proportion of row errors introduced. No interaction was found with the proportion of overlap between the datasets. Finally, in Panel G, the proportion of true matches in the blocked record pairs increased as the proportion of overlapping records between datasets increased.

thumbnail
Table 2. Fit statistics for scored XGBoost models with 50,000 rows per dataset.

https://doi.org/10.1371/journal.pone.0291581.t002

Results were replicated with 10,000 and 100,000 rows per dataset pair and provided comparable results, as presented in S4 Text and S2 and S3 Tables.

Applied analysis

Training accuracy.

Training data consisted of 5,771 record pairs of both known matches and nonmatches. These pairs included various prototypes of matches that would provide notably discriminable data for identifying high probability of a positive match (and conversely, pairs providing information about how to discern nonmatches). Parameters for each algorithm were selected through a grid search process, and parameters that maximized accuracy were selected. Hyperparameters are specific to the algorithm selected. PRLF includes the capability to train additional classifiers available in SciKit Learn with model-specific grid search parameters and selects the hyperparameters producing the best fit. Examples of other trainable models include logistic regression, random forests, XGBoost, and artificial neural networks. The XGBoost algorithm with match probability threshold set at 80% produced a solution that identified 93.8% of true matches, with 1.4% of nonmatches mislabeled a match. This is similar in performance to ChoiceMaker (match probability threshold of 80%, 94.7% of true matches identified, and 2.0% of nonmatches mislabeled as a match) and neural networks (match probability threshold of 75%, 96.8% of matches correctly identified, and 1.7% of nonmatches mislabeled as a match). Logistic regression (match probability threshold of 80%, 89.9% of true matches identified, and 4.3% of nonmatches mislabeled as a match) and random forests (match probability threshold of 40%, 95.0% of true matches identified, and 4.3% of nonmatches mislabeled a match) exhibited inflated levels of false positives compared to the other algorithms. As a result of these comparisons, the applied data example compared the performance of PRLF with XGBoost to Link Plus and ChoiceMaker.

Applied linkage accuracy.

Records from birth records and CPS records were linked using the three linkage tools: Link Plus, ChoiceMaker, and PRLF with XGBoost. The manual review of Link Plus record pairs yielded a linkage rate of 12.0% of the total birth record population (n = 58,377) to CPS before age 3 years, and ChoiceMaker provided a 12.2% (n = 59,685) linkage rate with a 80% threshold for match probability as suggested by ChoiceMaker and in line with prior research using current linkage solutions. These two methods overlapped with 97.18% of Link Plus pairs and 95.05% of ChoiceMaker pairs. PRLF using XGBoost identified 12.1% (n = 59,102) birth records as a match to a CPS record. When compared to the other solutions, 95.6% of PRLF pairs matched with Link Plus record pairs and 97.2% PRLF pairs matched with ChoiceMaker record pairs. As shown in Table 3, when the threshold for identifying a match pair varied (40%–95%), the proportion of overlap trended upward as the threshold for identifying match pairs increased (94.8%–96.1%). That is, as the threshold for a positive match increased, only more certain matches remained and the two datasets fell more in line with each other. The tradeoff with increasing the PRLF threshold is a lower percentage of Link Plus pairs matched (99.8% to 97.8% over the range of PRLF match percentage thresholds). A similar trend in solution comparisons is shown between ChoiceMaker and Link Plus in S4 Table, where higher match percentage thresholds yielded more certainty in overlapping pairs with a lower total of match pairs.

thumbnail
Table 3. PRLF match pair overlap with Link Plus match pairs.

https://doi.org/10.1371/journal.pone.0291581.t003

Applied linkage analysis.

The generalized linear model was fit to each linked data source: birth to CPS using PRLF, Link Plus, and ChoiceMaker. The birth cohort proportions and subsequent CPS subpopulations in the cohort are shown for each linkage method (see Table 3). These proportions in each linkage groups (i.e., children identified as having a CPS report before age 3) are similar across each linkage method, indicating that the small discrepancy of match pairs between linkage methods did not significantly alter the underlying analytic characteristics. Further, Table 4 compares the cohort distribution of birth record variables to the distributions in each linkage subpopulation. Significant differences occurred between the subpopulations on all demographic variable distributions and that of the overall cohort. All comparisons in Linkage subpopulations showed no significant distributional differences except maternal nativity ().

thumbnail
Table 4. Demographic characteristics of the 2016 California birth cohort.

https://doi.org/10.1371/journal.pone.0291581.t004

Table 5 displays the estimated parameters for each predictor with referral by age 3 as the outcome. These estimates and 99% confidence intervals show agreement between solutions where different linkages identified the subpopulation of children with a CPS referral. These findings largely mirror those previously found for similar linkages with previous years of data [2224]. That is, several factors were associated with increased risk of CPS referral (e.g., low birthweight, public insurance, later prenatal care), whereas other factors were associated with reduced risk (e.g., paternity established and high school graduation). Fig 1 shows how varying the threshold of acceptable match probabilities affected model prediction (with Link Plus and ChoiceMaker parameter estimates to show relative differences). In the range of probability thresholds examined (40%–95%), there does not appear to be significant bias in parameter estimates using PRLF with an XGBoost algorithm to link data. The confidence intervals of the established linkage solutions encompass the range of values the XGBoost algorithm estimated across all birth characteristics.

thumbnail
Fig 1. Comparison of PRLF with varied match thresholds to Link Plus and ChoiceMaker model estimates with 95% confidence intervals.

Estimated parameters for each predictor is shown with a solid dot and confidence intervals are represented by error bars. Blue identifies Link Plus parameter estimates and confidence intervals and red identifies ChoiceMaker estimates and confidence intervals. Black dots and grey confidence intervals show the estimates for the XGBoost algorithm of PRLF when the threshold for a pair identified as a match varied (0.40–0.95) from left to right. Values above exp(b) = 1 indicate an increase in likelihood of child welfare referral, whereas values below exp(b) = 1 indicate a decrease in the likelihood of child welfare referral.

https://doi.org/10.1371/journal.pone.0291581.g001

thumbnail
Table 5. Generalized linear modeling parameter estimates of birth characteristics predicting CPS referral by age 3.

https://doi.org/10.1371/journal.pone.0291581.t005

Discussion

Based on the simulation analysis, we found that the PRLF linkage solution performed comparably to other open-source and widely adopted linkage solutions, including ChoiceMaker and Link Plus. Linking data involves several stages in which researchers may have different requirements for the process, including cleaning functions to apply to available personally identifiable information fields to be compared; type and strictness of blocking algorithms; which algorithms are implemented in scoring comparison features of potential pairs; and at what probability to accept a match (i.e., how high the bar is set for matching people between sources). The PRLF solution allows researchers to adjust these settings and compare results across various linkage assumptions. The applied analysis identified a typical research use case of PRLF—the linkage of a population-based vital records dataset with an administrative dataset and analysis of program interaction for a birth cohort.

The simulation analysis identified the various metrics affected by row errors and dataset overlap. Many performance metrics improved as the theoretical overlap between the dataset pairs increased, in line with expectations for well-defined linkage solutions. Accuracy and precision showed the largest improvements, with other metrics stable for a specified proportion of row errors. That is, as the proportion of overlapping records increased between two datasets, the proportion of false positives fell while holding the match threshold constant.

Alternatively, the introduction of row errors reduced the ability of blocking to successfully retain true matches between datasets. This is shown in the proportion of matches identified in the blocked data (Table 2, Panel F). Although the prior metrics showed negligible effects from this degraded signal, it necessary to note that data linkage relies on accurate records from sources. Random error across datasets indicates a robust linkage solution can use the correct information across fields to obtain linkages in the range of: Proportion Blocked = 1−meanrow error. Two data sources with relatively clean records (i.e., free from error) would produce a blocked set with nearly all true matched record pairs.

Researchers may adjust the threshold for an acceptable match pair between datasets (or similarly, implement several algorithms for estimating the probability of match pairs). The applied analysis presented here shows how threshold selection, examined over a large range (40%–95%), can be tested in the context of PRLF if other linkage tools have been used. Similarly, the development of PRLF allows for multiple algorithms to be deployed simultaneously. Machine learning encompasses a wide variety of methods for classification problems, and these methods may be suitable across varying types of classification problems and combined to create an overall voting system for detecting classification accuracy [2527]. This allows researchers to produce ensembles based on several algorithm decisions for blocked pairs. In effect, training several algorithms allows further triangulation of identifying matched pairs between data sources.

Limitations

There are several limitations to note in our current work. First, the variation in unique name values available in our simulation was lower than the variation encountered in real-world data (i.e., birth cohort data showed more unique names when relative size was considered). This seems to be a product of the data generation functions used, and future work should test how increasing this source of variance affects results. Second, our simulation linkage used person and address information, whereas our applied analysis included guardian information in addition to the fields used in the simulation analysis. The introduction of more fields for comparing record pairs allows for a more nuanced and accurate scoring model based on increasing the number of points of comparison and thus, points where records with more than one match pair can differ.

Conclusion

The PRLF record linkage solution encapsulates all aspects of data cleaning, blocking, match probability modeling and scoring, and final linkage production. Our linkage solution, PRLF, is adaptable and allows researchers to adjust each step to best fit their needs. That is, if accuracy is of highest import (e.g., researchers would only like to identify true positives and reduce false positives), then a high threshold for selecting a match should be prioritized. These requirements will vary by use case, and PRLF has many ways to handle a wide range of applications for research making use of administrative data.

Supporting information

S1 Fig. Step by step workflow for PRLF model training and data scoring.

The inner two columns show the two workflows of PRLF: 1. Assigning match probabilities to potential shared pairs between two datasets, and 2. Generating models based on user-provided training data which can be used as score functions for Workflow 1.

https://doi.org/10.1371/journal.pone.0291581.s001

(TIF)

S1 Text. Probabilistic record linkage for families (PRLF) workflow process.

https://doi.org/10.1371/journal.pone.0291581.s002

(DOCX)

S3 Text. Sample of features generated during feature extraction.

https://doi.org/10.1371/journal.pone.0291581.s004

(DOCX)

S4 Text. Simulated data linkage for 10,000 and 100,000.

https://doi.org/10.1371/journal.pone.0291581.s005

(DOCX)

S1 Table. Sample of features generated when comparing records from two different sources.

https://doi.org/10.1371/journal.pone.0291581.s006

(DOCX)

S2 Table. Fit statistics for scored XGBoost models with 10,000 rows per dataset.

https://doi.org/10.1371/journal.pone.0291581.s007

(DOCX)

S3 Table. Fit statistics for scored XGBoost models with 100,000 rows per dataset.

https://doi.org/10.1371/journal.pone.0291581.s008

(DOCX)

S4 Table. Overlap of LinkPlus and ChoiceMaker matched pairs with varied thresholds for ChoiceMaker.

https://doi.org/10.1371/journal.pone.0291581.s009

(DOCX)

Acknowledgments

This project was possible thanks to an ongoing collaboration with colleagues at the California Health and Human Services Agency. Although the findings reported and conclusions drawn from these data are solely those of the authors and should not be considered to reflect those of any agency of the California government, this analysis would not have been possible without this parnership, reflecting their ongoing commitment to data-driven program and policy development. Colleagues at the Children’s Data Network contributed greatly to this work: Aarya Suryavanshi was involved in developing feature comparison functions for PRLF; Huy Nghiem provided assistance in developing the ChoiceMaker linkage; and Andy Xiang, Rachel Chen, and Jonathan Hoonhout provided review of Link Plus matched pairs. The Children’s Data Network thanks First 5 Los Angeles, the Conrad N. Hilton Foundation, the Reissa Foundation, and the Heising-Simons Foundation for infrastructure support.

References

  1. 1. Brownell MD, Jutte DP. Administrative data linkage as a tool for child maltreatment research. Child Abuse Negl. 2013;37:120–124. pmid:23260116
  2. 2. Campbell KM, Deck D, Krupski A. Record linkage software in the public domain: a comparison of LinkPlus, The Link King, and abasic’deterministic algorithm. Health Informatics J. 2008;14:5–15. pmid:18258671
  3. 3. Enamorado T, Fifield B, Imai K. Using a probabilistic model to assist merging of large-scale administrative records. Am Polit Sci Rev. 2019;113:353–371.
  4. 4. Tromp M, Ravelli AC, Bonsel GJ, Hasman A, Reitsma JB. Results from simulated data sets: probabilistic record linkage outperforms deterministic record linkage. J Clin Epidemiol. 2011;64:565–572. pmid:20952162
  5. 5. Grannis SJ, Overhage JM, Hui S, McDonald CJ. Analysis of a probabilistic record linkage technique without human review. In: AMIA annual symposium proceedings. Bethesda: American Medical Informatics Association; 2003. p. 259.
  6. 6. Christen P. Data matching: concepts and techniques for record linkage, entity resolution, and duplicate detection. Berlin; Springer-Verlag; 2012.
  7. 7. Sayers A, Ben-Shlomo Y, Blom AW, Steele F. Probabilistic record linkage. Int J Epidemiol. 2016;45:954–964. pmid:26686842
  8. 8. Harron K, Dibben C, Boyd J, Hjern A, Azimaee M, Barreto ML, et al. Challenges in administrative data linkage for research. Big Data Soc. 2017;4(2). pmid:30381794
  9. 9. National Program of Cancer Registries. Link Plus [revised 23 March 17; cited 2023 June 01]. Available from: http://www.cdc.gov/cancer/npcr/tools/registryplus/lp.htm
  10. 10. Borthwick A, Buechi M, Goldberg A. Key concepts in the ChoiceMaker 2 record matching system. In Proceedings of the KDD-2003 Workshop on Data Cleaning, Record Linkage, and Object Consolidation. New York: Association for Computing Machinery; 2003. p. 28160.
  11. 11. Campbell K, Deck D, Cox A, Broderick C. The Link King user manual. Version V5.2. 2015.
  12. 12. Gregg F, Eder D. Dedupe [revised 2023 Feb 7; cited 2023 June 01]. Available from: https://github.com/dedupeio/dedupe
  13. 13. National Cancer Institute. Match*Pro Software [revised 2023, July 6; cited 2023 June 01]. Available from: https://seer.cancer.gov/tools/matchpro/
  14. 14. Anzia S, Jares J, Malhotra N. Does receiving government assistance shape political attitudes? evidence from agricultural producers. Am Polit Sci Rev. 2022;116:1389–1406.
  15. 15. Sariyar M, Borg A. The RecordLinkage package: detecting errors in data. R J. 2010;2:61–67.
  16. 16. Putnam-Hornstein E, Foust R, Cuccaro-Alamin S, Prindle J, Nghiem H, Ahn E, Palmer L. A population-based study of mental health diagnoses and child protection system involvement among medicaid-insured children. The Journal of Pediatrics. 2023 Jan 1;252:117–23. pmid:36027974
  17. 17. Havrlant L, Kreinovich V. A simple probabilistic explanation of term frequency-inverse document frequency (TF-IDF) heuristic (and variations motivated by this explanation). Int J Gen Syst. 2017;46:27–36.
  18. 18. Buttan Y, Chaudhary A, Saxena K. An improved model for breast cancer classification using random forest with grid search method. In: Proceedings of the Second International Conference on Smart Energy and Communication. Singapore: Springer; 2021. pp. 407–415.
  19. 19. Putnam-Hornstein E, Cleves MA, Licht R, Needell B. Risk of fatal injury in young children following abuse allegations: evidence from a prospective, population-based study. Am J Public Health. 2013;103:e39–e44. pmid:23947328
  20. 20. Putnam-Hornstein E, Ghaly M, Wilkening M. Integrating data to advance research, operations, and client-centered services in California: Integrating millions of administrative records across California’s health and human services programs to improve operations, coordinate services, develop targeted interventions, and more. Health Aff. 2020;39(4):655–661.
  21. 21. Zou G. A modified Poisson regression approach to prospective studies with binary data. Am J Epidemiol. 2004;159:702–706. pmid:15033648
  22. 22. Putnam-Hornstein E, Needell B. Predictors of child protective service contact between birth and age five: an examination of California’s 2002 birth cohort. Child Youth Serv Rev. 2011;33:1337–1344.
  23. 23. Putnam-Hornstein E, Needell B, King B, Johnson-Motoyama M. Racial and ethnic disparities: a population-based examination of risk factors for involvement with child protective services. Child Abuse Negl. 2013;37:33–46. pmid:23317921
  24. 24. Putnam-Hornstein E, Prindle JJ, Rebbe R. Community disadvantage, family socioeconomic status, and racial/ethnic differences in maltreatment reporting risk during infancy. Child Abuse Negl. 2022;130:105446. pmid:35144838
  25. 25. Ardabili S, Mosavi A, Várkonyi-Kóczy AR. Advances in machine learning modeling reviewing hybrid and ensemble methods. In: Várkonyi-Kóczy AR, editor. Engineering for sustainable future: selected papers of the 18th International Conference on Global Research and Education Inter-Academia– 2019. Cham: Springer; 2020. pp. 215–227.
  26. 26. Jiang T, Gradus JL, Rosellini AJ. Supervised machine learning: a brief primer. Behav Ther. 2020;51:675–687. pmid:32800297
  27. 27. Sagi O, Rokach L. Ensemble learning: a survey. Data Min Knowl Discov. 2018;8:e1249.