- Research article
- Open Access
- Open Peer Review
The effect of data cleaning on record linkage quality
- Sean M Randall1Email author,
- Anna M Ferrante1,
- James H Boyd1 and
- James B Semmens1
https://doi.org/10.1186/1472-6947-13-64
© Randall et al.; licensee BioMed Central Ltd. 2013
- Received: 17 March 2013
- Accepted: 29 May 2013
- Published: 5 June 2013
Abstract
Background
Within the field of record linkage, numerous data cleaning and standardisation techniques are employed to ensure the highest quality of links. While these facilities are common in record linkage software packages and are regularly deployed across record linkage units, little work has been published demonstrating the impact of data cleaning on linkage quality.
Methods
A range of cleaning techniques was applied to both a synthetically generated dataset and a large administrative dataset previously linked to a high standard. The effect of these changes on linkage quality was investigated using pairwise F-measure to determine quality.
Results
Data cleaning made little difference to the overall linkage quality, with heavy cleaning leading to a decrease in quality. Further examination showed that decreases in linkage quality were due to cleaning techniques typically reducing the variability – although correct records were now more likely to match, incorrect records were also more likely to match, and these incorrect matches outweighed the correct matches, reducing quality overall.
Conclusions
Data cleaning techniques have minimal effect on linkage quality. Care should be taken during the data cleaning process.
Keywords
- Data cleaning
- Data quality
- Medical record linkage
Background
Record linkage in context
Record linkage is the process of bringing together data relating to the same individual from within or between datasets. This process is non-trivial when unique person based identifiers do not exist, and linkage is instead performed using probabilistic or other techniques that compare personally identifying information such as name and address, which may include error or change over time.
While record linkage is frequently performed in a business or administrative context to remove duplicate entries from person based datasets, it has also been widely used to enable health researchers to gain event based longitudinal information for entire populations. In Australia, research carried out using linked health data has led to numerous health policy changes [1, 2], and the success of previous linkage efforts has led to the development of national linkage infrastructure [3].
Record linkage methodology
Approaches used in record linkage fall across a spectrum between deterministic and probabilistic methods. Deterministic linkage methods range from simple joins of datasets by a consistent entity identifier to sophisticated stepwise algorithmic linkage which includes additional information to allow variation between records that match i.e. it does not rely on an exact match of the entity identifier. Probabilistic methods, on the other hand, use various fields between data sets to calculate the odds that two records belong together [4]. These odds are represented as probability weights or scores which are calculated (summed) for each pair of records as they are compared. If the total score for a record pair is greater than a set matching threshold, then they are deemed to be a match – the records belong to the same person. The probabilistic approach allows for inconsistencies between records with missing matches i.e. it has the capacity to link records with errors in the linking fields.
Several studies have demonstrated that probabilistic linkage techniques are more robust against errors, and result in better linkage quality than deterministic methods [5–7]. Probabilistic methods are also more adaptable when large amounts of data require linkage [8].
Data cleaning in record linkage
Irrespective of which linkage approach is being used, the linkage process is usually preceded by a data cleaning phase. Data cleaning (sometimes called standardisation or data cleansing) involves correcting, removing or in some way changing fields based on their values. These new values are assumed to improve data quality and thus be more useful in the linkage process.
There is evidence that improvements in the quality of the underlying data lead to improvements in the quality of the linkage process. For example, early studies of probabilistic linkage in health research demonstrated that greater amounts of personal identifying data greatly improved the accuracy of linkage results [9, 10]. Studies have also shown that data items with more discriminating power lead to better linkage results [11, 12].
In the absence of strongly identifying personal information, data cleaning has been recognised as one of the key ways to improve the quality of linkage [13]. The record linkage literature identifies data cleaning as one of the key steps in the linkage process [14–17], which can take up to 75% of the effort of record linkage itself [18].
Data cleaning techniques
A variety of data cleaning techniques are used in record linkage [18–20]. Some data cleaning techniques seek to increase the number of variables by splitting apart free text fields. Others seek to simply transform variables into a specific representation, without actually changing the information. Further techniques aim to change the information in the fields, either by removing invalid values, changing values, or imputing blank values. Based on a review of five institutions conducting linkage in Australia and eight linkage software packages [19], the following data cleaning techniques were identified.
Reformatting values
Data values can be simply changed to a new format without actually creating or removing information. This ensures that all data is in a common standard for comparison during linkage. For example, two datasets which store dates in a different format (such as ‘11/08/86’ and ‘11th August 1986’), would need to be changed to a common format for comparison. No data is changed by this transformation, only the representation of the data. This technique is essential for ensuring matching fields can be compared [18].
Removing punctuation
Unusual characters and punctuation are typically removed from alphabetic variables. Names with spaces, hyphens or apostrophes may be more likely to be misrepresented, and removing these values can remove any differences between these values.
Removing alternative missing values and uninformative values
Datasets can often contain specially coded input values when no information is available – for instance ‘9999’ for a missing postcode. Other datasets may contain information that is not useful to the linkage process - hospital admission records may contain ‘Baby of Rachael’ in a forename field, or ‘NO FIXED ADDRESS’ in an address field. These are commonly removed [18]. In traditional probabilistic linkage, two variables that agree on a value (for instance, both are marked ‘UNKNOWN ADDRESS’) will receive a positive score, which in this case, may be inappropriate. A comparison involving a missing or blank value will typically not result in any positive or negative score.
Phonetic encoding
By creating an encoding of the phonetic information encapsulated in an alphabetic variable (such as a surname) names that are recorded as different spellings but sound the same will be brought together. Phonetic encoding is a common technique in record linkage. Common encoding algorithms used in record linkage include Soundex [21], NYSIIS [22] and Metaphone [23]. NYSIIS has been used for record linkage in Canada [13], while in the Oxford Record Linkage Study the Soundex value of the NYSIIS code is used in their linkage [18].
Name and address standardisation
Name standardisation or name parsing is the process of breaking down a person’s full name into its individual components. For instance, a name field with the entry ‘Dr John Harry Williams’ could be broken down into title, first name, middle name and last name, and these components could be individually compared.
Similarly, an address can be broken down into its constituents such as street number, street name and street type. By creating multiple variables in this way, small differences between records such as a different order may have less effect in bringing these records together. Typically the process of breaking the address into separate components has been carried out using a set of rules [24], but the application of statistical methods has also proved useful [25].
Nickname lookups
A nickname file, containing common nicknames and diminutive names for given names can be used to translate forenames to a common value. Using a nickname lookup, a person recorded as Bill on one dataset and William on another could be given the same first name, potentially bringing these records together [18].
Sex imputation
A record with a missing sex value can have this value imputed based on their first name. This requires a lookup table which equates common first names with sex.
Variable and field consistency
Records containing variables which are inconsistent can be edited to remove this inconsistency [20]. For instance, a record with suburb of Sydney and postcode of 6000 is inconsistent, as this is the incorrect postcode for this suburb. It is not often clear which variable to change in order to resolve this inconsistency.
Prevalence of data cleaning
These techniques encapsulate those found in linkage software packages or in use by dedicated linkage units in Australia during our environmental scan. All techniques listed here were either in use or under consideration by at least one institution conducting linkage in Australia, and all institutions asked used at least one of these techniques to clean their data.
Availability of data cleaning functionality across a sample of linkage packages
Linkage Wiz | Febrl | BigMatch | Link king | FRIL | LINKS | ChoiceMaker | QualityStage | |
---|---|---|---|---|---|---|---|---|
Reformat values | Yes | Yes | No | Yes | Yes | No | Yes | Yes |
Remove punctuation | Yes | Yes | No | Yes | Yes | No | Yes | Yes |
Remove alt. missing values | Yes | Yes | No | Yes | Yes | No | Yes | Yes |
Phonetic encoding | Yes | Yes | No | Yes | Yes | No | Yes | Yes |
Name/Address Standardisation | Yes | Yes | No | No | No | No | Yes | Yes |
Nickname lookup | Yes | Yes | No | Yes | No | No | No | Yes |
Sex imputation | Yes | Yes | No | Yes | No | No | No | Yes |
Data cleaning functionality in linkage software packages ranges from non-existent (BigMatch, LINKS) to comprehensive (Febrl. QualityStage, Linkage Wiz). Techniques available for reformatting variables typically include trimming, splitting and merging fields, classifying values, and reformatting dates.
Packages which remove specific values typically use a default invalid value list, which can then be added to by the user (for example Febrl, Link King, QualityStage, Linkage Wiz). Phonetic encoding algorithms available typically include Soundex at a minimum, with NYSIIS also common. Additional available techniques include ‘backwards NYSIIS’, metaphone and double metaphone. The lack of data cleaning functionality in some packages tended to be the result of a design decision to split this functionality into a separate software package rather than a value judgement about its usefulness.
Advantages of data cleaning
In a record linkage context, the aim of data cleaning is to improve linkage quality [18, 34]– that is, reduce the number of false positives (two records incorrectly identified as belonging to the same person) and false negatives (two records incorrectly identified as not belonging to the same person). Without data cleaning, many true matches would not be found, as the associated attributes would not be sufficiently similar [35].
Despite its widespread availability in linkage software packages, its use by numerous linkage groups, and its recognition as a key step in the record linkage process, the record linkage literature has not extensively explored data cleaning in its own right. Particular methods of cleaning data variables have been evaluated previously. Churches et al. [25] compared rule based methods of name and address standardisation to methods based on probabilistic models, finding more accurate address information when cleaned using probabilistic models. Wilson [36] compared phonetic algorithms and hand curated mappings on a genealogical database, finding the hand-curated mappings more appropriate for name matching. To our knowledge there has been no systematic investigation of the extent to which data cleaning improves linkage quality, or which techniques are most effective.
Objectives
Implicit in the data cleaning process is the assumption that data cleaning will improve linkage quality. However there is limited literature that has quantified the extent of improvement arising from data cleaning. Moreover, little is known about the relative effectiveness of various techniques. The current study attempts to answer these questions through a systematic investigation of the effect of data cleaning on linkage quality using two datasets – a ‘synthetic’ dataset and a large-scale ‘real world’ administrative dataset.
Since real world datasets for which the ‘answers’ are known are both difficult to source and virtually impossible to share, we opted to generate and use a synthetic dataset. The synthetic data files contain artificially created records that have characteristics that closely resemble the attributes of real world datasets. Such datasets are typically use in benchmarking or systems testing.
Methods
Road map for measuring overall linkage quality.
Datasets
The synthetically generated data set consisted of 400,000 records, containing multiple records belonging to the same person. The synthetic data was generated using an amended version of the FEBRL data generator [37]. As a first step, the generator creates a user specified number of original records. These are created randomly, based on frequency lookup tables. Duplicate records are created in a second step, based on the original records. Duplicate records are created by randomly selecting an original record, then randomly choosing the number of duplicates to be created from it, and then randomly introducing errors according to user-specified parameters. An additional probability distribution specifies how likely data items or attributes are selected for introducing errors (it is possible for data items to have no errors at all).
The synthetic data file was based on frequency distributions obtained from the Western Australian electoral roll. As voting is compulsory in Australia, the electoral rolls are highly representative of the population. To avoid the potential of identifying individuals from the electoral data, the frequency list was truncated so that frequency counts below five were excluded.
A comparison of the most common fields in the created synthetic data and the original data it was based on
Surname (top 5) | Synthetic Per cent | Original Per cent | Male forename (top 5) | Synthetic Per cent | Original Per cent |
---|---|---|---|---|---|
Missing value | 1.98 | Missing value | 1.99 | ||
Smith | 0.92 | 0.94 | John | 3.44 | 3.47 |
Jones | 0.55 | 0.55 | David | 3.09 | 3.09 |
Brown | 0.46 | 0.46 | Michael | 2.95 | 2.95 |
Williams | 0.46 | 0.46 | Peter | 2.87 | 2.88 |
Taylor | 0.44 | 0.44 | Robert | 2.47 | 2.47 |
Female forename (top 5) | Synthetic Per cent | Original Per cent | Postcode (top 5) | Synthetic Per cent | Original Per cent |
Missing value | 1.99 | Missing value | 1.01 | ||
Margaret | 1.57 | 1.56 | 6210 | 2.84 | 2.84 |
Susan | 1.35 | 1.34 | 6163 | 2.33 | 2.34 |
Patricia | 1.22 | 1.22 | 6027 | 2.06 | 2.05 |
Jennifer | 1.19 | 1.20 | 6155 | 2.02 | 2.02 |
Elizabeth | 1.05 | 1.05 | 6065 | 2.00 | 1.98 |
This dataset had previously been used for an evaluation of linkage software [38]. An advantage to the use of synthetic datasets is that they are transportable, and so allow easier validation, and the ‘answers’ as to which records belong to the same person are available, unlike in real administrative data. This dataset is freely available (see Additional file 1).
Ten years of ‘real world’ hospital admissions data was sourced from one Australian state. This consisted of almost 7 million records. This dataset comprised the following fields: first name, middle name, surname, date of birth, sex, address, suburb, postcode and state. This data had previously been linked to a very high standard using probabilistic linkage along with a rigorous manual review of created links, and a quality assurance program to analyse and manually review likely errors. Based on quality assurance procedures, the estimated error rate of this linkage is 0.3% [39]. Furthermore, these links have been validated through this datasets use in a large number of research projects and published research articles [1]. The links created during this original linkage allowed us to evaluate our linkage quality in comparison.
Both synthetic data and real administrative data have advantages and disadvantages comparison data sets. Synthetic data may not manage to capture all the complexity of errors that real administrative data can. Using real administrative data requires relying on the results of previous linkages as a standard by which to compare which may not be entirely accurate, whereas synthetic data gives a known, accurate standard. By using both of these datasets in our analysis, we hope to avoid both of these issues, and gain the best of both worlds.
Cleaning techniques
Specific data cleaning techniques used on each dataset
Synthetic data | ||
---|---|---|
Fields available for linkage: forename, surname, date of birth, sex, postcode | ||
No cleaning | Minimal cleaning | High cleaning |
Reformat values: | Reformat values: | Reformat values: |
Not required | Not required | Not required |
Remove alt. missing values and uninformative values: | Remove alt. missing values and uninformative values: | |
Invalid dates of birth removed | Invalid dates of birth removed | |
Invalid postal code values removed | Invalid post code values removed | |
Remove punctuation: | Remove punctuation: | |
Both forename and surname fields had all punctuation and spaces removed | Both forename and surname fields had all punctuation and spaces removed | |
Nickname lookup: | ||
Nicknames were changed to their more common variant. | ||
Sex Imputation | ||
Records with missing sex had a value imputed based on their first name. | ||
Hospital admissions data | ||
Fields available for linkage: forename, middle name, surname, sex, date of birth, address, suburb, postcode, state | ||
No cleaning | Minimal cleaning | High cleaning |
Reformat values: | Reformat values: | Reformat values: |
Date of birth reformatted. | Date of birth reformatted | Date of birth reformatted. |
Remove alt. missing values and uninformative values: | Remove alt. missing values and uninformative values: | |
Invalid dates of birth were removed | Invalid dates of birth were removed | |
Invalid postcode values were removed (‘9999’ etc.) | Invalid postcode values were removed (‘9999’ etc.) | |
Uninformative address and suburb values removed (‘NO FIXED ADDRESS’, ‘UNKNOWN’ etc.) | Uninformative address and suburb values removed (‘NO FIXED ADDRESS’, ‘UNKNOWN’ etc.) | |
Birth information encoded in first name removed (‘TWIN ONE OF MARTHA’ etc.) | Birth information encoded in first name removed (‘TWIN ONE OF MARTHA’ etc.) | |
Remove punctuation: | Remove punctuation: | |
Forename, middle name surname and suburb fields had all punctuation and spaces removed | Forename, middle name surname and suburb fields had all punctuation and spaces removed | |
Nickname lookup: | ||
Nicknames were changed to their more common variant. |
A nickname lookup table was developed based on similar nickname lookup tables found in linkage packages and as used by Australian linkage units. A sex imputation table was developed by examining the frequency of each given name in the data files and calculating the probability of the person being male or female. A record with a missing sex value was then given the most common gender value for this name.
Linkage strategy
The linkage strategy chosen was based on a previously published default strategy used for an evaluation of linkage software [38]. A probabilistic linkage approach was used with two blocks (Soundex of surname with first initial, and date of birth) and all possible comparison variables were computed in each block. A String similarity measure (the Jaro-Winkler string comparator [40]) was used for all alphabetic variables (names, address and suburb) with exact matches being carried out on all other variables. Day, month and year of birth were all compared separately. Correct agreement and disagreement weights for probabilistic linkage [41] were calculated for each variable and used in linkage. The threshold setting was adjusted multiple times with the linkage quality computed for each adjustment, with the highest result (i.e. the largest F-measure) reported. The threshold was adjusted in both directions in increments of 0.5, until it was clear all future adjustments would continue to worsen the F-measure. This linkage strategy was based on a previously published ‘default’ linkage strategy [38].
Linkage methods
As probabilistic record linkage techniques provide robust matching results for data which contain inconsistencies or incomplete data, these have been used throughout the study to match both the synthetic and ‘real world’ data sets. Following the traditional probabilistic linkage approach, pairs of records were compared and classified as matches if the matching score is above the threshold.
To calculate the matching score reached by a pair of records, each field (for instance first name or postcode) has been compared. Scores for each individual field were computed using agreement and disagreement weights. The agreement weight expresses the likelihood that records which belong to the same person have the same value for this field. The disagreement weight expresses the likelihood that records which do not belong to the same person have the same value on this field. The sum of these individual field scores has been computed and compared to the matching threshold to determine matches or non-matches [15].
Linkage engine
BigMatch, developed by the US Bureau of Census [42] was used as the linkage engine for the analysis. BigMatch was chosen as it is fast, can handle large volumes, has a transparent linkage process based on probabilistic methods, and importantly, does not contain any automatic inbuilt data cleaning. The software had previously been evaluated and found to perform well against other linkage software packages [38].
Measuring linkage quality
Measuring the quality of a single variable
A similar approach to the one described above can be used when measuring the quality of a single variable. A variable which nearly always has the same value for all records belonging to the same person, but nearly always has a different value than all records belonging to other people, would be much more useful in the linkage process than one which seldom had these properties. Put in another way, a variable with a high precision (here measured as the proportion of times that two variables which have the same value belong to the same person) and a high recall (the proportion of times two records matching each other had the same value of the variable in question) will be more useful than one with lower precision and recall.
As some data cleaning techniques may increase precision and lower recall, we can determine which technique will have the overall best effect on predictive accuracy by using the F-measure of these two values. Furthermore we can measure the relative improvement of a data cleaning technique by comparing its individual F-measure before and after data cleaning.
Results
Overall linkage quality results
Synthetic data | |
---|---|
F-measure | |
No cleaning | 0.883 |
Minimal cleaning | 0.882 |
High cleaning | 0.875 |
Hospital admissions data | |
F-measure | |
No cleaning | 0.993 |
Minimal cleaning | 0.993 |
High cleaning | 0.992 |
Improvement in predictive ability of data cleaning techniques
Hospital admissions data | Synthetic data | |
---|---|---|
Remove punctuation | −a0.08% | +0.08% |
Remove alt. missing values | +0.5% | 0% |
Nickname lookup | −28% | −33% |
Sex Imputation | NA | −5% |
While removing missing values and uninformative values seemed to increased predictive ability, all other techniques displayed mixed or worse results. Using name variables that had nicknames and diminutive names replaced with their original names resulted in a large 30% decrease in that variable’s predictive value.
Examples of single variable changes in predictive ability for individual cleaning techniques in hospital admission data
Hospital admissions data | |||
---|---|---|---|
Precision | Recall | F-measure | |
Percentage difference from original variable | |||
Given name original | 0.006575 | 0.946085 | 0.013059 |
Given name with removed punctuation | 0.006573↓ b 0.03% | 0.947188↑0.11% | 0.013056↓0.02% |
Given name with nicknames removed | 0.004357↓33.7% | 0.953738↑0.81% | 0.008675↓33.5% |
Surname original | 0.025265 | 0.98824 | 0.049271 |
Soundex of surname | 0.008845↓65% | 0.994926↑0.67% | 0.017533↓64.4% |
Address original | 0.687066 | 0.669649 | 0.678246 |
Address with alternate missing values and uninformative values removed | 0.687398↑0.05% | 0.709426↑5.9% | 0.698238↑2.9% |
Discussion
Overall, it was found that the effect of data cleaning on linkage quality was very small. If there was any effect at all, it appeared to decrease linkage quality. While some techniques led to small improvements, many others led to a large decrease in quality.
These results were not as expected. Data cleaning is assumed to improve data quality and thus to increase linkage quality. Examining the effect individual transformations had on a single variable’s predictive ability allows us to explain why this occurred. While the number of correct matches that were brought together increased with data cleaning, the number of incorrect matches also increased, in most cases dramatically. By removing the variability between records we are reducing our ability to distinguish one record from another.
Data cleaning techniques typically reduce the variability between values of the field in question. By removing nicknames, a smaller variety of names will be found in the dataset. By removing differences created by punctuation, this variability will be removed. As anticipated [7] this leads to a greater number of correct matches found; however this also leads to the identification of more incorrect matches.
Strengths and limitations
Given the acceptance of data cleaning as an integral part of the linkage process, it was assumed that data cleaning would improve quality in general. The results obtained appear to contradict the conventional wisdom that data cleaning is a worthwhile procedure due to its ability to improve linkage quality.
Through the use of multiple representative datasets and the analysis of both linkage quality and individual transformations, these results seem robust. Measuring the effect of data cleaning in linkage is complex, as there are a multitude of parameters which can be altered that could affect the outcome of linkage quality. A potential concern is that some untested threshold value or other linkage parameter changes could drastically change these results. However, when analysed on their own, individual variables showed decreased predictive ability. If we accept that record linkage variables are independent (something which is an assumption of probabilistic record linkage) then it seems unlikely that any changes to linkage parameters will lead to linkage quality greater than that found in uncleaned data. On the other hand, the independence of variables used in linkage is often questionable, in which case the lower predictive ability of the individual variables is at the very least supportive of our conclusion.
The linkage strategy adopted here made heavy use of string similarity metrics. String similarity metrics may reduce the need for data cleaning, as they allow finer grained measures of similarity compared to exact matching, where variables with very slight differences will be treated as non-matches. A linkage strategy using exact matching only will have more need for data cleaning to bring correct records together, and this linkage strategy was not tested. However, the analysis of predictive ability of individual variables and their cleaned versions was carried out with exact matching only, which showed a decrease in predictive ability. This suggests data cleaning would not affect results any differently for those using an exact matching linkage strategy.
The linkages conducted simply replaced the original variables with the cleaned variables. An alternative method may be to use both the original and cleaned versions as variables in linkage. While this method violates the assumptions of independence underlying probabilistic record linkage [41], linkage variables are almost never independent, and such techniques have been implemented in some linkage packages. Further work would be required to determine the effect of using cleaned variables in conjunction with original uncleaned variables.
The f-measure was used as the sole measure of linkage quality. An underlying assumption of using this measure is that a single false positive is as equivalently undesirable as a single false negative. While this seems a sensible starting point, it should be noted that in numerous practical applications of record linkage this is not the case. For instance, if linking registry information to inform patients of their condition, it is much more important to reduce false negatives than false positives. Further analysis using additional metrics may be required to ensure these results hold using other linkage quality metrics. The key reason why cleaning failed to improve quality was the reduced variability of each field. Other data cleaning techniques not investigated here such as address standardisation increase the number of variables available for comparison and these techniques may improve quality.
Avenues for further research
From this work it is clear that data cleaning does not always lead to increased linkage quality. Without further testing on a wide variety of datasets, it is hard to draw any further conclusions about the use of data cleaning in record linkage. Repeating this research on a wide variety of datasets is important. Further research into the use of cleaned as well as uncleaned variables together in the same linkage, into the use of further cleaning technique such as name and address standardisation is required. This research suggests that there are some situations where data cleaning transformations are helpful and others where they are not – determining a way of identifying when a transformation is likely to be helpful would be an important and useful finding.
Conclusion
Data cleaning encompasses a variety of techniques which will be appropriate in specific circumstances. Care should be taken when using these techniques.
Declarations
Acknowledgements
This project is supported by the Australian Government National Collaborative Research Infrastructure Strategy’s Population Health Research Network. The authors would like to thank the reviewers for their invaluable comments.
Authors’ Affiliations
References
- Brook EL, Rosman DL, Holman CDAJ: Public good through data linkage: measuring research outputs from the Western Australian data linkage system. Aust N Z J Public Health. 2008, 32: 19-23. 10.1111/j.1753-6405.2008.00160.x.View ArticlePubMedGoogle Scholar
- Hall SE, Holman CDAJ, Finn J, Semmens JB: Improving the evidence base for promoting quality and equity of surgical care using population-based linkage of administrative health records. Int J Qual Health Care. 2005, 17: 375-381.View ArticleGoogle Scholar
- Boyd JH, Ferrante AM, O’Keefe CM, Bass AJ, Randall SM, Semmens JB: Data linkage infrastructure for cross-jurisdictional health-related research in Australia. BMC Health Serv Res. 2012, 12: 480-10.1186/1472-6963-12-480.View ArticlePubMedPubMed CentralGoogle Scholar
- Fellegi IP, Sunter AB: A theory for record linkage. J Am Stat Assoc. 1969, 64: 1183-1210. 10.1080/01621459.1969.10501049.View ArticleGoogle Scholar
- Pinder R, Chong N: Record linkage for registries: current approaches and innovative applications.http://www.naaccr.org/LinkClick.aspx?fileticket=wtyP5M23ymA%3D,
- Gomatam S, Carter R, Ariet M, Mitchell G: An empirical comparison of record linkage procedures. Stat Med. 2002, 21: 1485-1496. 10.1002/sim.1147.View ArticlePubMedGoogle Scholar
- Tromp M, Ravelli AC, Bonsel GJ, Hasman A, Reitsma JB: Results from simulated data sets: probabilistic record linkage outperforms deterministic record linkage. J Clin Epidemiol. 2011, 64: 565-572. 10.1016/j.jclinepi.2010.05.008.View ArticlePubMedGoogle Scholar
- Clark DE, Hahn DR: Comparison of probabilistic and deterministic record linkage in the development of a statewide trauma registry. Proceedings of the annual symposium on computer application in medical care. 1995, Maryland USA: American Medical Informatics Association, 397-Google Scholar
- Newcombe HB, Smith ME, Howe GR, Mingay J, Strugnell A, Adbatt JD: Reliability of computerized versus manual death searches in a study of the health of Eldorado uranium workers. Comput Biol Med. 1983, 13: 13-View ArticleGoogle Scholar
- Roos LL JRAW, Nicol JP: He art and science of record linkage: methods that work with few identifiers. Book the art and science of record linkage: methods that work with few identifiers. 1985, Winnipeg, Canada: Departments of Business Administration and Social and Preventive Medicine University of ManitobaGoogle Scholar
- Roos L, Wajda A: Record linkage strategies. Part I: Estimating information and evaluating approaches. Methods Inf Med. 1991, 30: 117-PubMedGoogle Scholar
- Quantin C, Bouzelat H, Allaert F, Benhamiche A-M, Faivre J, Dusserre L: How to ensure data security of an epidemiological follow-up: quality assessment of an anonymous record linkage procedure. Int J Med Inform. 1998, 49: 117-122. 10.1016/S1386-5056(98)00019-7.View ArticlePubMedGoogle Scholar
- Wajda A, Roos LL: Simplifying record linkage: software and strategy. Comput Biol Med. 1987, 17: 239-248. 10.1016/0010-4825(87)90010-2.View ArticlePubMedGoogle Scholar
- Gu L, Baxter R, Vickers D, Rainsford C: Record linkage: current practice and future directions. CSIRO Mathematical and Information Sciences Technical Report. 2003, 3: 83-Google Scholar
- Herzog TN, Scheuren FJ, Winkler WE: Data quality and record linkage techniques. 2007, New York: SpringerGoogle Scholar
- Winkler WE: Record linkage software and methods for merging administrative lists. 2001, : Statistical research division Technical Report RR01—03 US Bureau of CensusGoogle Scholar
- Jaro MA: Advances in record-linkage methodology as applied to matching the 1985 census of Tampa, Florida. J Am Stat Assoc. 1989, 89: 414-420.View ArticleGoogle Scholar
- Gill L: Methods for automatic record matching and linkage and their use in national statistics. 2001, London, UK: Office for National StatisticsGoogle Scholar
- Ferrante A, Boyd J: Data linkage software evaluation: a first report (part I). Perth. Book data linkage software evaluation: A first report (part I). 2010, Perth: Curtin UniversityGoogle Scholar
- Christen P: Data matching. 2012, New York: SpringerView ArticleGoogle Scholar
- Odell KM, Russell RC: Soundex phonetic comparison system. 1918, : US Patent 1261167, vol. 1261167Google Scholar
- Taft RL: Name search techniques. 1970, New York: Bureau of Systems DevelopmentGoogle Scholar
- Philips L: Hanging on the metaphone. Computer Language. 1990, 7 (23): 39-42.Google Scholar
- Day C: Record linkage II: experience using AUTOMATCH for record linkage in NASS. 1996, USA: US Department of AgricultureGoogle Scholar
- Churches T, Christen P, Lim K, Zhu JX: Preparation of name and address data for record linkage using hidden markov models. BMC Med Inform Decis Mak. 2002, 2: 9-10.1186/1472-6947-2-9.View ArticlePubMedPubMed CentralGoogle Scholar
- IBM Infosphere QualityStage.http://www-01.ibm.com/software/data/infosphere/qualitystage/,
- Linkage Wiz data matching software.http://www.linkagewiz.net/,
- Borthwick A, Buechi M, Goldberg A: Key concepts in the choicemaker 2 record matching system. Procs first workshop on data cleaning, record linkage, and object consolidation, in conjunction with KDD. 2003, Washington DC: SIGKIDDGoogle Scholar
- Christen P, Churches T, Hegland M: Febrl–a parallel open source data linkage system. Advances in knowledge discovery and data mining. 2004, New York: Springer, 638-647.View ArticleGoogle Scholar
- Jurczyk P, Lu JJ, Xiong L, Cragan JD, Correa A: FRIL: A tool for comparative record linkage. AMIA annual symposium proceedings. 2008, Maryland, USA: American Medical Informatics Association, 440-Google Scholar
- Campbell KM, Deck D, Krupski A: Record linkage software in the public domain: a comparison of link plus. The link king and a ‘basic’ deterministic algorithm. Health Informatics. 2008, 14: 5-15. 10.1177/1460458208088855.View ArticleGoogle Scholar
- Howe G, Lindsay J: A generalized iterative record linkage computer system for use in medical follow-up studies. Comput Biomed Res. 1981, 14: 327-340. 10.1016/0010-4809(81)90004-5.View ArticlePubMedGoogle Scholar
- Yancey WE: BigMatch: a program for extracting probable matches from a large file for record linkage. Computing. 2002, 01: 1-8.Google Scholar
- Tuoto T, Cibella N, Fortini M, Scannapieco M, Tosco L: RELAIS: Don’t Get lost in a record linkage project. Proc of the federal committee on statistical methodologies (FCSM 2007) research conference. 2007, Arlington, VA, USA: Federal Committee on Statistical MethodologiesGoogle Scholar
- Matching and record linkage. Edited by: Winkler WE. 1995, New Jersey, USA: John Wiley & SonsGoogle Scholar
- Wilson DR: Name standardization for genealogical record linkage. Proc of the 5th Annual family history technology workshop. 2005, USA: Brigham Young UniversityGoogle Scholar
- Pudjijono A, Christen P: Accurate synthetic generation of realistic personal information. Proceedings of the 13th pacific-asia conference on advances in knowledge discovery and data mining. 2009, USA: SpringerGoogle Scholar
- Ferrante A, Boyd J: A transparent and transportable methodology for evaluating data linkage software. J Biomed Inform. 2012, 45: 165-172. 10.1016/j.jbi.2011.10.006.View ArticlePubMedGoogle Scholar
- Rosman D, Garfield C, Fuller S, Stoney A, Owen T, Gawthorne G: Measuring data and link quality in a dynamic multi-set linkage system. Book measuring data and link quality in a dynamic multi-set linkage system. 2001, WA: Data Linkage Unit, Department of HealthGoogle Scholar
- Winkler WE: String comparator metrics and enhanced decision rules in the fellegi-sunter model of record linkage. 1990, :Google Scholar
- Newcombe HB: Handbook of record linkage: methods for health and statistical studies, administration and business. 1988, New York: Oxford University PressGoogle Scholar
- Yancey WE: BigMatch: a program for extracting probable matches from a large file for record linkage. 2002, Maryland USA: Statistical Research Division U.S. Bureau of the Census, 01-Google Scholar
- Christen P, Goiser K: Quality and complexity measures for data linkage and deduplication. Quality measures for data mining. Volume 43. 2007, Berlin: Springer, 127-151.Studies in Computational Intelligence,View ArticleGoogle Scholar
- The pre-publication history for this paper can be accessed here:http://www.biomedcentral.com/1472-6947/13/64/prepub
Pre-publication history
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.