 Research article
 Open access
 Published:
A deterministic approach for protecting privacy in sensitive personal data
BMC Medical Informatics and Decision Making volume 22, Article number: 24 (2022)
Abstract
Background
Data privacy is one of the biggest challenges for any organisation which processes personal data, especially in the area of medical research where data include sensitive information about patients and study participants. Sharing of data is therefore problematic, which is at odds with the principle of open data that is so important to the advancement of society and science. Several statistical methods and computational tools have been developed to help data custodians and analysts overcome this challenge.
Methods
In this paper, we propose a new deterministic approach for anonymising personal data. The method stratifies the underlying data by the categorical variables and redistributes the continuous variables through a k nearest neighbours based algorithm.
Results
We demonstrate the use of the deterministic anonymisation on real data, including data from a sample of Titanic passengers, and data from participants in the 1958 Birth Cohort.
Conclusions
The proposed procedure makes data reidentification difficult while minimising the loss of utility (by preserving the spatial properties of the underlying data); the latter means that informative statistical analysis can still be conducted.
Introduction
Protection of privacy in personal data is a key societal concern with implications for scientific, commercial and government communities. Recent legislation such as the EU General Data Protection Regulation highlights this need to handle sensitive personal data carefully. At the same time, however, modern science emphasises the importance of transparency and reproducibility of scientific research [1]; indeed, open data and open science is arguably one of the biggest gains of the data revolution [2]. This has resulted in increasing demands from funders and publishers for the release of data, among other things, to accompany grant proposals and journal articles (see for example the proposition from the International Committee of Medical Journal Editors [3]).
This potential conflict between data sharing and protection of sensitive data means that data custodians must use disclosure controls and anonymisation methodologies to maintain individual confidentiality when sharing data. This is especially so in disciplines where collected data include sensitive information protected by ethicolegal regulations. Achieving anonymity is complicated however, as granular data can be mapped back to the individual who provided them relatively easily. Such demands and requirements have therefore led to the development of a range of data sharing solutions, commonly achieving privacy protection through pseudonymisation, encryption or confusionbased approaches [4,5,6,7,8]. However, no single solution can fully protect a dataset and often combinations of methods and multifaceted systems are essential.
Existing methods for sharing sensitive data
A number of approaches to sharing sensitive data are in existence. Some of these are theoretically straightforward in their approach, for example simple adjustments of direct identifiers (such as names and other forms of direct identifiers), or trusted data sharing within secure policy and procedural frameworks such as the Data Safe Havens and the Trusted Research Environments [9, 10]. Even in the latter scenario, the stripping of direct identifiers is often required and constitutes good practice.
However, other more technical or statistical solutions may be considered as alternatives. At one end of the spectrum, ‘black box’ computing tools like DataSHIELD [11,12,13] and ViPAR [14] can be used to control accessibility and allow secure federated data analysis. Such innovative tools mitigate disclosure risk while preserving data utility and allowing statistical analysis to be performed on the original data. Here the term ‘original data’ refers to the real data for analysis but they may already have been preprocessed (e.g., cleaned, harmonised, transformed, etc.) and/or pseudonymised (i.e., direct identifiers removed or disguised) before data storage. Black box approaches often restrict the analysis to a limited range of statistical methods and on specific outputs that the analyst can receive and publish (often nondisclosive lowdimensional summary statistics are returned to the analyst).
At the other end of the spectrum, simulated synthetic data [15] can provide full protection of privacy and eliminate any reidentification risk. However, synthetic data often do not capture the higherorder statistical properties of the underlying data, as the synthesis relies on simulation of entirely new values from a simplified statistical model. The quality of synthetic data depends strongly on the ability of the selected models to capture important relationships found in the original data. Thus the loss of information in simulations of synthetic datasets can be significant but these data can be very useful for designing research studies, developing software tools and methodologies, and for training purposes (see for example, the generation of synthetic data for the development of a proofofconcept biomedical data exploration and visualisation tool in Virtual Reality [16]).
However, analysis of anonymised data can often be preferable to working via a tightlycontrolled disclosuremitigating blackbox or to the analysis of synthetic data. In contrast to the former, the analyst can then use any statistical methods he/she chooses and unlike the latter the analytic data preserves all of the underlying statistical properties of the original data. This has given rise to a number of privacy protection techniques that do not require black boxes or the creation of synthetic data. As well as suppressing direct identifiers (pseudonymisation), examples of such extended techniques include (1) the control of direct identifiers, unique combinations of records, outlier values, or small cell counts, (2) the aggregation of granular detailed data to wider groups or categories, (3) the rounding of numerical values to numbers with fewer significant figures, (4) the perturbation of data with addition or multiplication of random stochastic noise, etc. Many of these focus on minimising disclosure risk and utility loss.
Strategically, each method or tool for disclosure control aims to reduce the risk of reidentification and to minimise the loss of data utility. However, no single solution is adequate to fully protect the data and even the most complex and sophisticated of solutions never offer informative data with zero risk of disclosure. The choice of the optimal approach is often assessed in terms of retaining the desired statistical properties of the data and is therefore heavily contextdependent [17]. Consequently, the merits of each approach should be carefully considered for any given dataset.
For example, the kanonymisation process [18], which falls into categories (1)–(3) above, preserves privacy by reducing the granularity of the data by combining the suppression of records and aggregation of variables. It cannot be used in cases where unique patterns or outliers (e.g., a rare disease indicator) are essential for the statistical analysis because the change or suppression of this information can lead to an incorrect conclusion (e.g., an incorrect medical diagnosis or decision). Other weaknesses of kanonymisation are the biases and skewing effects that it introduces to the data [19], and the high amount of information loss when it is applied to highdimensional datasets [20].
Meanwhile probabilistic anonymisation, which falls into category (4) above, is based on randomisation and therefore each data custodian can choose the level of noise used to perturb their data. There are a number of modelling techniques that can ‘remove’ the effect of added noise from a regression outcome and recover the true model parameters [21,22,23]; this is, of course, a desirable property of the procedure. However, care needs to be taken with this technique: the release of many ‘noisy’ datasets generated from the same input dataset, but using a different random number generator seed, can give rise to inferences for the observed underlying data (i.e., the average of many noisy datasets converges to the expected value of the input).
Contribution of this paper
In this paper we present a new deterministic procedure for protecting privacy in personal data. The proposed method obscures the microlevel information from a set of personal data while retaining the spatial statistical properties of the numerical variables. In contrast to kanonymisation, this approach does not reduce the granularity of the data. Furthermore, it does not depend on randomisation and so the same output will always be produced given particular input parameters (in contrast to probabilistic anonymisation). This procedure has been recently used for the generation of privacypreserving data visualisations [24].
The paper is structured as follows. In the “Methodology” section we introduce the deterministic anonymisation procedure and we explicitly describe its five basic steps. We also outline the metrics that we use to measure the disclosure risk of the anonymised data and the information loss generated due to the data perturbation during anonymisation. In the “Applications” section we apply the method to two real datasets: (1) a publicly available sample of data about passengers on the Titanic; and (2) a set of data from participants in the National Child Development Study [25]. This second set of data had previously been created and harmonised for use—by a large research team including us—in the analysis of “healthy obesity” as part of the BioSHaREEU project [26]. As a second part of the same project, we also used these same data to support the methodological development of DataSHIELD as a blackbox solution to disclosure control [11]. In the “Applications” section we also conduct a sensitivity analysis to examine the effect of the number of nearest neighbours used in the procedure (parameter k) and the effect of the sample size to the information loss and disclosure risk of deterministically anonymised data. Finally, we discuss the strengths and limitations of the proposed procedure.
Methodology
Anonymisation procedure
The anonymisation method we propose in this paper stratifies the underlying data by all combinations of the levels in the categorical variables and then redistributes the continuous variables by replacing each record with the centroid of itself and its \(k1\) nearest neighbours for some specified k. It is important to note here that our procedure means that the categorical variables remain untouched by the anonymisation process; we ensure that this is safe in step (2) below. It is only the continuous data that change during the proposed anonymisation process.
The procedure consists of five basic steps:

(1)
Standardisation;

(2)
Stratification;

(3)
Location of k nearest neighbours;

(4)
Estimation of centroids;

(5)
Scaling and recentralisation of the data.
These steps are detailed below, but before their application we check that the dataset of individual records includes both continuous and categorical variables. If the given dataset includes only continuous variables, then step (2)—stratification—does not take place and the entire dataset is considered as a unique subset. Note that we are assuming that there is at least one continuous variable in the dataset.
Step (1): standardisation
We standardise each continuous variable using zscore transformation such that each variable is centralised to zero mean and scaled to one standard deviation.
Step (2): stratification
We stratify the data in relation to all possible combinations of the levels of categorical variables: each subgroup is now homogeneous in terms of the categorical data. For example, if a dataset includes three categorical variables where two of them are binary and the third has four categories, then the dataset is classified into \(2 \times 2 \times 4=16\) different subsets.
To make reidentification harder, we ensure that the data are kanonymous in each subset (i.e., each possible combination of levels of the categorical variables exists in the data at least k times). If kanonymity does not occur in one or more subsets (i.e., at least one subset has less than k observations), then we apply existing methods to achieve the kanonymity. This may include generalisation (i.e., individual values are aggregated in broader categories) and/or suppression (i.e., unique or rare patterns of individual records are removed from the dataset).
Note that the k used here is the same as the value used for the location of each k nearest neighbours (see the next step of the algorithm). The choice of k is open here, but is most likely that when choosing k large enough, the risk of disclosure is sufficiently small. However, choosing a very large k will result in substantive loss of information: for example, choosing k equal to the size of the smallest stratum will mean that each continuous variable for observations in this stratum will all be mapped to the same value in the next step.
Step (3): location of each observation’s \(k\)1 nearest neighbours
This step of the method locates the \(k1\) nearest neighbours of each observation, for which we use the socalled kNN (k nearest neighbours) algorithm. This procedure is based on a technique in machine learning and is widely used in several disciplines such as computational geometry [27], diagnostic medicine [28], cryptography [29], data mining [30], and as here, in solutions for data privacy [31].
The kNN algorithm is implemented separately in each stratum as formed in the previous step. The nearest neighbours are defined as the data points with minimum distances from each point of interest. Here we use the Euclidean distance as the distance metric between each two standardised coordinate vectors.
Step (4): estimation of centroids
When the \(k1\) nearest neighbours of each data point are identified, the algorithm computes the coordinates of the centroid of each k data points (i.e., the \(k1\) neighbours and the point of interest). The coordinates of each centroid are calculated as the average of the coordinates of the nearest neighbours in each dimension separately. For example, in a 2D space the xcoordinate of the centroid is the mean of the xcoordinates of the k nearest data points and the ycoordinate of the centroid is the mean of the ycoordinates accordingly.
The continuous values of each individual data point are then replaced by the coordinates of the corresponding centroid. These coordinates are then pooled together into a single data frame, which includes values for all strata (i.e., a reverse of the stratification process). The contents of this new data frame are the masked data.
Step (5): scaling and recentralisation
We apply scaling and shift to relocate the masked data back to the observed scale and location of the original data using an inverse of the zscore transformations applied in step (1). We multiply each continuous variable from the masked dataset (i.e., the centroids) with the ratio between the standard deviation of the original variable and the standard deviation of the masked variable. Adding the observed means of the original data to each continuous variable from the masked dataset completes the recentralisation.
Scaling is desirable because the third and the fourth steps typically shrink the overall variability of the data. However, it should be noted that when the scaling factor is sufficiently large, outliers or influential points may be displaced out of the convex hull of the original data. In particular, the masked data are not restricted to the space occupied by the original data.
Riskutility evaluation
Two key features should be addressed before releasing any anonymised data. Firstly, what is the risk of disclosure using the anonymised data? This will determine if the (anonymised) dataset is secure enough to be released. Secondly, how much information in the data was lost during the anonymisation process? That is, what is the utility loss? Of course, there is a tradeoff between the risk of disclosure and utility loss: methods that lower the risk of disclosure tend to increase utility loss.
Utility loss and disclosure risk evaluation is highly contextspecific and no metric can act as a catchall for either concern. In this paper we use three metrics to evaluate information loss, while a measure based on robust Mahalanobis distance will serve as a metric for disclosure risk.
Metrics for utility loss
The three metrics that we employ, look at different aspects of what it means to lose information due to the perturbation of the original values during anonymisation.
The first metric is a global metric that compares the entire anonymised dataset with the original data. We use a summary statistic of the propensity scores as the global measure of data utility of the anonymised data as proposed by Woo et al. [32]. To calculate the propensity scores, we first merge the original and the anonymised datasets vertically and we add a binary indicator which is equal to one for all records from the anonymised dataset and zero for all records from the original dataset. Then we estimate the propensity score which is the probability of a record being in the anonymised dataset. The estimated propensity score \({\hat{p}}_{i}\) of each record i is the predicted probability from a logistic model of the generated binary indicator regressed on all the variables in the merged dataset.
We then compare the distributions of the propensity scores in the original and masked data. If the two distributions are similar, we conclude that the data utility loss due to anonymisation is small. This global information loss [32] is given by
where N is the total number of rows in the merged dataset, \({\hat{p}}_{i}\) is the estimated propensity score for unit i, and c is the proportion of anonymised data in the merged dataset (in our procedure \(c=0.5\) as the original and anonymised datasets have the same number of observations). When the anonymised and original datasets are identical (no loss of utility, and we cannot distinguish between the two datasets), the distributions of their propensity scores are identical and U is equal to zero (i.e., no utility loss). If on the other hand the anonymised dataset can be entirely identified through the logistic regression above, then we have lost all utility and U is 1/4. Note, however, that this metric does not take into consideration the natural pairing of the original data values and their corresponding centroids. This metric may therefore not capture the more nuanced utility loss.
The second metric that we use to measure information loss is variablespecific, and unlike the first metric, takes into account the pairing between the real and anonymised observations. For each continuous variable we calculate the Euclidean distance between each original value and its corresponding anonymised value. We then estimate the mean of the squares of the Euclidean distances and divide it by the variance of the original variable. This standardised metric of the Euclidean distances of a variable \({\varvec{x}}\) is given by
where \({\varvec{x}}^{or}=(x_1^{or}, \ldots , x_n^{or})\) is the variable in the original dataset, \({\varvec{x}}^{an}=(x_1^{an}, \ldots , x_n^{an})\) is the corresponding variable in the anonymised dataset (both presented here in vector formats), and n is the length of the variables. This fraction indicates how much of the variability of the anonymised variable is introduced through the anonymisation. For example, if \(\delta _{{\varvec{x}}}\) = 0.4, we conclude that 60% of the variability of \({\varvec{x}}^{an}\) is inherited from the observed variability of \({\varvec{x}}^{or}\) while 40% is introduced due to anonymisation.
The third metric is an analysisspecific measure of utility loss. To make the analysisspecific comparison we must assume an intended analysis for the data; here we suppose that a regression model is required. The proposed regression model is applied to both datasets separately. A standardised difference between the estimate of the regression coefficients derived from the original data and the estimates derived from the anonymised data is then computed. The standardised difference is calculated as
where \({\hat{\beta }}_{or}\) and \({\hat{\beta }}_{an}\) are the coefficients of the same model estimated from the original and the anonymised data respectively and \(\text {SE}({\hat{\beta }}_{or})\) is the estimated standard error of the coefficients from the original data [33]. This metric can be thought of as a standardised bias between the estimates from the original and anonymised data. Checking for overlap between the confidence intervals of the regression coefficients from the original and anonymised data has also been suggested [33, 34].
Metrics for disclosure risk
A risk assessment is based on the specific data disclosure risk scenario that pertains and can be tackled from the point of view of the individual level data (either to assess for identity or attribute disclosure) or from the point of view of the global level data (risk related to the entire dataset or risk related to the protection method).
For the evaluation of the disclosure risk of anonymised data we use a robust Mahalanobis distancebased metric [35], which is estimated by the dRiskRMD function from the sdcMicro package in R [36]. This approach counts the number of ‘risky’ observations associated with higher probability of reidentification. Risky observations are those that have not been sufficiently perturbed and are located in the same vicinity as their original values (i.e., in very close proximity, and below an accepted threshold). In particular, it takes into account outliers and rare combinations of variables; these are usually at higher risk of identification in comparison to, for example, observations closer to centre of the mass of a dataset.
The algorithm that estimates the disclosure risk first calculates the robust Mahalanobis distance between each observation in the original dataset (a vector) and the mean vector of the original data. It then generates an interval around each (multivariate) original observation, where the length of the interval is defined by its squared robust Mahalanobis distance scaled by a constant, \(w_1\). These intervals create (possibly multidimensional) boxes around each observation. This scalar is a weight for adjusting the influence of the robust Mahalanobis distance and is chosen by the data custodian. Using this algorithm, we estimate two risks of disclosure.
The first risk of disclosure, risk1, is defined as the proportion of the anonymised values that fall into the intervals of their corresponding original values. Observations that fall within the intervals are considered potentially unsafe. The second metric, risk2, looks further at each observation flagged by risk1: it checks whether there are m other observations from the anonymised dataset very close to it, and computes the proportion of observations that are both at risk (according to risk1) and do not have m other observations close by. Note that risk2 will always therefore be at least as small as risk1. We use Euclidean distances here, and ‘close’ means being within some scalar \(w_2\) of the observation. The idea here is that if more than m points are ‘close’ to the observation then we consider this observation as safe. The parameters m and \(w_2\), are chosen by the data custodian.
Applications
The Titanic passengers’ data
Our first application illustrates the deterministic procedure through the anonymisation of a sample of Titanic passengers data. The dataset was obtained from Kaggle (https://www.kaggle.com/c/titanic/data) and includes 12 variables for 891 Titanic passengers. Our aim is to model the survival of passengers onboard the Titanic. We exclude three variables that are not considered important in predicting survival: ticket and the cabin numbers, and the trichotomous indicator of where the passengers embarked on the ship (Cherbourg, Southampton, or Queenstown).
The remaining nine variables are listed in Table 1 and include a unique ID for each passenger; a binary indicator showing if the passenger survived (1) or died (0); the passenger’s class (1st, 2nd, 3rd); the passenger’s name; sex; age; the number of siblings and spouses aboard the Titanic (SibSp); the number of parents and children aboard the Titanic (ParCh); and the fare paid for the ticket. In total there are therefore five categorical variables and a further two continuous variables.
We apply the deterministic procedure described in the “Methodology” section to anonymise the data and then apply a logistic regression model to predict the survival of the Titanic passengers in the given sample.
Anonymisation
In a realistic risk assessment scenario, the data custodian should first pseudonymise the data by replacing the passengers names (which is the only direct identifier) with nonidentifiable pseudo codes. For simplicity, we assume that the generated pseudo codes are the contents of the PassengerId variable already included in the data. The features that are left for the analysis are the variables Sex, Age, Pclass, SibSp, ParCh, Fare and the outcome variable Survived. From those variables, Age includes 177 missing values which are filled by the median age of the rest 714 passengers. We then convert the two discrete numerical variables—SibSp and ParCh—to a binary indicator (Family) denoting whether each passenger has any family members aboard the ship or not. Excluding any names and/or passenger identifiers, we now have four categorical variables (Survived, Class, Sex and Family) and two numeric variables (Age and Fare).
After preparing the data we follow the steps of the anonymisation process. Given that there are only two continuous variables, and it is only the continuous variables that go through the anonymisation procedure, the five steps of the anonymisation procedure as applied in the Titanic passengers data are visualised in Fig. 1. In particular, step (2) of the procedure checks if kanonymity holds for the categorical variables for some k. Table 2 shows the 12 distinct strata that can be formed by all combinations of categories of the variables Pclass, Sex and Family. From the Frequency column we can see that the minimum stratum has 32 observations which means that 32anonymity is applied. We will use k = 3 for the rest of this example, and given that 32anonymity applies then trivially the data satisfy the requirement of being 3anonymous.
Table 2 also presents the survival ratio that is the number of survived passengers over the total number of passengers in each stratum. By assuming that the Survived indicator is a sensitive attribute of the Titanic passengers we observe that there is significant uncertainty on disclosing this information for any particular individual (i.e., in no strata did all passengers survive or all died). This makes any attempt at inferential disclosure harder. For example, even if we know that a female was travelling in the first class of the Titanic with no family members, then we can only infer that the female died with probability 1/34.
With confirmation that the categorical variables in the dataset are 3anonymous after stratification, we apply the anonymisation procedure to mask the real values of the continuous variables Age and Fare.
Disclosure risk and utility loss
Following the anonymisation procedure we calculate the utility loss and disclosure risk of the anonymised continuous data. The disclosure risks as measured by the robust Mahalanobis distancebased approach, with \(w_1\) = 0.01 and \(w_2\) = 0.05, are risk1 = 0.0426 and risk2 = 0.009, which are indicators of the percentage of sensitive observations. In particular, the risk1 measure indicates that around 4% of the observations are ‘risky’—38 observations in total—in that they fall into the defined intervals around their corresponding original value. The risk2 metric indicates that only 0.9% of the observations—8 observations in total—are actually ‘unsafe’ in that they fall into the defined interval around their original value and no other observations are in close proximity. Note that here we use the default values set in dRiskRMD function for the parameters \(w_1\) and \(w_2\) and we further investigate the impact of their variation in the next section.
We can view these observations to better understand the nature of the potential disclosure, as shown in Fig. 2. It is observed that the points that are considered risky are points within the mass of the data points while extreme points (e.g., Fare>300) are unexpectedly considered ‘safe’. This happens because points in the center of the mass of the dataset tend to be replaced by centroids with high proximity to them due to the deteministic anonymisation, while extreme points such as outliers tend to be replaced by more distant centroids.
The global utility loss is U = 0.000117—close to zero—indicating some, but not much, loss of information. Looking individually at the propensity scores for each observation, most are very close to 0.5 (see Fig. 3).
The variablespecific information loss is \(\delta _{Age}\) = 0.0114 and \(\delta _{Fare}\) = 0.0473 which indicate that around 1.1% and 4.7% of the variability of the anonymised variables Age and Fare respectively is due to the anonymisation.
We then fit a logistic regression model with the variable Survived as the outcome and Pclass, Sex, Age, Fare and Family as covariates. We do this twice: here we have the luxury of being able to apply it to the original data as well as the anonymised data to compare the similarity of the estimates. The estimated coefficients of the model are shown in Table 3. We observe that the estimated coefficients obtained by the regression model applied to anonymised data are very close to those estimated using the original data. The standardised differences between the coefficients of the model applied to original and anonymised data are also given in Table 3. The confidence intervals of the regression coefficients from the original and anonymised data are also overlapping; see Fig. 4.
The 1958 National Child Development Study
Our second application is to a sample dataset of participants in the National Child Development Study (NCDS), also known as the 1958 Birth Cohort. The NCDS follows the lives of 17,415 people born in England, Scotland and Wales in a single week in 1958 [25]. The data consist of a set of risk factors and phenotypic variables from the 1958 Birth Cohort harmonised for the BioSHaREEU Healthy Obese Project, designed to examine the consequences of healthy obesity across several European biobanks and largescale cohort studies [26]. The sample dataset includes 99 variables for 1469 individuals, and a unique ID. This sample dataset was collected when the birth cohort’s participants were 45 years old.
For the purpose of this paper, we select four continuous and two categorical variables to anonymise and analyse. The continuous variables are the fasting glucose level; the highdensity lipoprotein (HDL) cholesterol; and the height and the weight of each participant. The categorical variables are gender (0 = Male and 1 = Female) and a binary variable indicating current cigarette smoking status (0 = Not a current cigarette smoker and 1 = Current cigarette smoker). The dataset has missing values and to avoid imputation this time we select only the complete cases which gives us data on 1211 individuals.
Anonymisation
We follow the five steps of anonymisation for the four continuous variables. For step (2)—stratification—we note that the two binary categorical variables form four possible strata: (a) 492 males are not smokers (\(\sim\)40.6% of the data), (b) 111 males are smokers (\(\sim\)9.2% of the data), (c) 468 females are not smokers (\(\sim\)38.6% of the data) and (d) 140 females are smokers (\(\sim\)11.6% of the data). The minimum stratum has 111 records, confirming that the data are 3anonymous. We follow the steps of the anonymisation procedure to obfuscate the four continuous variables, using k = 3.
Disclosure risk and utility loss
Once the (continuous) data have been anonymised, we calculate the information loss and disclosure risk metrics. The disclosure risks as estimated by the robust Mahalanobis distance metric with \(w_1\) = 0.01 and \(w_2\) = 0.05, are risk1 = 0.0140 and risk2 = 0.0099. This implies that around 17 of the 1211 observations (1.4%) are risky and 12 observations (0.99%) are potentially unsafe. We have also conducted a sensitivity analysis for the parameters \(w_1\) and \(w_2\). Figure 5A illustrates that when \(w_1\) increases the risk1 metric also increases as the intervals around the original observations become bigger and there is higher frequency of anonymised data points falling into those intervals. The risk1 metric is independent of the \(w_2\) parameter. Figure 5B shows that when \(w_2\) increases, risk2 decreases because more risky observations fall within bigger intervals thus obscuring the risky data points.
The global information loss of the anonymised data as measured by the summary of propensity scores is U = 0.00005: a low score suggesting that there is little in the way of overall loss of utility in anonymising the data. The variablespecific utility losses are \(\delta _{Glucose}\) = 0.0466, \(\delta _{HDL}\) = 0.0229, \(\delta _{Height}\) = 0.0296 and \(\delta _{Weight}\) = 0.0247 indicating that less than 5% of the variability of any of the anonymised variables is due to the anonymisation process itself.
We then apply a linear regression model to predict the fasting glucose level, with Sex, Smoker, HDL, Height and Weight as covariates. The estimated parameters of the model fitted to the original and the anonymised data are shown in Table 4. The estimated coefficients using the anonymised data are very close to the estimated coefficients using the original data. This yields small standardised differences between the coefficients of the model applied to original and anonymised data, which are also shown in Table 4. The similarity is perhaps easier to see in Fig. 6 in the overlapping of the confidence intervals.
Sensitivity analysis
We perform a sensitivity analysis to examine the effect of the choice of k (which dictates the number of nearest neighbours we seek for each observation) and the effect of the sample size on the disclosure risk and information loss of the nonstochastic anonymised data. We base the generation of synthetic data on 1958 Birth Cohort data in the previous section, using the observed distributions of the variables from the complete cases.
Choice of k
We set the size of the sample to 500 records. The four possible strata are created with the same proportions as seen in the real data, yielding 203 non smoking males (40.6% of 500), a stratum with 46 male smokers (9.2% of 500), a stratum with 193 non smoking females (38.6% of 500) and a stratum with 58 female smokers (11.6% of 500). For each stratum we simulate the four continuous variables. To simulate these nonnormal data we use the Vale and Maurelli method [37]. The simulation of the non normal multivariate distribution is based on the observed mean, skewness and kurtosis of each continuous variable in the original data, as well as all bivariate covariances.
For each simulated dataset we apply the deterministic anonymisation and calculate the disclosure risks and utility loss metrics. We apply the analysis for a k of 3, 5, 7, 10, 15, 20, 25, 30, 35, 40, and 46. The value of 46 is the maximum value that we can use as this is the number of records in the smallest stratum of the simulated data. In that case all data belonging to that stratum are replaced by the same centroid (and we would expect substantial loss of information).
For each value of k we generate 100 datasets, which will differ in their values of the continuous variables. The results of the mean of the metrics considered across all 100 datasets for each k are shown in Fig. 7. Figure 7A shows the increase in the global utility loss as measured by the propensity scores based metric as k increases; Fig. 7B shows the increase of the standardised Euclidean distances between original and anonymised data points as k increases; Fig. 7C presents the increase of the standardised difference between estimated regression coefficients with the original and anonymised data, as k increases; and Fig. 7D shows the reduction in the disclosure risk as k increases. All three measures of utility show information loss as we increase k, which is to be expected. Unsurprisingly they also show the typical tradeoff between disclosure risk and utility loss: lower values of k are preferable in terms of utility but this increases the risk of disclosure.
Sample size
Our second sensitivity analysis looks at the effect of sample size on the disclosure risk and information loss when using anonymised data. This analysis also accounts for the impact of the size of the minimum stratum: a smaller sample size will yield smaller minimum stratum.
For this analysis we fix the value of k to 5 and simulate sample sizes of 50, 100, 200, 500, 1000, 1500, 2000, 2500, and 3000. We follow the same data generating mechanism as described in the previous section, but this time with fixed k and varying sample sizes. With just 50 observations, the four strata have 20, 5, 19 and 6 records respectively. Setting k = 5 is therefore permissible, though we should expect major loss of information in this case especially for the two smallest strata.
For each sample size we generate 100 datasets (these will differ in terms of their ‘original’ continuous variables). The results from anonymising these 100 datasets, for each sample size considered, are shown in Fig. 8. Unsurprisingly all three metrics for utility improve as we increase sample size while the disclosure risk also increases as we increase sample size. The reason for this lies with the algorithm for anonymisation: as the sample size increases, so do the number of observations lying close to some other observation. So, for fixed k, the k nearest neighbours will potentially be closer compared to the situation where the dataset is much smaller. This results in using centroids that are closer to the original observation, which subsequently increases the risk. However, this needs to be tradedoff with the idea that if an observation has ‘neighbours’ that are relatively close, then the risk of actual disclosure is reduced.
Discussion
The evolution of data science exploits the availability of large, granular datasets with detailed information, and enhances the advancement of technologies, methods and applications that are used to translate such data into tools that can benefit society. However, with great potential in using data comes major challenges. One of those challenges is the restricted access to personal data especially in disciplines where data include sensitive information (from healthcare to banking).
The demand for access to such personal data has fuelled the exploration and development of several tools and approaches that allow data to be shared and analysed within secure frameworks that preserve privacy and maintain data subjects’ confidentiality. Deciding on the framework for mitigating disclosure risk of personal data is the responsibility of data custodians, who must make decisions based on a careful assessment of real risks that are typically context dependent. Suggested solutions have different strengths and weaknesses, and the final choice of the optimal method should depend on a rigorous assessment of the real risk of disclosure and the gravity of it occurring, versus the magnitude of the drawbacks of applying privacy protection including formal evaluation of potential information loss that must be tailored to the specific context that applies. In many cases simple data obfuscation through suppression and aggregation may be enough to mitigate the risk of disclosure while in other contexts more complex techniques might be required.
In this paper we propose a new deterministic approach for anonymising personal data that increases the uncertainty of data reidentification while limiting the loss of utility. This strategy therefore allows informative statistical analysis, but relies on having at least one continuous or numeric variable in the dataset. The method stratifies an underlying dataset by all possible combinations of the levels of categorical variables. The values of continuous variables are then replaced by taking into consideration the observation’s \((k1)\) nearest neighbours. This redistribution of the continuous variables in a dataset preserves their spatial correlations.
Our demonstration of the algorithm in “Applications” section shows that the procedure preserves the structure of the data well in cases where the required analysis is a regression model. The sensitivity analysis revealed the risk of disclosure and the utility loss are both dependent on sample size—which also tells us indirectly that it is dependent on the size of the strata—and also on our choice of k.
The final choice of the parameters in kanonymity and knearest neighbours depends on a rigorous assessment of the real risks of disclosure and the magnitude of the information loss generated by applying the anonymisation process. In general, this association is highly context dependent, and therefore the selection of the parameters must be specified based on each specific data situation. In many cases data protection through small values of the parameters in kanonymity and knearest neighbours may be enough to mitigate the risk of disclosure and preserve data utility, while in other contexts bigger values may be preferable [24].
This procedure has several advantages over other methods of anonymisation. The most important is that we introduce three layers of protection. The first layer occurs due to stratification: in cases where the strata themselves do not give rise to welldefined clusters of observations (which is often the case for real data), the search for the nearest neighbours within a stratum obfuscates the actual proximity of the data. That is, the nearest neighbours of a data point within a stratum might be different to its nearest neighbours if we consider the entire dataset. The nearest neighbours within a stratum will be at least as far as an observation’s nearest neighbours globally, so this increases the disturbance of each data point. This effect is greater for ‘outliers’ than for data points within a mass of observations. For example, if the \(k1\) nearest neighbours of an unusual observation (that is distant from the majority of the other observations in the dataset) are located within the cloud of the mass of the data, then the corresponding centroid is located closer to the data mass in contrast to the original outlier point. This does not mean that the anonymised data lie exclusively within the convex hull of the original data: the scaling that occurs in step (5) of the algorithm ensures that it is possible to obtain anonymised data that are outside this region.
The second layer of protection is the one to one replacement of each data point with the centroid of the point itself and its \((k1)\) nearest neighbours. Note that any value of \(k>2\) makes the calculation of the exact original values almost impossible (with k = 2, a malicious ‘attacker’ could potentially disclose the exact value of an observation but only if they knew the centroid replacing a data point, and the mean and the variance of the three original observations that created the centroid). The optimal value for k is contextspecific to minimise loss of utility as much as possible while ensuring a tolerable level of risk disclosure. A discussion on the optimal value of k could be linked to a wider review on the threshold rules used in other privacy protection techniques; for example, a similar disclosure control applied to a 2dimensional contingency table where if at least one cell has fewer counts than an agreed threshold, the information in the rest of the table is not released. Simple threshold rules like these for the proposed algorithm in this paper would help to ensure that the anonymised dataset contained no obviously disclosive information. Of course, such thresholds could never universally guarantee safety.
The third layer of protection is a result of the scaling of the centroids introduced in step (5) that introduces extra uncertainty to the anonymised data. The scaling is applied to the entire dataset and not within each stratum separately and shifts the centroid back to the observed variance of each variable. The direction in which the scaling factor shifts each coordinate of a centroid depends on its position in relation to the origin, \({\mathbf {0}}=(0,\ldots ,0)\). For example, in a 2Dspace the xcoordinate of a centroid is multiplied by the scaling factor of variable x and the ycoordinate of the centroid is multiplied by the scaling factor of the variable y. Therefore, the direction of the shift of each centroid due to scaling depends on the quadrant within which the centroid exists around the origin, (0, 0).
Another advantage of the proposed method is that its nonstochastic behaviour ensures that it generates the same anonymised data for any given underlying dataset and a fixed value of k. This gives data managers full control of data relocation at each step of the anonymisation procedure (in contrast to stochastic perturbations) and therefore based on their risk assessments to choose the optimal value of k for their specific situation. It also prevents disclosure attacks based on generating multiple datasets and then using the law of large numbers to obtain the underlying values of the data before perturbation.
No anonymisation procedure provides a perfect solution, and some limitations with the proposed procedure are evident. The first is how we deal with missing data (note that in other methods, e.g. purely synthetic data, this is not a problem). The decision is related to the sample size of the data and the information loss if we reduce the dataset to its complete case format. If the decision is to keep the sample size unchanged then we will handle missing values in continuous variables differently to those missing values in categorical variables. For missing values in continuous variables we can either use sophisticated imputation methods or we can simply replace the empty cells with the median (or the mean) of the nonmissing values within each column separately and independently from the others. For the categorical variables we can either impute missing values using a binomial distribution fitted to the nonmissing values or we can consider missing values as another level of the categories of each variable.
Aside from issues with missing data, note that kanonymity should be applied across all strata and in some cases this is an advantage: it adds an extra level of protection to the categorical variables. However, in big datasets with many categorical variables, or in smaller datasets where categorical variables have many levels, achieving kanonymity might be challenging. Ways around these issues include combining categories to reduce the number of strata, or in extreme cases removing one or more categorical variable may be necessary. Another limitation is the potential instability of the kNN classification in highdimensional data [38,39,40,41]. This instability can be reflected in both the performance of the algorithm and on the level of the loss of utility on the anonymised data. In such cases, the data custodian might decide to use the algorithm in a subset of variables (like a pseudonymisation process), especially on those associated with higher disclosure risk (e.g. direct and indirect identifiers).
Conclusion
In this paper, we propose a deterministic algorithm for data anonymisation, as a possible solution to eliminate some of the barriers to data access, and in doing so ensure that the scientific principles of transparency and reproducibility are maintained beyond the immediate research domain. The deterministic approach can be used across different domains where data are viewed as sensitive either because of concerns relating to information governance and/or data ethics, or because of their value in terms of intellectual property. The proposed approach would be of interest to a wide range of data stakeholders: not just data custodians, but also data analysts, study participants and the general public who all have an interest in ensuring that data are used as widely as possible in ways that are of value to society and yet protect their confidentiality. Unlike many other privacy preserving methods, the procedure of deterministic anonymisation still allows researchers access to informative data that retain the characteristics of the original data so that there is no restriction on how the data can be analysed. Therefore, this approach has the potential to enhance the discoverability and utility of individuallevel data across all disciplines where data are sensitive. The use of such approach is essential to support the collection and appropriate use of personal data, to maintain the trust of participants joining research studies, and to enhance the compliance with the increasingly robust frameworks. These underpin the governance and ethicolegal stipulations and regulations oversight management, process and use of sensitive personal data.
Availability of data and materials
This work made use of data and samples generated by the 1958 Birth Cohort (NCDS), which is managed by the Centre for Longitudinal Studies at the UCL Institute of Education, funded by the Economic and Social Research Council (Grant No. ES/M001660/1). Access to these resources was enabled via the Wellcome Trust & MRC: 58FORWARDS Grant [108439/Z/15/Z] (The 1958 Birth Cohort: Fostering new Opportunities for Research via Wider Access to Reliable Data and Samples). Before 2015 biomedical resources were maintained under the Wellcome Trust and Medical Research Council 58READIE Project (Grant Nos. WT095219MA and G1001799). In addition, this work made use of a sample of Titanic passengers data which is publicly available on the webpage https://www.kaggle.com/c/titanic/data.
References
Laine C, Goodman SN, Griswold ME, Sox HC. Reproducible research: moving toward research the public can really trust. Ann Intern Med. 2007;146(6):450–3. https://doi.org/10.7326/00034819146620070320000154.
Kitchin R. The data revolution: big data, open data data infrastructures & their consequences. Thousand Oaks: SAGE Publications Ltd; 2014. https://doi.org/10.4135/9781473909472.
Taichman DB, Backus J, Baethge C, Bauchner H, de Leeuw PW, Drazen JM, Fletcher J, Frizelle FA, Groves T, Haileamlak A, James A, Laine C, Peiperl L, Pinborg A, Sahni P, Wu S. Sharing clinical trial data—a proposal from the international committee of medical journal editors. N Engl J Med. 2016;374(4):384–6. https://doi.org/10.1056/NEJMe1515172.
Wu Z, Shen S, Zhou H, Li H, Lu C, Zou D. An effective approach for the protection of user commodity viewing privacy in ecommerce website. KnowlBased Syst. 2021;220:106952. https://doi.org/10.1016/j.knosys.2021.106952.
Wu Z, Li G, Shen S, Lian X, Chen E, Xu G. Constructing dummy query sequences to protect location privacy and query privacy in locationbased services. World Wide Web. 2020. https://doi.org/10.1007/s1128002000830x.
Wu Z, Wang R, Li Q, Lian X, Xu G, Chen E, Liu X. A location privacypreserving system based on query range coverup or locationbased services. IEEE Trans Veh Technol. 2020;69(5):5244–54. https://doi.org/10.1109/TVT.2020.2981633.
Wu Z, Shen S, Lian X, Su X, Chen E. A dummybased user privacy protection approach for text information retrieval. KnowlBased Syst. 2020;195:105679. https://doi.org/10.1016/j.knosys.2020.105679.
Wu Z, Xu G, Lu C, Chen E, Jiang F, Li G. An effective approach for the protection of privacy text data in the CloudDB. World Wide Web. 2017;21:915–38. https://doi.org/10.1007/s1128001704918.
Burton PR, Murtagh MJ, Boyd A, Williams JB, Dove ES, Wallace SE, Tassé AM, Little J, Chisholm RL, Gaye A, Hveem K, Brookes AJ, Goodwin P, Fistein J, Bobrow M, Knoppers BM. Data safe havens in health research and healthcare. Bioinformatics. 2015;31(20):3241–8. https://doi.org/10.1093/bioinformatics/btv279.
Lea NC, Nicholls J, Dobbs C, Sethi N, Cunningham J, Ainsworth J, Heaven M, Peacock T, Peacock A, Jones K, Laurie G, Kalra D. Data safe havens and trust: toward a common understanding of trusted research platforms for governing secure and ethical health research. JMIR Med Inform. 2016;4(2):22. https://doi.org/10.2196/medinform.5571.
Gaye A, Marcon Y, Isaeva J, LaFlamme P, Turner A, Jones EM, Minion J, Boyd AW, Newby CJ, Nuotio ML, Wilson R, Butters O, Murtagh B, Demir I, Doiron D, Giepmans L, Wallace SE, BudinLjøsne I, Schmidt CO, Boffetta P, Boniol M, Bota M, Carter KW, deKlerk N, Dibben C, Francis RW, Hiekkalinna T, Hveem K, Kvaløy K, Millar S, Perry IJ, Peters A, Phillips CM, Popham F, Raab G, Reischl E, Sheehan N, Waldenberger M, Perola M, van den Heuvel E, Macleod J, Knoppers BM, Stolk RP, Fortier I, Harris JR, Woffenbuttel BH, Murtagh MJ, Ferretti V, Burton PR. DataSHIELD: taking the analysis to the data, not the data to the analysis. Int J Epidemiol. 2014;43(6):1929–44. https://doi.org/10.1093/ije/dyu188.
Wilson R, Butters O, Avraam D, Baker J, Tedds J, Turner A, Murtagh M, Burton P. DataSHIELD—new directions and dimensions. Data Sci J. 2017;16:21. https://doi.org/10.5334/dsj2017021.
Marcon Y, Bishop T, Avraam D, EscribaMontagut X, RyserWelch P, Wheater S, Burton P, González JR. Orchestrating privacyprotected big data analyses of data from different resources with R and DataSHIELD. PLoS Comput Biol. 2021;17(3):1–17. https://doi.org/10.1371/journal.pcbi.1008880.
Carter KW, Francis RW, Carter KW, Francis RW, Bresnahan M, Gissler M, Grønborg TK, Gross R, Gunnes N, Hammond G, Hornig M, Hultman CM, Huttunen J, Langridge A, Leonard H, Newman S, Parner ET, Petersson G, Reichenberg A, Sandin S, Schendel DE, Schalkwyk L, Sourander A, Steadman C, Stoltenberg C, Suominen A, Surén P, Susser E, Sylvester Vethanayagam A, Yusof Z. International collaboration for autism registry epidemiology: ViPAR: a software platform for the virtual pooling and analysis of research data. Int J Epidemiol. 2016;45(2):408–16. https://doi.org/10.1093/ije/dyv193.
Rubin DB. Discussion statistical disclosure limitation. J Off Stat. 1993;9(2):461–8.
Avraam D, Wilson R, Burton P. Synthetic ALSPAC longitudinal datasets for the Big Data VR project. Wellcome Open Res. 2017. https://doi.org/10.12688/wellcomeopenres.12441.1.
Elliot M, Mackey E, O’Hara K, Tudor C. The anonymisation decisionmaking framework. Manchester: UKAN; 2016.
Sweeney L. Kanonymity: a model for protecting privacy. Int J Uncertain Fuzziness Knowlbased Syst. 2002;10(5):557–70. https://doi.org/10.1142/S0218488502001648.
Angiuli, O., Waldo, J.: Statistical tradeoffs between generalization and suppression in the deidentification of largescale data sets. In: 2016 IEEE 40th annual computer software and applications conference (COMPSAC), vol. 2; 2016. pp. 589–93. https://doi.org/10.1109/COMPSAC.2016.198.
Aggarwal, C.C.: On kanonymity and the curse of dimensionality. In: Proceedings of the 31st International Conference on Very Large Data Bases. VLDB ’05. VLDB Endowment; 2005. pp. 901–9.
Goldstein H, Shlomo N. A probabilistic procedure for anonymisation, for assessing the risk of reidentification and for the analysis of perturbed data sets. J Off Stat. 2020;36:89115. https://doi.org/10.2478/jos20200005.
Avraam D, Boyd A, Goldstein H, Burton P. A software package for the application of probabilistic anonymisation to sensitive individuallevel data: a proof of principle with an example from the ALSPAC birth cohort study. Longitud Life Course Stud. 2018;9(4):433–46. https://doi.org/10.14301/llcs.v9i4.478.
Goldstein H, Browne WJ, Charlton C. A Bayesian model for measurement and misclassification errors alongside missing data, with an application to higher education participation in australia. J Appl Stat. 2018;45(5):918–31. https://doi.org/10.1080/02664763.2017.1322558.
Avraam D, Wilson R, Butters O, Burton T, Nicolaides C, Jones E, Boyd A, Burton P. Privacy preserving data visualizations. EPJ Data Sci. 2021;10. https://doi.org/10.1140/epjds/s13688020002574.
Power C, Elliott J. Cohort profile: 1958 British birth cohort (National Child Development Study). Int J Epidemiol. 2005;35(1):34–41. https://doi.org/10.1093/ije/dyi183.
van VlietOstaptchouk JV, Nuotio ML, Slagter SN, Doiron D, Fischer K, Foco L, Gaye A, Gögele M, Heier M, Hiekkalinna T, Joensuu A, Newby C, Pang C, Partinen E, Reischl E, Schwienbacher C, Tammesoo ML, Swertz MA, Burton P, Ferretti V, Fortier I, Giepmans L, Harris JR, Hillege HL, Holmen J, Jula A, KootstraRos JE, Kvaløy K, Holmen TL, Männistö S, Metspalu A, Midthjell K, Murtagh MJ, Peters A, Pramstaller PP, Saaristo T, Salomaa V, Stolk RP, Uusitupa M, van der Harst P, van der Klauw MM, Waldenberger M, Perola M, Wolffenbuttel BH. The prevalence of metabolic syndrome and metabolically healthy obesity in Europe: a collaborative analysis of ten large cohort studies. BMC Endocr Disord. 2014. https://doi.org/10.1186/14726823149.
Nandy SC, Das S, Goswami PP. An efficient k nearest neighbors searching algorithm for a query line. Theoret Comput Sci. 2003;299(1):273–88. https://doi.org/10.1016/S03043975(02)003225.
Mayer BW, Rangwala HS, Gupta R, Srivastava J, Karypis G, Kumar V, de Groen PC. Feature mining for prediction of degree of liver fibrosis. AMIA Ann Symp Proc. 2005;2005:1048.
Singh, M.D., Krishna, P.R., Saxena, A.: A cryptography based privacy preserving solution to mine cloud data. In: Proceedings of the third annual ACM Bangalore conference. COMPUTE ’10. ACM, New York; 2010. pp. 14–1144. https://doi.org/10.1145/1754288.1754302.
Wu X, Kumar V, Ross Quinlan J, Ghosh J, Yang Q, Motoda H, McLachlan GJ, Ng A, Liu B, Yu PS, Zhou ZH, Steinbach M, Hand DJ, Steinberg D. Top 10 algorithms in data mining. Knowl Inf Syst. 2008;14(1):1–37. https://doi.org/10.1007/s1011500701142.
Amirbekyan, A., EstivillCastro, V.: Privacypreserving kNN for small and large data sets. In: Seventh IEEE international conference on data mining workshops (ICDMW 2007); 2007. pp. 699–704. https://doi.org/10.1109/ICDMW.2007.67.
Woo MJ, Reiter JP, Oganian A, Karr AF. Global measures of data utility for microdata masked for disclosure limitation. J Priv Confid. 2009;1(1):111–24. https://doi.org/10.29012/jpc.v1i1.568.
Snoke J, Raab GM, Nowok B, Dibben C, Slavkovic A. General and specific utility measures for synthetic data. J R Stat Soc A Stat Soc. 2018;181(3):663–88. https://doi.org/10.1111/rssa.12358.
Karr AF, Kohnen CN, Oganian A, Reiter JP, Sanil AP. A framework for evaluating the utility of data altered to protect confidentiality. Am Stat. 2006;60(3):224–32. https://doi.org/10.1198/000313006X124640.
Templ, M., Meindl, B.: Robust statistics meets SDC: new disclosure risk measures for continuous microdata masking. In: Privacy in statistical databases. Springer; 2008. pp. 177–89.
Templ M, Kowarik A, Meindl B. Statistical disclosure control for microdata using the R Package sdcMicro. J Stat Softw. 2015. https://doi.org/10.18637/jss.v067.i04.
Vale CD, Maurelli VA. Simulating multivariate nonnormal distributions. Psychometrika. 1983;48(3):465–71. https://doi.org/10.1007/bf02293687.
Pestov V. Is the kNN classifier in high dimensions affected by the curse of dimensionality? Comput Math Appl. 2013;65(10):1427–37. https://doi.org/10.1016/j.camwa.2012.09.011.
Indyk P. Nearest neighbors in highdimensional spaces. In: Goodman JE, O’Rourke J, editors. Handbook of discrete and computational geometry. 2nd ed. Boca Raton: CRC; 2004. p. 877–92. https://doi.org/10.1201/9781420035315.ch39.
Kushilevitz, E., Ostrovsky, R., Rabani, Y.: Efficient search for approximate nearest neighbor in high dimensional spaces. In: Proceedings of the thirtieth annual ACM symposium on theory of computing. STOC ’98. Association for Computing Machinery, New York; 1998. pp. 614–623. https://doi.org/10.1145/276698.276877.
Cérou F, Guyader A. Nearest neighbor classification in infinite dimension. ESAIM Probab Stat. 2006;10:340–55. https://doi.org/10.1051/ps:2006014.
Acknowledgements
We would like to acknowledge all the cohort members of the 1958 National Child Development Study and all the people who have supported the study over the course of its life.
Funding
This work was supported by a group of projects that underpin a program of development and application of secure methods for coanalysis, data sharing and visualisation in the Population Health Sciences Institute at Newcastle University (UK). These include: the Connected Health Cities project (North East and North Cumbria) funded by the UK Department of Health (RES/0150/7943/202); the EUCanConnect project funded by the European Commission H2020 Flagship Collaboration with Canada (No 824989); the ATHLETE project (Advancing Tools for Human Early Lifecourse Exposome Research and Translation) funded by the European Union’s Horizon 2020 research and innovation programme (No 874583); the 58FORWARDS project (Fostering new Opportunities for Researchers via Wider Access to Research Data and Samples) funded jointly by the Wellcome Trust and the Medical Research Council (108439/Z/15/Z); and the METADAC project (Managing Ethicosocial, Technical and Administrative issues in Data ACcess) funded jointly by the Medical Research Council, the Wellcome Trust and the Economic and Social Research Council (MR/N01104X/1 and MR/N01104X/2). The funders had no role in the design and implementation of the study, and no role in the decision to publish and preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
DA designed the methodology, developed the algorithms, generated the illustrations and the visualisations and wrote the manuscript with substantive input and support from EJ and PB. All authors contributed to the design of the research and to editing the original and the final manuscripts. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Data governance for the 1958 Birth Cohort data was provided by the METADAC data access committee, funded by ESRC, Wellcome, and MRC (20152018: Grant No. MR/N01104X/1, 20182020: Grant No. ES/S008349/1).
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Avraam, D., Jones, E. & Burton, P. A deterministic approach for protecting privacy in sensitive personal data. BMC Med Inform Decis Mak 22, 24 (2022). https://doi.org/10.1186/s12911022017544
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12911022017544