 Proceedings
 Open Access
 Published:
Secure distributed genome analysis for GWAS and sequence comparison computation
BMC Medical Informatics and Decision Making volume 15, Article number: S4 (2015)
Abstract
Background
The rapid increase in the availability and volume of genomic data makes significant advances in biomedical research possible, but sharing of genomic data poses challenges due to the highly sensitive nature of such data. To address the challenges, a competition for secure distributed processing of genomic data was organized by the iDASH research center.
Methods
In this work we propose techniques for securing computation with reallife genomic data for minor allele frequency and chisquared statistics computation, as well as distance computation between two genomic sequences, as specified by the iDASH competition tasks. We put forward novel optimizations, including a generalization of a version of mergesort, which might be of independent interest.
Results
We provide implementation results of our techniques based on secret sharing that demonstrate practicality of the suggested protocols and also report on performance improvements due to our optimization techniques.
Conclusions
This work describes our techniques, findings, and experimental results developed and obtained as part of iDASH 2015 research competition to secure reallife genomic computations and shows feasibility of securely computing with genomic data in practice.
Background
Introduction
The iDASH (Integrating Data for Analysis, Anonymization and SHaring) research center at the University of California, San Diego hosts an annual competition, which in 2015 was dedicated to secure genome analysis. The two challenges corresponded to secure noninteractive analysis of genomic data based on homomorphic encryption and secure interactive analysis using secure multiparty computation techniques. We focus on the second challenge and report our design and implementation of the competitions tasks, which consisted of distributed GWAS (GenomeWide Association Study) computation and secure sequence comparisons in the form of the Hamming distance or edit distance. Here, GWAS is a study of common genetic variants in different individuals using case and control groups to determine if any variant is associated with a specific trait or genetic conditions.
We utilize secure multiparty computation (SMC) techniques based on secret sharing with lightweight computational footprints. This requires that all computation carried out jointly by the parties (i.e., computation that cannot be performed locally by data owners) is dataoblivious, which means that all instructions and accessed memory locations must be independent of the data. While this does not pose a challenge for some simpler computational tasks, meeting this objective often involves using nontrivial techniques for more complex functionalities. In particular, computing both Hamming and edit distances of two genome sequences involves a form of aligning the input sequences which is not straightforward to achieve in secure setting. A logical tool to resort to is to utilize secure set intersection for computing chromosome positions that appear in both input sequences, for which both twoparty and multiparty implementations are known. Due to the specifics of our setting, we rely on the ideas from [1] for computing the set of positions common to both sequences, which in turn utilize oblivious sorting. The fastest oblivious sorting mechanism available to us at the time of competition preparation was sorting based on Batcher's mergesort [2], which works only on input sets, the size of which is a power of 2. This posed a problem because padding an input set of a large size to have the size equal to a power of 2 often can result in a significant performance slowdown which we wanted to overcome. Thus, the most challenging component of distance computation was generalizing the mergesort (more precisely, the merging step as described later) component of the computation to work with inputs of arbitrary sizes, which might be of independent interest. This and other optimizations and design decisions constitute the main contribution of this work. We also report on performance of our algorithms on real genome data.
Tasks of the challenge
The challenge for secure multiparty computation based genomic data analysis had two tasks:

1
The first task was to develop secure distributed protocols for GWAS computations. The input consists of the genotypes of two groups of individuals (one case and one control group) over a number of Single Nucleotide Polymorphisms (SNP) with each of them being a DNA sequence variation, where a single nucleotide (A, T, C, or G) in the genome differs between individuals. The input is horizontally partitioned among two sites (e.g., two institutions, medical facilities, etc.), where each site cannot reveal its input to other parties. The task consists of securely computing minor allele frequencies (MAF) and chisquared statistics for each of the SNPs in the case and control groups distributed across the two input parties. We provide the details of the computation below.

2
The second task was to develop secure distributed protocols for genomic sequence comparisons. The input consists of two genomic datasets, one from each individual, which are organized as the genotypes over many SNPs across the whole human genome. Each genomic dataset belongs to a different entity, and the data owner cannot reveal any information about its data to other parties. The task consists of securely computing either the Hamming distance or edit distance between the two genomic datasets, and we concentrate on Hamming distance computation. The computation involved in the Hamming distance computation of two genomic datasets differs from the traditional formulation of the Hamming distance and we describe the computation used in determining the distance later in this section.
Before we proceed with a detailed description of the tasks, we would like to note that the specification of the tasks, including all information about the participants' datasets that should be treated as public (such as the number records in one's dataset), was provided by the competition organizers. The goal was thus to provide a secure evaluation of the specified functionality using at least the semihonest security model (see below for detail), and the extent of the information about the other participant's data that can be deduced from the output is beyond the scope of this work.
We next describe the computation involved in the first task in more detail. The input comes as a list of SNPs, where for each SNP a number of genotypes corresponding to the individuals from the case and control groups are given. Let P denote the number of SNPs in the input and N_{ c } (N_{ t } ) denote the number of individuals in the case (resp., control) group, whose genomic data is provided for each SNP. For each SNP, a genotype corresponding to an individual consists of two nucleotides with three possible variations denoted by AA, AB, and BB, where A and B each represent a character from the set {A, T, C, G} and are alleles in our context.
For the purpose of MAF computation, there is no distinction between case and control groups, and all individuals in both groups are treated in the same way. We denote the total number of individuals by N = N_{ c } + N_{ t }. To determine MAF for a given SNP, one first needs to count the number of occurrences of alleles A and B as n_{ A } = 2n_{ AA } + n_{ AB } and n_{ B } = 2n_{ BB } + n_{ AB }, respectively, where n_{ AA }, n_{ AB }, and n_{ BB } denote the number of individuals with genotypes AA, AB, and BB for the given SNP. For an allele A or B, we compute its frequency as n_{ A }/2N or n_{ B }/2N , respectively, where 2 is the length of each genotype. To simplify notation, we also let N^{'} = 2N (and ${N}_{c}^{\prime}=2{N}_{c}$, ${N}_{t}^{\prime}=2{N}_{t}$). The smaller frequency corresponds to the minor allele and constitutes the output of MAF computation. We obtain the following:
Definition 1 Minor allele frequency (MAF) refers to the frequency at which the least common allele occurs in a given population and is computed as
We can simplify the computation by directly obtaining n_{ A } and n_{ B } after counting the number of times each of the two nucleotide values appears in the provided genotypes. In the case when the individuals are partitioned into two (case and control) groups, we will have n_{ A } = n_{ cA } + n_{ tA } and n_{ B } = n_{ cB } + n_{ tB } , where n_{ cA } and n_{ tA } represent the value of n_{ A } in the case and control groups, respectively, and similarly n_{ cB } and n_{ tB } represent the value of n_{ B } in the case and control groups. Furthermore, in our case the data are partitioned among two different entities and thus each of n_{ A } and n_{ B } need to be computed as the sum of the corresponding values at the respective sites. If we let superscripts (1) and (2) represent the values present in the genotypes of the individuals at sites 1 and 2, respectively, we now obtain and ${n}_{B}={n}_{cB}^{\left(1\right)}+{n}_{tB}^{\left(1\right)}+{n}_{cB}^{\left(2\right)}+{n}_{tB}^{\left(2\right)}$. Also, now N corresponds to the total number of individuals in the data at both sites and in both case and control groups. Using notation similar to the above, we let ${N}_{c}={N}_{c}^{\left(1\right)}+{N}_{c}^{\left(2\right)}$ and ${N}_{t}={N}_{t}^{\left(1\right)}+{N}_{t}^{\left(2\right)}$.
The chisquared test is also performed for each SNP independently, but now the data of the individuals in the case and control groups play different roles.
Definition 2 Chisquared (χ^{2}) test is a statistical test for comparing observed data with those expected according to a specific hypothesis and is represented as ${\chi}^{2}={\sum}_{i=1}^{m}\frac{{\left(ob{s}_{i}ex{p}_{i}\right)}^{2}}{ex{p}_{i}}$ for some settingdependent m.
In our case (for a single dataset), m = 4 and the observed values obs_{ i }'s correspond to the observed allele counts for a SNP, namely, n_{ cA }, n_{ cB }, n_{ tA }, n_{ tB }. The corresponding expected allele counts expi's are $\left({n}_{cA}+{n}_{tA}\right){N}_{c}^{\prime}\mathsf{\text{/}}{N}^{\prime}$, $\left({n}_{CB}+{n}_{tB}\right){N}_{c}^{\prime}\mathsf{\text{/}}{N}^{\prime}$, $\left({n}_{cA}+{n}_{tA}\right){N}_{c}^{\prime}/{N}^{\prime}$, $\left({n}_{cB}+{n}_{tB}\right){N}_{t}^{\prime}/{N}^{\prime}$. The resulting computation can be simplified to become
When the dataset is horizontally partitioned among two sites, the counts n_{ cA }, n_{ tA }, n_{ cB } , n_{ tB }, N_{ c }, N_{ t }, N become the sum of their respective values at both sites.
We can now proceed with the description of the computation involved in the second task, namely, the Hamming distance of two genomic datasets. In the traditional formulation of the Hamming distance, on input of two sequences of equal length, the distance is defined as the number of positions at which the corresponding symbols in the input sequences are different. This is not directly applicable to genomic sequences because they are not represented as perfectly aligned strings of the same length and thus the computation is more complex. Before we proceed with the details of the computation, we need to specify how the input (i.e., genomic datasets) are represented.
Genomic sequences are represented in the Variant Call Format (VCF), where each genomic sequence is a set of records. In each record, chromosome CHROM represents an identifier from the reference genome and position POS represents the reference position within the reference sequence CHROM. In other words, the pair ⟨CHROM, POS⟩ represents the location of the data associated with this record in the genome. The fields REF and ALT represent the reference and alternate bases, respectively, expressed as a sequence of one or more nucleotides. The field SVTYPE represents the type of the record, which is one of SUB, SNP, DEL, or INS. Only records of type SUB and SNP are used in the computation of the Hamming distance. In records of type SNP, both REF and ALT fields are one character long, while in records of type SUB, both fields can be longer. We also found that in two different inputs records at the same location ⟨CHROM, POS⟩ may be represented using different types (SUB and SNP).
Algorithm 1 HD(S_{1}, S_{2})
1: create an empty map M with keys as pair ⟨CHROM, POS⟩
2: dist = 0
3: for each record R in S_{1} do
4: if (R.SVTYPE = SUB or R.SVTYPE = SNP) then
5: M.put(⟨R.CHROM, R.POS⟩, R)
6: dist = dist + 1
7: end if
8: end for
9: for each record R in S_{2} do
10: if (R.SVTYPE = SUB or R.SVTYPE = SNP) then
11: if (M.containsKey(⟨R.CHROM, R.POS⟩) = false) then
12: dist = dist + 1
13: else
14: dist = dist − 1
15: if (M.get(⟨R.CHROM, R.POS⟩).REF = R.REF and M.get(⟨R.CHROM, R.POS⟩).ALT ≠ R.ALT) then
16: dist = dist + 1
17: end if
18: end if
19: end if
20: end for
21: return dist
To compute the Hamming distance between two genomic datasets, we initially set the distance to 0. Then for all records in the datasets with type SUB or SNP, if a location ⟨CHROM, POS⟩ is found only in one of the datasets (and is absent in the other), the Hamming distance is incremented by 1. Also, if the location is found in both datasets and the corresponding values of the REF fields are the same, but the values of the ALT fields are different, the Hamming distance is also incremented by 1. A more detailed specification of how this procedure may be implemented is given in Algorithm 1. The algorithm uses a map to store all records of type SUB and SNP from the first dataset and (optimistically) increments the distance by 1 for each record placed in the map (lines 37). Then for each record of type SUB or SNP from the second dataset, if there was no record with the same location ⟨CHROM, POS⟩ in the first dataset, the distance is incremented by 1 (lines 1012). If, however, the location is present in both datasets, the distance is first decremented by 1 (line 14). The algorithm then compares the fields REF and ALT of the records from the two datasets with the same location. If the former are equal and the latter differ, the distance is incremented by 1 (lines 1517).
Additional information about the tasks, including examples that illustrate the computation, can be found on the competition web site [3] as well as in an article [4] being prepared by the competition organizers.
Secure multiparty computation background
Secure multiparty computation allows two or more participants to jointly evaluate a function on their private inputs without revealing any information about the private data other than the output of the agreedupon function. There are two standard security models used for secure function evaluation on private data that differ with respect to the types of adversaries they can tolerate. The first security model known as semihonest (or honestbutcurious or passive) requires that all computation participants follow the computation as prescribed, but might save any information observed throughout the computation and compute with it with the goal to discover additional information about private input values. A protocol is said to be secure if no coalition of semihonest participants (adversaries) can learn any additional information about private inputs of other parties other than what they can already compute from their legitimate output. It also follows from the security properties that any outside party is unable to learn any information about the participants' data and protocol output corresponds to evaluating the correct function on the provided data. Security in presence of semihonest participants was a minimum security requirement for this competition.
The second, stronger, security model permits malicious (also known as active) participants who can arbitrarily deviate from the prescribed computation (and coordinate their actions). Security in this model holds if the same data protection and output correctness properties are achieved as before. Known techniques for achieving security in this model typically involve substantially larger overhead than in the semihonest model, and we do not use it in our implementation.
There are a variety of available techniques on which secure multiparty computation protocols can be based. For the purposes of this competition, we utilize an (n, t)threshold linear secret sharing scheme for representation of and secure computation over private values. With such a scheme, each private value is split into n secret shares (using n computational parties each of whom receives a share), such that combining t or fewer shares informationtheoretically reveals no information about the private value, but combining t + 1 or more shares allows the value to be reconstructed exactly. All computation proceeds on secret shares, which means that before the computation commences each participant distributes her private data among n computational parties and at the end of the computation reconstructs the result from the shares obtained from at least t+1 computation participants. We utilize Shamir's secret sharing scheme [5] and a typical way of conducting computation using this scheme requires that t < n/2. Thus we use n = 3 computational parties and set t = 1 (i.e., the parties are assumed not to collude).
In both tasks of the challenge, there are two parties who contribute their input. They will play the role of computational parties together with another party who contributes no input. As mentioned above, each input provider produces secret shares of her data and distributes them among the participants and reconstructs the output at the end of the computation. We assume that the three computational parties are connected by pairwise secure authenticated channels (that provide secrecy and authenticity) with each other, which can be achieved using standard means.
The underlying secret sharing scheme [5] requires that shares are represented as elements of a field, which means that the input needs to be provided in the form of integer values. With a linear secret sharing scheme, a linear combination of secretshared values can be performed by each computational party locally, without any interaction. Multiplication of two secretshared values, on the other hand, requires communication between all of them and is treated as an elementary building block of secure protocols (we assume the multiplication protocol from [6]). These operations are typically used as the main building blocks in more complex computations, e.g., comparisons and divisions.
We utilize a number of efficient protocols for integer computation that have previously been shown secure in the standard security model. It is also known that by invoking the composition theorem [7], secure building blocks can be combined together to achieve security of the overall computation. The building blocks that will be used in the computation of the two tasks are listed next (as mentioned before, all correspond to integer computation). When performance of a building block depends on the size of the arguments provided into the function, the size is listed as a separate argument.

[z] ← Mult([x], [y]) is a multiplication protocol that on input two secretshared values x and y outputs a secretshared product z = xy.

[b] ← EQ([x], [y], ℓ ) is an equality protocol that on input two secretshared values x and y of bitlength at most ℓ outputs a bit b which is set to 1 iff x = y.

[b] ← LT([x], [y], ℓ ) is a comparison protocol that on input two secretshared values x and y of bitlength at most ℓ outputs a bit b which is set to 1 iff x < y.

[z] ← Div([x], [y], f) is a division protocol that on input two secretshared values x and y of bitlength at most f outputs a secretshared quotient z that satisfies z = ⌊ x/y⌋.
As shown above, each protocol takes shares of its input and produces shares of the output. It means that these protocols can be naturally and securely invoked as part of larger computation and we use them as steps in larger computation. In our implementation, we use Mult from [6], EQ and LT from [8], and Div from [9], and we refer the reader for the details of these protocols to the respective publications.
Performance of secure computation protocols is of a paramount importance for their practical use. In the case of techniques based on secret sharing, the computation is normally lightweight and thus performance is measured in terms of two parameters: (i) the number of interactive operations (e.g., multiplications) necessary to perform the computation and (ii) the number of sequential interactions, i.e., rounds. Our goal is to minimize both of these parameters for the computation performed for each task.
Before we conclude this section, we would like to say that other options for securely evaluating the functions of the competition tasks are possible. In particular, the garbled circuit evaluation approach [10] allows any function to be securely evaluated in the twoparty setting. Similarly, any function can be evaluated using homomorphic encryption, or specialpurpose building blocks such as private set intersection (e.g., [11]) can be used as the basis for building a custom solution for a task of the competition. Furthermore, secure computation compilers such as Fairplay [12], Sharemind [13], PICCO [14], etc. are able to produce secure implementations given function specification in a form of a program. This competition, however, allowed for custom solutions that can tune general building blocks to the needs of the tasks and result in improved performance. Because no secure implementations of the competition tasks were available to us prior to the competition, we are unable to directly compare performance of different approaches in this paper.
Methods
Secure distributed GWAS computation
In this section we describe our approach to securely computing the task of distributed GWAS computation, namely, computing minor allele frequencies and chisquared statistics.
According to the task specification, the size of the input at each site, i.e., the number of SNPs and the number of individuals in the case and control groups, are treated as public and are not protected. This means that parameters P, ${N}_{c}^{\left(1\right)}$, ${N}_{c}^{\left(2\right)}$, ${N}_{t}^{\left(1\right)}$ and ${N}_{t}^{\left(2\right)}$ are known to all computation participants. All remaining data (i.e., the genotypes themselves) are private.
In what follows, we first describe a basic version of our solution and then provide optimization techniques that improve the runtime of program execution.
Basic solution
For each SNP in the input, the computation is identical (and independent of other SNPs) and thus it suffices to describe the computation for a single SNP.
We divide the overall computation into three phases: input preparation, computation execution, and output reconstruction, which proceed as follows. Observe that each input site i can locally compute ${n}_{cA}^{\left(i\right)}$, ${n}_{tA}^{\left(i\right)}$, ${n}_{cB}^{\left(i\right)}$, ${n}_{tB}^{\left(i\right)}$ for each SNP. This is what is done as part of input preparation, after which each input site secret shares each of its computed values and distributes the shares among all three computational parties. We use notation [a] to denote that the value of a is secretshared among the computational parties (i.e., each party holds a different share of a).
During computation execution, the computation proceeds on the shares to compute MAF and chisquared statistics using equations 1 and 2 and secure building blocks from the previous section. We choose to perform only the private portion of the computation on secret shares, while postponing the computation with public constants to the output reconstruction phase. This is done for performance reasons to reduce the size of values used in the computation.
To calculate the MAF for each SNP in parallel, the computation follows equation 1 with provisions to make the computation dataoblivious. That is, each computational party performs the following steps: In this section we describe our approach to securely
1 $\left[{n}_{A}\right]=\left[{n}_{cA}^{\left(1\right)}\right]+\left[{n}_{tA}^{\left(1\right)}\right]+\left[{n}_{cA}^{\left(2\right)}\right]+\left[{n}_{tA}^{\left(2\right)}\right];$
2 $\left[{n}_{B}\right]=\left[{n}_{cB}^{\left(1\right)}\right]+\left[{n}_{tB}^{\left(1\right)}\right]+\left[{n}_{cB}^{\left(2\right)}\right]+\left[{n}_{tB}^{\left(2\right)}\right];$
3 $\left[b\right]=\mathsf{\text{LT(}}\left[{n}_{A}\right],\left[{n}_{B}\right],{\ell}_{1}\mathsf{\text{);}}$
4 $\left[re{s}_{1}\right]=\mathsf{\text{Mult}}\left(\left[b\right],\left[{n}_{A}\right]\left[{n}_{B}\right]\right)+\left[{n}_{B}\right];$
The first two steps that aggregate the input values are local to each computational party, but steps 3 and 4 that produce the minimum of n_{ A } and n_{ B } involve joint computation by all of them. We subsequently discuss the choice of the parameter ℓ _{1}.
To compute the chisquared statistics for each SNP in parallel, we similarly follow the computation in equation 2 using the following steps:
1 $\left[{n}_{cA}\right]=\left[{n}_{cA}^{\left(1\right)}\right]+\left[{n}_{cA}^{\left(2\right)}\right];$
2 $\left[{n}_{tA}\right]=\left[{n}_{tA}^{\left(1\right)}\right]+\left[{n}_{tA}^{\left(2\right)}\right];$
3 $\left[{n}_{cB}\right]=\left[{n}_{cB}^{\left(1\right)}\right]+\left[{n}_{cB}^{\left(2\right)}\right];$
4 $\left[{n}_{tB}\right]=\left[{n}_{tB}^{\left(1\right)}\right]+\left[{n}_{tB}^{\left(2\right)}\right];$
5 [a] = Mul([n_{ cA }], [n_{ tB }]);
6 [b] = Mul([n_{ cB }], [n_{ tA }]);
7 [c] = Mul([n_{ cA }] + [n_{ tA }], [n_{ cB }] + [n_{ tB }]);
8 [d] = Mul([a] − [b], [a] − [b]);
9 [res_{2}] = Div(k · [d], [c], ℓ _{2});
Lines 5, 6, and 8 compute the numerator in equation 2 and line 7 its denominator (multiplication by public N , N_{ c }, and N_{ t } is omitted). The numerator is then scaled up by a factor of k to ensure that using integer division will provide sufficient precision of the result. The bitlength of k will be on the order of the precision of the answer in bits. We defer discussion of the choice of ℓ _{2} to the next section.
At the end of the computation, all computational parties send their shares of the result res_{1} and res_{2} for each SNP to one of the input sites who reconstruct the values. The output party then sets the result of MAF computation to res_{1}/N^{'}and the result of the chisquared computation to $\left(re{s}_{2}\cdot {N}^{\prime}\right)/\left(k{N}_{c}^{\prime}{N}_{t}^{\prime}\right)$.
Optimizations
We applied several optimizations to the computation to improve its runtime.

1
The nature of the computation in this task allows all interactive operations to run in parallel in a single batch for all SNPs. That is, all P comparisons corresponding to line 3 of MAF computation are executed simultaneously. The same applies to line 4 of MAF computation and lines 59 of chisquared computation.
We can further reduce the number of rounds in chisquared computation by running interactive independent operations at the same time. In particular, this means that lines 57 of the computation can be executed in a single round.

2
We modify chisquared computation to use floating point instead of integer division after converting both operands d and c to floating point representation. This is primarily driven by the fact that performance of division we rely on (described in [15, 9]) depends on the maximum of the bitlengths of its arguments and we can use substantially shorter values with floating point division compared to integer division (i.e., the bitlength can be comparable to that of k instead of the sum of the bitlengths of d and k). The savings noticeably outweigh the cost of integertofloating point conversion, or normalization (to use floating point division we need to normalize two values, while integer division needs to compute normalization of one of its arguments). We additionally slightly optimize integer to floating point conversion and floating point division compared to those given in [9] using information known about d and c (e.g., the fact that they are positive).

3
For performance reasons, we want to set parameters ℓ _{1} and ℓ _{2} (as well as the bitlength of secret shared values) to their minimum values that guarantee correctness. When the bitlength of the arguments to both comparison and division differ, the larger value is to be used. In particular, for ℓ _{1}, the largest value of n_{ A } or n_{ B } in the LT protocol appears when only one nucleotide is present in all genotypes in both case and control groups (i.e., max(n_{ A }, n_{ B }) = N^{'}and min(n_{ A }, n_{ B }) = 0), and we set ℓ_{1} = ⌈log N^{'}⌉ (where the extra 1 is due to the specifics of the LT operation). For ℓ _{2}, the largest value of d or c appears when n_{ cB } = n_{ tA } = 0, which leads to ${n}_{cA}={N}_{c}^{\prime}$, ${n}_{tB}={N}_{t}^{\prime}$, and $d={\left({N}_{c}^{\prime}\right)}^{2}{\left({N}_{t}^{\prime}\right)}^{2}$, and we set ${n}_{tB}={N}_{t}^{\prime}$, and $d={\left({N}_{c}^{\prime}\right)}^{2}{\left({N}_{t}^{\prime}\right)}^{2}$, and we set ${\ell}_{2}=\u23082\left(\mathsf{\text{log}}{{N}^{\prime}}_{c}+\mathsf{\text{log}}{{N}^{\prime}}_{t}\right)\u2309$. For integer division, this value of ℓ _{2} needs to be additionally incremented by the bitlength of precision k, but fortunately after we switch to floating point representation, we can reduce the bitlength to the desired precision of the result because the values are represented in a normalized form.
Secure distributed genomic Hamming distance computation
We next concentrate on the second task of securely computing the Hamming distance between a pair of genomic datasets in a distributed setting.
According to the task specification, the number of records in each of the two datasets are known to all parties and we denote them as N_{1} and N_{2}, respectively. The content of the records, however, is private (in particular, the values that fields CHROM, POS, REF, ALT, and SVTYPE take). Because only records with SVTYPE equal to SUB and SNP are relevant for the computation, for ease of notation we refer to them as SUB and SNP records, respectively.
The highlevel idea behind our solution is as follows: we first let each input site extract SUB and SNP records from its dataset and pad the resulting set with dummy records to hide its size. After each input site secret shares its records across all computational parties, the parties then run a set operation to identify all records that appear in both dataset (conceptually similar to set intersection) using ⟨CHROM, POS⟩ as the key as well as all records that appear only in one dataset (conceptually similar to symmetric difference). We accomplish this by obliviously sorting all records from both datasets using Batcher's mergesort [2] and scanning the sorted set examining every two adjacent elements in it to determine if the Hamming distance needs to be incremented by one for that pair.
At the time of competition preparation, Batcher's mergesort was available to us as one of the best options for oblivious sorting (based on the overall amount work as well as its round complexity). It is particularly well suited to this task because it is a recursive algorithm that works by first sorting the first and the second half of its input set and then merging the sorted halves. In our setup this means that the input datasets can be presorted by each input site locally and only the merge step needs to be run jointly. Unfortunately, Batcher's mergesort (including the merge step) has the drawback that the number of elements in the input set has to be a power of 2, which may unnecessarily increase the runtime.
In what follows, we start by describing in detail a basic solution in the first subsection and then discuss two optimizations in the two consecutive subsections.
Basic solution
As before, we divide the overall computation into three phases: input preparation, joint computation execution, and output reconstruction.
Input preparation. Each input site i extracts all SUB and SNP records from its dataset and pads them with dummy records to size N_{ i } + 1 (we require at least one dummy record). (If the combined fraction of SUB and SNP records is guaranteed to be within a certain fraction α < 1 of the total size for typical genomic datasets, then the datasets can be padded to αN_{ i } + 1 records. For this competition, α could not be lower than 1.) Furthermore, to meet Batcher's mergesort requirements, the input parties additionally pad the sets with dummy records so that the combined size of the two datasets is 2^{q}, where q = ⌈log2(N1 + N2 + 2)⌉. We use this newly formed dataset as the input into the computation and refer to it as a "dataset".
Next, the values in each record need to be converted to integers, which we accomplish as follows:

1
The location ⟨CHROM, POS⟩ is represented as V_{1} = CHROM · L + POS, where L is the maximum length of any existing human chromosome. CHROM ranges from 1 to 24 (22 autosomes, plus × and Y), and for dummy records we set V1 = 25L + 1 to avoid overlap with real records.

2
REF and ALT fields are represented as strings of nucleotides in the input. To produce their numeric counterparts, we map each nucleotide value to a twobit integer (e.g., A = 0, C = 1, G = 2, and T = 3) and concatenate twobit integers from a string to form a single number. To hide information about the size of the fields, the values need to be represented using the same bitlength for all records based on the maximum string length M. Because shorter strings need to be padded to the maximum size, we need to ensure that strings of different sizes will always be different (i.e., the padding character cannot be one of 03).
Instead of introducing a separate padding character, which increases the bitlength of one character from 2 to 3 bits, we append the string length in bits at the end of the string and use 0 for padding. Thus, all strings are represented using 2M +log M bits. Let V_{2} and V_{3} denote numeric values of REF and ALT fields in a record. V_{2} and V_{3} are set to 0 for dummy records.
In our implementation with M = 100, we partition representation of V_{2} and V_{3} into three blocks of size (2M + log M )/3 each. This still requires comparing all 2M + log M bits when two such values need to be compared, but reduces the size of secret shared values and thus the cost of the corresponding arithmetic. When M is large, V_{2} and V_{3} can instead be set to the hash of REF and ALT strings. This would guarantee constant size representation regardless of the value of M.
After computing a 3tuple (V_{1}, V_{2}, V_{3}) for each record in its dataset, an input site i sorts the records by the V_{1} field to form set S_{ i }, generates shares of all records in S_{ i }, and distributes them to the computational parties (we slightly abuse notation and use [S_{ i }] to denote shares of all values in S_{ i }). It also distributes shares of the number of dummy records d_{ i } in S_{ i }.
Computation execution. After receiving two sorted sets of ([V_{1}], [V_{2}], [V_{3}]) triples from both input sites, the computational parties run oblivious merge using [V1] as the key. The algorithm is built using an inputindependent sequence of compareandexchange operations. Each operation takes two integers and either swaps them or leaves them unchanged so that the first output (min) is always smaller than the second (max). In our framework, it is implemented as follows:
1 [c] = LT([a], [b], ℓ );
2 [min] = [c]([a] − [b]) + [b];
3 [max] = [c]([b] − [a]) + [a];
Note that lines 2 and 3 involve only a single multiplication (i.e., first compute [c]([a] − [b]) and then set [min] and [max] with no additional interaction). When applying this operation to our setting, comparisons on line 1 are performed using [V_{1}]'s, but the entire records ([V_{1}], [V_{2}], [V_{3}]) are swapped (or left unchanged) using comparison results [c].
The computational parties next compute the Hamming distance as specified in Algorithm 2. Sets S_{1} and S_{2} represent sorted input triples of the input parties and parameters ℓ_{1} and ℓ_{2} correspond to the bitlengths of fields V_{1} and V_{2} (or V_{3}), as discussed previously.
Because a specific location V_{1} appears only once in each of the input datasets (except for dummy records), there will be at most two records with the same V_{1} in the combined set. The algorithm works by looking at each pair of two consecutive elements in the combined sorted set and adds 1 to dist if this is the first time the location appears on the list (i.e., a_{ i } = 0 on line 4). The distance is incremented automatically for the first record (dist = 1 on line 2). Then, if a location appears for the second time (a_{ i } = 1 on line 4), the algorithm examines the values of V_{2} and V_{3} fields (lines 56) to determine whether the condition for incrementing the distance is satisfied (i.e., b_{ i } = 1 and c_{ i } = 0). If not (b_{ i } = 0 or c_{ i } = 1), dist is decremented by 1 to compensate for the fact that it was increased during previous loop iteration. All dummy records collectively contribute distance −d_{1}−d_{2} +2 (i.e., 0 for the first two records and −1 for each additional record) and this is why we adjust the computed distance at the end (line 9). We note that all loop iterations and all comparisons within a loop iteration can be carried out in parallel.
Algorithm 2 SecureHD([S_{1}], [S_{2}], [d_{1}], [d_{2}])
1: ${\left(\left[{V}_{1}^{\left(i\right)}\right],\left[{V}_{2}^{\left(i\right)}\right],\left[{V}_{3}^{\left(i\right)}\right]\right)}_{i=1}^{{2}^{q}}=\mathsf{\text{Merge}}\left(\left[{S}_{1}\right],\left[{S}_{2}\right]\right)$
2: dist = 1
3: for i = 2 to 2^{q} do
4: $\left[{a}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right],\left[{V}_{1}^{\left(i\right)}\right],{\ell}_{1}\right)$
5: $\left[{b}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{2}^{\left(i1\right)}\right],\left[{V}_{2}^{\left(i\right)}\right],{\ell}_{2}\right)$
6: $\left[{c}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{3}^{\left(i1\right)}\right],\left[{V}_{3}^{\left(i\right)}\right],{\ell}_{2}\right)$
7: dist = dist + (1 − [a_{ i }]) + [a_{ i }]([b_{ i }](1 − [c_{ i }]) − 1)
8: end for
9: dist = dist + [d_{1}] + [d_{2}] −2
10: return dist
Output reconstruction is straightforward and consists only of receiving and combining shares of the computed Hamming distance.
Separating SUB and SNP records
As the first significant optimization, we separate computation of the distance for SNP and SUB records and consequently reconstruct the overall distance from the two values. The main reason for this is to reduce the time comparisons of V_{2} and V_{3} take. Recall that SNP records contain a single character in REF and ALT fields, while SUB records can contain longer strings. In the genomic datasets we worked with, a great majority of all records were SNP records that can be processed using 2bit comparisons for V_{2} and V_{3} (i.e., ℓ_{2} = 2). In the basic solution, however, the bitlength had to be unnecessarily increased by two orders of magnitude for most records to meet privacy requirements. Thus, the idea consists of extracting two sets from each input dataset: one consisting of SNP records and another consisting of SUB records. Then the distance for SUB records is computed separately from the distance for SNP records and the sum is returned as the result.
This strategy works well if all records with the same ⟨CHROM, POS⟩ pair are always marked with the same type across all datasets. It is, however, possible for two datasets to contain SUB and SNP records corresponding to the same location. Because of the existence of such records, the Hamming distance will not be computed correctly if we simply add the two distances together. That is, if one record appears in the SUB set and another with the same location appears in the SNP set, they collectively will contribute 2 to the overall distance instead of correct 0 or 1 (depending on other attributes). To address this, we need to find all such pairs and compensate for the difference they introduced, which is the most subtle part of our solution. We next provide more detail about the solution and highlight the differences from the basic scheme.
Input preparation. Given a dataset, an input entity produces two subsets: one composed of SUB records and one composed of both SUB and SNP records from the dataset. As before, both sets need to be padded with dummy records to hide their number and make the size to be a power of 2 to the combined size of 2^{qs} and 2^{q}, where q_{ s } = ⌈log(α_{ s }(N_{1} + N_{2}) + 2)⌉ and q = ⌈log(α(N_{1} + N_{2}) + 2⌉ and α_{ s } (α) denotes the maximum fraction of SUB (resp., SNP and SUB) records in genomic datasets (we were given α = 1 and α_{ s } = 0.3). All records in the SUB set are converted to (V_{1}, V_{2}, V_{3}) triples as before. For the SNP&SUB set, onecharacter REF and ALT fields in SNP or SUB records are represented using integers 03, while these fields of longer length in SUB records are represented using integer 4 (i.e., V_{2} and V_{3} fields are 3 bits long). This will guarantee that comparison of a onecharacter long REF or ALT field in a SNP record with a longer REF or ALT field in a SUB record will result in their inequality. We also add another binary attribute V_{4} to each record of the SNP&SUB set that indicates whether the record is of SUB type (V_{4} = 0) or SNP type (V_{4} = 1). We set V_{4} = 0 in dummy records.
Each input entity now produces shares of (V_{1}, V_{2}, V_{3}) in its SUB set and (V_{1}, V_{2}, V_{3}, V_{4}) in its SNP&SUB set (together with the number of dummy records in each set) and distributes them to the computational parties. We note that computation with SNP&SUB sets can be performed on shorter bitlengths, which results in faster arithmetic, and thus we setup two different instances of the secret sharing scheme and process SUB sets separately from SNP&SUB sets.
Computation execution. To compute the Hamming distance correctly, we now distinguish between different cases: (i) SUB records that don't have a SNP record with identical location in the other dataset, (ii) SNP records that don't have a SUB record with identical location in the other dataset, and (iii) records that have another record with identical location but different type present in the other dataset. Let N0 denote the number of records in the third category.
Algorithm 3 SecureHD2([S_{1}], [S_{2}])
1: ${\left(\left[{V}_{1}^{\left(i\right)}\right],\left[{V}_{2}^{\left(i\right)}\right],\left[{V}_{3}^{\left(i\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)}_{i=1}^{{2}^{q}}=\mathsf{\text{Merge}}\left(\left[{S}_{1}\right],\left[{S}_{2}\right]\right)$
2: $\mathsf{\text{dist}}=\left[{V}_{4}^{\left(1\right)}\right]$
3: for i = 2 to 2^{q} do
4: $\left[{a}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right],\left[{V}_{1}^{\left(i\right)}\right]\right)$
5: $\left[{b}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{2}^{\left(i1\right)}\right],\left[{V}_{2}^{\left(i\right)}\right]\right)$
6: $\left[{c}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{3}^{\left(i1\right)}\right],\left[{V}_{3}^{\left(i\right)}\right]\right)$
7: $\left[{d}_{i}\right]=\mathsf{\text{OR}}\left(\left[{V}_{4}^{\left(i1\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)$
8: $\left[{e}_{i}\right]=\mathsf{\text{XOR}}\left(\left[{V}_{4}^{\left(i1\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)$
9: $\mathsf{\text{dist}}+=\left[{d}_{i}\right]\left(\left(1\left[{a}_{i}\right]\right)\left[{V}_{4}^{\left(i\right)}\right]+\left[{a}_{i}\right]\left(\left[{b}_{i}\right]\left(1\left[{c}_{i}\right]\right)\left[{V}_{4}^{\left(i1\right)}\right]\right)\right)\left[{a}_{i}\right]\left[{e}_{i}\right]$
10: end for
11: return dist
The computational parties execute Algorithm 2 on two SUB datasets. This computes the distance corresponding to the records in category 1, but also introduces offset N_{0}. The parties then execute Algorithm 3 on two SNP&SUB sets that computes the distance corresponding to categories 2 and 3 and additionally compensates for the offset. The output will then be the sum of the distances computed by both algorithms.
In Algorithm 3, when examining each pair of consecutive records, we only consider those that contain at least one SNP record within the pair (d_{ i } = 1 on line 9). Furthermore, similar Algorithm 2, when observing a location for the first time, we add 1 to the Hamming distance, but only if it is a SNP record (${V}_{4}^{\left(1\right)}=1$ on line 2 and ${V}_{4}^{\left(i\right)}=1$ on line 9). If a location appears for the second time, we undo the previous increment if b_{ i } = 0 or c_{ i } = 1 as before, but only if the record preceding the current one is of type SNP (i.e., ${V}_{4}^{\left(i1\right)}=1$ on line 9). By doing that, we are able compute the distance corresponding to records of second and third types without introducing errors. The offset N_{0} is compensated by the last term a_{ i }e_{ i } on line 9, that counts the number of pairs of consecutive records that have the same location (a_{ i } = 1), but different types (e_{ i } = 1). OR([x], [y]) and XOR([x], [y]) are implemented as [x]+[y]−Mult([x], [y]) and [x]+[y]−2Mult([x], [y]), respectively (computation of d_{ i } and e_{ i } reuses the same multiplication result).
Note that dummy records do not introduce any error in Algorithm 3. That is, d_{ i } = 0 and e_{ i } = 0 when both records i and i − 1 are dummy and the expression on line 9 evaluates to 0. Similarly, when record i−1 is real while record i is fake that expression also evaluates to 0 because a_{ i } = 0 and ${V}_{4}^{\left(i\right)}=0$.
After computing the distances corresponding to SUB and SNP&SUB sets, the parties need to convert shares of one of them into shares of the same value in the secret sharing setup used by the other algorithm. Then the distances can be locally added to compute the overall result. Output reconstruction is performed as before by exchanging the shares and recovering the result.
The performance gain achieved by this optimization highly depends on the values of public parameters α_{ s }, α, and M. While the gain stems from using shorter values for V_{2} and V_{3} with SNP&SUB sets, the total number of records processed using this solution $\left({2}^{q}+{2}^{{q}_{s}}\right)$ is greater than in the basic scheme (2_{ q }). Therefore, this optimization is recommended with relatively small α_{ s } and large M. In our experiments with α_{ s } = 0.3, α = 1, and M = 100, we observed approximately 30% performance improvement compared to the basic scheme.
Reducing set size
Our second optimization is with respect to oblivious sort and removing the requirement that the input size has to be a power of 2 for the merge step of Batcher's mergesort. To explain how our optimization works, we need to provide additional details about Batcher's mergesort algorithm.
Recall that the merge step takes two sorted sets L_{1} = (a_{1}, a_{2}, ..., a_{ m }) and L_{2} = (b_{1}, b_{2},..., b_{ n }), where m + n is a power of 2. It first combines them into a single sequence that first monotonically increases and then decreases as L = (a_{1}, a_{2}, ..., a_{ m }, b_{ n }, ..., b_{2}, b_{1}), after which a sequence of compareandexchange operation is executed as specified by the following pseudocode:
After executing the first iteration of the outer loop, the first (second) half of L will contain (m + n)/2 smallest (resp., largest) elements of L although they are not necessarily sorted. After its second iteration, the ith quarter of L will contain the ith quarter of elements in the final sorted list for i = 1, ..., 4. This process continues until each sublist contains one element and L becomes sorted. Notice that the algorithm uses log(m + n) iterations of the outer loop, and in each iteration every element in the list is used in a compareandexchange operation, which is the reason for requiring the size of the list to be a power of 2.
Consider an example with input L_{1} = (3, 4, 6, 9, 12) and L_{2} = (2, 5, 10), which is combined into L = (3, 4, 6, 9, 12, 10, 5, 2). In the first iteration of the outer loop, compareandexchange operations are performed on pairs (3, 12), (4, 10), (6, 5), and (9, 2), and the resulting list is (3, 4, 5, 2, 12, 10, 6, 9). In the second iteration, comparisons are performed on (3, 5), (4, 2), (12, 6), (10, 9) and produce (3, 2, 5, 4, 6, 9, 12, 10). In the last iteration, we compare every pair of consecutive elements (3, 2), (5, 4), (6, 9), (12, 10), which results in the final sorted list (2, 3, 4, 5, 6, 9, 10, 12). If L2 = (2, 5) instead, after all iterations 2 will remain at the end of the list making it unsorted, as the element does not have any pair to use in a comparison.
We next proceed with describing our strategy for generalizing the merge operation to work with inputs of arbitrary sizes, which might be of independent interest. There will be no need to pad the input in the beginning to make the overall input size to be a power of 2, but dummy records are now added throughout algorithm execution as needed. This means that earlier loop iterations use a smaller number of elements and are therefore faster than in the original algorithm. In particular, at each loop iteration, if the size of a sublist is odd, we append a copy of its last element to the end. This will ensure that comparisons can be performed at the current level while still preserving the necessary properties of the (partially) sorted list. For example, suppose we want to merge (3, 6, 8) and (5, 7). Before the first iteration 5 will be added to the list (3, 6, 8, 7, 5) because the number of elements in it is odd, and we obtain (3, 5, 5, 7, 6, 8) at the end of that iteration. At the time of second iteration, the size of sublists (3, 5, 5) and (7, 6, 8) is also odd and they are modified to be (3, 5, 5, 5) and (7, 6, 8, 8). In the next iteration no additional padding is used and we obtain (3, 5, 5, 5, 6, 7, 8, 8) at the end of the algorithm.
In the context of Hamming distance computation, we similarly make a copy of the entire last record of a subset as needed during the merge step. More importantly, after having the list sorted, we need to ensure the Hamming distance is computed correctly because the introduction of repeated records creates inaccuracies in Algorithm 2. Now two consecutive records with the same location in the sorted set may correspond to (i) two records in the original datasets, (ii) two copied records, or (iii) one original and one copied record. Let a_{ i } and a_{ j } be two different records with the same location in the original datasets. If they get copied during the merge as ${a}_{i}^{\prime}$ and ${a}_{j}^{\prime}$, the relative order of these records in the sorted list can be arbitrary (e.g., $\left({a}_{i}^{\prime},{a}_{i},{a}_{j},{a}_{j}^{\prime}\right)$, $\left({a}_{i}^{\prime},{a}_{j},{a}_{i},{a}_{j}^{\prime}\right)$, etc.) and they may contribute more than 1 to the computed distance.
We address the problem by modifying locations of records in the datasets so that (i) two records originally with the same location are assigned locations that differ by 1 and (ii) two records originally with different locations are assigned locations that differ by more than 1. By doing that, a pair of consecutive records with the same location in the sorted set is guaranteed to correspond to either two copied records or one original record and its copy. In either case, the Hamming distance should not get affected. We implement this change by setting the location to 4V_{1}, where V_{1} is the original location, for records from the first input site and to 4V_{1} + 1 for records from the second input site.
Algorithm 4 SecureHD3([S_{1}], [S_{2}])
1: ${\left(\left[{V}_{1}^{\left(i\right)}\right],\left[{V}_{2}^{\left(i\right)}\right],\left[{V}_{3}^{\left(i\right)}\right]\right)}_{i=1}^{{2}^{q}}=\mathsf{\text{NewMerge}}\left(\left[{S}_{1}\right],\left[{S}_{2}\right]\right)$
2: dist = 1
3: for i = 2 to 2^{q} do
4: $\left[{a}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right],\left[{V}_{1}^{\left(i\right)}\right]\right)$
5: $\left[{b}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{2}^{\left(i1\right)}\right],\left[{V}_{2}^{\left(i\right)}\right]\right)$
6: $\left[{c}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{3}^{\left(i1\right)}\right],\left[{V}_{3}^{\left(i\right)}\right]\right)$
7: $\left[{d}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right]+1,\left[{V}_{1}^{\left(i\right)}\right]\right)$
8: $\mathsf{\text{dist}}+=\left(1\left[{a}_{i}\right]\right)\left(1\left[{d}_{i}\right]+\left[{d}_{i}\right]\left(\left[{b}_{i}\right]\left(1\left[{c}_{i}\right]\right)1\right)\right)$
9: end for
10: return dist
With this solution, the input sites prepare their input datasets as in the basic scheme, but pad a set to size N_{ i } + 1 instead of requiring the combined size to be a power of 2. The computational parties can locally modify V_{1}'s in the input records run the improved merge and compute the Hamming distance as specified in Algorithm 4. The algorithm has two major differences compared to Algorithm 2: (i) when examining each pair of consecutive records, only pairs with different locations can contribute to the distance (a_{ i } = 0 on line 8) and (ii) when locations of two consecutive records differ by 1 (d_{ i } = 1 on line 8), they are treated as having the same location in Algorithm 2, and when the locations differ by neither 1 nor 0 (d_{ i } = 0 and a_{ i } = 0 on line 8), they are treated as having different locations in Algorithm 2.
Dummy records inserted by each input site into their input datasets do not affect correctness of the Hamming distance that uses this optimization (including the combined solution in Algorithm 5). This is because the first dummy record from the first input set will result in the distance incremented by 1, while the first dummy record from the second input set will result in the distance decremented by 1. All consecutive dummy records from the first or the second input datasets do not modify the distance (because all records with the same V_{1} are ignored).
We recently became aware of a sorting algorithm [16] that generalizes Batcher's bitonic sort to input sizes which are not a power of 2 without adding extra records during the sorting procedure. The algorithm results in the same asymptotic complexity as our solution, but performs fewer compareandexchange operations in each iteration (because dummy records are not added), which is expected to lead to better performance than our algorithm. We plan to provide both theoretical and empirical comparison of this algorithm with our solution as future work.
Our final solution consists of using both optimizations from the previous and current subsections, and we summarize it in Algorithm 5. We omit its explanation due to space considerations.
Algorithm 5 SecureHD4 $\left(\left[{S}_{1}^{\mathsf{\text{SUB}}}\right],\left[{S}_{1}^{\mathsf{\text{SNP}}}\right],\left[{S}_{2}^{\mathsf{\text{SUB}}}\right],\left[{S}_{2}^{\mathsf{\text{SNP}}}\right]\right)$
1: dist_{1} = SecureHD3$\left(\left[{S}_{1}^{\mathsf{\text{SUB}}}\right],\left[{S}_{2}^{\mathsf{\text{SUB}}}\right]\right)$
2: ${\left(\left[{V}_{1}^{\left(i\right)}\right],\left[{V}_{2}^{\left(i\right)}\right],\left[{V}_{3}^{\left(i\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)}_{i=1}^{{2}^{q}}=\mathsf{\text{NewMerge}}\left(\left[{S}_{1}^{\mathsf{\text{SNP}}}\right],\left[{S}_{2}^{\mathsf{\text{SNP}}}\right]\right)$
3: $\mathsf{\text{dis}}{\mathsf{\text{t}}}_{2}=\left[{V}_{4}^{\left(0\right)}\right]$
4: for i = 2 to 2^{q} do
5: $\left[{a}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right],\left[{V}_{1}^{\left(i\right)}\right]\right)$
6: $\left[{b}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{2}^{\left(i1\right)}\right],\left[{V}_{2}^{\left(i\right)}\right]\right)$
7: $\left[{c}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{3}^{\left(i1\right)}\right],\left[{V}_{3}^{\left(i\right)}\right]\right)$
8: $\left[{d}_{i}\right]=\mathsf{\text{OR}}\left(\left[{V}_{4}^{\left(i1\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)$
9: $\left[{e}_{i}\right]=\mathsf{\text{XOR}}\left(\left[{V}_{4}^{\left(i1\right)}\right],\left[{V}_{4}^{\left(i\right)}\right]\right)$
10: $\left[{g}_{i}\right]=\mathsf{\text{EQ}}\left(\left[{V}_{1}^{\left(i1\right)}\right]+1,\left[{V}_{1}^{\left(i\right)}\right]\right)$
11: $\mathsf{\text{dis}}{\mathsf{\text{t}}}_{2}+=\left(1\left[{a}_{i}\right]\right)\left(\left[{d}_{i}\right]\left(\left(1\left[{g}_{i}\right]\right)\left[{V}_{4}^{\left(i\right)}\right]+\left[{g}_{i}\right]\left(\left[{b}_{i}\right]\left(1\left[{c}_{i}\right]\right){V}_{4}^{\left(i1\right)}\right)\right)\left[{g}_{i}\right]\left[{e}_{i}\right]\right)$
12: end for
13: return dist_{1} + Convert(dist_{2})
Results
In this section, we provide experimental results of securely computing GWAS statistics and the Hamming distance in a distributed setting. We ran experiments in LAN and WAN settings with three computational parties connected by pairwise secure authenticated channels with each other. The LAN experiments were conducted using 2.4 GHz 6core Red Hat Linux machines connected through 1 Gb/s Ethernet with pairwise roundtrip times 0.3 msec. Our WAN experiments used two machines from the LAN setting and employed another 2.1 GHz 8core machine from the GENI infrastructure [17] at a different geographic location with Red Hat Linux. The pairwise roundtrip times between these machines were 0.3 msec, 9.2 msec, and 9.2 msec. Each GWAS experiment was run 20 times and each Hamming distance experiment was run 5 times and the median over all runs is reported.
For GWAS computation, case and control groups consisted of genotypes of 200 individuals each (100 individuals at each input site in each group). We measured the runtime of the MAF and chisquared computation by varying the number of SNPs in the input. The results are given in Table 1. Modulus size (Mod) corresponds to the bitlength of secret shared values.
Our implementation incorporates all optimizations and uses parameters ℓ _{1} = 11 and ℓ _{2} = 21 computed as described in the optimizations subsection of the secure GWAS computation section (ℓ _{2} = 35 + k would be required for integer division, but a lower parameter is requested precision). We can see from the table that sufficient with floating point operation to obtain the the execution time is linear in the number of SNPs in both settings, and the overhead in WAN is almost three times as large as that in LAN, which is primarily due to larger communication delays in WAN. Another observation not present in the table is that division performed in chisquared computation contributes almost the entire runtime (close to 99%) and thus any optimizations applied to division can lead to direct improvement of chisquared performance.
For the Hamming distance computation, we conducted four sets of experiments in the LAN setting, that correspond to the basic scheme (i) with no optimizations, (ii) with the first optimization, (iii) with the second optimization, and (iv) with both optimizations. By comparing execution times of different schemes, we can see performance gains from different optimizations. In the WAN setting, we only report the timings of the best (last) scheme. For each set of experiments, we varied the number of records in the genomic dataset at each input site. The results are presented in Table 2.
We used two different secret sharing bitlengths for schemes that apply the first optimization (one for the computation with SUB records and another for the computation with SNP&SUB records). The complexity of the merge is O(n log n) for combined sequences of size n and computing the distance itself is linear in n (with larger constants), which the runtimes in Table 2 follow. In the LAN setting, the two optimizations result in performance improvement up to 27.9% and 13.1% on our set of parameters when applied separately, and 40.9% when applied together. Performance gain of the first optimization heavily depends on parameters (α, α_{ s } , and M ), while the gain of the second optimization depends on the difference between the combined input size N and 2^{⌈log N⌉}. The the smaller the difference is, the smaller improvement is expected.
Conclusions
In this work we report on our experience with participation in the 2015 iDASH secure genomic computation competition. We show how to securely compute MAF and chisquared statistics in the context of GWAS computation and the Hamming distance between two genomic datasets and report on their performance results. We develop a number of novel optimizations, some of which may be of independent interest.
References
 1.
Blanton M, Aguiar E: Private and oblivious set and multiset operations. ACM Symposium on Information, Computer and Communications Security (ASIACCS). 2012
 2.
Batcher K: Sorting networks and their applications. AFIPS Spring Joint Computer Conference. 1968, 307314.
 3.
iDASH 2015 Secure Genome Analysis Competition: Competition Tasks. [http://www.humangenomeprivacy.org/2015/competitiontasks.html]
 4.
Tang H, Jiang X, Wang X, Wang S, Sofia H, Fox D, Lauter K, Malin B, Telenti A, Xiong L, OhnoMachado L: Protecting Genomic Data Analytics in the Cloud: State of the Art and Opportunities.
 5.
Shamir A: How to share a secret. Communications of the ACM. 1979, 22 (11): 612613.
 6.
Gennaro R, Rabin M, Rabin T: Simplified VSS and fasttrack multiparty computations with applications to threshold cryptography. ACM Symposium on Principles of Distributed Computing (PODC). 1998, 101111.
 7.
Canetti R: Security and composition of multiparty cryptographic protocols. Journal of Cryptology. 2000, 13 (1): 143202.
 8.
Catrina O, de Hoogh S: Improved primitives for secure multiparty integer computation. Security and Cryptography for Networks (SCN). 2010, 182199.
 9.
Aliasgari M, Blanton M, Zhang Y, Steele A: Secure computation on floating point numbers. Network and Distributed System Security Symposuim (NDSS). 2013
 10.
Yao A: How to generate and exchange secrets. IEEE Symposium on Foundations of Computer Science. 1986, 162167.
 11.
De Cristofaro E, Tsudik G: Practical private set intersection protocols with linear complexity. Financial Cryptography and Data Security (FC). 2010, 143159.
 12.
Malkhi D, Nisan N, Pinkas B, Sella Y: Fairplay  a secure twoparty computation system. USENIX Security Symposium. 2004, 287302.
 13.
Bogdanov D, Laur S, Willemson J: Sharemind: A framework for fast privacypreserving computations. European Symposium On Research In Computer Security (ESORICS). 2008, 192206.
 14.
Zhang Y, Steele A, Blanton M: PICCO: A generalpurpose compiler for private distributed computation. ACM Conference on Computer and Communications Security (CCS). 2013, 813826.
 15.
Catrina O, Saxena A: Secure computation with fixedpoint numbers. Financial Cryptography and Data Security (FC). 2010, 3550.
 16.
Bitonic Sorting Network for n Not a Power Of 2. [http://www.iti.fhflensburg.de/lang/algorithmen/sortieren/bitonic/oddn.htm]
 17.
GENI: Global Environment for Network Innovations. [http://www.geni.net]
Acknowledgements
This work was supported in part by grants CNS1223699 and CNS1319090 from the National Science Foundation and FA95501310066 from the Air Force Office of Scientific Research. Any opinions, findings, and conclusions or recommendations expressed in this publication are those of the authors and do not necessarily reflect the views of the funding agencies. We also acknowledge the NSFsponsored Global Environment for Network Innovations (GENI) test bed, which allowed us to run WAN experiments.
This article has been published as part of BMC Medical Informatics and Decision Making Volume 15 Supplement 5, 2015: Proceedings of the 4th iDASH Privacy Workshop: Critical Assessment of Data Privacy and Protection (CADPP) challenge. The full contents of the supplement are available online at http://www.biomedcentral.com/14726947/15/S5.
Declarations
Publication funding for this supplement was supported by iDASH U54HL108460, iDASH linked R01HG007078 (Indiana University), NHGRI K99HG008175 and NLM R00LM011392.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
Marina Blanton and Yihua Zhang designed the protocols for distributed GWAS computation and secure sequence comparison in the form of Hamming distance. Yihua Zhang and Ghada Almashaqbeh implemented the protocols and reported evaluation results. Marina Blanton and Yihua Zhang drafted the manuscript for publication. Ghada Almashaqbeh contributed to the work while at the University of Notre Dame.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Zhang, Y., Blanton, M. & Almashaqbeh, G. Secure distributed genome analysis for GWAS and sequence comparison computation. BMC Med Inform Decis Mak 15, S4 (2015). https://doi.org/10.1186/1472694715S5S4
Published:
Keywords
 secure genome analysis
 iDASH competition
 Hamming distance
 oblivious sorting
 oblivious merge
 GWAS computation
 secret sharing