In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by a patient's public keys at a storage and processing unit (SPU). A part of the corresponding private key is also stored on the SPU. At the time of the test by a medical unit (MU), the patient provides the second part of the private key to the MU. A test with its associated markers is determined by the MU and sent to the SPU. The test is carried out on the encrypted values thanks to homomorphic operation and returned back to the MU. The latter uses the second part of the private key to access the result.

Patent
   9270446
Priority
Sep 14 2012
Filed
Sep 10 2013
Issued
Feb 23 2016
Expiry
Sep 10 2033
Assg.orig
Entity
Small
8
1
EXPIRED<2yrs
1. A method to process genomic data comprising the steps of:
associating, by a certified institution, a patient identification for a given patient;
generating, by the certified institution, a pair of asymmetric keys related to said patient comprising a private and a public key;
analyzing, by the certified institution, an output of a deoxyribonucleic acid (DNA) sequencer and preparing an aligned genomic data for said patient comprising approved variants, said approved variants being approved by medical authorities, each approved variant representing a position in the genome and a value representing a nucleotide that varies between individuals;
extracting, by the certified institution, real and potential variants from said approved variants, said real and potential variants having each a position, said real variants being a subset of the approved variants and being different for each human being, said potential variants being the remaining part of the approved variants;
selecting, by the certified institution, all or part of the potential variants;
analyzing, by the certified institution, the correlation between the selected potential variants and a privacy sensitivity of the real variants;
selecting, by the certified institution, a number of other potential variants, said number being determined according to the previous analysis and a level of privacy required;
encrypting, by the certified institution, the value of each real variant and selected potential variants with the public key of the patient;
storing, by the certified institution, the encrypted values with their respective positions and the patient identification into a storage and processing unit;
dividing, by the certified institution, the private key into at least a first and a second part;
storing, by the certified institution, the second part of the private key in the certified institution or in a patient device;
transmitting, by the certified institution, the first part of the private key to the storage and processing unit;
selecting by a medical unit a personalized clinical test to be carried out and related genetic markers, each marker having a position and a contribution;
determining, by the medical unit, the contribution of each marker according to the personalized clinical test selected;
receiving, by the storage and processing unit from the medical unit, genetic markers related to the personalized clinical test, the respective contributions of the related genetic markers and the patient identification of the patient;
retrieving by the storage and processing unit the encrypted values for said patient matching the position of the related genetic markers;
executing by the storage and processing unit a genetic test by using the retrieved values, and the contribution of the respective genetic markers thanks to homomorphic operations;
partially decrypting by the storage and processing unit the result of the genetic test using the first part of the private key;
sending by the storage and processing unit the decrypted result to the medical unit;
whereby the medical unit can use the second part of the private key to obtain the result of the performed personalized clinical test.
2. The method to process genomic data according to claim 1, further comprising:
generating, by the certified institution, a dummy variant comprising a dummy position and a dummy value, said dummy position being outside of the overall variant positions of a sequence;
encrypting, by the certified institution, the positions of the real variants with the symmetric key of the patient;
encrypting, by the certified institution, the dummy value with the public key of the patient;
encrypting, by the certified institution, the position of the dummy variant with the symmetric key of the patient;
storing, by the certified institution, together with the encrypted variants, the dummy variant as well as the encrypted positions and the encrypted dummy position into a storage and processing unit;
storing, by the certified institution, the position of the dummy variant into the patient device;
determining by the certified institution a set of positions which are common between the marker's position and the real variant's positions;
receiving by the certified institution from the medical unit an encrypted set of positions with the symmetric key of said patient, and for the marker's positions not present in the variant's position, dummy positions;
sending by the certified institution to the storage and processing unit the encrypted marker's positions as well as the patient identification; and
retrieving by the storage and processing unit the encrypted values for said patient at these encrypted locations and sending them to the medical unit.

This application is a U.S. National Stage Application of International Application No. PCT/EP2013/068658 filed Sep. 10, 2013, which claims priority to European Application No. 12184372.6 filed Sep. 14, 2012 and U.S. Provisional Patent Application No. 61/700,897 filed Sep. 14, 2012.

In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. First, we highlight the potential privacy threats on genomic data and the challenges of providing privacy-preserving algorithms. Then, focusing specifically on a typical disease-susceptibility test, we develop a new architecture (between the patient and the medical unit) and propose privacy-preserving algorithms.

Privacy control can be defined as the ability of individuals to determine when, how, and to what extent information about themselves is revealed to others. In this way, the usage of private data will remain in context and it will be used exclusively for the purpose the data owner has in mind. Privacy is usually protected by both legal and technological means. By using legal actions, such as data protection directives and fair information practices, privacy regulations can enforce privacy protection on a large scale. Yet, this approach is mostly reactive, as it defines regulations after technologies are put in place. To avoid this issue, Privacy-Enhancing Technologies (PETs) [1-3] can be incorporated into the design of new systems in order to protect individuals' data. PETs protect privacy by eliminating or obfuscating personal data, thereby preventing misuse or involuntary loss of data, without affecting the functionality of the information system [4].

Their objective is to make it difficult for a malicious entity to link information to specific users. In order to obfuscate personal data, PETs often rely on cryptographic primitives, such as anonymous authentication and encryption.

Genomics is becoming the next significant challenge for privacy. The price of a complete genome profile has plummeted below $200 for genome-wide genotyping (i.e., the characterization of about one million common genetic variants), which is offered by a number of companies (located mostly in the US). Whole genome sequencing is also offered through the same direct-to-consumer model (but at a higher price). This low cost of DNA sequencing will break the physician/patient connection, because private citizens (from anywhere in the world) can have their genome sequenced without involving their family doctor. This can open the door to all kinds of abuse, not yet fully understood.

As a result of the rapid evolution in genomic research, substantial progress is expected in terms of improved diagnosis and better preventive medicine. However, the impact on privacy is unprecedented, because (i) genetic diseases can be unveiled, (ii) the propensity to develop specific diseases (such as Alzheimer's) can be revealed, (iii) a volunteer accepting de facto to have his genomic code made public (as it already happened) can leak substantial information about his ethnic heritage and genomic data of his relatives (possibly against their will), and (iv) complex privacy issues can arise if DNA analysis is used for criminal investigations and insurance purposes. Such issues could lead to genetic discrimination (e.g., ancestry discrimination or discrimination due to geographic mapping of people). Even though the Genetic Information Non-discrimination Act (GINA), which prohibits the use of genomic information in health insurance and employment, attempted to solve some of these problems in the US, these types of laws are very difficult to enforce.

An even more severe case, currently not widely considered, is where a malicious party initiates a cross-layer attack by utilizing privacy-sensitive information belonging to a person retrieved from different sources (e.g., genomic data, location, online social network, etc.), thus creating the opportunity for a large variety of fraudulent uses of such data. For example, as stated in the Personal Genome Project (PGP) consent form [5], a malicious party could make synthetic DNA of a person and plant it at a crime scene to falsely accuse him.

In this hypothetical situation, the attacker can make his accusation stronger if he has the location patterns of the person to be blamed, and hence, knows that the person was close to the crime scene at the time of the crime. Similarly, an attacker can easily obtain information on close relatives of a target from online social network data, thus effectively increasing the potential access to target's genomic data if his relatives' DNA has been sequenced. In other words, even if the person has perfect privacy on his own genome, if the attacker has access to the DNA sequence of the relatives, he can obtain significant information about the person's DNA sequence.

Even though, at this stage, the field of genomics is generally free from serious attacks, it is likely that the above threats will become more serious as the number of sequenced individuals becomes larger. Such was the case of the Internet that was initially run and used by well-intentioned researchers. However, once it became more widely used, it became plagued by uncountable attacks such as spyware, viruses, spam, botnets, Denial-of-Service attacks, etc. Therefore, the need to adapt PETs to personal genomic data will only grow with time, as they are key tools for preventing an adversary from linking particular genomic data to a specific person or from inferring privacy-sensitive genomic data about a person.

It is obvious that users need to reveal personal and even privacy-sensitive information for genomic tests (e.g., paternity tests, disease-susceptibility tests, etc.). Nevertheless, they want to control how this information is used by the service providers (e.g., medical units such as healthcare centers or pharmaceutical companies, depending on the type of the test). Currently, the companies and hospitals that perform DNA sequencing store the genomic data of their customers and patients. Of course, tight legislation regulates their activities, but it is extremely difficult for them to protect themselves against the misdeeds of a hacker or a disgruntled employee. In a non-adversarial scenario, however, making use of this data requires legitimate professionals (e.g., physicians and pharmacists) to access the data in some way. Therefore, new architectures and protocols are needed to store and process this privacy-sensitive genomic data, while still enabling its utilization by the service providers (e.g., medical units).

In this work, our goal is to protect the privacy of users' genomic data while enabling medical units to access the genomic data in order to conduct medical tests or develop personalized medicine methods. In a medical test, a medical unit checks for different health risks (e.g., disease susceptibilities) of a user by using specific parts of his genome. Similarly, to provide personalized medicine, a pharmaceutical company tests the compatibility of a user on a particular medicine, or a pharmacist checks the compatibility of a given medicine (e.g., over-the-counter drug) to a given user. In both scenarios, in order to preserve his privacy, the user does not want to reveal his complete genome to the medical unit or to the pharmaceutical company. Moreover, in some scenarios, it is the pharmaceutical companies who do not want to reveal the genetic properties of their drugs. To achieve these goals, we propose to store the genomic data at a Storage and Processing Unit (SPU) and conduct the computations on genomic data utilizing homomorphic encryption and proxy encryption to preserve the privacy of the genomic data.

The rest of the paper is organized as follows. In the rest of this section, we discuss the challenges in genomic privacy and summarize the related work on genomic privacy. In Section 2, we describe our proposed schemes for privacy-preserving medical tests and personalized medicine. Furthermore, we analyze the level of privacy provided by the proposed schemes for different design and genomic criteria. Then, in Section 3, we discuss the implementation of the proposed schemes and present their complexity and security evaluations.

Finally, in Section 4, we conclude the paper and discuss new research directions on genomic privacy.

Obviously, there are certain obstacles for achieving our goals on genomic privacy. These are mostly due to (i) the balance between privacy and reliability of the genomic data, (ii) the structure of the human genome, and (iii) the evolution of the genomic research.

PETs generally protect users' privacy by either breaking the link between individuals' identities and the data they provide (e.g., removing user's identities from the published genomic data), or by decreasing the information provided (e.g., by using cryptographic tools or obfuscation techniques). Both techniques might reduce the reliability and the accuracy of the genomic data. Thus, a major issue to be addressed when designing PETs is limiting private information leakage while keeping an acceptable level of reliability and accuracy of the genomic data for the researchers and medical units.

Moreover, developing PETs for genomic data has many unique challenges, due to the architecture of the human genome. The human genome is encoded in double stranded DNA molecules consisting of two complementary polymer chains. Each chain consists of simple units called nucleotides (A, C, G, T). The human genome consists of approximately three billion letters. Existing privacy-preserving methods do not scale to these large genomic data sizes; hence current algorithms are inadequate for privacy protection on the genomic level.

Finally, the rapid evolution in the field of genomics produces many new discoveries every year, which cause significant changes in the known facts. For example, the sensitivity of certain genomic information will change over time; hence it is crucial to develop dynamic algorithms that can smoothly adapt to this rapid evolution.

Due to the sensitivity of genomic data, research on the privacy of genomic data has considerably accelerated over the past few years. We can put the research on genomic privacy in three main categories: (i) private string searching and comparison, (ii) private release of aggregate data, and (iii) private clinical genomics.

In [6], Troncoso-Pastoriza et al. propose a protocol for string searching, which is then improved by Blanton and Aliasgari [7]. In this approach, one party with his own DNA snippet can verify the existence of a short template within his snippet by using a Finite State Machine in an oblivious manner. To compute the similarity of DNA sequences, in [8], Jha et al. propose techniques for privately computing the edit distance of two strings by using garbled circuits. In [9], Bruekers et al. propose privacy-enhanced comparison of DNA profiles for identity, paternity and ancestry tests using homomorphic encryption. Similar to our work, in [10], Kantarcioglu et al. propose using homomorphic encryption to perform scientific investigations on integrated genomic data. In their scheme, all genomic data is encrypted by the same public key of the data storage site, and there is a single key holder site which can decrypt everything. Thus, a curious party at the key holder site can obtain the genomic information of all users in case of a possible data leakage from the data storage site. Moreover, in [10], only the encrypted variants (i.e., positions in the genome holding a nucleotide that varies between individuals) of the users are stored at the data storage site along with their plaintext locations (on the DNA), which can leak substantial information to the data storage site about the genomic sequences of the users, as we discuss in Section 2.4. As opposed to [10], we focus on personal use of genomic data (e.g., in medical tests and personalized medicine methods), propose methods in which each user's genomic data is encrypted via his own cryptographic key, and prevent the leakage of genomic data due to statistical dependence between the variants. In one of the recent works [11], Baldi et al. make use of both medical and cryptographic tools for privacy-preserving paternity tests, personalized medicine, and genetic compatibility tests. Instead of utilizing public key encryption protocols, in [12], Canim et al. propose securing the biomedical data using cryptographic hardware. Finally, in [13], Eppstein et al. propose a privacy-enhanced method for comparing two compressed DNA sequences by using Invertible Bloom Filter [14].

When releasing databases consisting of aggregate genomic data (e.g., for research purposes), it is shown that known privacy-preserving approaches (e.g., de-identification) are ineffective on (un-encrypted) genomic data [15, 16]. Homer et al. [17] prove that the presence of an individual in a case group can be determined using aggregate allele frequencies and his DNA profile. In another recent study [18], Gitschier shows that a combination of information, from genealogical registries and a haplotype analysis of the Y chromosome collected for the HapMap project, allows for the prediction of the surnames of a number of individuals held in the HapMap database. Thus, releasing genomic data (even in aggregate form) is currently banned by many institutions due to this privacy risk. In [19], Zhou et al. study the privacy risks of releasing the aggregate genomic data. They propose a risk-scale system to classify aggregate data and a guide for the release of such data. Recently, using differential privacy was proposed by Fienberg et al. [20]; they aim to ensure that two aggregated databases, differing from each other by only one individual's data (e.g., DNA sequence), have indistinguishable statistical features.

Recently, in [21], utilizing a public cloud, Chen et al. propose a secure and efficient algorithm to align short DNA sequences to a reference (human) DNA sequence (i.e., read mapping). Finally, in [22], Wang et al. propose a privacy-protection framework for important classes of genomic computations (e.g., search for homologous genes), in which they partition a genomic computation, distributing sensitive data to the data provider and the public data to the data user.

In this work, we focus on medical tests (e.g., disease-susceptibility test) and personalized medicine methods by using users' genomic data while protecting user's genomic privacy. As a result of our extensive collaboration with geneticists, clinicians, and biologists, we conclude that DNA string comparison is insufficient in many medical tests (that use genomic data) and would not be enough to pave the way to personalized medicine. As it will become clearer in the next sections, specific variants must be considered individually for each genetic test. Thus, as opposed to the above private string search and comparison techniques, which focus on privately comparing the distance between the genomic sequences, we use the individual variants of the users to conduct genetic disease susceptibility tests and develop personalized medicine methods. We consider the individual contribution of each variant to a particular disease, for which a string comparison algorithm (such as Private Set Intersection[23, 24]) would not work. Further, in our proposed algorithms, we consider the statistical relationship between the variants for the genomic privacy of the users. In addition, we make use of a Storage and Processing Unit (SPU) between the user (patient) and the medical unit to store the genomic data in encrypted form and make computations on it using homomorphic encryption and proxy encryption.

The aim of the present invention is to propose a privacy-enhancing method for medical tests and personalized medicine methods, which utilize patients' genomic data. It is proposed a method to process genomic data comprising the steps of:

at an initialization stage:

The method of the invention is split into a first phase in which the DNA sequence is processed and stored in the SPU and a second phase during which a test is carried out.

During the first phase, the DNA sequence, produced by an authorized laboratory, is processed and encrypted as explained above. During this second phase, the medical test selected by the medical unit is carried out without having the possibility to retrieve all information of the patient.

The method proposed by the invention is based on the use of homomorphic encryption and proxy encryption. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by their public keys at a Storage and Processing Unit (SPU).

The proposed algorithm lets the SPU (or the medical unit) process the encrypted genomic data for medical tests and personalized medicine methods while preserving the privacy of patients' genomic data. We extensively analyze the relationship between the storage cost (of the genomic data), the level of genomic privacy (of the patient), and the characteristics of the genomic data. Furthermore, we implement and show via a complexity analysis the practicality of the proposed schemes. Finally, we evaluate the security of the proposed schemes and propose new research directions on genomic privacy.

The invention will be better understood thanks to the attached figures in which:

The FIG. 1 illustrates the General architecture between the patient, SPU, and the medical unit.

The FIG. 2 illustrates the Privacy-preserving protocol for disease-susceptibility test using Method 1 or Method 2.

The FIG. 3 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the number of LD pairs per SNP (i.e., μ(k)) and storage redundancy.

The FIG. 4 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the LD strength between two SNPs (i.e., μ(l)) and storage redundancy.

The FIG. 5 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying standard deviation and mean values of the number of LD pairs per SNP (i.e., σ(k) and μ(k)) and storage redundancy.

The FIG. 6 illustrates the Average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying standard deviation and mean values of the LD strength between two SNPs (i.e., σ(l) and μ(l)) and storage redundancy.

The FIG. 7 illustrates the Increase in genomic privacy of different types of patients with 100% increments in the storage redundancy. For example, increasing the storage redundancy from 400% to 500% would increase the privacy of Patient A (who carries mostly low severity real SNPs) by 5%, whereas the same scenario increases the privacy of Patient B (who carries mostly high severity SNPs) by 13%.

The FIG. 8 illustrates the Level of genomic privacy, as defined by (8), for different types of patients with varying storage redundancy.

The FIG. 9 illustrates the Privacy-preserving protocol for disease-susceptibility test using Method 3.

The FIG. 10 illustrates the Privacy, practicality, and storage overhead comparison of the proposed methods.

In the present case, we study the privacy issues of medical tests and personalized medicine methods. Most medical tests and personalized medicine methods (that use genomic data) involve a patient and a medical unit. The patient is identified by a patient identification (ID), which could be a user name or a pseudonym (e.g., hash value of his social security number). In general, the medical unit is the family doctor, a physician, a pharmacist, or a medical council. In this study, we consider a malicious medical unit as the potential attacker. That is, a medical unit can be a malicious institution trying to obtain private information about a patient. Even if the medical unit is non-malicious, it is extremely difficult for medical units to protect themselves against the misdeeds of a hacker or a disgruntled employee. Similarly, the genomic data is too sensitive to be stored on users' personal devices (mostly due to security, availability, and storage issues), hence it is risky to leave the users' genomic data in their own hands. In addition, extreme precaution is needed between the patient and the medical unit due to the sensitivity of genomic data. Thus, we believe that a Storage and Processing Unit (SPU) should be used to store and process the genomic data. We note that a private company (e.g., cloud storage service), the government, or a non-profit organization could play the role of the SPU. We also assume that the SPU is an honest organization, but it might be curious (e.g., existence of a curious party at the SPU), hence genomic data should be stored at the SPU in encrypted form (i.e., the SPU should not be able to access the content of patients' genomic data). This general architecture is illustrated in FIG. 1. For the simplicity of presentation, in the rest of this section, we will focus on a particular medical test (namely, computing genetic disease susceptibility). We note that similar techniques would apply for other medical tests and personalized medicine methods. In a typical disease-susceptibility test, a medical unit (MU) wants to check the susceptibility of a patient (P) to a particular disease X (i.e., probability that the patient P will develop disease X). It is shown that a genetic disease-susceptibility test can be realized by analyzing particular Single Nucleotide Polymorphisms (SNPs) of the patient via some operations, such as weighted averaging [25] or Likelihood Ratio (LR) test [26]. A SNP is a position in the genome holding a nucleotide (A, T, C or G), which varies between individuals. For example, it is reported that there are three particular genes bearing a total of ten particular SNPs necessary to analyze a patient's susceptibility to Alzheimer's disease[27]. Each SNP contributes to the susceptibility in a different amount and the contribution amount of each SNP is determined by previous studies on case and control groups (these studies are published in several papers). Furthermore, some of the SNPs contribute to the development of a disease, whereas some are protective.

In general, there are two alleles observed at a given SNP position: (i) The major allele is the most frequently observed nucleotide, and (ii) the minor allele is the rare nucleotide. Everyone inherits one allele of every SNP location from each of his parents. If an individual receives the same allele from both parents, he is said to have a homozygous variant for that SNP location. If, however, he inherits a different allele from each parent (one minor and one major), he has a heterozygous variant. There are approximately 40 million approved variants (SNPs) in the human population as of now (according to the NCBI dbSNP [28]) and each patient carries on average 4 million SNPs (i.e., real variants) out of this 40 million. Moreover, this set of 4 million SNPs is different for each patient. From now on, to avoid confusion, for each patient, we refer to these 4 million variants as the real SNPs and the remaining non-variants (approved SNPs that do not exist for the considered patient) as the potential SNPs of the patient; when we only say “SNPs”, we mean both the real and potential SNPs.

At this point, it can be argued that these 4 million real SNPs (nucleotides) could be easily stored on the patient's computer or mobile device, instead of the SPU. However, we assert that this should be avoided due to the following issues. On one hand, the number of approved SNPs in human population continues to increase with new discoveries. Further, types of variations in human population are not limited to SNPs, and there are other types of variations such as Copy-Number Variations (CNVs), rearrangements, or translocations (our proposed privacy-preserving mechanisms can be smoothly adapted for these alternative variations), consequently the required storage per patient is likely to be considerably more than only 4 million nucleotides. This higher storage cost might still be affordable to an average patient (via desktop computers or USB drives), however, genomic data of the patient should be available any time (e.g., for emergencies), thus it should be stored at a reliable source such as the SPU. On the other hand, as we discussed before, leaving the patient's genomic data in his own hands and letting him store it on his computer or mobile device is risky, because his mobile device can be stolen or his computer can be hacked.

A potential attacker can learn about the susceptibilities of the patient to privacy-sensitive diseases if he obtains some specific real SNPs of the patient. Moreover, the knowledge of 75 real SNPs (out of approximately 4 million), if not fewer, will enable the attacker to identify a person[29]. These situations could lead to genetic discrimination such as denying a person's access to health (or life) insurance or obstructing his employment opportunities. As we discussed before, in our setting, both the MU and SPU pose a threat to the patient's privacy. On one hand, the MU can either be a malicious institution trying to obtain private information about the patient or it can be hacked by another malicious entity. On the other hand, the SPU is considered as an honest but curious entity. Thus, our goal is to build mechanisms in which the patient can preserve the privacy of his genomic sequence (his real SNPs) while enabling the MU to access his genomic data and conduct genetic tests.

We assume that the whole genome sequencing is done by a Certified Institution (CI) with the consent of the patient. Moreover, the genomic data of the patient is encrypted by the same CI (using the patient's public key) and uploaded to the SPU so that only the patient can decrypt the stored (potential or real) SNPs, and the SPU cannot access the SNPs of the patient. We are aware that the number of discovered SNPs increases with time. Thus, the patient's complete DNA sequence is also encrypted as a single vector file (via symmetric encryption using the patient's key) and stored at the SPU, thus when new SNPs are discovered, these can be included in the pool of the previously stored SNPs of the patient. We also assume the SPU does not have access to the real identities of the patients and data is stored at the SPU by using pseudonyms; this way, the SPU cannot associate the conducted genomic tests to the real identities of the patients. As an alternative, the privacy of the genomic data at the SPU can be further increased using privacy enhanced access control [30] or Oblivious RAM (G-RAM) storage [31] techniques, in which the data access patterns are completely hidden from the server (SPU). Note that even the most efficient implementation of O-RAM introduces high storage overhead to the client (patient), and it introduces 20˜25 times more overhead with respect to non-oblivious storage. Thus once it becomes more efficient, O-RAM storage could be considered as a future add-on to the proposed privacy-preserving mechanisms.

Depending on the access rights of the MU, the SPU can either (i) compute Pr(X), the probability that the patient will develop the disease X by checking the patient's encrypted SNPs via homomorphic encryption techniques [33] (In one of our proposed schemes, see Method 3 in Section 2.4, Pr(X) is computed at the MC via homomorphic operations), or (ii) provide the relevant SNPs to the MU (e.g., for complex diseases that cannot be interpreted using homomorphic operations). These access rights are defined either jointly by the MU and the patient or by the medical authorities. Further, access rights can be enforced by using a secure attributebased system as in [34]. We note that homomorphic encryption lets the SPU (or MU) compute Pr(X) using encrypted SNPs of the patient P. In other words, the SPU (or MU) does not access P's SNPs to compute his predicted disease susceptibility. We use a modification of the Paillier cryptosystem (described in Section 2.1) to support the homomorphic operations at the SPU (or MU).

We propose four different techniques for the storage and process of the SNPs at the SPU and the preservation of the patient's privacy: (i) Method 0 in Section 2.2, (ii) Method 1 in Section 2.3, (iii) Method 2 in Section 2.4, and (vi) Method 3 in Section 2.5. We describe these proposed techniques in detail in the following subsections. We also discuss the computation of genetic disease susceptibility by using homomorphic operations in Section 2.6.

In the rest of this work, for simplicity of the presentation, we do not consider the type of the variant at a real SNP location (i.e., whether the variation is homozygous or heterozygous for that real SNP); we only consider whether the patient has a real SNP or not at a particular location. However, the proposed approaches and the analysis (in Section 2.4) can easily be extended to cover the types of the variants. In order to facilitate future references, frequently used notations are listed in Table I for the different stages of the proposed schemes.

TABLE I
NOTATIONS AND DEFINITIONS.
General Notations
SNPiP Type of SNP i, SNPi, of the patient P. SNPiP ε {0, 1}, 0
representing a potential SNP (i.e., non-variant) for P, and 1
representing a real SNP (i.e., a variant) for P.
custom characterPx Predicted susceptibility of the patient P to disease X.
custom characterP Set of real SNPs of the patient P (SNPs at which P has a
variant: around 4 million at each patient).
ΩP Set of potential SNPs of the patient P (SNPs at which P does
not have a variant: around 36 million at each patient).
Cryptographic Notations
n, g Public parameters of modified Paillier cryptosystem.
x Weak private key of the patient P.
x(i) ith share of the patient P's private key.
gx Public key of the patient P.
E(m, gx) Encryption of message m with the patient P's public key.
Susceptibility Test via Weighted Averaging
pji(X) Probability that P would develop disease X, given SNPiP = j,
Pr(X|SNPiP = j).
CiX Contribution of SNPi to the susceptibility to disease X.
Susceptibility Test via Likelihood Ratios
IXP Initial risk of the patient P for disease X.
LXi(j) Likelihood Ratio (LR) when SNPi = j for disease X.

In this section, we briefly review the modified Paillier cryptosystem (described in detail in [33, 35]), which we use in this work, and its homomorphic properties. We note that the usual notation in Paillier cryptosystem is to use a pair of keys named public and secret key. However, for the present description, we will use the denote the keys as public and private.

The public key of the patient P is represented as (n, g, h=gx), where the strong private key is the factorization of n=pq (p, q are safe primes), the weak private key is xε [1, n2/2], and g of order (p−1)(q−1)/2. Such a g can be easily found by selecting a random a εZn2*, and computing g=−a2n.

Encryption of a Message:

To encrypt a message mεZn2*, we first select a random rε[1, n/4] and generate the ciphertext pair (T1, T2) as below:
T1=gr mod n2 and T2=hr(1+mn)mod n2  (1)
Re-Encryption of a Message:

An encrypted message (T1, T2) can be re-encrypted under the same public key, using a new random number r1ε[1, n/4] as below:
{circumflex over (T)}1=gr1T1 mod n2 and {circumflex over (T)}2=hr1T2 mod n2.  (2)
Decryption of a Message:

The message m can be recovered as follows:
m=Λ(T2/T1X),  (3)
where

Λ ( u ) = ( u - 1 ) mod n 2 n , for u { u < n 2 | u = 1 mod u } .
Homomorphic Properties:

Assume two messages m1 and m2 are encrypted using two different random numbers r1 and r2, under the same public key, (n, g, h=gx), such that E(m1, r1, gx)=(T11, T21) and E(m2, r2, gx)=(T12, T22) Assume also that c is a constant number. Then the below-mentioned homomorphic properties are supported by Paillier cryptosystem:

These homomorphic operations are conducted at the SPU (or MU depending on which approach is used) to compute the predicted susceptibility of the patient P to disease X, as will be discussed in Section 2.6.

Proxy encryption: The patient's weak private key x is divided (preferably randomly or by any other rule) into two shares: x(1) and x(2) (such that x=x(1)+x(2)). x(1) is given to the SPU and x(2) is given to the MU. Using the above Paillier cryptosystem, an encrypted message (T1, T2) (under the patient's public key) can be partially decrypted by the SPU (using x(1)) to generate the ciphertext pair (T1%, T2%) as below:
T1%=T2 and T2%=T2/T1x(1)mod n2.  (6)

Now, (T1%, T2%) can be decrypted at the MU using x(2) to recover the original message. x(2) can be provided to the MU once the patient is registered to the medical unit or through the patient's digital ID card. Further details about the distribution of shares are out of the scope of this paper. We note that this approach is not proxy re-encryption; it is based on secret-sharing.

Overall, this modified Paillier cryptosystem is not key optimal, because the size of the MU's and SPU's secret storages do not remain constant. That is, both the MU and SPU need to store a secret for every patient.

However, this storage cost can be considered negligible when compared to the storage of the genomic data. Further, the shares (e.g., x(1) and x(2)) can be stored by the patient and sent to the MU and SPU only when it is needed in order to resolve this storage issue at the expense of extra communication overhead. Furthermore, the above modified Paillier cryptosystem is not proxy invisible, because all participants of the systems (i.e., P, MU and SPU) should be aware of the existence of the proxy. We discuss the security evaluation of this cryptosystem in Section 3.2.

In this approach, the real SNPs of the patient are stored encrypted (via the patient's public key) and the locations of the corresponding real SNPs are stored in plaintext at the SPU.

We assume that SNP, at the patient P is represented as SNPiP and SNPiP=1, if P has a real SNP (i.e., variant) at this location, and SNPiP=0, if P does not have a variant at this location. We let γP be the set of real SNPs of the patient P (at which SNPiP=1). We also let P represent the set of potential SNPs (at which SNPiP=0).

Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique.

Method 0 in Section 2.2 might leak private information to the curious party at the SPU. As the locations of the SNPs are stored in plaintext, if the SPU only stores the real SNPs in γP, a curious party at the SPU can learn all real SNP locations of the patient, and hence, much about his genomic sequence. The nucleotides corresponding to variants at particular locations of the DNA sequence are public knowledge. Thus, even though the contents of patient's real SNPs are encrypted, a curious party at the SPU can infer the nucleotides corresponding to these SNPs from their plaintext locations. Therefore, in this method, the SPU stores the contents of both real and potential SNP locations (in {γP∪ΩP}) in order to preserve the privacy of the patient. The locations of the corresponding SNPs are again stored in plaintext at the SPU. This is because, when a particular SNP (or set of SNPs) are queried by the MU, the SPU should know which SNPs to process (or send to the MU).

As before, we assume that SNPi at the patient P is represented as SNPiP and SNPiP=1, if P has a real SNP (i.e., variant) at this location, and SNPiP=0, if P does not have a variant at this location. We let γP be the set of real SNPs of the patient P (at which SNPiP=1). We also let P represent the set of potential SNPs (at which SNPiP=0). Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique. This approach is illustrated in FIG. 2.

The above technique provides a high level of privacy and practicality for the patient, because (i) from the view point of a curious party at the SPU, inferring the locations of the patient's real SNPs with the stored information is equivalent to inferring them with no information about the patient, and (ii) the patient is not involved in the protocol after the sequencing (except for the consent between the patient and the MU for a particular test). However, this level of privacy and practicality comes at the cost of extra storage overhead at the SPU (due to the storage of both real and potential SNPs as discussed in Section 3.1).

Due to the significant storage overhead mentioned in Section 2.3, here we propose another technique that reduces the storage overhead at the SPU at the expense of decrease in privacy. In a nutshell, we leave everything the same as in Section 2.3, but, instead of storing the contents of all potential and real SNP locations, we store the real SNPs of the patient along with a certain level of redundancy (i.e., contents of some potential SNP locations). In other words, to mislead a curious party at the SPU, among the 40 million discovered SNPs, we store the approximately 4 million real SNPs (for which SNPiP=1, iεγp) along with some redundant content from Ωp (with SNPjP=0), for each patient.

Again, we assume that the location of the encrypted (real or potential) SNPs are stored in plaintext at the SPU and there exists a potential curious party at the SPU trying to infer the real SNPs of the patient (in γp). An important issue to consider in this approach is the Linkage Disequilibrium (LD) between SNPs [36].

LD occurs when SNPs at two loci (SNP positions) are not independent of each other. For simplicity, we represent the LD relationship between two SNPs i and j as Pr(SNPi|SNPj), where SNP, (or SNPj) takes values from the set {0, 1}. In compliance with genetic observations, we assume that the LD between two SNPs are not symmetric, i.e., Pr(SNPi|SNPj)≠Pr(SNPj|SNPi). We note that LD relationships are defined among all 40 million discovered SNPs, regardless of their type (i.e., real or potential) at a particular patient.

As in Section 2.3, the SPU provides the end-result of a disease-susceptibility test or the relevant SNPs to the MU. However, in this case, if a particular potential SNP (requested by the MU or needed in the susceptibility test) is not stored at the SPU (i.e., SNPjP=0), one of the following two scenarios occurs: (i) If the SPU provides the relevant SNPs to the MU, MU infers the missing potential SNPs from the reference genome (since it is known that the missing potential SNPs are not a variant for P), or (ii) if the SPU provides the end-result of the susceptibility test, the SPU uses the fact that SNPjP=0 for each missing potential SNPj.

As expected, the amount of storage redundancy (due to the storage of the content from Ωp), along with the LD between the SNPs and their characteristics, determine the level of a patient's genomic privacy.

Therefore, in the rest of this section, we analyze the relationship between the amount of redundancy, LD values, characteristics of the SNPs, and the level of privacy. To do so, first, we observe the average probability of correctly inferring the locations of P's real SNPs (in γp) considering varying amounts of redundancy and the LD values between the SNPs. That is, how much information from a patient's un-stored potential SNPs is revealed to the curious party at the SPU about the locations of his real SNPs ? This problem can also be formulated similarly if the goal of the attacker is to determine the type of the variant at a real SNP location (e.g., homozygous or heterozygous). In this case, SNPiP can take three different values from the set {0, 1, 2}, 0 representing a potential SNP (i.e., non-variant) 1 representing a real homozygous SNP, and 2 representing a real heterozygous SNP for P. It is worth noting that for this study, we create realistic models for the LD values and the characteristics of the SNPs. Further, for the created models, we try a wide range of parameters and observe a wide range of results to address most potential scenarios. However, as the field of genomics becomes more mature, our models can be replaced by the values obtained from the medical research.

We let Ωps and Ωpu denote the set of P's potential SNPs that are stored (for redundancy) and not stored at the SPU, respectively (Ωps∪Ωpup). Further, Ki is the set of SNPs with which a particular SNPi has LD, and |Ki|=k (for each SNP, these k SNPs are chosen among approximately 40 million SNPs). We assume that k≧0 and it is a truncated Gaussian random variable with only discrete values and obtained from a distribution with mean σ(k) and standard deviation (k).

Initially, we compute Pr(SNPiP) for all (real and potential) SNPs in {γp∪Ωps} by using the LD relationships between these SNPs and those in Ωpu. As all SNPs in {γp∪Ωps} are encrypted and stored at the SPU, only the LD relationships between these SNPs and the un-stored SNPs in Ωpu are helpful for the curious party.

Therefore, for each real SNPiεγp, we observe Pr(SNPiP=1|SNPmP=0) for all mε{K(i)∩≡Ωpt s}┤}, get the average of these values, and compute Pr(SNPiP=1). Similarly, for each potential SNPjεΩps, we observe Pr(SNPjP=0| SNPmP=0) for all mε{Kjcustom characterΩpu}custom character, average these values, and compute Pr(SNPjP=0). We let/be the indicator of the LD strength between two SNPs. Thus, we represent Pr(SNPiP=1|SNPmP=0)=(εγp, mε{Kicustom characterΩpu}custom character) and Pr(SNPjP=0|SNPmP=0)=/(jεΩpt s, mε{K(j)∩Ωptu}) as truncated Gaussian random variables with range [0.5, 1], obtained from a distribution with mean μ(l) and standard deviation σ(l).

Finally, if |Ki|=k=0 or |Ki∩Ωpu|=0 for a SNP i in {γp∪Ωps}, we update Pr(SNPiP=1) considering the fact that the expected value of all stored SNPs is known by the curious party as below:

1 Υ P U Ω P s j Υ P U Ω P s ( SNP j P ) × Pr ( SNP j P ) = Υ P Υ P U Ω P s . ( 7 )

In the following, we illustrate our numerical results that represent the relationship between storage, inference power of the curious party at the SPU, and LD values. We assume |γP|=4 million and |γP∪ΩP|=40 million. We define the percentage of storage redundancy at the SPU as

Ω s s Y p I × 100
and compute the average value of Pr(SNPiP=1) for a SNP in γP for varying values of μ(k), (k), μ(l), and σ(l). Higher values of Pr(SNPiP=1) indicate a higher inference power for the curious party at the SPU. We repeat each simulation 100 times to obtain an average. Note that Method 1 (in Section 2.3) is a special case of Method 2 (when the storage redundancy at the SPU is 900%), hence its privacy is the same as 900% redundancy in the following results.

In FIG. 3, we illustrate the variance in the average value of Pr(SNPiP=1) for different values of μ(k), when μ(l)=0.8, σ(l)=0.15, and σ(k)=0.75. We note that “no LD” curve in the figure represents the case in which the LD values between the SNPs are ignored. We observe that as the available side information (i.e., number of un-stored potential SNPs in Ωpu having LD with the stored ones) increases, the inference power of the curious party increases, especially for low values of storage redundancy. For example, to have the same inference power for the curious party, 200% storage redundancy is required when μ(k)=0, whereas it is 700% when μ(k)=4. Furthermore, even at the maximum (i.e, 900%) storage redundancy, the curious party still has a slight probability of inferring the variants of the patient, because it knows that 4 out of 40 million of the stored content are variants. Next, in FIG. 4, we illustrate the variance in the same probability, this time for different values of μ(l), when μ(k)=2, σ(k)=0.75, and σ(l)=0.25. For higher values of σ(l), the gap between the different μ(l) curves becomes negligible, because the distribution becomes almost uniform, rather than truncated Gaussian. As expected, the inference power of the curious party increases when the strength of LD between the SNPs increases (i.e., when μ(l) increases).

We observe that the strength of LD, however, does not affect the inference power as strong as k. Then, FIG. 3 illustrates the average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the number of LD pairs per SNP (i.e., μ(k)) and storage redundancy.

The FIG. 4 illustrates the average probability to correctly infer the locations of patient's real SNPs (for the curious party at the SPU) with varying mean values of the LD strength between two SNPs (i.e., μ(l)) and storage redundancy.

In the FIGS. 5 and 6, we show the Average{Pr(SNPiP=1)} for varying standard deviations of k and l, and with 500% storage redundancy as follows: (i) in FIG. 5, we vary σ(k) and μ(k), when μ(l)=0.8 and σ(l)=0.15, and (ii) in FIG. 6, we vary σ(l) and μ(l), when μ(k)=2 and σ(k)=0.75. We observe that the inference power of the curious party varies (either increases or decreases) with an increasing value of σ(k) (σ(l)) depending on μ(k) (μ(l)), and, as expected, all curves converge to a single value for higher values of σ(k) (σ(l)).

Next, considering the individual characteristics of the real SNPs (i.e., their severity levels), we analyze the level of genomic privacy of a patient against a curious party at the SPU. By the level of genomic privacy, we understand the level of information that a third party can infer about the real variants of a patient. The severity of a SNPi can be defined as the privacy-sensitivity of the SNP when SNPiP=1 (i.e., when it exists as a variant at the patient P). For example, a real SNP revealing the predisposition of a patient for Alzheimer's disease can be considered more severe than another real SNP revealing his predisposition to a more benign disease. Severity values of the SNPs are determined as a result of medical studies (depending on their contributions to various diseases) and tables of disease severities provided by insurance companies (e.g., percentage of invalidity). We denote the severity of a real SNP i as Vi, and 0≦Vi≦1 (1 denotes the highest severity). Thus, we define the genomic privacy of the patient P as below:

Φ p = - i Υ P ` log 2 ( Pr ( SNP i P = 1 ) ) × V i · ( 8 )

We do not use the traditional entropy metric [37, 38] to quantify privacy, as only one state of SNPiP poses privacy risks (i.e. SNPiP=1), as discussed before.

First, we study the relationship between the storage redundancy and the severity of the real SNPs by focusing on three types of patients: (i) patient A, carrying mostly low severity real SNPs (in γA), (ii) patient B, carrying mostly high severity real SNPs (in γB), and (iii) patient C, carrying mixed severity real SNPs (in γC). For each patient, the highest level of privacy is achieved when the storage redundancy is maximum (as in Method 1 in Section 2.3). Thus, we recognize this level as 100% genomic privacy for the patient. For the evaluation, we take the highest privacy level of patient C as the base and normalize everything with respect to this value. We use the following parameters for the simulation. The severities of patient A's and patient B's real SNPs are represented as truncated Gaussian random variables with (μA, σA)=(0.25, 0.15) and (μB, σB)=(0.75, 0.15), respectively. Furthermore, the severity of patient C's real SNPs are represented as a uniform distribution between 0 and 1. We also set μ(l)=0.8, σ(l)=0.25, μ(k)=2, and σ(k)=0.75. In FIG. 7, we illustrate the increase in privacy with increments in the storage redundancy for these three types of patients (A, B, and C).

We observe that by increasing the storage redundancy, a patient with high severity real SNPs gains more privacy than a patient with lower severity real SNPs, hence the storage redundancy can be customized for each patient differently based on the types of his real SNPs. It can be argued that the amount of storage redundancy for a patient can leak information (to the curious party the SPU) about the seventies of his real SNPs. However, the severity of the SNPs is not the only criteria to determine the storage redundancy for a desired level of genomic privacy as we discuss next.

Finally, we study the relationship between the severity of the real SNPs, the number of LD pairs per SNP (number of SNPs with which a particular SNP has LD, i.e., k), and the storage redundancy. We assign the Vi values of the real SNPs (in γP) following a uniform distribution between 0 and 1. We set the LD parameters as μ(l)=0.8, σ(l)=0.25, μ(k)=2, and σ(k)=1.5. Then, we observe and compare the following potential scenarios in different types of patients: (i) The real low severity SNPs of the patient (i.e., his real SNPs with low Vi values) have a higher number of LD pairs (i.e., higher k values) with respect to his high severity real SNPs (we note that, in all cases, k values are obtained from the same truncated Gaussian distribution with μ(k)=2, and σ(k)=1.5); (ii) k values are assigned randomly to the SNPs; and (iii) the real high severity SNPs of the patient (i.e., his real SNPs with high Vi values) have a higher number of LD pairs (i.e., higher k values) with respect to his low severity real SNPs. Again, we set a patient's genomic privacy to 100% when the storage redundancy is maximum at the SPU (as in Method 1 in Section 2.3). We illustrate our results in FIG. 8, and show different storage redundancy requirements for different types of patients (to provide the same level of privacy). For example, to achieve 40% genomic privacy, the SPU requires 400% storage redundancy for a patient whose less severe real SNPs have more LD pairs, whereas it requires 600% storage redundancy for another patient whose more severe real SNPs have more LD pairs (which means more storage per patient, as discussed in Section 3.1). This result also supports our belief to customize the storage redundancy for each patient.

We obtained similar patterns for further variations of the variables but we do not present these results due to the space limitation. In summary, depending on the actual μ(k), σ(k), μ(l), σ(l), and Vi values (which will be determined as a result of the medical research), the storage redundancy can be determined (and customized for each patient based on the types of his variations) for this approach to keep the genomic privacy of the patient at a desired level. Note that the curious party at the SPU cannot infer the real SNPs of the patient (or the severities of the patient's real SNPs) from the amount of customized storage redundancy, because the storage redundancy (for a desired level of genomic privacy) depends on various factors. For example, a patient with low storage redundancy (for a desired level of genomic privacy) could mean that (i) he carries mostly low severity real SNP (as in FIG. 7), (ii) he carries mixed severity real SNPs, but his less severe real SNPs have more LD pairs (as in FIG. 8), (iii) his real SNPs (regardless of their severities) have low number of LD pairs (as in FIG. 3), or (iv) his real SNPs (regardless of their seventies) have low LD strengths (as in FIG. 4).

Let LP={Li:iεγP} represent the set of locations (on the DNA sequence) of the patient P's real SNPs (in γP). As opposed to the previous two approaches, here, we propose to encrypt the locations of the SNPs along with their contents. By doing so, we save storage costs by storing only the real SNPs in γP at the SPU (around 4 million) while providing the highest level of privacy (as in Section 2.3). These benefits, however, come with a cost in the practicality of the algorithm, introducing extra steps for the patient (P) during the protocol. Although we can assume that these extra steps can easily be handled via the patient's device such as smart card or mobile device, this approach still requires more message exchanges (as will be described next) between the parties, compared to the previous two approaches.

In some environments, dividing the weak private of the patient, and distributing two shares of the weak private key to the SPU and MU might not be acceptable (e.g., when it is likely that the SPU and MU will collaborate to retrieve patient's weak private). Therefore, for the sake of completeness, in the following, we present Method 3 with and without proxy encryption (i.e., without distributing the patient's private key to other parties). The Method 1 and Method 2 can also be modified similarly to avoid proxy encryption.

The initial steps of the protocol are the same as in Section 2.3, except for Steps 2 and 3 in which the locations of the SNPs are encrypted and a Bloom filter [39] is generated. Below, we summarize the different steps of this approach (the unchanged steps are not repeated). These steps are illustrated in FIG. 9.

A Bloom filter is a simple space-efficient randomized data structure for representing a set in order to support membership queries [39]. A Bloom filter for representing a set LP is described by an array of K bits, initially all set to 0. It employs independent hash functions H1, . . . , Hγ with range {1, . . . , κ}. For every element LiεLP, the bits H1(Li), . . . , Hγ (Li) in the array are set to 1. A location can be set to 1 multiple times, but only the first change has an effect.

After constructing the Bloom filter, the CI encrypts each element in LP by using a symmetric key shared between the CI and P (established during Step 0 as in Section 2.3) and generates LEP={E(Li):iεγP}. The CI also encrypts a dummy variant (representing the potential SNPs in ΩP) along with the real SNPs of the patient (using P's public key). Furthermore, the CI associates a dummy position L0 for this dummy variant and encrypts L0 using the symmetric key between the CI and P to obtain the encrypted dummy position E(L0).

To check if Lj belongs to LP, the patient checks whether all H1(Lj), . . . , Hγ(Lj) are set to 1. If not, Lj definitely does not belong to LP. Otherwise, the patient assumes LjεLP, although this may be wrong with some probability. That is, a Bloom filter could yield a false positive, where it suggests that Lj is in LP even though it is not. This probability can be decreased at the expense of increasing Bloom filter length (i.e., κ). Further, the false positive probability can be significantly reduced by using some proposed techniques such as [40, 41]. As a result of this process

(a) If the location is in his Bloom filter (i.e., if he has a real SNP at the corresponding location), P encrypts the location with the symmetric key between the CI and P.

(b) If the location is not in his Bloom filter (i.e., if he does not have a real SNP at the corresponding location), P uses E(L0) as the encrypted location.

We note that the above operations can be easily done via the patient's device (e.g., by reading the patient's device at the MU as a consent to the test) or mobile device (e.g., by consenting via a smart phone application) by using the stored Bloom filter output, E(L0), and symmetric key between the CI and P.

In this approach, the SPU stores only the encrypted SNPs and encrypted locations. Genomic data encrypted by P's public key is only decrypted at P, and the weak private key of P remains only at P (i.e., shares of the weak private key are not distributed to the SPU or MU). Most of this approach is the same as Method 3 with proxy encryption. Indeed, the first 8 steps of the algorithm are the same, except for the distribution of parts of P's private key. The only difference is the transfer of the end-result or the relevant SNPs to the MU as follows:

We note that the security of the communication between P and the MU is provided by symmetric keys as discussed before. The above operations put some more burdens on the patient during the protocol. However, we emphasize that these operations can be smoothly done on the patient's device without requiring a substantial effort from the patient himself.

In summary, as the locations of the real SNPs are encrypted, a curious party at the SPU cannot infer the contents of the SNPs from their locations (as in Section 2.3), hence it is enough to store only the real SNPs in γP. Furthermore, the privacy provided by this approach (with or without proxy encryption) is the same as 900% redundancy in Method 2 (i.e., similar to Method 1), hence we do not discuss it again. Another advantage of this approach (i.e., Method 3 in general) is that individual contributions of the genetic variant markers remain secret at the MU, because the homomorphic operations are conducted at the MU. This advantage might become more significant when this approach is used for personalized medicine methods in which the pharmaceutical company (embodied in this case as the medical unit) does not want to reveal the genetic properties of its drugs. Thus, if introducing the described extra steps for the patient and few additional message exchanges between the parties are tolerated, this approach operates with relatively modest storage and yet provides very good privacy.

We now present the disease-susceptibility test via homomorphic operations at the SPU for Method 1 (Section 2.3) and Method 2 (Section 2.4). Similar techniques can be used for Method 3 at the MU, as discussed in Section 2.5.

The SPU uses a proper function to compute P's predicted disease susceptibility via homomorphic encryption. There are different functions for computing the predicted susceptibility. In [25], focusing on one example of many diseases that require a susceptibility test involving multiple SNPs, Kathiresan et al. propose to count the number of unfavorable alleles carried by the patient for each SNP related to a particular disease. Similarly, in [26], Ashley et al. propose to multiply the Likelihood Ratios (LRs) of the most important SNPs for a particular disease in order to compute a patient's predicted susceptibility. LR values are determined as a result of medical studies. Furthermore, a weighted averaging function can also be used, which computes the predicted susceptibility by weighting the contributions of SNPs by their contributions (e.g., LR values of the SNPs). Note that our proposed privacy-preserving mechanisms are not limited by the types of the functions (used to test the disease susceptibility). It is expected that these functions will evolve over time; hence the proposed algorithms can be developed to keep up with this evolution.

In the following, we discuss how to compute the predicted disease susceptibility at the SPU by using a toy example to show how the homomorphic encryption is used at the SPU. Initially, we assume that the function at the SPU is weighted averaging (which is an advanced version of the function proposed in [25]) and show how the predicted susceptibility is computed using encrypted SNPs. Then, we show how the function proposed in [26] (i.e., multiplication of LR values) can be utilized at the SPU.

Assume that (for simplicity) the susceptibility to disease X is determined by the set of SNPs Ω={SNPm,SNPn}, which occur at particular locations of the DNA sequence. SNPmP and SNPnP are not necessarily among the real SNPs of the patient P (i.e., P does not need to have a variant at those locations). The contributions of different states of SNPiP for iε{m, n} to the susceptibility to disease X are computed via previous studies (on case and control populations) and they are already known by the MU. That is, p0i(X) @Pr(X|SNPiP=0) and p1i(X) @Pr(X|SNPiP=1) (iε{m, n}) are determined and known by the MU. Further, the contribution (e.g., LR value) of SNPi to the susceptibility to disease X is denoted by CiX. Note that these contributions are also computed by previous studies on case and control groups and they are known by the MU.

As we have discussed before, the SPU stores the set of SNPs of the patient P, encrypted by P's public key (n, g, h=gx). Encryption is done using the modified Paillier cryptosystem as discussed in Section 2.1. Thus, the SPU uses E(SNPmP, gx) and E(SNPnP, gx) for the computation of predicted susceptibility of P to disease X. From now on, we drop the r values in the above encrypted messages for the clarity of the presentation (r values are chosen randomly from the set [1, n/4] for every encrypted message as discussed in Section 2.1). Similarly, the MU provides the following to the SPU in plaintext: (i) the markers for disease X (SNPm and SNPn), (ii) corresponding probabilities pji(X), iε{m, n} and jε{0,1}, and (iii) the contributions of each SNP CiX.

Next, the SPU encrypts j(jε{0,1}) using P's public key to obtain E(0, gx) and E(1, gx) for the homomorphic computations. This encryption can also be done at the MU and sent to the SPU. Alternatively, we might assume that SNPs of a patient are stored at the SPU in pairs of {E(|SNPiP−0|, gx), E(|SNPiP−1|, gx)} for each SNPiP, instead of the actual values of the SNPs. In this case, the above encryption at the SPU would not be required.

The SPU computes the predicted susceptibility of the patient P to disease X by using weighted averaging.

This can be computed in plaintext as below:

S P X = 1 C m X + C n X × i m , n C i X { P 0 i ( X ) ( 0 - 1 ) [ SNP i P - 1 ] + P 1 i ( X ) ( 1 - 0 ) [ SNP i P - 0 ] } . ( 9 )

The computation in (9) can be realized using the encrypted SNPs of the patient (and utilizing the homomorphic properties of the Paillier cryptosystem) to compute the encrypted disease susceptibility, E(SPX, gx) as below:

E ( S P X , g x ) = { i { m , n } { [ E ( SNP i P , g x ) E ( 1 , g x ) - 1 ] V i 1 × [ E ( SNP i P , g x ) E ( 0 , g x ) - 1 ] V i 2 } C i X } θ , where ( 10 ) Δ i 1 = P 0 i ( X ) 0 - 1 , ( 11 a ) Δ i 2 = P 1 i ( X ) 1 - 0 , ( 11 b ) Θ = 1 C m X + C m X . ( 11 c )

We note that the end-result in (10) is encrypted by P's public key.

Then, the SPU partially decrypts the end-result E(SPX, gx) using its share (x(1)) of P's private key (x) as discussed in Section 2.1 to obtain E(SPX, gx(2)) and sends it to the MU. Finally, the MU decrypts E(SPX, gx(2)) using its share (x(2)) of P's private key to recover the end-result SPX.

In some genetic tests, the types of the real SNPs (e.g., homozygous or heterozygous) become also important. In this case, SNPiP can take three different values from the set {0, 1, 2} to represent a potential SNP (i.e., nonvariant), a real homozygous SNP, and a real heterozygous SNP, respectively. In such a scenario, to conduct the disease-susceptibility test via homomorphic operations, the SPU should store the squared values of the SNPs. That is, for each SNPiP of the patient P, the SPU should store E((SNPiP)2, gx). Depending on the types of genomic tests that would be supported by the SPU (and the functions required for these tests), the format of storage of patient's SNPs can be determined beforehand, and SNPs can be stored accordingly just after the sequencing process.

We now assume that the predicted disease susceptibility is computed from the multiplication of Likelihood Ratios (LRs) of the corresponding SNPs as in [26] and show how such a computation would be handled at the SPU by using homomorphic operations.

In this approach, the predicted disease susceptibility is computed by multiplying the initial risk of the patient (e.g., for disease X) by the LR value of each SNP related to that disease (LR value of a SNP i depends on the value of SNPiP at the patient P). The initial risk of the patient P for the disease X is represented as IXP. We note that IXP is determined by considering several factors (other than patient's genomic data) such as patient's age, gender, height, weight, and environment. Thus, this initial risk can be computed directly by the MU. We also note that if the LR value corresponding to a particular SNP is less than one, the risk for the disease decreases. Otherwise, if the LR value is greater than one, the risk increases for the corresponding disease.

Similar to before, we assume that the susceptibility to disease X is determined by the set of SNPs in Ω={SNPm,SNPn}. We denote the LR values due to SNPiP=0 and SNPiP=1 for disease X as LXi (0) and LXi (1), respectively.

The SPU stores the SNPs of the patient P, encrypted by P's public key. The MU sends the following to the SPU: (i) LXi (j) values (iε{m, n} and jε{0,1}) in plaintext, and (ii) the markers for disease X. The MU also encrypts the log of initial risk value, ln(IXP), by P's public key and sends E(ln(IXP), gx) to the SPU. Alternatively, the contribution of the initial risk to the disease susceptibility can be included to the end-result at the end, at the MU.

The Paillier cryptosystem does not support multiplicative homomorphism in ciphertext (it only supports the multiplication of a ciphertext with a constant as discussed in Section 2.1). Thus, instead of multiplying the LR values, we propose using addition in log-domain at the SPU. Thus, the SPU computes the predicted susceptibility of P to disease X as below:

E ( ln ( S P X , g x ) = E ( ln ( I P X , g x ) × i m , n { [ E ( SNP i P , g x ) · E ( 1 , g x ) - 1 ] Ξ i 1 × [ E ( SNP i P , g x ) · E ( 0 , g x ) - 1 ] Ξ i 2 } where ( 12 ) Ξ i 1 = ln ( L X i ( 0 ) ) ( 0 - 1 ) ( 13 a ) Ξ i 2 = ln ( L X i ( 1 ) ) ( 1 - 0 ) ( 13 b )

We note that (12) corresponds to the below computation in plaintext:

ln ( S P X ) = ln ( I X P ) + i m , n { [ SNP i P - 1 ] × ln ( L X i ( 0 ) ) ( 0 - 1 ) + [ SNP i P - 0 ] × ln ( L X i ( 1 ) ) ( 1 - 0 ) } . ( 14 )

As before, the SPU partially decrypts E(ln(SX P), gx) using x(1) (its share of P's private key) to obtain E(ln(SPX), gx(2)) and sends it to the MU. Finally, the MU decrypts E(ln(SPX), gx(2)) using x(2) (its share of P's private key) to recover ln(SPX), and computes

( ln ( S P x ) )
to obtain SPX. Similar to weighted averaging, if the types of the real SNPs are used for the test (in which there are three possible states for SNPiP), squared values of the SNPs should be stored at the SPU for each patient.

In FIG. 10, based on the discussion in the previous sections, we graphically compare the proposed methods considering the level of privacy they provide, their practicality (for the patient), and their storage requirements (at the SPU). In this section, we report our findings about the complexity and security of the proposed methods.

To evaluate the practicality of the proposed privacy-preserving algorithms, we implemented them, and assessed their storage requirements and computational complexities on Intel Core i7-2620M CPU with 2.70 GHz processor under Windows 7 Enterprise 64-bit Operating System. We set the size of the security parameter (n in Paillier cryptosystem in Section 2.1) to 1024 bits. We computed the disease susceptibility using weighted averaging (at the SPU or MU, see Section 2.6.1 as well as LR test in Section 2.6.2 which also has similar complexity) and real SNP profiles from [42]. Our implementation relies on a MySQL 5.5 database managed by the open source tool MySQL Workbench. To provide a platform-independent implementation, we used the Java programming language along with the open-source Integrated Development Environment, NetBeans IDE 7.1.1., for the implementation of the Java code. We note that our code for the implementation is not optimized, and better results can be expected with an optimized implementation.

In Table II, we summarize the computational and storage complexities of the proposed methods at (i) Certified Institution (CI), (ii) SPU, (iii) MU, and (iv) P. We evaluate the proposed methods considering the following costs: (i) encryption of patient's variants, (ii) disease-susceptibility test at the SPU via homomorphic operations (using ten variants), (iii) decryption of the end-result (or relevant SNPs), (iv) proxy encryption, and (v) storage costs, in which θ represent the percentage of storage redundancy at the SPU. We did not explicitly implement the Bloom filter (for Method 3) and symmetric encryption/decryption between the parties for the security of the communication. However, the computational costs due to these operations are negligible compared to Paillier encryption/decryption and homomorphic operations.

We emphasize that the encryption of the variants at the CI is a one-time operation and is significantly faster than the sequencing and analysis of the sequence (which takes days). Further, this encryption can be conducted much more efficiently by computing some parameters, such as (gr, hr) pairs, offline for various r values, for each patient. Indeed, by computing (gr, hr) pairs offline, we observe that the encryption takes only 0.017 ms per variant at the CI.

TABLE 2
Computational and Storage Complexities of the Proposed Methods
Method 1 and Method 2
@CI @SPU @MU
Paillier Encryption Homomorphic Operations Proxy Encryption Storage Paillier Decryption
30 ms./variant 1 sec. (10 variants) 2 ms. 500 × ( 1 + ϑ 100 ) MB / patient 26 ms.
Method 3 with proxy encryption
@CI @SPU @MU
Paillier Encryption Proxy Encryption Storage Homomorphic Operations Paillier Decryption
30 ms./variant 2 ms. 500 MB/patient 1 sec. (10 variants) 26 ms.
Method 3 without proxy encryption
@CI @SPU @MU @P
Paillier Encryption Storage Homomorphic Operations Paillier Decryption
30 ms./variant 500 MB/patient 1 sec. (10 variants) 26 ms.

It is also possible to conduct private statistical tests (by a medical researcher) on the data stored at the SPU in order to get statistics about the variants of multiple patients. Conducting such a statistical test for a variant (about its type) on 100K patients takes around 55 minutes at the SPU and scales linearly with the number of patients. Note that such a statistical test is only possible with Method 1 or Method 2; using Method 3 and querying the encrypted locations of SNPs from 100K patients is not practical for this application.

In summary, all these numbers show the practicality of our privacy-preserving algorithms.

The proposed schemes preserve the privacy of patients' genomic data relying on the security strength of modified Paillier cryptosystem (in Section 2.1). The extensive security evaluation of the modified Paillier cryptosystem can be found in [33]. Below we summarize two important security features of this cryptosystem.

Finally, if the weak private key of the patient, x, is randomly divided and distributed to the Storage and Processing Unit (SPU) and Medical unit (MU) as in Method 1, this weak private key could be revealed if the MU colludes with the SPU, but the factors n, p, and q remain secret. We note that such a collusion is not considered in this study. However, for the sake of completeness, in Section 2.5.2, we present an alternative approach (Method 3 without proxy encryption) that avoids distributing the patient's weak private key to other parties, hence is robust against such a collusion.

Hubaux, Jean-Pierre, Telenti, Amalio, Fellay, Jacques, Ayday, Erman, Raisaro, Jean Louis, McLaren, Paul Jack, Rougemont, Jacques, Humbert, Mathias

Patent Priority Assignee Title
10402588, Jun 19 2013 SOPHIA GENETICS S.A. Method to manage raw genomic data in a privacy preserving manner in a biobank
10528758, May 02 2014 Intertrust Technologies Corporation Genomic informatics service
10599753, Nov 11 2013 Amazon Technologies, Inc Document version control in collaborative environment
10686788, Feb 07 2014 Amazon Technologies, Inc. Developer based document collaboration
10691877, Feb 07 2014 Amazon Technologies, Inc Homogenous insertion of interactions into documents
10877953, Nov 11 2013 Amazon Technologies, Inc. Processing service requests for non-transactional databases
11240033, Sep 26 2019 MERATIVE US L P Secure DNA-based password
11336648, Nov 11 2013 Amazon Technologies, Inc Document management and collaboration system
Patent Priority Assignee Title
20100121872,
/////////
Executed onAssignorAssigneeConveyanceFrameReelDoc
Sep 10 2013ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE (EPFL)(assignment on the face of the patent)
Mar 05 2015FELLAY, JACQUESECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 06 2015HUBAUX, JEAN-PIERREECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 06 2015RAISARO, JEAN LOUISECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 06 2015TELENTI, AMALIOECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 06 2015HUMBERT, MATHIASECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 13 2015MCLAREN, PAUL JACKECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Mar 13 2015ROUGEMONT, JACQUESECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Apr 16 2015AYDAY, ERMANECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS 0366350790 pdf
Date Maintenance Fee Events
Aug 12 2019M2551: Payment of Maintenance Fee, 4th Yr, Small Entity.
Oct 16 2023REM: Maintenance Fee Reminder Mailed.
Apr 01 2024EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
Feb 23 20194 years fee payment window open
Aug 23 20196 months grace period start (w surcharge)
Feb 23 2020patent expiry (for year 4)
Feb 23 20222 years to revive unintentionally abandoned end. (for year 4)
Feb 23 20238 years fee payment window open
Aug 23 20236 months grace period start (w surcharge)
Feb 23 2024patent expiry (for year 8)
Feb 23 20262 years to revive unintentionally abandoned end. (for year 8)
Feb 23 202712 years fee payment window open
Aug 23 20276 months grace period start (w surcharge)
Feb 23 2028patent expiry (for year 12)
Feb 23 20302 years to revive unintentionally abandoned end. (for year 12)