In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by a patient's public keys at a storage and processing unit (SPU). A part of the corresponding private key is also stored on the SPU. At the time of the test by a medical unit (MU), the patient provides the second part of the private key to the MU. A test with its associated markers is determined by the MU and sent to the SPU. The test is carried out on the encrypted values thanks to homomorphic operation and returned back to the MU. The latter uses the second part of the private key to access the result.
|
1. A method to process genomic data comprising the steps of:
associating, by a certified institution, a patient identification for a given patient;
generating, by the certified institution, a pair of asymmetric keys related to said patient comprising a private and a public key;
analyzing, by the certified institution, an output of a deoxyribonucleic acid (DNA) sequencer and preparing an aligned genomic data for said patient comprising approved variants, said approved variants being approved by medical authorities, each approved variant representing a position in the genome and a value representing a nucleotide that varies between individuals;
extracting, by the certified institution, real and potential variants from said approved variants, said real and potential variants having each a position, said real variants being a subset of the approved variants and being different for each human being, said potential variants being the remaining part of the approved variants;
selecting, by the certified institution, all or part of the potential variants;
analyzing, by the certified institution, the correlation between the selected potential variants and a privacy sensitivity of the real variants;
selecting, by the certified institution, a number of other potential variants, said number being determined according to the previous analysis and a level of privacy required;
encrypting, by the certified institution, the value of each real variant and selected potential variants with the public key of the patient;
storing, by the certified institution, the encrypted values with their respective positions and the patient identification into a storage and processing unit;
dividing, by the certified institution, the private key into at least a first and a second part;
storing, by the certified institution, the second part of the private key in the certified institution or in a patient device;
transmitting, by the certified institution, the first part of the private key to the storage and processing unit;
selecting by a medical unit a personalized clinical test to be carried out and related genetic markers, each marker having a position and a contribution;
determining, by the medical unit, the contribution of each marker according to the personalized clinical test selected;
receiving, by the storage and processing unit from the medical unit, genetic markers related to the personalized clinical test, the respective contributions of the related genetic markers and the patient identification of the patient;
retrieving by the storage and processing unit the encrypted values for said patient matching the position of the related genetic markers;
executing by the storage and processing unit a genetic test by using the retrieved values, and the contribution of the respective genetic markers thanks to homomorphic operations;
partially decrypting by the storage and processing unit the result of the genetic test using the first part of the private key;
sending by the storage and processing unit the decrypted result to the medical unit;
whereby the medical unit can use the second part of the private key to obtain the result of the performed personalized clinical test.
2. The method to process genomic data according to
generating, by the certified institution, a dummy variant comprising a dummy position and a dummy value, said dummy position being outside of the overall variant positions of a sequence;
encrypting, by the certified institution, the positions of the real variants with the symmetric key of the patient;
encrypting, by the certified institution, the dummy value with the public key of the patient;
encrypting, by the certified institution, the position of the dummy variant with the symmetric key of the patient;
storing, by the certified institution, together with the encrypted variants, the dummy variant as well as the encrypted positions and the encrypted dummy position into a storage and processing unit;
storing, by the certified institution, the position of the dummy variant into the patient device;
determining by the certified institution a set of positions which are common between the marker's position and the real variant's positions;
receiving by the certified institution from the medical unit an encrypted set of positions with the symmetric key of said patient, and for the marker's positions not present in the variant's position, dummy positions;
sending by the certified institution to the storage and processing unit the encrypted marker's positions as well as the patient identification; and
retrieving by the storage and processing unit the encrypted values for said patient at these encrypted locations and sending them to the medical unit.
|
This application is a U.S. National Stage Application of International Application No. PCT/EP2013/068658 filed Sep. 10, 2013, which claims priority to European Application No. 12184372.6 filed Sep. 14, 2012 and U.S. Provisional Patent Application No. 61/700,897 filed Sep. 14, 2012.
In this invention, we propose privacy-enhancing technologies for medical tests and personalized medicine methods, which utilize patients' genomic data. First, we highlight the potential privacy threats on genomic data and the challenges of providing privacy-preserving algorithms. Then, focusing specifically on a typical disease-susceptibility test, we develop a new architecture (between the patient and the medical unit) and propose privacy-preserving algorithms.
Privacy control can be defined as the ability of individuals to determine when, how, and to what extent information about themselves is revealed to others. In this way, the usage of private data will remain in context and it will be used exclusively for the purpose the data owner has in mind. Privacy is usually protected by both legal and technological means. By using legal actions, such as data protection directives and fair information practices, privacy regulations can enforce privacy protection on a large scale. Yet, this approach is mostly reactive, as it defines regulations after technologies are put in place. To avoid this issue, Privacy-Enhancing Technologies (PETs) [1-3] can be incorporated into the design of new systems in order to protect individuals' data. PETs protect privacy by eliminating or obfuscating personal data, thereby preventing misuse or involuntary loss of data, without affecting the functionality of the information system [4].
Their objective is to make it difficult for a malicious entity to link information to specific users. In order to obfuscate personal data, PETs often rely on cryptographic primitives, such as anonymous authentication and encryption.
Genomics is becoming the next significant challenge for privacy. The price of a complete genome profile has plummeted below $200 for genome-wide genotyping (i.e., the characterization of about one million common genetic variants), which is offered by a number of companies (located mostly in the US). Whole genome sequencing is also offered through the same direct-to-consumer model (but at a higher price). This low cost of DNA sequencing will break the physician/patient connection, because private citizens (from anywhere in the world) can have their genome sequenced without involving their family doctor. This can open the door to all kinds of abuse, not yet fully understood.
As a result of the rapid evolution in genomic research, substantial progress is expected in terms of improved diagnosis and better preventive medicine. However, the impact on privacy is unprecedented, because (i) genetic diseases can be unveiled, (ii) the propensity to develop specific diseases (such as Alzheimer's) can be revealed, (iii) a volunteer accepting de facto to have his genomic code made public (as it already happened) can leak substantial information about his ethnic heritage and genomic data of his relatives (possibly against their will), and (iv) complex privacy issues can arise if DNA analysis is used for criminal investigations and insurance purposes. Such issues could lead to genetic discrimination (e.g., ancestry discrimination or discrimination due to geographic mapping of people). Even though the Genetic Information Non-discrimination Act (GINA), which prohibits the use of genomic information in health insurance and employment, attempted to solve some of these problems in the US, these types of laws are very difficult to enforce.
An even more severe case, currently not widely considered, is where a malicious party initiates a cross-layer attack by utilizing privacy-sensitive information belonging to a person retrieved from different sources (e.g., genomic data, location, online social network, etc.), thus creating the opportunity for a large variety of fraudulent uses of such data. For example, as stated in the Personal Genome Project (PGP) consent form [5], a malicious party could make synthetic DNA of a person and plant it at a crime scene to falsely accuse him.
In this hypothetical situation, the attacker can make his accusation stronger if he has the location patterns of the person to be blamed, and hence, knows that the person was close to the crime scene at the time of the crime. Similarly, an attacker can easily obtain information on close relatives of a target from online social network data, thus effectively increasing the potential access to target's genomic data if his relatives' DNA has been sequenced. In other words, even if the person has perfect privacy on his own genome, if the attacker has access to the DNA sequence of the relatives, he can obtain significant information about the person's DNA sequence.
Even though, at this stage, the field of genomics is generally free from serious attacks, it is likely that the above threats will become more serious as the number of sequenced individuals becomes larger. Such was the case of the Internet that was initially run and used by well-intentioned researchers. However, once it became more widely used, it became plagued by uncountable attacks such as spyware, viruses, spam, botnets, Denial-of-Service attacks, etc. Therefore, the need to adapt PETs to personal genomic data will only grow with time, as they are key tools for preventing an adversary from linking particular genomic data to a specific person or from inferring privacy-sensitive genomic data about a person.
It is obvious that users need to reveal personal and even privacy-sensitive information for genomic tests (e.g., paternity tests, disease-susceptibility tests, etc.). Nevertheless, they want to control how this information is used by the service providers (e.g., medical units such as healthcare centers or pharmaceutical companies, depending on the type of the test). Currently, the companies and hospitals that perform DNA sequencing store the genomic data of their customers and patients. Of course, tight legislation regulates their activities, but it is extremely difficult for them to protect themselves against the misdeeds of a hacker or a disgruntled employee. In a non-adversarial scenario, however, making use of this data requires legitimate professionals (e.g., physicians and pharmacists) to access the data in some way. Therefore, new architectures and protocols are needed to store and process this privacy-sensitive genomic data, while still enabling its utilization by the service providers (e.g., medical units).
In this work, our goal is to protect the privacy of users' genomic data while enabling medical units to access the genomic data in order to conduct medical tests or develop personalized medicine methods. In a medical test, a medical unit checks for different health risks (e.g., disease susceptibilities) of a user by using specific parts of his genome. Similarly, to provide personalized medicine, a pharmaceutical company tests the compatibility of a user on a particular medicine, or a pharmacist checks the compatibility of a given medicine (e.g., over-the-counter drug) to a given user. In both scenarios, in order to preserve his privacy, the user does not want to reveal his complete genome to the medical unit or to the pharmaceutical company. Moreover, in some scenarios, it is the pharmaceutical companies who do not want to reveal the genetic properties of their drugs. To achieve these goals, we propose to store the genomic data at a Storage and Processing Unit (SPU) and conduct the computations on genomic data utilizing homomorphic encryption and proxy encryption to preserve the privacy of the genomic data.
The rest of the paper is organized as follows. In the rest of this section, we discuss the challenges in genomic privacy and summarize the related work on genomic privacy. In Section 2, we describe our proposed schemes for privacy-preserving medical tests and personalized medicine. Furthermore, we analyze the level of privacy provided by the proposed schemes for different design and genomic criteria. Then, in Section 3, we discuss the implementation of the proposed schemes and present their complexity and security evaluations.
Finally, in Section 4, we conclude the paper and discuss new research directions on genomic privacy.
Obviously, there are certain obstacles for achieving our goals on genomic privacy. These are mostly due to (i) the balance between privacy and reliability of the genomic data, (ii) the structure of the human genome, and (iii) the evolution of the genomic research.
PETs generally protect users' privacy by either breaking the link between individuals' identities and the data they provide (e.g., removing user's identities from the published genomic data), or by decreasing the information provided (e.g., by using cryptographic tools or obfuscation techniques). Both techniques might reduce the reliability and the accuracy of the genomic data. Thus, a major issue to be addressed when designing PETs is limiting private information leakage while keeping an acceptable level of reliability and accuracy of the genomic data for the researchers and medical units.
Moreover, developing PETs for genomic data has many unique challenges, due to the architecture of the human genome. The human genome is encoded in double stranded DNA molecules consisting of two complementary polymer chains. Each chain consists of simple units called nucleotides (A, C, G, T). The human genome consists of approximately three billion letters. Existing privacy-preserving methods do not scale to these large genomic data sizes; hence current algorithms are inadequate for privacy protection on the genomic level.
Finally, the rapid evolution in the field of genomics produces many new discoveries every year, which cause significant changes in the known facts. For example, the sensitivity of certain genomic information will change over time; hence it is crucial to develop dynamic algorithms that can smoothly adapt to this rapid evolution.
Due to the sensitivity of genomic data, research on the privacy of genomic data has considerably accelerated over the past few years. We can put the research on genomic privacy in three main categories: (i) private string searching and comparison, (ii) private release of aggregate data, and (iii) private clinical genomics.
In [6], Troncoso-Pastoriza et al. propose a protocol for string searching, which is then improved by Blanton and Aliasgari [7]. In this approach, one party with his own DNA snippet can verify the existence of a short template within his snippet by using a Finite State Machine in an oblivious manner. To compute the similarity of DNA sequences, in [8], Jha et al. propose techniques for privately computing the edit distance of two strings by using garbled circuits. In [9], Bruekers et al. propose privacy-enhanced comparison of DNA profiles for identity, paternity and ancestry tests using homomorphic encryption. Similar to our work, in [10], Kantarcioglu et al. propose using homomorphic encryption to perform scientific investigations on integrated genomic data. In their scheme, all genomic data is encrypted by the same public key of the data storage site, and there is a single key holder site which can decrypt everything. Thus, a curious party at the key holder site can obtain the genomic information of all users in case of a possible data leakage from the data storage site. Moreover, in [10], only the encrypted variants (i.e., positions in the genome holding a nucleotide that varies between individuals) of the users are stored at the data storage site along with their plaintext locations (on the DNA), which can leak substantial information to the data storage site about the genomic sequences of the users, as we discuss in Section 2.4. As opposed to [10], we focus on personal use of genomic data (e.g., in medical tests and personalized medicine methods), propose methods in which each user's genomic data is encrypted via his own cryptographic key, and prevent the leakage of genomic data due to statistical dependence between the variants. In one of the recent works [11], Baldi et al. make use of both medical and cryptographic tools for privacy-preserving paternity tests, personalized medicine, and genetic compatibility tests. Instead of utilizing public key encryption protocols, in [12], Canim et al. propose securing the biomedical data using cryptographic hardware. Finally, in [13], Eppstein et al. propose a privacy-enhanced method for comparing two compressed DNA sequences by using Invertible Bloom Filter [14].
When releasing databases consisting of aggregate genomic data (e.g., for research purposes), it is shown that known privacy-preserving approaches (e.g., de-identification) are ineffective on (un-encrypted) genomic data [15, 16]. Homer et al. [17] prove that the presence of an individual in a case group can be determined using aggregate allele frequencies and his DNA profile. In another recent study [18], Gitschier shows that a combination of information, from genealogical registries and a haplotype analysis of the Y chromosome collected for the HapMap project, allows for the prediction of the surnames of a number of individuals held in the HapMap database. Thus, releasing genomic data (even in aggregate form) is currently banned by many institutions due to this privacy risk. In [19], Zhou et al. study the privacy risks of releasing the aggregate genomic data. They propose a risk-scale system to classify aggregate data and a guide for the release of such data. Recently, using differential privacy was proposed by Fienberg et al. [20]; they aim to ensure that two aggregated databases, differing from each other by only one individual's data (e.g., DNA sequence), have indistinguishable statistical features.
Recently, in [21], utilizing a public cloud, Chen et al. propose a secure and efficient algorithm to align short DNA sequences to a reference (human) DNA sequence (i.e., read mapping). Finally, in [22], Wang et al. propose a privacy-protection framework for important classes of genomic computations (e.g., search for homologous genes), in which they partition a genomic computation, distributing sensitive data to the data provider and the public data to the data user.
In this work, we focus on medical tests (e.g., disease-susceptibility test) and personalized medicine methods by using users' genomic data while protecting user's genomic privacy. As a result of our extensive collaboration with geneticists, clinicians, and biologists, we conclude that DNA string comparison is insufficient in many medical tests (that use genomic data) and would not be enough to pave the way to personalized medicine. As it will become clearer in the next sections, specific variants must be considered individually for each genetic test. Thus, as opposed to the above private string search and comparison techniques, which focus on privately comparing the distance between the genomic sequences, we use the individual variants of the users to conduct genetic disease susceptibility tests and develop personalized medicine methods. We consider the individual contribution of each variant to a particular disease, for which a string comparison algorithm (such as Private Set Intersection[23, 24]) would not work. Further, in our proposed algorithms, we consider the statistical relationship between the variants for the genomic privacy of the users. In addition, we make use of a Storage and Processing Unit (SPU) between the user (patient) and the medical unit to store the genomic data in encrypted form and make computations on it using homomorphic encryption and proxy encryption.
The aim of the present invention is to propose a privacy-enhancing method for medical tests and personalized medicine methods, which utilize patients' genomic data. It is proposed a method to process genomic data comprising the steps of:
at an initialization stage:
The method of the invention is split into a first phase in which the DNA sequence is processed and stored in the SPU and a second phase during which a test is carried out.
During the first phase, the DNA sequence, produced by an authorized laboratory, is processed and encrypted as explained above. During this second phase, the medical test selected by the medical unit is carried out without having the possibility to retrieve all information of the patient.
The method proposed by the invention is based on the use of homomorphic encryption and proxy encryption. Assuming the whole genome sequencing is done by a certified institution, we propose to store patients' genomic data encrypted by their public keys at a Storage and Processing Unit (SPU).
The proposed algorithm lets the SPU (or the medical unit) process the encrypted genomic data for medical tests and personalized medicine methods while preserving the privacy of patients' genomic data. We extensively analyze the relationship between the storage cost (of the genomic data), the level of genomic privacy (of the patient), and the characteristics of the genomic data. Furthermore, we implement and show via a complexity analysis the practicality of the proposed schemes. Finally, we evaluate the security of the proposed schemes and propose new research directions on genomic privacy.
The invention will be better understood thanks to the attached figures in which:
The
The
The
The
The
The
The
The
The
The
In the present case, we study the privacy issues of medical tests and personalized medicine methods. Most medical tests and personalized medicine methods (that use genomic data) involve a patient and a medical unit. The patient is identified by a patient identification (ID), which could be a user name or a pseudonym (e.g., hash value of his social security number). In general, the medical unit is the family doctor, a physician, a pharmacist, or a medical council. In this study, we consider a malicious medical unit as the potential attacker. That is, a medical unit can be a malicious institution trying to obtain private information about a patient. Even if the medical unit is non-malicious, it is extremely difficult for medical units to protect themselves against the misdeeds of a hacker or a disgruntled employee. Similarly, the genomic data is too sensitive to be stored on users' personal devices (mostly due to security, availability, and storage issues), hence it is risky to leave the users' genomic data in their own hands. In addition, extreme precaution is needed between the patient and the medical unit due to the sensitivity of genomic data. Thus, we believe that a Storage and Processing Unit (SPU) should be used to store and process the genomic data. We note that a private company (e.g., cloud storage service), the government, or a non-profit organization could play the role of the SPU. We also assume that the SPU is an honest organization, but it might be curious (e.g., existence of a curious party at the SPU), hence genomic data should be stored at the SPU in encrypted form (i.e., the SPU should not be able to access the content of patients' genomic data). This general architecture is illustrated in
In general, there are two alleles observed at a given SNP position: (i) The major allele is the most frequently observed nucleotide, and (ii) the minor allele is the rare nucleotide. Everyone inherits one allele of every SNP location from each of his parents. If an individual receives the same allele from both parents, he is said to have a homozygous variant for that SNP location. If, however, he inherits a different allele from each parent (one minor and one major), he has a heterozygous variant. There are approximately 40 million approved variants (SNPs) in the human population as of now (according to the NCBI dbSNP [28]) and each patient carries on average 4 million SNPs (i.e., real variants) out of this 40 million. Moreover, this set of 4 million SNPs is different for each patient. From now on, to avoid confusion, for each patient, we refer to these 4 million variants as the real SNPs and the remaining non-variants (approved SNPs that do not exist for the considered patient) as the potential SNPs of the patient; when we only say “SNPs”, we mean both the real and potential SNPs.
At this point, it can be argued that these 4 million real SNPs (nucleotides) could be easily stored on the patient's computer or mobile device, instead of the SPU. However, we assert that this should be avoided due to the following issues. On one hand, the number of approved SNPs in human population continues to increase with new discoveries. Further, types of variations in human population are not limited to SNPs, and there are other types of variations such as Copy-Number Variations (CNVs), rearrangements, or translocations (our proposed privacy-preserving mechanisms can be smoothly adapted for these alternative variations), consequently the required storage per patient is likely to be considerably more than only 4 million nucleotides. This higher storage cost might still be affordable to an average patient (via desktop computers or USB drives), however, genomic data of the patient should be available any time (e.g., for emergencies), thus it should be stored at a reliable source such as the SPU. On the other hand, as we discussed before, leaving the patient's genomic data in his own hands and letting him store it on his computer or mobile device is risky, because his mobile device can be stolen or his computer can be hacked.
A potential attacker can learn about the susceptibilities of the patient to privacy-sensitive diseases if he obtains some specific real SNPs of the patient. Moreover, the knowledge of 75 real SNPs (out of approximately 4 million), if not fewer, will enable the attacker to identify a person[29]. These situations could lead to genetic discrimination such as denying a person's access to health (or life) insurance or obstructing his employment opportunities. As we discussed before, in our setting, both the MU and SPU pose a threat to the patient's privacy. On one hand, the MU can either be a malicious institution trying to obtain private information about the patient or it can be hacked by another malicious entity. On the other hand, the SPU is considered as an honest but curious entity. Thus, our goal is to build mechanisms in which the patient can preserve the privacy of his genomic sequence (his real SNPs) while enabling the MU to access his genomic data and conduct genetic tests.
We assume that the whole genome sequencing is done by a Certified Institution (CI) with the consent of the patient. Moreover, the genomic data of the patient is encrypted by the same CI (using the patient's public key) and uploaded to the SPU so that only the patient can decrypt the stored (potential or real) SNPs, and the SPU cannot access the SNPs of the patient. We are aware that the number of discovered SNPs increases with time. Thus, the patient's complete DNA sequence is also encrypted as a single vector file (via symmetric encryption using the patient's key) and stored at the SPU, thus when new SNPs are discovered, these can be included in the pool of the previously stored SNPs of the patient. We also assume the SPU does not have access to the real identities of the patients and data is stored at the SPU by using pseudonyms; this way, the SPU cannot associate the conducted genomic tests to the real identities of the patients. As an alternative, the privacy of the genomic data at the SPU can be further increased using privacy enhanced access control [30] or Oblivious RAM (G-RAM) storage [31] techniques, in which the data access patterns are completely hidden from the server (SPU). Note that even the most efficient implementation of O-RAM introduces high storage overhead to the client (patient), and it introduces 20˜25 times more overhead with respect to non-oblivious storage. Thus once it becomes more efficient, O-RAM storage could be considered as a future add-on to the proposed privacy-preserving mechanisms.
Depending on the access rights of the MU, the SPU can either (i) compute Pr(X), the probability that the patient will develop the disease X by checking the patient's encrypted SNPs via homomorphic encryption techniques [33] (In one of our proposed schemes, see Method 3 in Section 2.4, Pr(X) is computed at the MC via homomorphic operations), or (ii) provide the relevant SNPs to the MU (e.g., for complex diseases that cannot be interpreted using homomorphic operations). These access rights are defined either jointly by the MU and the patient or by the medical authorities. Further, access rights can be enforced by using a secure attributebased system as in [34]. We note that homomorphic encryption lets the SPU (or MU) compute Pr(X) using encrypted SNPs of the patient P. In other words, the SPU (or MU) does not access P's SNPs to compute his predicted disease susceptibility. We use a modification of the Paillier cryptosystem (described in Section 2.1) to support the homomorphic operations at the SPU (or MU).
We propose four different techniques for the storage and process of the SNPs at the SPU and the preservation of the patient's privacy: (i) Method 0 in Section 2.2, (ii) Method 1 in Section 2.3, (iii) Method 2 in Section 2.4, and (vi) Method 3 in Section 2.5. We describe these proposed techniques in detail in the following subsections. We also discuss the computation of genetic disease susceptibility by using homomorphic operations in Section 2.6.
In the rest of this work, for simplicity of the presentation, we do not consider the type of the variant at a real SNP location (i.e., whether the variation is homozygous or heterozygous for that real SNP); we only consider whether the patient has a real SNP or not at a particular location. However, the proposed approaches and the analysis (in Section 2.4) can easily be extended to cover the types of the variants. In order to facilitate future references, frequently used notations are listed in Table I for the different stages of the proposed schemes.
TABLE I
NOTATIONS AND DEFINITIONS.
General Notations
SNPiP
Type of SNP i, SNPi, of the patient P. SNPiP ε {0, 1}, 0
representing a potential SNP (i.e., non-variant) for P, and 1
representing a real SNP (i.e., a variant) for P.
Px
Predicted susceptibility of the patient P to disease X.
P
Set of real SNPs of the patient P (SNPs at which P has a
variant: around 4 million at each patient).
ΩP
Set of potential SNPs of the patient P (SNPs at which P does
not have a variant: around 36 million at each patient).
Cryptographic Notations
n, g
Public parameters of modified Paillier cryptosystem.
x
Weak private key of the patient P.
x(i)
ith share of the patient P's private key.
gx
Public key of the patient P.
E(m, gx)
Encryption of message m with the patient P's public key.
Susceptibility Test via Weighted Averaging
pji(X)
Probability that P would develop disease X, given SNPiP = j,
Pr(X|SNPiP = j).
CiX
Contribution of SNPi to the susceptibility to disease X.
Susceptibility Test via Likelihood Ratios
IXP
Initial risk of the patient P for disease X.
LXi(j)
Likelihood Ratio (LR) when SNPi = j for disease X.
In this section, we briefly review the modified Paillier cryptosystem (described in detail in [33, 35]), which we use in this work, and its homomorphic properties. We note that the usual notation in Paillier cryptosystem is to use a pair of keys named public and secret key. However, for the present description, we will use the denote the keys as public and private.
The public key of the patient P is represented as (n, g, h=gx), where the strong private key is the factorization of n=pq (p, q are safe primes), the weak private key is xε [1, n2/2], and g of order (p−1)(q−1)/2. Such a g can be easily found by selecting a random a εZn
Encryption of a Message:
To encrypt a message mεZn
T1=gr mod n2 and T2=hr(1+mn)mod n2 (1)
Re-Encryption of a Message:
An encrypted message (T1, T2) can be re-encrypted under the same public key, using a new random number r1ε[1, n/4] as below:
{circumflex over (T)}1=gr
Decryption of a Message:
The message m can be recovered as follows:
m=Λ(T2/T1X), (3)
where
Homomorphic Properties:
Assume two messages m1 and m2 are encrypted using two different random numbers r1 and r2, under the same public key, (n, g, h=gx), such that E(m1, r1, gx)=(T11, T21) and E(m2, r2, gx)=(T12, T22) Assume also that c is a constant number. Then the below-mentioned homomorphic properties are supported by Paillier cryptosystem:
These homomorphic operations are conducted at the SPU (or MU depending on which approach is used) to compute the predicted susceptibility of the patient P to disease X, as will be discussed in Section 2.6.
Proxy encryption: The patient's weak private key x is divided (preferably randomly or by any other rule) into two shares: x(1) and x(2) (such that x=x(1)+x(2)). x(1) is given to the SPU and x(2) is given to the MU. Using the above Paillier cryptosystem, an encrypted message (T1, T2) (under the patient's public key) can be partially decrypted by the SPU (using x(1)) to generate the ciphertext pair (T1%, T2%) as below:
T1%=T2 and T2%=T2/T1x
Now, (T1%, T2%) can be decrypted at the MU using x(2) to recover the original message. x(2) can be provided to the MU once the patient is registered to the medical unit or through the patient's digital ID card. Further details about the distribution of shares are out of the scope of this paper. We note that this approach is not proxy re-encryption; it is based on secret-sharing.
Overall, this modified Paillier cryptosystem is not key optimal, because the size of the MU's and SPU's secret storages do not remain constant. That is, both the MU and SPU need to store a secret for every patient.
However, this storage cost can be considered negligible when compared to the storage of the genomic data. Further, the shares (e.g., x(1) and x(2)) can be stored by the patient and sent to the MU and SPU only when it is needed in order to resolve this storage issue at the expense of extra communication overhead. Furthermore, the above modified Paillier cryptosystem is not proxy invisible, because all participants of the systems (i.e., P, MU and SPU) should be aware of the existence of the proxy. We discuss the security evaluation of this cryptosystem in Section 3.2.
In this approach, the real SNPs of the patient are stored encrypted (via the patient's public key) and the locations of the corresponding real SNPs are stored in plaintext at the SPU.
We assume that SNP, at the patient P is represented as SNPiP and SNPiP=1, if P has a real SNP (i.e., variant) at this location, and SNPiP=0, if P does not have a variant at this location. We let γP be the set of real SNPs of the patient P (at which SNPiP=1). We also let P represent the set of potential SNPs (at which SNPiP=0).
Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique.
Method 0 in Section 2.2 might leak private information to the curious party at the SPU. As the locations of the SNPs are stored in plaintext, if the SPU only stores the real SNPs in γP, a curious party at the SPU can learn all real SNP locations of the patient, and hence, much about his genomic sequence. The nucleotides corresponding to variants at particular locations of the DNA sequence are public knowledge. Thus, even though the contents of patient's real SNPs are encrypted, a curious party at the SPU can infer the nucleotides corresponding to these SNPs from their plaintext locations. Therefore, in this method, the SPU stores the contents of both real and potential SNP locations (in {γP∪ΩP}) in order to preserve the privacy of the patient. The locations of the corresponding SNPs are again stored in plaintext at the SPU. This is because, when a particular SNP (or set of SNPs) are queried by the MU, the SPU should know which SNPs to process (or send to the MU).
As before, we assume that SNPi at the patient P is represented as SNPiP and SNPiP=1, if P has a real SNP (i.e., variant) at this location, and SNPiP=0, if P does not have a variant at this location. We let γP be the set of real SNPs of the patient P (at which SNPiP=1). We also let P represent the set of potential SNPs (at which SNPiP=0). Below, we summarize the proposed approach for the privacy protecting disease-susceptibility test by using this particular storage technique. This approach is illustrated in
The above technique provides a high level of privacy and practicality for the patient, because (i) from the view point of a curious party at the SPU, inferring the locations of the patient's real SNPs with the stored information is equivalent to inferring them with no information about the patient, and (ii) the patient is not involved in the protocol after the sequencing (except for the consent between the patient and the MU for a particular test). However, this level of privacy and practicality comes at the cost of extra storage overhead at the SPU (due to the storage of both real and potential SNPs as discussed in Section 3.1).
Due to the significant storage overhead mentioned in Section 2.3, here we propose another technique that reduces the storage overhead at the SPU at the expense of decrease in privacy. In a nutshell, we leave everything the same as in Section 2.3, but, instead of storing the contents of all potential and real SNP locations, we store the real SNPs of the patient along with a certain level of redundancy (i.e., contents of some potential SNP locations). In other words, to mislead a curious party at the SPU, among the 40 million discovered SNPs, we store the approximately 4 million real SNPs (for which SNPiP=1, iεγp) along with some redundant content from Ωp (with SNPjP=0), for each patient.
Again, we assume that the location of the encrypted (real or potential) SNPs are stored in plaintext at the SPU and there exists a potential curious party at the SPU trying to infer the real SNPs of the patient (in γp). An important issue to consider in this approach is the Linkage Disequilibrium (LD) between SNPs [36].
LD occurs when SNPs at two loci (SNP positions) are not independent of each other. For simplicity, we represent the LD relationship between two SNPs i and j as Pr(SNPi|SNPj), where SNP, (or SNPj) takes values from the set {0, 1}. In compliance with genetic observations, we assume that the LD between two SNPs are not symmetric, i.e., Pr(SNPi|SNPj)≠Pr(SNPj|SNPi). We note that LD relationships are defined among all 40 million discovered SNPs, regardless of their type (i.e., real or potential) at a particular patient.
As in Section 2.3, the SPU provides the end-result of a disease-susceptibility test or the relevant SNPs to the MU. However, in this case, if a particular potential SNP (requested by the MU or needed in the susceptibility test) is not stored at the SPU (i.e., SNPjP=0), one of the following two scenarios occurs: (i) If the SPU provides the relevant SNPs to the MU, MU infers the missing potential SNPs from the reference genome (since it is known that the missing potential SNPs are not a variant for P), or (ii) if the SPU provides the end-result of the susceptibility test, the SPU uses the fact that SNPjP=0 for each missing potential SNPj.
As expected, the amount of storage redundancy (due to the storage of the content from Ωp), along with the LD between the SNPs and their characteristics, determine the level of a patient's genomic privacy.
Therefore, in the rest of this section, we analyze the relationship between the amount of redundancy, LD values, characteristics of the SNPs, and the level of privacy. To do so, first, we observe the average probability of correctly inferring the locations of P's real SNPs (in γp) considering varying amounts of redundancy and the LD values between the SNPs. That is, how much information from a patient's un-stored potential SNPs is revealed to the curious party at the SPU about the locations of his real SNPs ? This problem can also be formulated similarly if the goal of the attacker is to determine the type of the variant at a real SNP location (e.g., homozygous or heterozygous). In this case, SNPiP can take three different values from the set {0, 1, 2}, 0 representing a potential SNP (i.e., non-variant) 1 representing a real homozygous SNP, and 2 representing a real heterozygous SNP for P. It is worth noting that for this study, we create realistic models for the LD values and the characteristics of the SNPs. Further, for the created models, we try a wide range of parameters and observe a wide range of results to address most potential scenarios. However, as the field of genomics becomes more mature, our models can be replaced by the values obtained from the medical research.
We let Ωps and Ωpu denote the set of P's potential SNPs that are stored (for redundancy) and not stored at the SPU, respectively (Ωps∪Ωpu=Ωp). Further, Ki is the set of SNPs with which a particular SNPi has LD, and |Ki|=k (for each SNP, these k SNPs are chosen among approximately 40 million SNPs). We assume that k≧0 and it is a truncated Gaussian random variable with only discrete values and obtained from a distribution with mean σ(k) and standard deviation (k).
Initially, we compute Pr(SNPiP) for all (real and potential) SNPs in {γp∪Ωps} by using the LD relationships between these SNPs and those in Ωpu. As all SNPs in {γp∪Ωps} are encrypted and stored at the SPU, only the LD relationships between these SNPs and the un-stored SNPs in Ωpu are helpful for the curious party.
Therefore, for each real SNPiεγp, we observe Pr(SNPiP=1|SNPmP=0) for all mε{K⊥(i)∩≡Ω⊥pt s}┤}, get the average of these values, and compute Pr(SNPiP=1). Similarly, for each potential SNPjεΩps, we observe Pr(SNPjP=0| SNPmP=0) for all mε{Kj∩Ωpu}, average these values, and compute Pr(SNPjP=0). We let/be the indicator of the LD strength between two SNPs. Thus, we represent Pr(SNPiP=1|SNPmP=0)=(εγp, mε{Ki∩Ωpu}) and Pr(SNPjP=0|SNPmP=0)=/(jεΩ⊥pt s, mε{K⊥(j)∩Ω⊥ptu}) as truncated Gaussian random variables with range [0.5, 1], obtained from a distribution with mean μ(l) and standard deviation σ(l).
Finally, if |Ki|=k=0 or |Ki∩Ωpu|=0 for a SNP i in {γp∪Ωps}, we update Pr(SNPiP=1) considering the fact that the expected value of all stored SNPs is known by the curious party as below:
In the following, we illustrate our numerical results that represent the relationship between storage, inference power of the curious party at the SPU, and LD values. We assume |γP|=4 million and |γP∪ΩP|=40 million. We define the percentage of storage redundancy at the SPU as
and compute the average value of Pr(SNPiP=1) for a SNP in γP for varying values of μ(k), (k), μ(l), and σ(l). Higher values of Pr(SNPiP=1) indicate a higher inference power for the curious party at the SPU. We repeat each simulation 100 times to obtain an average. Note that Method 1 (in Section 2.3) is a special case of Method 2 (when the storage redundancy at the SPU is 900%), hence its privacy is the same as 900% redundancy in the following results.
In
We observe that the strength of LD, however, does not affect the inference power as strong as k. Then,
The
In the
Next, considering the individual characteristics of the real SNPs (i.e., their severity levels), we analyze the level of genomic privacy of a patient against a curious party at the SPU. By the level of genomic privacy, we understand the level of information that a third party can infer about the real variants of a patient. The severity of a SNPi can be defined as the privacy-sensitivity of the SNP when SNPiP=1 (i.e., when it exists as a variant at the patient P). For example, a real SNP revealing the predisposition of a patient for Alzheimer's disease can be considered more severe than another real SNP revealing his predisposition to a more benign disease. Severity values of the SNPs are determined as a result of medical studies (depending on their contributions to various diseases) and tables of disease severities provided by insurance companies (e.g., percentage of invalidity). We denote the severity of a real SNP i as Vi, and 0≦Vi≦1 (1 denotes the highest severity). Thus, we define the genomic privacy of the patient P as below:
We do not use the traditional entropy metric [37, 38] to quantify privacy, as only one state of SNPiP poses privacy risks (i.e. SNPiP=1), as discussed before.
First, we study the relationship between the storage redundancy and the severity of the real SNPs by focusing on three types of patients: (i) patient A, carrying mostly low severity real SNPs (in γA), (ii) patient B, carrying mostly high severity real SNPs (in γB), and (iii) patient C, carrying mixed severity real SNPs (in γC). For each patient, the highest level of privacy is achieved when the storage redundancy is maximum (as in Method 1 in Section 2.3). Thus, we recognize this level as 100% genomic privacy for the patient. For the evaluation, we take the highest privacy level of patient C as the base and normalize everything with respect to this value. We use the following parameters for the simulation. The severities of patient A's and patient B's real SNPs are represented as truncated Gaussian random variables with (μA, σA)=(0.25, 0.15) and (μB, σB)=(0.75, 0.15), respectively. Furthermore, the severity of patient C's real SNPs are represented as a uniform distribution between 0 and 1. We also set μ(l)=0.8, σ(l)=0.25, μ(k)=2, and σ(k)=0.75. In
We observe that by increasing the storage redundancy, a patient with high severity real SNPs gains more privacy than a patient with lower severity real SNPs, hence the storage redundancy can be customized for each patient differently based on the types of his real SNPs. It can be argued that the amount of storage redundancy for a patient can leak information (to the curious party the SPU) about the seventies of his real SNPs. However, the severity of the SNPs is not the only criteria to determine the storage redundancy for a desired level of genomic privacy as we discuss next.
Finally, we study the relationship between the severity of the real SNPs, the number of LD pairs per SNP (number of SNPs with which a particular SNP has LD, i.e., k), and the storage redundancy. We assign the Vi values of the real SNPs (in γP) following a uniform distribution between 0 and 1. We set the LD parameters as μ(l)=0.8, σ(l)=0.25, μ(k)=2, and σ(k)=1.5. Then, we observe and compare the following potential scenarios in different types of patients: (i) The real low severity SNPs of the patient (i.e., his real SNPs with low Vi values) have a higher number of LD pairs (i.e., higher k values) with respect to his high severity real SNPs (we note that, in all cases, k values are obtained from the same truncated Gaussian distribution with μ(k)=2, and σ(k)=1.5); (ii) k values are assigned randomly to the SNPs; and (iii) the real high severity SNPs of the patient (i.e., his real SNPs with high Vi values) have a higher number of LD pairs (i.e., higher k values) with respect to his low severity real SNPs. Again, we set a patient's genomic privacy to 100% when the storage redundancy is maximum at the SPU (as in Method 1 in Section 2.3). We illustrate our results in
We obtained similar patterns for further variations of the variables but we do not present these results due to the space limitation. In summary, depending on the actual μ(k), σ(k), μ(l), σ(l), and Vi values (which will be determined as a result of the medical research), the storage redundancy can be determined (and customized for each patient based on the types of his variations) for this approach to keep the genomic privacy of the patient at a desired level. Note that the curious party at the SPU cannot infer the real SNPs of the patient (or the severities of the patient's real SNPs) from the amount of customized storage redundancy, because the storage redundancy (for a desired level of genomic privacy) depends on various factors. For example, a patient with low storage redundancy (for a desired level of genomic privacy) could mean that (i) he carries mostly low severity real SNP (as in
Let LP={Li:iεγP} represent the set of locations (on the DNA sequence) of the patient P's real SNPs (in γP). As opposed to the previous two approaches, here, we propose to encrypt the locations of the SNPs along with their contents. By doing so, we save storage costs by storing only the real SNPs in γP at the SPU (around 4 million) while providing the highest level of privacy (as in Section 2.3). These benefits, however, come with a cost in the practicality of the algorithm, introducing extra steps for the patient (P) during the protocol. Although we can assume that these extra steps can easily be handled via the patient's device such as smart card or mobile device, this approach still requires more message exchanges (as will be described next) between the parties, compared to the previous two approaches.
In some environments, dividing the weak private of the patient, and distributing two shares of the weak private key to the SPU and MU might not be acceptable (e.g., when it is likely that the SPU and MU will collaborate to retrieve patient's weak private). Therefore, for the sake of completeness, in the following, we present Method 3 with and without proxy encryption (i.e., without distributing the patient's private key to other parties). The Method 1 and Method 2 can also be modified similarly to avoid proxy encryption.
The initial steps of the protocol are the same as in Section 2.3, except for Steps 2 and 3 in which the locations of the SNPs are encrypted and a Bloom filter [39] is generated. Below, we summarize the different steps of this approach (the unchanged steps are not repeated). These steps are illustrated in
A Bloom filter is a simple space-efficient randomized data structure for representing a set in order to support membership queries [39]. A Bloom filter for representing a set LP is described by an array of K bits, initially all set to 0. It employs independent hash functions H1, . . . , Hγ with range {1, . . . , κ}. For every element LiεLP, the bits H1(Li), . . . , Hγ (Li) in the array are set to 1. A location can be set to 1 multiple times, but only the first change has an effect.
After constructing the Bloom filter, the CI encrypts each element in LP by using a symmetric key shared between the CI and P (established during Step 0 as in Section 2.3) and generates LEP={E(Li):iεγP}. The CI also encrypts a dummy variant (representing the potential SNPs in ΩP) along with the real SNPs of the patient (using P's public key). Furthermore, the CI associates a dummy position L0 for this dummy variant and encrypts L0 using the symmetric key between the CI and P to obtain the encrypted dummy position E(L0).
To check if Lj belongs to LP, the patient checks whether all H1(Lj), . . . , Hγ(Lj) are set to 1. If not, Lj definitely does not belong to LP. Otherwise, the patient assumes LjεLP, although this may be wrong with some probability. That is, a Bloom filter could yield a false positive, where it suggests that Lj is in LP even though it is not. This probability can be decreased at the expense of increasing Bloom filter length (i.e., κ). Further, the false positive probability can be significantly reduced by using some proposed techniques such as [40, 41]. As a result of this process
(a) If the location is in his Bloom filter (i.e., if he has a real SNP at the corresponding location), P encrypts the location with the symmetric key between the CI and P.
(b) If the location is not in his Bloom filter (i.e., if he does not have a real SNP at the corresponding location), P uses E(L0) as the encrypted location.
We note that the above operations can be easily done via the patient's device (e.g., by reading the patient's device at the MU as a consent to the test) or mobile device (e.g., by consenting via a smart phone application) by using the stored Bloom filter output, E(L0), and symmetric key between the CI and P.
In this approach, the SPU stores only the encrypted SNPs and encrypted locations. Genomic data encrypted by P's public key is only decrypted at P, and the weak private key of P remains only at P (i.e., shares of the weak private key are not distributed to the SPU or MU). Most of this approach is the same as Method 3 with proxy encryption. Indeed, the first 8 steps of the algorithm are the same, except for the distribution of parts of P's private key. The only difference is the transfer of the end-result or the relevant SNPs to the MU as follows:
We note that the security of the communication between P and the MU is provided by symmetric keys as discussed before. The above operations put some more burdens on the patient during the protocol. However, we emphasize that these operations can be smoothly done on the patient's device without requiring a substantial effort from the patient himself.
In summary, as the locations of the real SNPs are encrypted, a curious party at the SPU cannot infer the contents of the SNPs from their locations (as in Section 2.3), hence it is enough to store only the real SNPs in γP. Furthermore, the privacy provided by this approach (with or without proxy encryption) is the same as 900% redundancy in Method 2 (i.e., similar to Method 1), hence we do not discuss it again. Another advantage of this approach (i.e., Method 3 in general) is that individual contributions of the genetic variant markers remain secret at the MU, because the homomorphic operations are conducted at the MU. This advantage might become more significant when this approach is used for personalized medicine methods in which the pharmaceutical company (embodied in this case as the medical unit) does not want to reveal the genetic properties of its drugs. Thus, if introducing the described extra steps for the patient and few additional message exchanges between the parties are tolerated, this approach operates with relatively modest storage and yet provides very good privacy.
We now present the disease-susceptibility test via homomorphic operations at the SPU for Method 1 (Section 2.3) and Method 2 (Section 2.4). Similar techniques can be used for Method 3 at the MU, as discussed in Section 2.5.
The SPU uses a proper function to compute P's predicted disease susceptibility via homomorphic encryption. There are different functions for computing the predicted susceptibility. In [25], focusing on one example of many diseases that require a susceptibility test involving multiple SNPs, Kathiresan et al. propose to count the number of unfavorable alleles carried by the patient for each SNP related to a particular disease. Similarly, in [26], Ashley et al. propose to multiply the Likelihood Ratios (LRs) of the most important SNPs for a particular disease in order to compute a patient's predicted susceptibility. LR values are determined as a result of medical studies. Furthermore, a weighted averaging function can also be used, which computes the predicted susceptibility by weighting the contributions of SNPs by their contributions (e.g., LR values of the SNPs). Note that our proposed privacy-preserving mechanisms are not limited by the types of the functions (used to test the disease susceptibility). It is expected that these functions will evolve over time; hence the proposed algorithms can be developed to keep up with this evolution.
In the following, we discuss how to compute the predicted disease susceptibility at the SPU by using a toy example to show how the homomorphic encryption is used at the SPU. Initially, we assume that the function at the SPU is weighted averaging (which is an advanced version of the function proposed in [25]) and show how the predicted susceptibility is computed using encrypted SNPs. Then, we show how the function proposed in [26] (i.e., multiplication of LR values) can be utilized at the SPU.
Assume that (for simplicity) the susceptibility to disease X is determined by the set of SNPs Ω={SNPm,SNPn}, which occur at particular locations of the DNA sequence. SNPmP and SNPnP are not necessarily among the real SNPs of the patient P (i.e., P does not need to have a variant at those locations). The contributions of different states of SNPiP for iε{m, n} to the susceptibility to disease X are computed via previous studies (on case and control populations) and they are already known by the MU. That is, p0i(X) @Pr(X|SNPiP=0) and p1i(X) @Pr(X|SNPiP=1) (iε{m, n}) are determined and known by the MU. Further, the contribution (e.g., LR value) of SNPi to the susceptibility to disease X is denoted by CiX. Note that these contributions are also computed by previous studies on case and control groups and they are known by the MU.
As we have discussed before, the SPU stores the set of SNPs of the patient P, encrypted by P's public key (n, g, h=gx). Encryption is done using the modified Paillier cryptosystem as discussed in Section 2.1. Thus, the SPU uses E(SNPmP, gx) and E(SNPnP, gx) for the computation of predicted susceptibility of P to disease X. From now on, we drop the r values in the above encrypted messages for the clarity of the presentation (r values are chosen randomly from the set [1, n/4] for every encrypted message as discussed in Section 2.1). Similarly, the MU provides the following to the SPU in plaintext: (i) the markers for disease X (SNPm and SNPn), (ii) corresponding probabilities pji(X), iε{m, n} and jε{0,1}, and (iii) the contributions of each SNP CiX.
Next, the SPU encrypts j(jε{0,1}) using P's public key to obtain E(0, gx) and E(1, gx) for the homomorphic computations. This encryption can also be done at the MU and sent to the SPU. Alternatively, we might assume that SNPs of a patient are stored at the SPU in pairs of {E(|SNPiP−0|, gx), E(|SNPiP−1|, gx)} for each SNPiP, instead of the actual values of the SNPs. In this case, the above encryption at the SPU would not be required.
The SPU computes the predicted susceptibility of the patient P to disease X by using weighted averaging.
This can be computed in plaintext as below:
The computation in (9) can be realized using the encrypted SNPs of the patient (and utilizing the homomorphic properties of the Paillier cryptosystem) to compute the encrypted disease susceptibility, E(SPX, gx) as below:
We note that the end-result in (10) is encrypted by P's public key.
Then, the SPU partially decrypts the end-result E(SPX, gx) using its share (x(1)) of P's private key (x) as discussed in Section 2.1 to obtain E(SPX, gx
In some genetic tests, the types of the real SNPs (e.g., homozygous or heterozygous) become also important. In this case, SNPiP can take three different values from the set {0, 1, 2} to represent a potential SNP (i.e., nonvariant), a real homozygous SNP, and a real heterozygous SNP, respectively. In such a scenario, to conduct the disease-susceptibility test via homomorphic operations, the SPU should store the squared values of the SNPs. That is, for each SNPiP of the patient P, the SPU should store E((SNPiP)2, gx). Depending on the types of genomic tests that would be supported by the SPU (and the functions required for these tests), the format of storage of patient's SNPs can be determined beforehand, and SNPs can be stored accordingly just after the sequencing process.
We now assume that the predicted disease susceptibility is computed from the multiplication of Likelihood Ratios (LRs) of the corresponding SNPs as in [26] and show how such a computation would be handled at the SPU by using homomorphic operations.
In this approach, the predicted disease susceptibility is computed by multiplying the initial risk of the patient (e.g., for disease X) by the LR value of each SNP related to that disease (LR value of a SNP i depends on the value of SNPiP at the patient P). The initial risk of the patient P for the disease X is represented as IXP. We note that IXP is determined by considering several factors (other than patient's genomic data) such as patient's age, gender, height, weight, and environment. Thus, this initial risk can be computed directly by the MU. We also note that if the LR value corresponding to a particular SNP is less than one, the risk for the disease decreases. Otherwise, if the LR value is greater than one, the risk increases for the corresponding disease.
Similar to before, we assume that the susceptibility to disease X is determined by the set of SNPs in Ω={SNPm,SNPn}. We denote the LR values due to SNPiP=0 and SNPiP=1 for disease X as LXi (0) and LXi (1), respectively.
The SPU stores the SNPs of the patient P, encrypted by P's public key. The MU sends the following to the SPU: (i) LXi (j) values (iε{m, n} and jε{0,1}) in plaintext, and (ii) the markers for disease X. The MU also encrypts the log of initial risk value, ln(IXP), by P's public key and sends E(ln(IXP), gx) to the SPU. Alternatively, the contribution of the initial risk to the disease susceptibility can be included to the end-result at the end, at the MU.
The Paillier cryptosystem does not support multiplicative homomorphism in ciphertext (it only supports the multiplication of a ciphertext with a constant as discussed in Section 2.1). Thus, instead of multiplying the LR values, we propose using addition in log-domain at the SPU. Thus, the SPU computes the predicted susceptibility of P to disease X as below:
We note that (12) corresponds to the below computation in plaintext:
As before, the SPU partially decrypts E(ln(SX P), gx) using x(1) (its share of P's private key) to obtain E(ln(SPX), gx
to obtain SPX. Similar to weighted averaging, if the types of the real SNPs are used for the test (in which there are three possible states for SNPiP), squared values of the SNPs should be stored at the SPU for each patient.
In
To evaluate the practicality of the proposed privacy-preserving algorithms, we implemented them, and assessed their storage requirements and computational complexities on Intel Core i7-2620M CPU with 2.70 GHz processor under Windows 7 Enterprise 64-bit Operating System. We set the size of the security parameter (n in Paillier cryptosystem in Section 2.1) to 1024 bits. We computed the disease susceptibility using weighted averaging (at the SPU or MU, see Section 2.6.1 as well as LR test in Section 2.6.2 which also has similar complexity) and real SNP profiles from [42]. Our implementation relies on a MySQL 5.5 database managed by the open source tool MySQL Workbench. To provide a platform-independent implementation, we used the Java programming language along with the open-source Integrated Development Environment, NetBeans IDE 7.1.1., for the implementation of the Java code. We note that our code for the implementation is not optimized, and better results can be expected with an optimized implementation.
In Table II, we summarize the computational and storage complexities of the proposed methods at (i) Certified Institution (CI), (ii) SPU, (iii) MU, and (iv) P. We evaluate the proposed methods considering the following costs: (i) encryption of patient's variants, (ii) disease-susceptibility test at the SPU via homomorphic operations (using ten variants), (iii) decryption of the end-result (or relevant SNPs), (iv) proxy encryption, and (v) storage costs, in which θ represent the percentage of storage redundancy at the SPU. We did not explicitly implement the Bloom filter (for Method 3) and symmetric encryption/decryption between the parties for the security of the communication. However, the computational costs due to these operations are negligible compared to Paillier encryption/decryption and homomorphic operations.
We emphasize that the encryption of the variants at the CI is a one-time operation and is significantly faster than the sequencing and analysis of the sequence (which takes days). Further, this encryption can be conducted much more efficiently by computing some parameters, such as (gr, hr) pairs, offline for various r values, for each patient. Indeed, by computing (gr, hr) pairs offline, we observe that the encryption takes only 0.017 ms per variant at the CI.
TABLE 2
Computational and Storage Complexities of the Proposed Methods
Method 1 and Method 2
@CI
@SPU
@MU
Paillier Encryption
Homomorphic Operations
Proxy Encryption
Storage
Paillier Decryption
30 ms./variant
1 sec. (10 variants)
2 ms.
26 ms.
Method 3 with proxy encryption
@CI
@SPU
@MU
Paillier Encryption
Proxy Encryption
Storage
Homomorphic Operations
Paillier Decryption
30 ms./variant
2 ms.
500 MB/patient
1 sec. (10 variants)
26 ms.
Method 3 without proxy encryption
@CI
@SPU
@MU
@P
Paillier Encryption
Storage
Homomorphic Operations
Paillier Decryption
30 ms./variant
500 MB/patient
1 sec. (10 variants)
26 ms.
It is also possible to conduct private statistical tests (by a medical researcher) on the data stored at the SPU in order to get statistics about the variants of multiple patients. Conducting such a statistical test for a variant (about its type) on 100K patients takes around 55 minutes at the SPU and scales linearly with the number of patients. Note that such a statistical test is only possible with Method 1 or Method 2; using Method 3 and querying the encrypted locations of SNPs from 100K patients is not practical for this application.
In summary, all these numbers show the practicality of our privacy-preserving algorithms.
The proposed schemes preserve the privacy of patients' genomic data relying on the security strength of modified Paillier cryptosystem (in Section 2.1). The extensive security evaluation of the modified Paillier cryptosystem can be found in [33]. Below we summarize two important security features of this cryptosystem.
Finally, if the weak private key of the patient, x, is randomly divided and distributed to the Storage and Processing Unit (SPU) and Medical unit (MU) as in Method 1, this weak private key could be revealed if the MU colludes with the SPU, but the factors n, p, and q remain secret. We note that such a collusion is not considered in this study. However, for the sake of completeness, in Section 2.5.2, we present an alternative approach (Method 3 without proxy encryption) that avoids distributing the patient's weak private key to other parties, hence is robust against such a collusion.
Hubaux, Jean-Pierre, Telenti, Amalio, Fellay, Jacques, Ayday, Erman, Raisaro, Jean Louis, McLaren, Paul Jack, Rougemont, Jacques, Humbert, Mathias
Patent | Priority | Assignee | Title |
10402588, | Jun 19 2013 | SOPHIA GENETICS S.A. | Method to manage raw genomic data in a privacy preserving manner in a biobank |
10528758, | May 02 2014 | Intertrust Technologies Corporation | Genomic informatics service |
10599753, | Nov 11 2013 | Amazon Technologies, Inc | Document version control in collaborative environment |
10686788, | Feb 07 2014 | Amazon Technologies, Inc. | Developer based document collaboration |
10691877, | Feb 07 2014 | Amazon Technologies, Inc | Homogenous insertion of interactions into documents |
10877953, | Nov 11 2013 | Amazon Technologies, Inc. | Processing service requests for non-transactional databases |
11240033, | Sep 26 2019 | MERATIVE US L P | Secure DNA-based password |
11336648, | Nov 11 2013 | Amazon Technologies, Inc | Document management and collaboration system |
Patent | Priority | Assignee | Title |
20100121872, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Sep 10 2013 | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE (EPFL) | (assignment on the face of the patent) | / | |||
Mar 05 2015 | FELLAY, JACQUES | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 06 2015 | HUBAUX, JEAN-PIERRE | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 06 2015 | RAISARO, JEAN LOUIS | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 06 2015 | TELENTI, AMALIO | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 06 2015 | HUMBERT, MATHIAS | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 13 2015 | MCLAREN, PAUL JACK | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Mar 13 2015 | ROUGEMONT, JACQUES | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 | |
Apr 16 2015 | AYDAY, ERMAN | ECOLE POLYTECHNIQUE FEDERALE DE LAUSANNE EPFL | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036635 | /0790 |
Date | Maintenance Fee Events |
Aug 12 2019 | M2551: Payment of Maintenance Fee, 4th Yr, Small Entity. |
Oct 16 2023 | REM: Maintenance Fee Reminder Mailed. |
Apr 01 2024 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Feb 23 2019 | 4 years fee payment window open |
Aug 23 2019 | 6 months grace period start (w surcharge) |
Feb 23 2020 | patent expiry (for year 4) |
Feb 23 2022 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 23 2023 | 8 years fee payment window open |
Aug 23 2023 | 6 months grace period start (w surcharge) |
Feb 23 2024 | patent expiry (for year 8) |
Feb 23 2026 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 23 2027 | 12 years fee payment window open |
Aug 23 2027 | 6 months grace period start (w surcharge) |
Feb 23 2028 | patent expiry (for year 12) |
Feb 23 2030 | 2 years to revive unintentionally abandoned end. (for year 12) |