This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
1. Introduction
In the last two years, the COVID-19 pandemic has become a major disaster in the world. As COVID-19 has a certain fatality rate and spreads very fast, prevention and control of this virus have become a top priority worldwide. However, compared to the prevention and control regarding SARS in 2003, information related technique is being widely used in all aspects regarding prevention and control of this COVID-19 pandemic. Hence, extensive collection, processing, and investigation of personal data has become an important part of the anti-pandemic work. Given the huge amount of data collected, it is necessary to store these data in the cloud to reduce the storage burden.
As a new storage paradigm, cloud storage collects different storage devices to provide users with massive data storage. Hospitals and patients can easily access data by connecting to the cloud anytime, anywhere, and through any networked device whenever needed.
The infrastructures supporting cloud storage are distributed and virtual. This brings some threats to users’ data security, such as network virus propagation, unauthorized access, denial of service attacks, information leakage, data loss, as well as network infrastructure that could damage data integrity during data transmission, etc. For pandemic prevention and control, it is clear that the loss of data, such as the patient’s recent whereabouts, will certainly cause enormous trouble and could even lead to further virus spread and may even favour the situation of successive waves of the new coronavirus.
Due to the threat of internal or external attacks, data stored in the cloud are easily damaged. In addition, the cloud service provider (CSP) may not notify the user of this event in consideration of its own reputation. The user has a mechanism to detect data corruption only after accessing the data [1–3]. Therefore, in order to improve the reputation of cloud storage and let users know the integrity of the hosted data in a timely manner, a mechanism is needed to verify the data integrity in the cloud. Hence, an integrity verification mechanism is very important in the cloud environment.
1.1. Related Works
Traditional methods need to download the entire data from the cloud when verifying data integrity, which brings unacceptable communication and computing costs and greatly consumes users’ resources. In order to satisfy the user’s remote checking of data integrity, the cloud data remote integrity check solution should need not to download the complete data in the cloud storage environment. Thus, the following solutions have been proposed.
Ateniese et al. [4] first proposed a provable data possession (PDP) scheme, an effective technology to audit the cloud storage. In the PDP protocol, data are encoded as blocks, and the user processes the block data to generate a verifiable authenticator, and then it outsources the data blocks and authenticators to the cloud. A public verifier with sufficient resources is also called a third-party auditor (TPA) and is trusted by users to check the data integrity. TPA creates a challenge to the server by randomly selecting a small group of block indexes. The server returns a proof that proves the integrity of the challenged blocks. TPA can effectively verify the proof without downloading data block. PDP has laid the foundation for the design of cloud storage audit schemes. In recent years, many researchers have conducted extensive and in-depth explorations around PDP [5–7].
In order to obtain better efficiency and performance, several improved PDP protocols have been proposed [8, 9]. The previously proposed schemes mostly use traditional public key cryptography, so a trusted certificate authority (CA) is required to issue a certificate to bind certain user identities and their public keys. Heavy certificate management, including certificate generation, distribution, and revocation, requires a lot of computing and storage resources. As the number of users increases, certificate management becomes extremely difficult. In addition, the verifier must retrieve the certificate from the CA and then check the validity of the public key certificate, which also brings heavy calculation and communication costs to the verifier. Therefore, the certificate-based PDP protocol is very inefficient when used in actual situations.
In order to overcome this problem, researchers considered applying identity-based cryptography to the PDP protocol and therefore proposed many ID-PDP solutions. Wang et al. [10] first introduced the concept of identity-based PDP (ID-PDP), which uses user names or emails instead of public keys. Then, ID-PDP is further extended to the multicloud storage environment [11] to check the integrity of remote data. In order to improve performance, Wang et al. [12] added a proxy server to the remote data integrity check scheme. The proxy server processes data instead of users. In the scheme, incentive and unconditional anonymous ID-PDP was first proposed to protect and encourage criminal whistleblowers. Yu et al. [13] used RSA signature technology to design an ID-based integrity cloud data check protocol. The protocol supports variable size file blocks and public verification. In order to further improve security, Yu et al. [14] combined the key homomorphic encryption technology in the cryptographic cloud audit system and proposed an improved scheme with perfect data privacy protection capabilities. The ID-based privacy-preserving integrity verification of shared data over untrusted cloud scheme is proposed, which can support users to update the data in cloud and protect users’ privacy in untrusted cloud servers. Li et al. [15] proposed identity-based privacy-preserving remote data integrity checking for cloud storage scheme, which uses homomorphic verifiable tags to reduce the computational complexity and uses random integer addition to mask the original data to protect the verifier from obtaining any knowledge about the data during the integrity checking process.
1.2. Contribution
Currently, using cloud storage audit protocols is regarded as an important cloud service. However, the existing audit protocols have certain shortcomings. On the one hand, most of them rely on expensive public key infrastructure (PKI), so certificate management/verification is very complicated. On the other hand, most cloud users have limited resources. Nowadays, ID-based cloud audit protocols have attracted the attention of researchers, but most of them require users with limited resources to perform expensive operations.
Recently, Rabaninejad et al. [16] proposed a lightweight identity-based provable data ownership cloud storage audit scheme, which supports privacy and traceability of user identities. They also proposed an online/offline ID-based PDP scheme [17]. However, we discovered that there are security flaws in the digital signature (OffTagGen) of their scheme. Attackers, such as malicious cloud servers, can destroy the privacy of user’s identity privacy and damage data privacy and integrity. In order to get a more secure protocol, based on the scheme presented in [16], we propose an improved one and discuss its application in pandemic data management. The main contributions of this paper are summarized as follows:
(1) We firstly point out the insecurity of Rabaninejad et al.’s lightweight identity-based provable data ownership cloud storage audit scheme. We give two attacks to show that data tags can be easily forged.
(2) We provide an improved secure cloud audit protocol that protects user privacy. This new protocol is effective yet resistant to attacks.
(3) Finally, we show how our scheme can be applied to the pandemic data management.
1.3. Organization
The rest of this article is organized as follows. In Section 2, we describe the system framework. In Section 3, we review the cloud audit scheme proposed by Rabaninejad et al. [17]. In Sections 4 and 5, we introduce the attack. In Section 6, we provide an improved privacy protection cloud audit protocol and conduct a rough analysis of its security. In Section 7, we apply our scheme to pandemic data management. Finally, in Section 8, we conclude the work and point out some directions for future work.
2. System Framework
In this section, we first describe the system model. Then, we give the goals of the design. After that, we introduce some necessary definitions. Finally, we show the security model.
2.1. System Model
The system model includes four entities, as shown in Figure 1, which involves the key generation center (KGC), the users, the third-party auditor (TPA), and the cloud server. The functions of each entity are summarized as follows:
(1) KGC. Based on the user’s identity, KGC generates its private key.
(2)
(3)
(4)
[figure omitted; refer to PDF]
As shown in Figure 1, the workflow of the four parties can be described as follows:
(1) The users generate the offline tags and store them locally.
(2) The users send their identity information to KGC.
(3) KGC uses the master key and the users’ identity information to generate the users’ private key and returns it to them.
(4) When users need to store data files in the cloud server, they generate online tags by using some lightweight computations based on offline tags.
(5) Users outsource the (block, online tag) pair to the cloud server.
(6) The user sends an audit request to TPA with some audit information attached.
(7) TPA sends the challenge message to the cloud server.
(8) The cloud server generates a proof based on the challenge message and sends it to TPA.
(9) TPA sends the results of the audit as the auditing report to the users.
2.2. Design Goal
The design goals are roughly as follows.
(i)
(ii)
(iii)
(iv)
(v)
2.3. Definition
The ID-PDP scheme includes the following algorithms:
(i)
(ii)
(iii)
(iv)
(v)
(vi)
2.4. Security Model
For maintaining their own reputation, cloud servers are generally unwilling to disclose data loss/damage to the verifier, so they are not completely credible in the PDP scheme. Here, we focus on the security model of auditing soundness of the cloud storage auditing protocol. The security model is described as follows: a game between an adversary server
(i)
(ii)
(a)
(b)
(c)
(iii)
(iv)
(v)
3. Review of Rabaninejad’s Scheme
In this section, we will review the specific scheme of Rabaninejad et al. [17]. First, we review the concept of a bilinear map.
(i)
(ii)
(iii) Computability. There exists an efficient algorithm to compute
Rabaninejad et al. [17] proposed an online/offline ID-based PDP scheme, which consists of the following algorithms. In addition to the definition and notation in the bilinear map, two hash functions
(1) Setup. The KGC chooses a random value
(2) Extract. The KGC uses
(3)
At last, the offline tags
(4)
Finally, the online
(5)
(a) Choose a random subset
(b) Send the challenge
(6)
(a) Computes a combination of the challenged blocks as
(b) Aggregates the tags as
(c) Sends back
(7) ProofVerify. When TPA receives the proof
4. Attack I on the OffTagGen Algorithm
The attack I is as follows:
(1) The adversary (which can be the malicious cloud server) can obtain many block-signature pairs, such as
(2) Let
(3) With these equations, the adversary can compute
and then computes
(4) Let
For the exponential prime modular, q is publicly known to all; thus, the adversary can compute
(5) With
(a)
Because the adversary knows
(b)
Finally, the online tags
We can check that the forged tag
Thus, OffTagGen algorithm is not secure. Even with two block-signature pairs, anyone can first modify the contents of the blocks and then forge the offline and online tags correspondingly.
5. Attack II on the Cloud Auditing Protocol
In our attack II, we show that the adversary (which can be the malicious cloud server) can forge proof while it can even delete all the outsourced data blocks. Concretely, the attack is the following:
(1) The first four steps of the attack are the same as attack I. The malicious cloud server can get
(2) When the cloud server receives the tags
(3)
(a) Choose a random c−element subset
(b) Send the challenge
(4)
(a) First, it randomly chooses
(b) For any
(c) It sends back
(5)
Here, we can verify that the forged proof, is a valid one because the below equation holds:
6. Our Improved Cloud Auditing Protocol
(1)
(2)
(3)
At last, it locally stores the offline tags,
(4)
Finally, the online tags
When the server receives the tags
(5) Challenge. In order to challenge the integrity of the file
(a) Choose a random c−element subset
(b) Send the challenge
(6) ProofGen. The server generates an auditing proof according to the received challenge
(a) Computes a combination of the challenged blocks as
(b) Aggregates the tags as
(c) Sends back
(7) ProofVerify. When TPA receives the proof
7. Security Analysis
In this section, we first prove the correctness of our improved scheme. Then, we prove that the audit proof in our proposed scheme cannot be forged, which proves that our proposed scheme can resist attacks I and II.
(1) Correctness. The correctness of verification equation (17) is proved below:
(2) Soundness. In our improved scheme, a malicious CSP cannot forge a correct audit proof by using our attacks.
Proof
(i)
(ii)
(a)
(b)
(c)
(iii)
(iv) ProofGen. Adversary A executes the algorithm and computes a Proof
(v) ProofVerify. If the proof is verified, A wins the game. Also, part of the proof represented by
In the original scheme, the adversary A can compute
However, in our improved scheme, the authentication tag is calculated as equation (15):
Before the malicious adversary forges an off tag, he needs to know the value of
(3) Data Privacy against TPA. While providing the integrity audit service to the user, TPA cannot obtain any information about the content of the user’s data from the information provided by the user or from the auditing process.
Proof. On the one hand, TPA received information from user before performing the auditing work are (root, σroot) and (γ, Fname). The user data cannot be accessed from root due to the one-way nature of the hash function. Meanwhile, γ = mpkx, and TPA cannot get user’s data from it.
On the other hand, in the auditing process, the TPA gains
Therefore, our improved scheme can preserve the user’s data privacy.
8. Application of Our Scheme
As an application, we consider the hospital’s data management in cloud setting as an example to demonstrate the effectiveness of our scheme to the actual pandemic data management. During the prevention and control of the epidemic, many medical data of patients need to be recorded, including the patients’ nucleic acid testing results, the doctor’s diagnosis, and if diagnosed as COVID-19, the recent whereabouts of the patients. In fact, the amount of these data is very huge; if the hospital stores them locally, it will consume a lot of storage resources, so it is better to outsource these huge data to cloud servers for storage. However, the cloud server is not completely reliable. Many important data may be lost due to various unexpected accidents. This will cause the treatment of many patients to be delayed due to the loss of data. Furthermore, the loss of some key data of diagnosed patients may lead to inadequate control of the epidemic, which may lead to the spread of the epidemic. If there is no data integrity audit mechanism, we cannot know whether the data are completely stored. So, an integrity audit mechanism is used to ensure the integrity of cloud data. At the same time, the public key is usually used to generate the authentication information of patient data. Due to the large number of patients and the continuous increase in the number of patients, the key management is a big problem. Our scheme uses identity-based way to generate public and private keys and directly uses the user’s identity information to generate public keys, which effectively avoids the difficulty of key management.
Figure 2 illustrates the system model. It includes five entities. The five entities are patients, hospital (which we termed as HSP), KGC, CSP, and TPA. Because the KGC generates the private keys for the patients, it is necessary for the hospital to select a trusted key generation server as the KGC. As the cloud server needs to store a large amount of medical data, servers with strong storage capacity are selected as the cloud storage servers. Since the TPA requires a lot of audit work, a server with powerful computing power is used as the TPA. There are three steps in this model, and they are key generation, data upload, and integrity verification. The concrete implementation is as follows:
Step 1 (key generation): first, the KGC in the HSP sets the parameters and calculates the master key in the Setup stage. When the patient comes to the hospital, the hospital generates the patient’s ID based on the patient’s identity information and sends the ID to KGC. Then, the KGC computes the identity-based private key for the patient according to the Extract algorithm.
Step 2 (data upload): after the hospital collected the patients’ data, these data need to be uploaded to the HSP’s storage server. Firstly, the hospital computes the corresponding tag for the patient based on the corresponding collected data. Then, it uploads the data blocks and the corresponding tags to the HSP’s storage server. According to the policy, the tags and auxiliary information based on patient’s identity are transmitted to TPA, ensuring that the TPA can implement the auditing for the data stored in the cloud server.
Step 3 (integrity verification): in order to guarantee the integrity of the data, HSP needs to check it regularly. First of all, the hospital or patients make a request for integrity verification to TPA, which in turn uses the challenge-response auditing protocol to verify the integrity of the data stored in the cloud, as requested. If the verification is successful, the patient’s data are considered to be good stored in the cloud server. Otherwise, the CSP does not store the patient’s data well, and other patient’s data blocks may also be lost. At this time, other important data need to be checked also, and if the data are lost, remedial measures such as backup and recovery of the lost data are needed in time.
9. Conclusions
In this paper, we review a lightweight ID-based verifiable data ownership cloud storage audit scheme proposed by Rabaninejad et al. [17]. Then, we point out the security vulnerabilities in the OffTagGen and OnTagGen part of the scheme and further demonstrate the insecurity of the original protocol by showing the attack. In order to protect the integrity of users’ data, an improved secure cloud audit protocol is proposed. The security analysis shows that the new protocol is secure.
Acknowledgments
This study was supported by the National Natural Science Foundation of China under grant nos. U1636114, 62102452, and 62172436, Open Project from Guizhou Provincial Key Laboratory of Public Big Data under grant no. 2019BDKFJJ008, Engineering University of PAP’s Funding for Scientific Research Innovation Team under grant no. KYTD201805, and Engineering University of PAP’s Funding for Key Researcher under grant no. KYGG202011.
[1] K. Ren, C. Wang, Q. Wang, "Security challenges for the public cloud," IEEE Internet Computing, vol. 16 no. 1, pp. 69-73, DOI: 10.1109/mic.2012.14, 2012.
[2] M. Khorshed, A. B. M. Ali, S. Wasimi, "A survey on gaps, threat remediation challenges and some thoughts for proactive attack detection in cloud computing," Future Generation Computer Systems, vol. 28, pp. 833-851, DOI: 10.1016/j.future.2012.01.006, 2012.
[3] J. Gudeme, S. Pasupuleti, R. Kandukuri, "Review of remote data integrity auditing schemes in cloud computing: taxonomy, analysis, and open issues," International Journal of Cloud Computing, vol. 8,DOI: 10.1504/ijcc.2019.10019207, 2019.
[4] G. Ateniese, R. Burns, R. Curtmola, J. Herring, L. Kissner, Z. Peterson, D. Song, "Provable data possession at untrusted stores," pp. 598-609, DOI: 10.1145/1315245.1315318, 2007.
[5] M. Shah, R. Swaminathan, M. Baker, "Privacy-preserving audit and extraction of digital contents," IACR Cryptology ePrint Archive, vol. 2008, 2008.
[6] H. Shacham, B. Waters, "Compact proofs of retrievability," Proceedings of the International Conference on the Theory and Application of Cryptology and Information Security, pp. 90-107, DOI: 10.1007/978-3-540-89255-7_7, .
[7] R. Curtmola, O. Khan, R. Burns, G. Ateniese, "MR-PDP: multiple-replica provable data possession," Proceedings of the 28th International Conference on Distributed Computing Systems,DOI: 10.1109/icdcs.2008.68, .
[8] Z. Hao, S. Zhong, N. Yu, "A privacy-preserving remote data integrity checking protocol with data dynamics and public verifiability," IEEE Transactions on Knowledge and Data Engineering, vol. 23, pp. 1432-1437, DOI: 10.1109/TKDE.2011.62, 2011.
[9] C. Wang, Q. Wang, K. Ren, W. Lou, "Privacy-preserving public auditing for data storage security in cloud computing," Proceedings of the 2010 Proceedings IEEE INFOCOM, pp. 525-533, DOI: 10.1109/infcom.2010.5462173, .
[10] H. Wang, Q. Wu, B. Qin, J. Domingo-Ferrer, "Identity-based remote data possession checking in public clouds," Information Security, vol. 8, pp. 114-121, DOI: 10.1049/iet-ifs.2012.0271, 2014.
[11] H. Wang, "Identity-based distributed provable data possession in multicloud storage," IEEE Transactions on Services Computing, vol. 8, pp. 328-340, DOI: 10.1109/tsc.2014.1, 2015.
[12] H. Wang, D. He, S. Tang, "Identity-based proxy-oriented data uploading and remote data integrity checking in public cloud," IEEE Transactions on Information Forensics and Security, vol. 11 no. 6, pp. 1165-1176, DOI: 10.1109/tifs.2016.2520886, 2016.
[13] Y. Yu, M. Au, G. Ateniese, X. Huang, W. Susilo, Y. Dai, G. Min, "Identity-based remote data integrity checking with perfect data privacy preserving for cloud storage," IEEE Transactions on Information Forensics and Security, vol. 12 no. 4, pp. 767-778, DOI: 10.1109/tifs.2016.2615853, 2017.
[14] Y. Yu, L. Xue, M. H. Au, W. Susilo, J. Ni, Y. Zhang, A. Vasilakos, J. Shen, "Cloud data integrity checking with an identity-based auditing mechanism from RSA," Future Generation Computer Systems, vol. 62,DOI: 10.1016/j.future.2016.02.003, 2016.
[15] J. Li, H. Yan, Y. Zhang, "Identity-based privacy preserving remote data integrity checking for cloud storage," IEEE Systems Journal, vol. 15 no. 1, pp. 577-585, DOI: 10.1109/jsyst.2020.2978146, 2020.
[16] R. Rabaninejad, S. M. Sedaghat, M. Ahmadian Attari, M. R. Aref, "An ID-based privacy-preserving integrity verification of shared data over untrusted cloud," Computer Society of Iran, vol. 2020,DOI: 10.1109/csicc49403.2020.9050098, 2020.
[17] R. Rabaninejad, M. R. Asaar, M. A. Attari, M. R. Aref, "An identity-based online/offline secure cloud storage auditing scheme," Cluster Computing, vol. 23, pp. 1455-1468, DOI: 10.1007/s10586-019-03000-5, 2020.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2022 Xu An Wang et al. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Data integrity verification mechanisms play an important role in cloud environments. Recently, a lightweight identity-based cloud storage audit scheme has been proposed; this paper points out security vulnerabilities of their OffTagGen algorithm. That is, the attackers such as malicious cloud servers can forge the tags, which can destroy data integrity. By improving the construction of OffTagGen algorithm, an improved security cloud auditing protocol is proposed in this work to better protect user’s privacy. The analysis shows that the new protocol is effective and resistant to attacks.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details



1 Engineering University of PAP, Xi’an, China; State Key Laboratory of Public Big Data, Guizhou University, Guiyang, China
2 Engineering University of PAP, Xi’an, China
3 Xianyang Vocational Technical College, Xianyang, China
4 Nanjing University of Posts and Telecommunications, Nanjing, China
5 Chongqing University Cancer Hospital, Chongqing, China