** Cloud Storage Security Deduplication Scheme Based on Dynamic Bloom Filter **

Xi-ai Yan* , Wei-qi Shi* and Hua Tian*

## Article Information

## Abstract

**Abstract:** Data deduplication is a common method to improve cloud storage efficiency and save network communication bandwidth, but it also brings a series of problems such as privacy disclosure and dictionary attacks. This paper proposes a secure deduplication scheme for cloud storage based on Bloom filter, and dynamically extends the standard Bloom filter. A public dynamic Bloom filter array (PDBFA) is constructed, which improves the efficiency of ownership proof, realizes the fast detection of duplicate data blocks and reduces the false positive rate of the system. In addition, in the process of file encryption and upload, the convergent key is encrypted twice, which can effectively prevent violent dictionary attacks. The experimental results show that the PDBFA scheme has the characteristics of low computational overhead and low false positive rate.

**Keywords:** Bloom Filter , Cloud Storage , Data Deduplication , Privacy Protection

## 1. Introduction

With the rapid development of cloud computing technology, more and more users choose to outsource data to cloud service providers. According to IDC analysis, cloud storage space is expected to reach 44 ZB [1] by 2020. How to effectively reduce storage space has become a hot research topic. Data deduplication technology provides an effective solution for minimizing storage capacity. It can automatically search for duplicate data, keep only one copy of all the same data, and make the actual data stored in the system decline geometrically.

In traditional storage systems, data are stored in plaintext. Users have full ownership and management rights. Deduplication technology of data is relatively mature, and more research results have been achieved, such as duplication detection of the same data and duplication detection of similar data. However, in cloud storage systems, the following challenges exist in data deduplication: (1) data is often stored in the ciphertext form, and different keys encrypt the same plaintext to generate different ciphertext; (2) data management and ownership are separated, users lose absolute control of data, and always worry about data security; (3) bandwidth resources and computing capital of cloud data are very valuable. How to detect duplicate data blocks efficiently and cost less is very important.

It uses virtualization technology to achieve on-demand allocation of multi-users storage resources in cloud computing. When cloud servers detect data duplication among multi-users, it is easy to cause user privacy disclosure. How to realize data deduplication and protect user privacy is an urgent problem, in this paper, it proposed a security deduplication scheme based on Bloom filter. A concise global Bloom filter array is constructed to realize data ownership proof and duplication comparison. The Bloom filter is dynamically extended to effectively alleviate the problem that the false positive rate of Bloom filter increases too fast.

## 2. Related Research

Research on data security deduplication in cloud environment has achieved certain results. Douceur et al. [2] first proposed convergent encryption (CE) technology, which is an encryption technology using hash value of data as the key. CE can effectively protect data confidentiality while deleting duplicate data. Li et al. [3] combined secret sharing AONT-RS with CE, adopted convergent diffusion mechanism and proposed a CDStore scheme, which replaced random information in diffusion algorithm with hash fingerprint of data, and ensured the certainty of the algorithm to achieve data security and deduplication. Stanek et al. [4] combined with CE, proposed a scheme to provide different levels of security encryption for data according to different levels of privacy. In view of the security and privacy problems faced by cloud data security deduplication schemes, Puzio et al. [5] proposed a secure and effective storage system, which provides data block level deduplication and confidentiality. On the basis of CE, additional semantic security encryption schemes and access control mechanisms are added to resist the existing CE attacks.

In view of the attack that the opponent obtains the fingerprint information of the user’s files and obtains the complete files from the server by using the client’s deduplication mechanism, scholars put forward the concept of ownership certification. Halevi et al. [6] first introduced the concept of “proof of ownership” (PoW), which can be verified only when the user has complete data, but can’t be validated only if he has partial data. Xu et al. [7] improved Halevi’s scheme, first encrypted the original file, then constructed MHT for PoW verification, and proved the scheme to be secure under the random oracle model. In order to reduce the computational overhead and the number of I/O reads and writes, Di Pietro and Sorniotti [8] proposed a PoW optimization scheme (s-PoW) to achieve efficient and information-theoretically secure document ownership certification. In order to further improve the efficiency of server-side in existing PoW schemes, Blasco et al. [9] proposed a flexible, scalable and provably secure cloud data deduplication scheme based on Bloom filter. By comparing the validation information of file blocks with Bloom filter, we can judge whether the user has data or not, which can greatly reduce the overhead of client and server-side. However, data security issue is not discussed in [9]. [10] combines CE with Bloom filter to ensure the security of experimental data, but adopts standard Bloom filter, without discussing the false positive rate of Bloom filter. In Bloom filter research, [11] proposes a high-precision multi-dimensional Bloom filter, which uses bijective function to transform multi-dimensional attributes of elements into one-dimensional values to represent the overall information of elements, which can effectively reduce the misjudgment rate. [12] propose a privacy-preserving reversible Bloom filter. With the help of homomorphic encryption functions, message aggregation can be accomplished by simple multiplication and addition of ciphertext. [13] proposes an efficient grouping classification algorithm based on counting Bloom filter, which applies counting Bloom filters to predict the failures of flow table lookups without traversing flow tables for most mask probing. [14] proposes a ciphertext retrieval ranking method based on counting Bloom filter (CRRM-CBF) constructs secure retrieval index through Bloom filter to achieve efficient retrieval of ciphertext data and dynamic update of index. [15] proposes a homomorphic encryption protocol based on Bloom filter, which improves the execution efficiency of the protocol by virtue of the advantages of low computational complexity, high spatial utilization and high query efficiency of Bloom filter.

## 3. Preparatory Knowledge

3.1 Standard Bloom filterBloom filter is a spatially efficient randomized data structure for the simplified representation of element sets and membership queries. Bloom filter is widely used in network packet classification, IP routing lookup, deep packet detection, and so on [16]. Standard Bloom filters are m-bit vectors V representing the aggregate information of queries. Assuming that the set of elements is [TeX:] $$S=\left\{s_{1}, s_{2} \dots s_{n}\right\}$$ , K hash functions [TeX:] $$h_{1}(x), h_{2}(x) \ldots h_{k}(x)$$ corresponding to each element, the range of values of the hash functions is [TeX:] $$[0, m-1]$$ . The initial state of the vector V is all 0. When the element is stored, the hash function [TeX:] $$h_{1}(x), h_{2}(x) \ldots h_{k}(x)$$ of the V is set to be all 1. When the element is queried, whether the search bits of [TeX:] $$h_{1}(x), h_{2}(x) \ldots h_{k}(x)$$ are all 1. If all the bits are 1, the element is in the set. Otherwise, the element is not in the set [17].

Example: Let’s assume that the length of the vector is [TeX:] $$m=6 b i t$$ , the number of hash functions is [TeX:] $$k=2,h_{1}(x)=x \bmod 6, h_{2}(x)=(2 x+1) \bmod 6$$ , respectively, [TeX:] $$S=\{13,14\}$$ ,whether the query element [TeX:] $$\{12,19\}$$ is in the set or not. The results of insertion and query are shown in Fig. 1.

Bloom filters may judge elements that do not belong to this set as elements belonging to this set. Let single Bloom filter false positive rate [TeX:] $$F P R(n, m, k)$$ , where n denotes the number of elements in set S, m denotes the number of filters, K denotes the number of hash functions.

The probability of one bit being zero in the filter is [TeX:] $$\left(1-\frac{1}{m}\right)^{k}$$ , the probability of one bit being 1 is [TeX:] $$1-\left(1-\frac{1}{m}\right)^{k n}$$ , if the hash function is uniformly distributed, and then [TeX:] $$F P R(n, m, k)$$ can be expressed as:

##### (1)

[TeX:] $$F P R(n, m, k)=\left(1-\left(1-\frac{1}{m}\right)^{k n}\right)^{k} \approx\left(1-e^{-k n / m}\right)^{k}=\exp \left(k \ln \left(1-e^{-k n / m}\right)\right)$$Let the number of Bloom filters is r, and use to express the false positive rate of global Bloom filters:

Formula (1) can be changed into:

Formula (1) shows that the false positive rate of the filter increases with the increase of the number of elements in the set. If the upper limit of the false positive rate of a single Bloom filter is f and the size of the elements in the set is n_1, it can be seen from formula (2) that if the false positive rate is within the controllable range, it must:

3.2 Convergent Encryption

To solve the problem that the traditional random encryption algorithm can’t detect repeatability and store multiple copies of the same file in the cloud at the same time, which seriously wastes storage space, Douceur et al. [2] proposed CE algorithm. The key generation algorithm is a deterministic algorithm, which is usually obtained by hashing the original data to ensure that the same data gets the same key.

The basic cryptographic primitives based on CE algorithm are as follows:(1) Key generation：[TeX:] $$K=H(F), H$$ is Hash function;

(2) Encryption：[TeX:] $$C=E_{H(F)}(F)$$ , the client encrypts the file;

(3) Decryption：[TeX:] $$F=D_{H(F)}(C)$$ , users get ciphertext from the server and decrypt it;

(4) Label generation：[TeX:] $$i d=G E N(F)$$ , generating file or data block labels.

The ciphertext verifiability of CE algorithm makes it widely used in the field of cloud data deduplication. Many research results combine CE with various mechanisms to achieve deduplication of ciphertext data. CE is widely used to construct secure data deduplication system, but it also faces many dangers such as data leakage, copy forgery attack, choice plaintext attack, and content distribution network attack, and so on [18,19].

## 4. System Model

The cloud storage security deduplication model based on dynamic Bloom filter (DBF) is shown in Fig. 2. The main entities include users, cloud service providers (CSP) and third-party servers (TPS).

The users divides file F into blocks of the same size, performs CE, and generates ciphertext blocks and block labels. A public dynamic Bloom filter is deployed on the TPS to realize user privilege detection and data block repeatability detection. According to the test results, it decides whether to issue the privilege label, upload data or return a storage pointer to the user. Another function of TPS is to achieve secondary encryption of convergent keys and effectively prevent violent dictionary attacks. CSP clothing is responsible for storing data blocks and block labels, and accepting user uploads and downloads of data blocks according to the privilege labels.

## 5. System Design and Implementation

5.1 Design of Dynamic Bloom FilterIn the design of Bloom filter, firstly a Bloom filter is constructed for permission query to achieve identity authentication, and then a Bloom filter is constructed for each storage node to achieve data deduplication. All Bloom filters are isomorphic and form a public Bloom filter array (PBFA) globally. Bloom filter array can achieve public weight removal, not limited to a single storage node. The PBFA is shown in Fig. 3.

In order to slow down the growth of false alarm rate, a DBF is designed for each storage node. DBF is dynamically composed of several standard sub-filters according to the size of the set elements, [TeX:] $$D B F=\left\{D B F_{1}, D B F_{2}, D B F_{3}, \ldots D B F_{t-1}, D B F_{t}\right\}$$ , a PDBFA is constructed globally.

The DBF creation algorithm is as follows:

Step 1: Initialize the standard Bloom filter, set the upper limit f and bit m of the false positive rate of the filter, and calculate n_1according to formula (3).

Step 2: Create the first Bloom filter vector [TeX:] $$D B F_{1}$$ .

Step 3: if [TeX:] $$\leq n_{1}$$ , map elements to [TeX:] $$D B F_{1}$$ .

Step 4: if [TeX:] $$n>n_{1}$$ , create [TeX:] $$D B F_{2}$$ (isomorphic to [TeX:] $$D B F_{1}$$ ), map elements to [TeX:] $$D B F_{2}$$ .

Step 5: if [TeX:] $$n>2 n_{1}$$ , create [TeX:] $$D B F_{3}$$ (isomorphic to [TeX:] $$D B F_{1}$$ ), map elements to [TeX:] $$D B F_{3}$$ .

Step 6: if [TeX:] $$n>(t-1) n_{1}$$ , where [TeX:] $$t=\left\lfloor n / n_{1}\right\rfloor+ 1$$ , create [TeX:] $$D B F_{t}$$ (isomorphic to [TeX:] $$D B F_{1}$$ ), map elements to [TeX:] $$D B F_{t}$$ .

The maximum number of elements that DBF can represent in a controllable range is n:

[TeX:] $$D F P R\left(n, n_{1} m, k\right)$$ denotes false positive rate of DBF，the number of last filter elements is [TeX:] $$i=n-n_{1} \times\left\lfloor n / n_{1}\right\rfloor$$ , According to formula (1), [TeX:] $$D F P R\left(n, n_{1} m, k\right)$$ can be expressed as:

##### (6)

[TeX:] $$\begin{aligned} D F P R\left(n, n_{1} m, k\right) &=1-\left(1-F P R\left(n_{1}, m, k\right)\right)^{\left.| n / n_{1}\right\rfloor}(1-F P R(i, m, k)) \\ &=1-\left(1-\left(1-e^{-k n_{1} / m}\right)^{k}\right)^{\left\lfloor n / n_{1}\right\rfloor}\left(1-\left(1-e^{-k i / m}\right)^{k}\right) \end{aligned}$$[TeX:] $$G D F P R\left(n, n_{1} m, k, r\right)$$ denotes false positive rate of DBFA:

##### (7)

[TeX:] $$G D F P R\left(n, n_{1} m, k, r\right)=1-\left(1-D F P R\left(n, n_{1} m, k\right)\right)^{r}$$Assuming [TeX:] $$k=4, m=6400$$ , the upper limit of false positive rate of single standard sub-filter is [TeX:] $$f=0.01$$ , it can be obtained by formula (3):

The number of elements that a single sub-filter can tolerate most is 405.

According to formulas (2) and (7), the false positive rate of PBFA and PDBFA increases with the increase of elements. Assuming r = 10, the change of false positive rate is shown in Fig. 4. From Fig. 4, PDBFA can dynamically build new sub-filters, which can effectively control the growth of false positive rate at the expense of a certain storage space.

5.2 Implementation of Safe Deduplication SchemeThe first section is system initialization as follows:

(1) Data preprocessing: dividing the file into blocks: [TeX:] $$F=\left(F_{1}, F_{2}, F_{3} \ldots F_{i}\right)$$ , initializing four algorithms of CE scheme.

(2) Privilege initialization: System privileges are entirely S, suppose the system has N users, each user's privileges is [TeX:] $$S_{U}$$ , the access privilege is [TeX:] $$S_{F_{i}}$$ to data block [TeX:] $$F_{i}$$ .

(3) Bloom filter initialization: Mapping access right of data block [TeX:] $$F_{i}$$ to the first Bloom filter through K hash functions, and mapping data blocks of each storage node to corresponding Bloom filters through K hash functions.

The second section is file upload as follows:

Step 1: Authentication request, user asks TSP for query [TeX:] $$S_{U}$$ in Bloom filters.

Step 2: Bloom filters queries and updates. Whether the corresponding K bit of [TeX:] $$S_{U}$$ is all 1, if all 1, the query passes; otherwise, the user is required to prove the ownership of the data, if prove ok, the corresponding K bit is set to 1 by updating the Bloom filter.

Step 3: Distribution of permission labels. After the permission query is passed, the corresponding permission key [TeX:] $$K_{S_{i}}$$ is selected, the permission label [TeX:] $$\phi_{i}=i d \oplus K_{S_{i}}$$ is generated and distributed to the user.

Step 4: Repetitive data block detection. User uses [TeX:] $$\phi_{i}$$ for repetitive detection through row manner in PDBFA. When duplication is found, a storage pointer is issued to user, and corresponding permission is allocated, so data is no longer uploaded to the cloud server.

Step 5: Non-duplicate data processing. When the user access all data in PDBFA and no duplicate block are found, it proves to be a new data block.

Step 6: Data block encryption upload. User randomly select [TeX:] $$r \in Z_{q}$$ as the blinding factor, and send the convergent key to TPS after blinding process, TPS encrypts the blinded convergent key twice to obtain [TeX:] $$K_{i}^{\prime}$$ . User process the blinded key [TeX:] $$K_{i}^{\prime}$$ to obtain [TeX:] $$K_{i}^{\prime \prime}$$ , and upload the ciphertext [TeX:] $$C_{i}=E_{K_{i}^{\prime \prime}} \left(F_{i}\right)$$ to CSP.

The third section is file download as follows:

Step 1: Users request to download data block [TeX:] $$F_{i}$$ , send data block ID and corresponding permissions [TeX:] $$S_{U}$$ to TPS.

Step 2: TPS queries the DBF for access rights. If not, the access fails. If yes, a privilege label [TeX:] $$\phi_{i}$$ is issued to the user, and then the user will send it to CSP.

Step 3: CSP will send the corresponding ciphertext block [TeX:] $$C_{i}$$ and [TeX:] $$K_{i}^{\prime}$$ to the user, the user will obtain [TeX:] $$K_{i}^{\prime \prime}$$ after de-blinding process of [TeX:] $$K_{i}^{\prime}$$ , decrypt [TeX:] $$F_{i}=D_{K_{i}^{\prime \prime}}\left(C_{i}\right)$$ .

## 6. Analysis of Experimental Performance

6.1 Experimental EnvironmentDatasets are collected from various portals by using web downloaders. The selected websites are Sina, Tencent, and Netease. The file formats are web pages, text, pictures, music, animation and video. The total number of files in the dataset is 562352, with a total size of 725.78 GB. The hardware environment of the experiment is shown in Table 1.

In the experiment, the SHA-1 algorithm is used to calculate the data block label, and the SHA-2 algorithm is used to calculate the convergence key of the data block. The capacity of single Bloom filter is m = 65536 bit, and the number of hash functions is K = 6.

6.2 Upload Time Overhead of Data BlockBaseline scheme proposed in [20] is a classical scheme for cloud storage security deduplication. Baseline scheme first calculates the file tag and blocks, then calculates the block convergence key to encrypt the block, calculates the block ciphertext hash value to prove ownership, and finally uses the private key to encrypt the convergence key and upload it to the storage server. This experiment compares PDBFA scheme with Baseline scheme, and compares the time cost of file upload from three aspects: no duplication, duplication, and partial duplication.

When there is no duplication of data blocks, the complexity of computing hash value increases with the increase of data blocks, and the upload time of data blocks also increases. In PDBFA, it takes a part of time to initialize the Bloom filter, so there is little difference between them. But because PDBFA uses DBF to prove ownership, the overall time overhead is slightly better, and the comparison results are shown in Fig. 5.

When data blocks are duplicated, baseline scheme needs to calculate block ciphertext hash value for ownership certification in the global scope. In PDBFA scheme, because of the use of DBF, users can only query within the scope of data with access rights efficiently, which reduces the time of authorization verification. With the increase of data blocks, the advantages of Baseline scheme are more obvious, and the results are compared as shown in Fig. 6.

When the repeatability of data blocks is different, the fixed size of data blocks is 100 MB. The data blocks are compared at different repeatability rates of 0%, 25%, 50%, 75%, and 100%, respectively. As the repetition rate increases, the encryption operation decreases, so the overall upload time overhead decreases. The key encryption time of baseline scheme is fixed. But in PDBFA scheme, the number of convergent key encryption decreases greatly due to the increasing of repetition rate. Compared with baseline scheme, the time overhead advantage is obvious. The comparison results are shown in Fig. 7.

##### 6.3 False Positive Rate Analysis

Set the bit of standard Bloom filter is m = 65536, the upper limit of false positive rate of standard filter is f = 0.001, according to formula (3), obtained [TeX:] $$n_{1}=4152$$ . The comparison of false positive rate between PBFA and PDBFA is shown in Fig. 8 when the number of fingerprints in coding block is different.

The experimental results show that when the number of data blocks [TeX:] $$n<4152$$ , the two mechanisms are the same, and similar false positive rate can be obtained. When [TeX:] $$n=2 n_{1}$$ , the false positive rate of PBFA was about 8 times that of PDBFA. The reason was that PDBFA expanded the capacity of Bloom filter and controlled the actual false positive rate of single Bloom filter within the effective threshold. Subsequently, as the number of data blocks continues to grow, the false positive rate difference between them further enlarges.

## 7. Summary

Data security deduplication in cloud environment is a very active research direction. At present, it is still in its infancy, and there is still a big gap between key technologies and practical applications. Bloom filter uses a very compact form to represent information. Hash lookup time is constant, storage space is low, and cache is quite low. It is especially suitable for the search and recognition of feature characters in massive data sets.

Aiming at the privacy protection and fast deduplication of cloud storage, this paper proposes a secure deduplication scheme of cloud storage based on DBF, which improves the efficiency of ownership certification and realizes fast detection of duplicate data blocks. In addition, the convergent key is encrypted twice by means of blind processing method, which can effectively resist violent dictionary attacks. In the scheme, a scalable DBF is designed to control the false positive rate within a certain threshold. The experimental results show that the computational overhead of file upload is relatively small, and it can effectively control the false positive rate of the system.

Compared with previous studies, the contribution of this paper is to design dynamic global Bloom filter array to complete de-duplication and proof of ownership, which greatly reduces the cost of query and proof time for duplicate data blocks. The disadvantage is that the sequence structure and time structure of the stream data are not designed in Bloom filter, and the dynamic stream data related to time cannot be processed directly. The future plan is to study the time dimension expansion of Bloom filter, reduce the storage space and computing overhead of stream data processing, and provide efficient security protection for stream data.

## Biography

## Biography

## References

- 1
*Cisco Global cloud index (Online). Available:*, https://www.cisco.com/c/dam/assets/sol/sp/gci/global-cloud-index-infographic.html - 2 J. R. Douceur, A. Adya, W. J. Bolosky, P. Simon, M. Theimer, "Reclaiming space from duplicate files in a serverless distributed file system," in
*Proceedings 22nd International Conference on Distributed Computing Systems*, Vienna, Austria, 2002;pp. 617-624. custom:[[[-]]] - 3 M. Li, C. Qin, P. P. Lee, "CDStore: toward reliable, secure, and cost-efficient cloud storage via convergent dispersal," in
*Proceedings of 2015 USENIX Annual Technical Conference*, Santa Clara, CA, 2015;pp. 111-124. custom:[[[-]]] - 4 J. Stanek, A. Sorniotti, E. Androulaki, L. Kencl, "A secure data deduplication scheme for cloud storage,"
*in Financial Cryptography and Data Security. Heidelberg: Springer*, pp. 99-118, 2014.doi:[[[10.1007/978-3-662-45472-5_8]]] - 5 P. Puzio, R. Molva, M. Onen, S. Loureiro, "PerfectDedup: secure data deduplication,"
*in Data Privacy Managementand Security Assurance. Cham: Springer*, pp. 150-166, 2015.doi:[[[10.1007/978-3-319-29883-2_10]]] - 6 S. Halevi, D. Harnik, B. Pinkas, A. Shulman-Peleg, "Proofs of ownership in remote storage systems," in
*Proceedings of the 18th ACM Conference on Computer and Communications Security*, Chicago, IL, 2011;pp. 491-500. custom:[[[-]]] - 7 J. Xu, E. C. Chang, J. Zhou, "Weak leakage-resilient client-side deduplication of encrypted data in cloud storage," in
*Proceedings of the 8th ACM SIGSAC Symposium on Information*, Computer and Communications Security, Hangzhou, China, 2013;pp. 195-206. custom:[[[-]]] - 8 R. Di Pietro, A. Sorniotti, "Boosting efficiency and security in proof of ownership for deduplication," in
*Proceedings of the 7th ACM Symposium on Information*, Computer and Communications Security, Seoul, Korea, 2012;pp. 81-82. custom:[[[-]]] - 9 J. Blasco, R. Di Pietro, A. Orfila, A. Sorniotti, "A tunable proof of ownership scheme for deduplication using bloom filters," in
*Proceedings of 2014 IEEE Conference on Communications and Network Security*, San Francisco, CA, 2014;pp. 481-489. custom:[[[-]]] - 10 Z. Liu, Z. Yang, "Efficient and secure deduplication cloud storage scheme based on proof of ownership by Bloom filter,"
*Journal of Computer Applications*, vol. 37, no. 3, pp. 766-770, 2017.custom:[[[-]]] - 11 W. Li, D. F. Zhang, K. Huang, K. Xie, "Accurate multi-dimension counting Bloom filter for big data processing,"
*Chinese Journal of Electronics*, vol. 43, no. 4, pp. 652-657, 2015.doi:[[[10.3969/j.issn.0372-2112.2015.04.005]]] - 12 K. Xie, W. Shi, "PPIBF: a privacy preservation invertible Bloom filter,"
*Computer Engineering & Science*, vol. 39, no. 6, pp. 1104-1111, 2017.custom:[[[-]]] - 13 J. Zhao, Z. Hu, B. Xiong, K. Li, "Accelerating packet classification with counting bloom filters for virtual openflow switching,"
*China Communications*, vol. 15, no. 10, pp. 117-128, 2018.doi:[[[10.1109/CC.2018.8485474]]] - 14 Y. Li, Z. Xiang, "Ciphertext retrieval ranking method based on counting Bloom filter,"
*Journal of Computer Applications*, vol. 38, no. 9, pp. 2554-2559, 2018.custom:[[[-]]] - 15 E. Zhang, G. Jin, "Cloud outsourcing multiparty private set intersection protocol based on homomorphic encryption and Bloom filter,"
*Journal of Computer Applications*, vol. 38, no. 8, pp. 2256-2260, 2018.custom:[[[-]]] - 16 S. Dharmapurikar, P. Krishnamurthy, D. E. Taylor, "Longest prefix matching using bloom filters,"
*IEEE/ACM Transactions on Networking*, vol. 14, no. 2, pp. 397-409, 2006.doi:[[[10.1109/TNET.2006.872576]]] - 17 C. M. Tseng, J. R. Ciou, T. J. Liu, "A cluster-based data de-duplication technology," in
*Proceedings of the 2nd International Symposium on Computing and Networking*, Shizuoka, Japan, 2014;pp. 226-230. custom:[[[-]]] - 18 J. Xiong, Y. Zhang, F. Li, S. Li, J. Ren, Z. Y ao, "Research progress on secure data deduplication in cloud,"
*Journal on Communications*, vol. 37, no. 11, pp. 169-180, 2016.doi:[[[10.11959/j.issn.1000-436x.2016238]]] - 19 P. Anderson, L. Zhang, "Fast and secure laptop backups with encrypted de-duplication," in
*Proceedings of the 24th Large Installation System Administration (LISA)*, San Jose, CA, 2010;custom:[[[-]]] - 20 L. Gonzalez-Manzano, A. Orfila, "An efficient confidentiality-preserving proof of ownership for deduplication,"
*Journal of Network and Computer Applications*, vol. 50, pp. 49-59, 2015.doi:[[[10.1016/j.jnca.2014.12.004]]]