Sat - 56.Pdf - Analysis On Voting Data Using Deduplication Techniques in Cloud
Sat - 56.Pdf - Analysis On Voting Data Using Deduplication Techniques in Cloud
Sat - 56.Pdf - Analysis On Voting Data Using Deduplication Techniques in Cloud
v
TABLE OF CONTENTS
ABSTRACT Vi
1. INTRODUCTION 1
1.1 OVERVIEW 1
2. LITERATURE SURVEY 3
3. SYSTEM DESIGN 6
3.2 OBJECTIVE 6
5. SYSTEM IMPLEMENTATION 25
vi
6. CONCLUSION 37
vii
LIST OF FIGURES
6.2 INTERFACE 39
6.3 REGISTRATION 40
6.4 LOGIN 41
viii
CHAPTER 1
INTRODUCTION
1.1 OVERVIEW
Cloud computing provides seemingly unlimited “virtualized” resources to
users as services across the whole Internet, while hiding platform and
implementation details. Today’s cloud service providers offer both highly available
storage and massively parallel computing resources at relatively low costs. As cloud
computing becomes prevalent, an increasing amount of data is being stored in the
cloud and shared by users with specified privileges, which define the access rights of
the stored data. One critical challenge of cloud storage services is the management
of the ever-increasing volume of data. To make data management scalable in cloud
computing, de duplication has been a well-known technique and has attracted more
and more attention recently.
1
It encrypts/decrypts a data copy with a convergent key, which is obtained by
computing the cryptographic hash value of the content of the data copy. After key
generation and data encryption, users retain the keys and send the cipher text to the
cloud. Since the encryption operation is Deterministic and is derived from the data
content, identical data copies will generate the same convergent key and hence the
same cipher text. To prevent unauthorized access, a secure proof of ownership
protocol is also needed to provide the proof that the user indeed owns the same file
when a duplicate is found. After the proof, subsequent users with the same file will be
provided a pointer from the server without needing to upload the same file. A user
can download the encrypted file with the pointer from the server, which can only be
decrypted by the corresponding data owners with their convergent keys.
In order to save cost and efficiently management, the data will be moved to
the storage server provider (SSP) in the public cloud with specified privileges and the
de duplication technique will be applied to store only one copy of the same file.
Because of privacy consideration, some files will be encrypted and allowed the
duplicate check by employees with specified privileges to realize the access control.
Traditional de duplication systems based on convergent encryption, although
providing confidentiality to some extent; do not support the duplicate check with
differential privileges. In other words, no differential privileges have been considered
in the de duplication based on convergent encryption technique.
2
CHAPTER 2
LITERATURE SURVEY
2. LITERATURE SURVEY:
J.Johin, R.K. Rohith, R.S.S. Yukesh Kumar, G. Paavi [1] in the paper on
“Deduplication in Cloud for Improvising Efficiency Towards Potential Practical
Usage” proposed that if the duplicate of files is available on the cloud server, the
application needs to map the original files to the current file that is uploaded so that
user will not get affected by the mapping but on backend no duplicate files will be
created but the files get mapped by the application, In this way the duplication files
on the cloud server will be avoided thereby reducing the storage space requirement
on the cloud server.
Dr. MK Jayanthi, P. Sri Vaibhavi, P.V. Naga Sathya [5] in the paper on
“Achieved Efficient Data Deduplication and Key Aggregation” in this the project
implemented a scenario where the cloud service can deduplicate the upload data.
The data was prevented from unwanted exposure and unauthorized access by
placing a proper access control mechanism. Authorized data deduplication aims at
data security to keep the data aims at data security to keep the data secured and
avoid unauthorized access. Deduplication at encryption level saves a lot of memory
and memory can be utilized efficiently.
Y. Li, T Wang, G. Wang, J. Liang and H. Chen [8] in the paper on “Efficient
data Collection in Sensor Cloud System Using Multiple Mobile Sinks” in this
paper the results shown that our algorithm can uploaded the data from WSNs to
4
cloud with the limited latency and minimized the energy consumption as well. The
outcome shows that our calculation can communicate information from WSN to the
cloud in a brief time frame and lessen power utilization. Distributed computing further
develops information handling and wireless sensor networks(WSNs).
5
CHAPTER 3
SYSTEM DESIGN
3.2 OBJECTIVE
The main goal is to enable de duplication and distributed storage of the data
across multiple storage servers.
6
De duplication can take place at either the file level or the block level. For file
level de duplication, it eliminates duplicate copies of the same file. De duplication can
also take place at the block level, which eliminates duplicate blocks of data that occur
in non-identical files. Identical data copies of different users will lead to different
cipher texts, making de duplication impossible.
3.4.1 DISADVANTAGES
Traditional encryption, while providing data confidentiality, is incompatible
with data de duplication.
Identical data copies of different users will lead to different cipher texts,
making de duplication impossible.