SlideShare a Scribd company logo
1 of 7
Download to read offline
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3800
Cloud Based Deduplication Using Middleware Approach
Sujit Tilak1, Shrey Jakhmola2, Arvind Hariharan Nair3, Bhabya Mishra4 , Isha Dwivedi5
1Professor, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India
2B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India
3B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India
4B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India
5B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India
---------------------------------------------------------------------***---------------------------------------------------------------------
Abstract – Data Deduplication is an integral part of cloud
storage. It efficiently reduces redundant data and thus
reducing storage space. Data deduplication can be doneinthe
server side or in the client side location. But, providing a
middleware will be beneficial for the users, as they can have
their preference for the storage solution. The benefit of
implementing through client side is a reduction in the use of
bandwidth and storage which would in turn results in user
satisfaction, but it induces load on the user system,
synchronization problem with the server and also increases
difficulties in the technical aspect of the overall architecture.
Data deduplication is done by splitting the file into small
chunks in the server side and further generating a hash value
which will be unique to the file content. The hash values are
stored in the database. The hashing can be effectively
performed by the SHA1 algorithm, since it has no noticeable
collision thereby maintaining security. If the same hash value
has been found in the storage server database then the file is
not stored again but a pointer is created that points to the
original file which will be used when a user requests for the
file. Here, the entire deduplication process is performed in a
middleware which connects the user and the server.
Key Words: Cloud storage, deduplication, middleware,
chunking, hashing, SHA1,
1.INTRODUCTION
Cloud storage is a cloud computing model that is used to
store data online which can be accessed anytime by the user
with active Internet connectivity. It can provide strong
protection for our data and can be beneficial for cases like
data backup and archival. It is also reliable during natural
disasters, as the data can be stored in a remote server. The
overall cost of using a cloud storage device is low as there is
no need to purchase hardware; therefore reducing
maintenance cost drastically compared to the traditional
storage systems. Cloud not only store data. It is also known
for being the shared pool of resources. In cloud computing,
duplication of data is the main problem. Data duplication, as
the name suggests is duplication of file while storing in the
cloud. Removing the duplicate file from the storage space is
termed as deduplication.
Data deduplicationcanbeperformedatvariouslocationsand
levels using several methodologies. Deduplication can be
either be implemented in a client sideor server sidelocation.
But implementing in either side has its' advantages and
disadvantages. Performing the process in the client side will
increase the load on the client system. Softwareisrequiredin
the client side to proceed for deduplication. While on the
server side, high bandwidth is utilized by the client. To
overcome some of these issues in the client and server side,
the middleware approach comes in to picture. The
middleware acts as a bridge for the user and the storage. It
allows the software to be isolated from the storage solution.
It also helps reduce storage space. Instead, the user can rely
on existing storage like Google Drive, Dropbox or any other
user preference as this application is platform independent.
Middleware is the place where the core deduplication
procedure takes place that involves fixed size chunking and
hashing using the SHA1 algorithm.
2. LITERATURE SURVEY
Many researchers have done commendable work on the
subject. This section contains the overview of the work done
by some of the researchers on deduplication.
1) Data Deduplication Cluster Based on Similarity-Locality
Approach [2] .
Data deduplication can be done in file level, block level and
byte level. The block level deduplication will reduce the
expenses of the system and detection of segment withinafile
which is redundant can be detected. The similarity and
locality concept is combined and used in this deduplication.
The similarity deduplication take the similar kind of
characteristics from backup stream. Whileontheotherhand,
locality deduplication means the order of the backup stream
chunks will remain same in other high probability backup
streams. The file is segmented and distributed to several
nodes. These incoming segments are checked withrespectto
the local finger print. The Bloom filters are used to store
these fingerprints. The incoming segments are identified by
using MD5 or SHA1 algorithm by the fingerprints. If the
fingerprint is similar then the segment isn't stored. But if it's
not, then based on the degree of similarity comparison is
done with the fingerprint summary or corresponding nodes.
2) A Hybrid Cloud Approach for Secure Authorized
Deduplication [3] .
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3801
Hybrid Cloud Approach is a combination of both private and
public cloud. The private cloud is responsible here for the
sharing the file token to the user on request and providing
with the privatekey. Through this waytheprivateclouddoes
the authentication of the user so that only authorized user
can check for deduplication of data, upload or download the
file. Once the user gets hold ofthe token, it then sharesitwith
the public cloud for upload or download purpose. The file
token is received by the public cloud and results are given
accordingly as per the request.
3) TryManagingYourDeduplicationFine-grained-ly:AMulti-
tieredandDynamicSLA-drivenDeduplicationFrameworkfor
Primary Storage [4].
The primary storage system is made up by two main
modules: proxy server and object storage cluster (OSC). The
client side requests are sent to the OSC storage node by the
proxy server. The newly entered files meta data is present in
the buffer zone present in the proxy server. OSC is a
distributed storage cluster that contains approximately
hundreds ofstorage nodes made up of variouschunks.These
storage nodes maintain the fingerprint in the index table for
all the chunks. Fingerprint is nothing but hash of the content
that is done by using SHA1 technique. Deduplication doesn't
interfere with the Mudder on going operations, instead it
performs the task in the background itslef.DCL-Module,SLA-
Handler, Deduplication Executor(Deduper) and Workload
Monitor (Monitor) are the main modules of the Mudder.
When new files meta data are filles in the buffer zoneand the
entries reach a threshold value N, DCL module fetches all
these entries and generate a Target list. DCL module is
operable insidethe proxy server. Initially the entry in the list
is predefined as dirty, meaning not duplicate. The DCL
module then calculated for deduplication. DCL-Module
pushes the list to Deduper which is a channel where
deduplication between storage nodes and proxy takes place.
Procedure of removing redundant data and index update is
similar to other systems.
4) DeduplicationincloudstorageusingHashingtechniquefor
encrypted data [6].
The data to be sent for uploading purpose is encrypted first
before sending it to the Cloud Service Provider. The
encryption is done using the AES 128 bit encryption that is
protected from various attacks, one being Birthday brute
force attack. The data is encryptedinthedataownermachine
by using a secure key. In this way the Cloud Service Provider
doesn’t have access to the plain text of the file to be uploaded
because it receives a encryptedcopyofthefile.Theencrypted
file received undergoes chunkingandhashing.Thehashingis
performed using the MD5 algorithm. If the hash value of the
data is same as that already present in the cloud, then the
data isn’t stored again. If the hash isn’t present in the cloud
then the file is stored in the cloud and the location is updated
in theindex table of Cloud Service Provider.
5) Improving the Performance of System in Cloud By Using
Selective Deduplication [7].
The user of the system can request for upload. This upload
takes place from LAN or WAN or MAN to thecloud.Itdepend
on the users whether they wanttocheck thededuplicationof
the data or not. The selective deduplication is applied on the
data to find the duplicate data if any. If any duplicate data is
found then the feedback is sent to user and the redundant
data is deleted from the cloud storage. If in case, the data is
unique, that is, not redundant, then user is asked for
encryption. Security key for the file is generated and
encryption is performend on the file. After encryption, the
file is stored in the cloud storage. When the user requires,
decryption can be performed on the file and thus a key is
generated for decrypting file from cloud.
6) PFP: Improving the Reliability of Deduplication-based
Storage Systems with Per-File Parity [8].
During the process of deduplication, the reliability gets
affected because of the removal of redundant chunks. These
identified duplicate chunks are removed because they are
common to other files too. At the end, there is just one chunk
that is shared by all the files after the deduplication. This
removal of chunks renders the sharing files unavailable.This
paper gave an idea of Per File Parity(PFP) to improve the
reliability. PFP performs the XOR parity within the parity
groups of data chunks which are part of each file after the
chunks are created but beforethesechunksarededuplicated.
PFP performs XOR parity within the data chunks parity
group. Therefore, PFP provides security from the parity
redundancy. The PFP doesn't hinder the flow of data
duplication.Thefingerprintcalculation,chunking,detectionof
duplicate chunks and removal of these duplicate chunks,
nothing changes.Parity generation is implemented on every
file to prevent deduplication of the data. The resulting parity
metadata is stored as the metadata of file. The incoming file
data chunks arecreatedandthefingerprintsarechecked.PFP
then divides these chunks into groups as per predefined
group size of N. For N data chunk group the parity is
calculated by usingXORoperation.Thisschemecantoleratea
single chunk failures.
3. PROPOSED SYSTEM
The proposed system comprises of components like the
Client, Middleware and the Server. Middleware is the core
component where all the processing takes place.
Fig -1: System Architecture
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3802
A middleware is a software thatacts as a bridgebetweentwo
endpoints, generally a client and a server. The middleware
gives the users an interface to upload the files. The user has
to create an account on the application in order to perform
functions like upload or download. Once the user has an
account, they can add files to their storage using the
Middleware. The Middleware will be handling the process of
data deduplication.
Fig -2: Middleware and associated modules
The middleware consists of the the following modules:
1. Authentication Module : It is used to authenticate the user
so that the user can access the middleware dashboard. The
user needs to supply their credentials in the login form to get
access to their account. The authentication module also
provided a service to sign up for the account of the
middleware.
2. Chunking Module: The chunking module is invoked by an
API call which chunks the file into small fixed size pieces in a
synchronized manner which is then passed on tothehashing
module.
3. Hashing Module: The hashing module takes the chunked
file input from the chunking module and generates SHA1
hash of each chunk and stores it in an array until the the
chunks of the current file being processed is completely
supplied to the module. After the individual hashing of the
chunks, the SHA1 of the wholearraywhichconsistsofhashof
the chunk is hashed and returned to the API.
4. Drive Module: The drive module is used for uploading the
user file to the user’s drive storage. It is also used by the cron
job process to sync drive storage content information with
the middleware by using the metadata of the files present on
the cloud of the user and alsoto download the file from users
cloud for deduplication processing.
5. API: The API Module is used by other modules to interact
to other middleware or topassthedatathroughouttheentire
middleware. It handles the incoming requests and data
response forother modules using HTTP requests. Thedatais
wrapped on the response body.
6. Cron Module: The cron module runs automatically
everyday once to sync withtheuserscloudstorage.Thejobof
the cron module is to check whether the user has uploaded
any file directly to the cloud storage by fetching the file’s
metadata and checking in the database. If the file has never
been processed by the middleware, then it fetches the file
from the cloud storage usingtheAPIanddrivemoduleforthe
process of deduplication.
The processing of the file can take place based on the two
scenarios with respect to the method of file upload. In the
first scenario, the user uploads the file to the middleware via
the uploader. In the second scenario, if the user uploads the
file directly to the cloud storage or for the old files on cloud
storage need to go through the process of deduplication
which is done using the cron job module.
The file which is uploaded tothe middleware or downloaded
from the user’s cloud storage will go through chunking and
hashing. Based on the value of hash generated it is
determined whether duplication is present or not by
checking in the database withrespect to that individual user.
If the hash of the file exists in the database then a pointer is
added to the location where the originalfileexistsandthefile
at hand is discarded. But, if the hash is not present in the
database then the current fileinformationalongwithitshash
value is added in the database. While syncing with the cloud
storage, using file ID from the metadata of the file, we check
whether the file has been ever processed or not. If not
processed then it goes through the above process otherwise
it is sent for batch processing.
3.1. Techniques Used
The system focuses on removal of the duplicatedataoncethe
file is uploaded by the user. The major techniques usedinthe
system are chunking and hashing that is performed in the
middleware. The chunking and hashing are used for the
deduplication process. The deduplication is done in order to
remove the redundancy. If the user uploads a file, and the
same is present in the storage server then a pointer is
assigned to the original file instead of saving the new file
again. Deduplication reduces use of storage space.
This is performed in two steps:
1. Chunking: This is a technique that breaks a large piece of
information say file into multiple pieces or chunks. These
chunks are helpful for further data deduplication. Chunking
can be done by either using fixed size chunking or file level
chunking.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3803
Fig -3: Chunking Process
2. Hashing: This is a method used to generate a value of the
incoming data ofa file through a mathematical function. This
is performed during the deduplication procedure in the
middleware. The hash values can be generated by either
using SHA1 or MD5 algorithm.SHA1ispreferredoverMD5as
the collision experience is less. Also, SHA1 is faster with
respect to it's successors.
Fig -4: Hashing Process
The chunks created from the chunking process are hashed
and pushed into an array. Finally afterallthechunkshashare
pushed into an array, the arrays' hash is generated that is
used fordeduplication. Themiddleware runsaservicewhich
is used to sync the data already present on the user's cloud
storage. The sync checks whether the file present on the
cloud storage is processed by the middleware or not using
the fileID from metadata. The file is fetched and processed
from the cloud server and same deduplication process is
done if the file is not processed ever otherwise the files are
sent for batch processing which is optional or not required if
the cloud is only used for storage and not for editing files.
3.2 Proposed System Procedure Flow
The user's file upload command is examined by the
middleware. The middleware fetches the file from the user
for chunking and hashing. The chunking process first takes
place which gives small chunks of the file whose hash is
generated. The number of chunks depend on the size of the
file with respect to the chunk size. One by one the hash of the
chunks are generated and pushed into an array. Once all the
chunks of the file are created, the hashing of the complete
array is generated. This generated hash of the file is now
compared with the hash values present in the database.
As shown in Fig 5, there are two scenarios as stated below:
Fig -5: Procedural Flow
Case I: Hash generated of the file is not present in the
database. In this case the database is updated along with the
new hash value. The file in this case is uploaded in the
storage server.
Case II: Hash generated of the file is present in the database.
In this case the file is not saved again. Rather a pointer is
added that will point to the original file presentinthestorage
server.
3.3 Use Case Diagram
There are three actors involved in the system as showninFig
6. Those can be described as:
1. User: The user is the one who will use the deduplication
application.
2. Middleware: The middleware is the system which will
check for deduplication and ensure that there is no duplicate
file in the storage server.
3. Storage: Storage is a system where the file of the users is
stored and retrieved.
The system validates the user credentials and logs in them
into the system. The middleware fetches the file from the
storage server using the appropriate credentials. The
middleware uploads the processed file to the storage server
usingappropriatecredentials. Theuseruploadsthefiletothe
middleware using the dashboard. The middleware create
chunks of the file which was uploaded by the user. The
chunks created will be hashed and stored in a hash digest.
The middleware will now store the hash generated to the
database.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3804
Fig -6: Use Case Diagram
4. PERFORMANCE EVALUATION
4.1 Datasets Used
The datasets used in this section are iso files. An ISO file is a
disc image of an optical disc and contains everythingthatcan
be written to the optical drive. In earlier days ISO was the
format used to burn CDs and DVDs whereas, in the present
days an ISO file is used to transfer a larger amount of files
compressed into a single ISO file which is then shared over
the internet using peer to peer or FTP servers. The name ISO
is taken from the ISO 9660 file system.Thedatasetstakenare
ISO image of Linux distributions that are most widely used.
These datasets are relatively large, usually, the size rangesin
Gigabytes. The large dataset will ensure harmonious and
distinct results between fixed size hashing and file level
hashing.
Table -1: Datasets Involved
Datasets Dataset Name Type Provide
d By
Size
(GiB)
Dataset 1 Antergos-18.12-
x86_64.is
ISO
File
Antergos
.com
2.1
Dataset 2 openSUSE-
Tubleweed-DVD-
x86_64Snapshot2
0181128-
Media.iso
ISO
File
Suse
Linux
4.1
Dataset 3 Popos_18.10_amd
64_intel_11.iso
ISO
File
System
76 inc.
2
Dataset 4 Ubuntu-18.04.1-
desktop-
amd64.iso
ISO
File
Canonica
l inc.
1.8
4.2 Tools Involved
The performance analysis done on the proposed system
involves the use of following tools:
1. Time command linux: The time command runs the
specified program with the provided arguments. When
command finishes the execution timewrites back a message
to standard error giving timing and processing statistics
about the program that was runned.
2. htop for cpu usage: Htop is an interactive system viewer
and process managing tool. Itis based on UNIX tool top. Htop
allows the user to view process or tasks statistics like CPU
usage, ram consumption, and the process uptime.
3. console.time() and console.timeEnd() for nodejs function:
These inbuilt methods are used to benchmark javascript
applications for response and average computation time
required to complete the given task. console.time(label) is
used to set the initial reference from where the timer would
begin.console.timeEnd(label)isusedastheclosingstatement
the tasks in between console.time() and console.timeEnd is
bench marked for time.
4.3. Evaluation Metrics
The metrics usedforbenchmarkingthesystemsperformance
with respect to the defined constraints is the Evaluation
metrics.
1. Hashing Time: The Hashing Time is the time required (in
Seconds) to produce the hash of a given input. The hash
function encodes each byte as two hexadecimal characters.
The hash time should be lower in order to increase the
efficiency of the application. Lower hash time also helps in
scaling the application if needed.
Table -2: Hashing Time in seconds
Datasets File Level Fixed Size
Dataset 1 67.3 24.19
Dataset 2 132.66 46.1
Dataset 3 82.59 22.6
Dataset 4 34.5 25
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3805
Fig -7. Time consumption graph for hashing
2. Chunking Time: The time required to divide a file into
chunks is termed as chunking time . It is usually measured in
milliseconds (ms). The lesser the chunking time, the faster
the file data will be processed and analyzed. It obtained
results won't be better than file level processing, as the
chunking time increase.
Table -3: Chunking Time in milliseconds
Datasets Fixed Size
Dataset 1 1.7
Dataset 2 2.47
Dataset 3 2.93
Dataset 4 2.41
Fig -8. Time consumption graph for chunking
3. CPUload: CPU load measures theamountofcomputational
work in percentage that the system performs in order chunk
and hash the input data. The load on the CPU should be
neutral and consistent. The more the load on the CPU the
more the chances of the application getting halt. Also, less
CPU load ensures less cost on dedicated hardware and
increased scalability of the application.
Table -4: CPU Load in Percentage
Datasets File Level Fixed Size
Dataset 1 40 23
Dataset 2 66 40
Dataset 3 38 37
Dataset 4 40 33
Fig -9: Tme consumption graph for CPU
5. CONCLUSION
In effect of all the results that were obtained in our proposed
system, the evaluation of the performance of our project, as
well as the extensive Literature survey conducted by us in
this project it can be concluded that the the middleware
approach forperforming the deduplication is moreeffiecient
and cost effective. The workload on the client and serverside
isreduced.Theperformanceevaluationrevealsthatchunking
is the most efficient form of file processing and hence it must
be used in data dedupication. Thehashingisperformedusing
SHA1 algorithm. We discovered that, SHA1 is significantly
more securethan MD5 as it isprone to less collisionsand itis
significantly faster than SHA256 and SHA512. Hence, it
provides the bestbalancebetweensecurityandperformance.
REFERENCES
[1] W. Litwin, Thomas Schwarz,"Combining Chunk
Boundary and Chunk Signature Calculations for
Deduplication ", IEEE, 2012.
[2] Jian Zhang, Xingyu Zhang, "Data Deduplication Cluster
Based on Similarity-Locality Approach",IEEE, 2013.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072
© 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3806
[3] Jin Li, Yan Kit Li, Xiaofeng Chen, Patrick P. C. Lee,
Wenjing Lou, "A Hybrid Cloud Approach for Secure
Authorized Deduplication", IEEE, 2014.
[4] Yan Tang, Jianwei Yin, and Zhaohui Wu, "Try Managing
Your Deduplication Fine-grained-ly: A Multi-tiered and
Dynamic SLA-driven Deduplication Framework for
Primary Storage", IEEE, 2016.
[5] Nancy Digra1 , Sandeep Sharma, "A Noval Approach of
Enhancing Security in Cloud Using Diffie Hellman
Algorithm",IJSRM, 2017.
[6] Vaishnavi Moorthy, Arpit Parwal and Udit
Rout,"Deduplication in Cloud Storage Using Hashing
Technique For Encrypted Data",IEEE, 2018.
[7] Mr. Nishant N.Pachpor, Dr. Prakash S.Prasad,
"Improving the Performance of System in Cloud by
Using Selective Deduplication", IEEE, 2018.
[8] Suzhen Wu, Bo Mao, Hong Jiang, Huagao Luan, Jindong
Zhou,"PFP: Improving the Reliability of Deduplication-
based Storage Systems with Per-File Parity",IEEE,2019.

More Related Content

What's hot

Data integrity proof techniques in cloud storage
Data integrity proof techniques in cloud storageData integrity proof techniques in cloud storage
Data integrity proof techniques in cloud storageIAEME Publication
 
International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER)International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER)ijceronline
 
Improving Efficiency of Security in Multi-Cloud
Improving Efficiency of Security in Multi-CloudImproving Efficiency of Security in Multi-Cloud
Improving Efficiency of Security in Multi-CloudIJTET Journal
 
Improved Data Integrity Protection Regenerating-Coding Based Cloud Storage
Improved Data Integrity Protection Regenerating-Coding Based Cloud StorageImproved Data Integrity Protection Regenerating-Coding Based Cloud Storage
Improved Data Integrity Protection Regenerating-Coding Based Cloud StorageIJSRD
 
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...neirew J
 
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...ijccsa
 
A hybrid cloud approach for secure authorized
A hybrid cloud approach for secure authorizedA hybrid cloud approach for secure authorized
A hybrid cloud approach for secure authorizedNinad Samel
 
F233842
F233842F233842
F233842irjes
 
A Hybrid Cloud Approach for Secure Authorized De-Duplication
A Hybrid Cloud Approach for Secure Authorized De-DuplicationA Hybrid Cloud Approach for Secure Authorized De-Duplication
A Hybrid Cloud Approach for Secure Authorized De-DuplicationEditor IJMTER
 
Distributed Algorithms with DDS
Distributed Algorithms with DDSDistributed Algorithms with DDS
Distributed Algorithms with DDSAngelo Corsaro
 
Ieeepro techno solutions 2014 ieee java project - distributed, concurrent, ...
Ieeepro techno solutions   2014 ieee java project - distributed, concurrent, ...Ieeepro techno solutions   2014 ieee java project - distributed, concurrent, ...
Ieeepro techno solutions 2014 ieee java project - distributed, concurrent, ...hemanthbbc
 
A Hybrid Cloud Approach for Secure Authorized Deduplication
A Hybrid Cloud Approach for Secure Authorized DeduplicationA Hybrid Cloud Approach for Secure Authorized Deduplication
A Hybrid Cloud Approach for Secure Authorized Deduplication1crore projects
 
DATA PROVENENCE IN PUBLIC CLOUD
DATA PROVENENCE IN PUBLIC CLOUDDATA PROVENENCE IN PUBLIC CLOUD
DATA PROVENENCE IN PUBLIC CLOUDijsrd.com
 
Enabling Integrity for the Compressed Files in Cloud Server
Enabling Integrity for the Compressed Files in Cloud ServerEnabling Integrity for the Compressed Files in Cloud Server
Enabling Integrity for the Compressed Files in Cloud ServerIOSR Journals
 
A hybrid cloud approach for secure authorized deduplication
A hybrid cloud approach for secure authorized deduplicationA hybrid cloud approach for secure authorized deduplication
A hybrid cloud approach for secure authorized deduplicationTmks Infotech
 
Getting Started with Vortex
Getting Started with VortexGetting Started with Vortex
Getting Started with VortexAngelo Corsaro
 
IRJET - Confidential Image De-Duplication in Cloud Storage
IRJET - Confidential Image De-Duplication in Cloud StorageIRJET - Confidential Image De-Duplication in Cloud Storage
IRJET - Confidential Image De-Duplication in Cloud StorageIRJET Journal
 
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...dbpublications
 
iaetsd Controlling data deuplication in cloud storage
iaetsd Controlling data deuplication in cloud storageiaetsd Controlling data deuplication in cloud storage
iaetsd Controlling data deuplication in cloud storageIaetsd Iaetsd
 

What's hot (20)

Data integrity proof techniques in cloud storage
Data integrity proof techniques in cloud storageData integrity proof techniques in cloud storage
Data integrity proof techniques in cloud storage
 
El35782786
El35782786El35782786
El35782786
 
International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER)International Journal of Computational Engineering Research(IJCER)
International Journal of Computational Engineering Research(IJCER)
 
Improving Efficiency of Security in Multi-Cloud
Improving Efficiency of Security in Multi-CloudImproving Efficiency of Security in Multi-Cloud
Improving Efficiency of Security in Multi-Cloud
 
Improved Data Integrity Protection Regenerating-Coding Based Cloud Storage
Improved Data Integrity Protection Regenerating-Coding Based Cloud StorageImproved Data Integrity Protection Regenerating-Coding Based Cloud Storage
Improved Data Integrity Protection Regenerating-Coding Based Cloud Storage
 
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...
Neuro-Fuzzy System Based Dynamic Resource Allocation in Collaborative Cloud C...
 
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...
NEURO-FUZZY SYSTEM BASED DYNAMIC RESOURCE ALLOCATION IN COLLABORATIVE CLOUD C...
 
A hybrid cloud approach for secure authorized
A hybrid cloud approach for secure authorizedA hybrid cloud approach for secure authorized
A hybrid cloud approach for secure authorized
 
F233842
F233842F233842
F233842
 
A Hybrid Cloud Approach for Secure Authorized De-Duplication
A Hybrid Cloud Approach for Secure Authorized De-DuplicationA Hybrid Cloud Approach for Secure Authorized De-Duplication
A Hybrid Cloud Approach for Secure Authorized De-Duplication
 
Distributed Algorithms with DDS
Distributed Algorithms with DDSDistributed Algorithms with DDS
Distributed Algorithms with DDS
 
Ieeepro techno solutions 2014 ieee java project - distributed, concurrent, ...
Ieeepro techno solutions   2014 ieee java project - distributed, concurrent, ...Ieeepro techno solutions   2014 ieee java project - distributed, concurrent, ...
Ieeepro techno solutions 2014 ieee java project - distributed, concurrent, ...
 
A Hybrid Cloud Approach for Secure Authorized Deduplication
A Hybrid Cloud Approach for Secure Authorized DeduplicationA Hybrid Cloud Approach for Secure Authorized Deduplication
A Hybrid Cloud Approach for Secure Authorized Deduplication
 
DATA PROVENENCE IN PUBLIC CLOUD
DATA PROVENENCE IN PUBLIC CLOUDDATA PROVENENCE IN PUBLIC CLOUD
DATA PROVENENCE IN PUBLIC CLOUD
 
Enabling Integrity for the Compressed Files in Cloud Server
Enabling Integrity for the Compressed Files in Cloud ServerEnabling Integrity for the Compressed Files in Cloud Server
Enabling Integrity for the Compressed Files in Cloud Server
 
A hybrid cloud approach for secure authorized deduplication
A hybrid cloud approach for secure authorized deduplicationA hybrid cloud approach for secure authorized deduplication
A hybrid cloud approach for secure authorized deduplication
 
Getting Started with Vortex
Getting Started with VortexGetting Started with Vortex
Getting Started with Vortex
 
IRJET - Confidential Image De-Duplication in Cloud Storage
IRJET - Confidential Image De-Duplication in Cloud StorageIRJET - Confidential Image De-Duplication in Cloud Storage
IRJET - Confidential Image De-Duplication in Cloud Storage
 
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...
Role Based Access Control Model (RBACM) With Efficient Genetic Algorithm (GA)...
 
iaetsd Controlling data deuplication in cloud storage
iaetsd Controlling data deuplication in cloud storageiaetsd Controlling data deuplication in cloud storage
iaetsd Controlling data deuplication in cloud storage
 

Similar to IRJET- Cloud based Deduplication using Middleware Approach

A Survey Paper on Removal of Data Duplication in a Hybrid Cloud
 A Survey Paper on Removal of Data Duplication in a Hybrid Cloud  A Survey Paper on Removal of Data Duplication in a Hybrid Cloud
A Survey Paper on Removal of Data Duplication in a Hybrid Cloud IRJET Journal
 
thilaganga journal 1
thilaganga journal 1thilaganga journal 1
thilaganga journal 1thilaganga
 
Review and Analysis of Self Destruction of Data in Cloud Computing
Review and Analysis of Self Destruction of Data in Cloud ComputingReview and Analysis of Self Destruction of Data in Cloud Computing
Review and Analysis of Self Destruction of Data in Cloud ComputingIRJET Journal
 
Dynamic Resource Allocation and Data Security for Cloud
Dynamic Resource Allocation and Data Security for CloudDynamic Resource Allocation and Data Security for Cloud
Dynamic Resource Allocation and Data Security for CloudAM Publications
 
Distributed Scheme to Authenticate Data Storage Security in Cloud Computing
Distributed Scheme to Authenticate Data Storage Security in Cloud ComputingDistributed Scheme to Authenticate Data Storage Security in Cloud Computing
Distributed Scheme to Authenticate Data Storage Security in Cloud ComputingAIRCC Publishing Corporation
 
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGDISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGAIRCC Publishing Corporation
 
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGDISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGijcsit
 
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET-  	  Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET-  	  Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET Journal
 
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET Journal
 
IRJET- Protection of Personal Data on Distributed Cloud using Biometrics
IRJET- Protection of Personal Data on Distributed Cloud using BiometricsIRJET- Protection of Personal Data on Distributed Cloud using Biometrics
IRJET- Protection of Personal Data on Distributed Cloud using BiometricsIRJET Journal
 
An Approach towards Shuffling of Data to Avoid Tampering in Cloud
An Approach towards Shuffling of Data to Avoid Tampering in CloudAn Approach towards Shuffling of Data to Avoid Tampering in Cloud
An Approach towards Shuffling of Data to Avoid Tampering in CloudIRJET Journal
 
IRJET- Improving Data Availability by using VPC Strategy in Cloud Environ...
IRJET-  	  Improving Data Availability by using VPC Strategy in Cloud Environ...IRJET-  	  Improving Data Availability by using VPC Strategy in Cloud Environ...
IRJET- Improving Data Availability by using VPC Strategy in Cloud Environ...IRJET Journal
 
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...IRJET Journal
 
Flaw less coding and authentication of user data using multiple clouds
Flaw less coding and authentication of user data using multiple cloudsFlaw less coding and authentication of user data using multiple clouds
Flaw less coding and authentication of user data using multiple cloudsIRJET Journal
 
IRJET- Redsc: Reliablity of Data Sharing in Cloud
IRJET- Redsc: Reliablity of Data Sharing in CloudIRJET- Redsc: Reliablity of Data Sharing in Cloud
IRJET- Redsc: Reliablity of Data Sharing in CloudIRJET Journal
 
IRJET- A Research Paper on Block Design-based Key Agreement for Group Dat...
IRJET-  	  A Research Paper on Block Design-based Key Agreement for Group Dat...IRJET-  	  A Research Paper on Block Design-based Key Agreement for Group Dat...
IRJET- A Research Paper on Block Design-based Key Agreement for Group Dat...IRJET Journal
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESijccsa
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESijccsa
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESijccsa
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESneirew J
 

Similar to IRJET- Cloud based Deduplication using Middleware Approach (20)

A Survey Paper on Removal of Data Duplication in a Hybrid Cloud
 A Survey Paper on Removal of Data Duplication in a Hybrid Cloud  A Survey Paper on Removal of Data Duplication in a Hybrid Cloud
A Survey Paper on Removal of Data Duplication in a Hybrid Cloud
 
thilaganga journal 1
thilaganga journal 1thilaganga journal 1
thilaganga journal 1
 
Review and Analysis of Self Destruction of Data in Cloud Computing
Review and Analysis of Self Destruction of Data in Cloud ComputingReview and Analysis of Self Destruction of Data in Cloud Computing
Review and Analysis of Self Destruction of Data in Cloud Computing
 
Dynamic Resource Allocation and Data Security for Cloud
Dynamic Resource Allocation and Data Security for CloudDynamic Resource Allocation and Data Security for Cloud
Dynamic Resource Allocation and Data Security for Cloud
 
Distributed Scheme to Authenticate Data Storage Security in Cloud Computing
Distributed Scheme to Authenticate Data Storage Security in Cloud ComputingDistributed Scheme to Authenticate Data Storage Security in Cloud Computing
Distributed Scheme to Authenticate Data Storage Security in Cloud Computing
 
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGDISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
 
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTINGDISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
DISTRIBUTED SCHEME TO AUTHENTICATE DATA STORAGE SECURITY IN CLOUD COMPUTING
 
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET-  	  Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET-  	  Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
 
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASCIRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
IRJET- Secure Data Sharing Scheme for Mobile Cloud Computing using SEDASC
 
IRJET- Protection of Personal Data on Distributed Cloud using Biometrics
IRJET- Protection of Personal Data on Distributed Cloud using BiometricsIRJET- Protection of Personal Data on Distributed Cloud using Biometrics
IRJET- Protection of Personal Data on Distributed Cloud using Biometrics
 
An Approach towards Shuffling of Data to Avoid Tampering in Cloud
An Approach towards Shuffling of Data to Avoid Tampering in CloudAn Approach towards Shuffling of Data to Avoid Tampering in Cloud
An Approach towards Shuffling of Data to Avoid Tampering in Cloud
 
IRJET- Improving Data Availability by using VPC Strategy in Cloud Environ...
IRJET-  	  Improving Data Availability by using VPC Strategy in Cloud Environ...IRJET-  	  Improving Data Availability by using VPC Strategy in Cloud Environ...
IRJET- Improving Data Availability by using VPC Strategy in Cloud Environ...
 
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...
IRJET- An Integrity Auditing &Data Dedupe withEffective Bandwidth in Cloud St...
 
Flaw less coding and authentication of user data using multiple clouds
Flaw less coding and authentication of user data using multiple cloudsFlaw less coding and authentication of user data using multiple clouds
Flaw less coding and authentication of user data using multiple clouds
 
IRJET- Redsc: Reliablity of Data Sharing in Cloud
IRJET- Redsc: Reliablity of Data Sharing in CloudIRJET- Redsc: Reliablity of Data Sharing in Cloud
IRJET- Redsc: Reliablity of Data Sharing in Cloud
 
IRJET- A Research Paper on Block Design-based Key Agreement for Group Dat...
IRJET-  	  A Research Paper on Block Design-based Key Agreement for Group Dat...IRJET-  	  A Research Paper on Block Design-based Key Agreement for Group Dat...
IRJET- A Research Paper on Block Design-based Key Agreement for Group Dat...
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
 
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUESANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
ANALYSIS OF ATTACK TECHNIQUES ON CLOUD BASED DATA DEDUPLICATION TECHNIQUES
 

More from IRJET Journal

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASIRJET Journal
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesIRJET Journal
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web applicationIRJET Journal
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
 

More from IRJET Journal (20)

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil Characteristics
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADAS
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare System
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridges
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web application
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
 

Recently uploaded

HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSRajkumarAkumalla
 
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSMANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSSIVASHANKAR N
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...Call Girls in Nagpur High Profile
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Christo Ananth
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxupamatechverse
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSISrknatarajan
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝soniya singh
 
Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxpranjaldaimarysona
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...ranjana rawat
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINESIVASHANKAR N
 
Extrusion Processes and Their Limitations
Extrusion Processes and Their LimitationsExtrusion Processes and Their Limitations
Extrusion Processes and Their Limitations120cr0395
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
 
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 

Recently uploaded (20)

HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
 
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLSMANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
MANUFACTURING PROCESS-II UNIT-5 NC MACHINE TOOLS
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptx
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
 
Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptx
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
 
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
Call Girls in Nagpur Suman Call 7001035870 Meet With Nagpur Escorts
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
 
Extrusion Processes and Their Limitations
Extrusion Processes and Their LimitationsExtrusion Processes and Their Limitations
Extrusion Processes and Their Limitations
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
★ CALL US 9953330565 ( HOT Young Call Girls In Badarpur delhi NCR
 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
 
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
 

IRJET- Cloud based Deduplication using Middleware Approach

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3800 Cloud Based Deduplication Using Middleware Approach Sujit Tilak1, Shrey Jakhmola2, Arvind Hariharan Nair3, Bhabya Mishra4 , Isha Dwivedi5 1Professor, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India 2B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India 3B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India 4B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India 5B.E. Student, Department of Computer Engineering, Pillai College of Engineering, Maharashtra, India ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract – Data Deduplication is an integral part of cloud storage. It efficiently reduces redundant data and thus reducing storage space. Data deduplication can be doneinthe server side or in the client side location. But, providing a middleware will be beneficial for the users, as they can have their preference for the storage solution. The benefit of implementing through client side is a reduction in the use of bandwidth and storage which would in turn results in user satisfaction, but it induces load on the user system, synchronization problem with the server and also increases difficulties in the technical aspect of the overall architecture. Data deduplication is done by splitting the file into small chunks in the server side and further generating a hash value which will be unique to the file content. The hash values are stored in the database. The hashing can be effectively performed by the SHA1 algorithm, since it has no noticeable collision thereby maintaining security. If the same hash value has been found in the storage server database then the file is not stored again but a pointer is created that points to the original file which will be used when a user requests for the file. Here, the entire deduplication process is performed in a middleware which connects the user and the server. Key Words: Cloud storage, deduplication, middleware, chunking, hashing, SHA1, 1.INTRODUCTION Cloud storage is a cloud computing model that is used to store data online which can be accessed anytime by the user with active Internet connectivity. It can provide strong protection for our data and can be beneficial for cases like data backup and archival. It is also reliable during natural disasters, as the data can be stored in a remote server. The overall cost of using a cloud storage device is low as there is no need to purchase hardware; therefore reducing maintenance cost drastically compared to the traditional storage systems. Cloud not only store data. It is also known for being the shared pool of resources. In cloud computing, duplication of data is the main problem. Data duplication, as the name suggests is duplication of file while storing in the cloud. Removing the duplicate file from the storage space is termed as deduplication. Data deduplicationcanbeperformedatvariouslocationsand levels using several methodologies. Deduplication can be either be implemented in a client sideor server sidelocation. But implementing in either side has its' advantages and disadvantages. Performing the process in the client side will increase the load on the client system. Softwareisrequiredin the client side to proceed for deduplication. While on the server side, high bandwidth is utilized by the client. To overcome some of these issues in the client and server side, the middleware approach comes in to picture. The middleware acts as a bridge for the user and the storage. It allows the software to be isolated from the storage solution. It also helps reduce storage space. Instead, the user can rely on existing storage like Google Drive, Dropbox or any other user preference as this application is platform independent. Middleware is the place where the core deduplication procedure takes place that involves fixed size chunking and hashing using the SHA1 algorithm. 2. LITERATURE SURVEY Many researchers have done commendable work on the subject. This section contains the overview of the work done by some of the researchers on deduplication. 1) Data Deduplication Cluster Based on Similarity-Locality Approach [2] . Data deduplication can be done in file level, block level and byte level. The block level deduplication will reduce the expenses of the system and detection of segment withinafile which is redundant can be detected. The similarity and locality concept is combined and used in this deduplication. The similarity deduplication take the similar kind of characteristics from backup stream. Whileontheotherhand, locality deduplication means the order of the backup stream chunks will remain same in other high probability backup streams. The file is segmented and distributed to several nodes. These incoming segments are checked withrespectto the local finger print. The Bloom filters are used to store these fingerprints. The incoming segments are identified by using MD5 or SHA1 algorithm by the fingerprints. If the fingerprint is similar then the segment isn't stored. But if it's not, then based on the degree of similarity comparison is done with the fingerprint summary or corresponding nodes. 2) A Hybrid Cloud Approach for Secure Authorized Deduplication [3] .
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3801 Hybrid Cloud Approach is a combination of both private and public cloud. The private cloud is responsible here for the sharing the file token to the user on request and providing with the privatekey. Through this waytheprivateclouddoes the authentication of the user so that only authorized user can check for deduplication of data, upload or download the file. Once the user gets hold ofthe token, it then sharesitwith the public cloud for upload or download purpose. The file token is received by the public cloud and results are given accordingly as per the request. 3) TryManagingYourDeduplicationFine-grained-ly:AMulti- tieredandDynamicSLA-drivenDeduplicationFrameworkfor Primary Storage [4]. The primary storage system is made up by two main modules: proxy server and object storage cluster (OSC). The client side requests are sent to the OSC storage node by the proxy server. The newly entered files meta data is present in the buffer zone present in the proxy server. OSC is a distributed storage cluster that contains approximately hundreds ofstorage nodes made up of variouschunks.These storage nodes maintain the fingerprint in the index table for all the chunks. Fingerprint is nothing but hash of the content that is done by using SHA1 technique. Deduplication doesn't interfere with the Mudder on going operations, instead it performs the task in the background itslef.DCL-Module,SLA- Handler, Deduplication Executor(Deduper) and Workload Monitor (Monitor) are the main modules of the Mudder. When new files meta data are filles in the buffer zoneand the entries reach a threshold value N, DCL module fetches all these entries and generate a Target list. DCL module is operable insidethe proxy server. Initially the entry in the list is predefined as dirty, meaning not duplicate. The DCL module then calculated for deduplication. DCL-Module pushes the list to Deduper which is a channel where deduplication between storage nodes and proxy takes place. Procedure of removing redundant data and index update is similar to other systems. 4) DeduplicationincloudstorageusingHashingtechniquefor encrypted data [6]. The data to be sent for uploading purpose is encrypted first before sending it to the Cloud Service Provider. The encryption is done using the AES 128 bit encryption that is protected from various attacks, one being Birthday brute force attack. The data is encryptedinthedataownermachine by using a secure key. In this way the Cloud Service Provider doesn’t have access to the plain text of the file to be uploaded because it receives a encryptedcopyofthefile.Theencrypted file received undergoes chunkingandhashing.Thehashingis performed using the MD5 algorithm. If the hash value of the data is same as that already present in the cloud, then the data isn’t stored again. If the hash isn’t present in the cloud then the file is stored in the cloud and the location is updated in theindex table of Cloud Service Provider. 5) Improving the Performance of System in Cloud By Using Selective Deduplication [7]. The user of the system can request for upload. This upload takes place from LAN or WAN or MAN to thecloud.Itdepend on the users whether they wanttocheck thededuplicationof the data or not. The selective deduplication is applied on the data to find the duplicate data if any. If any duplicate data is found then the feedback is sent to user and the redundant data is deleted from the cloud storage. If in case, the data is unique, that is, not redundant, then user is asked for encryption. Security key for the file is generated and encryption is performend on the file. After encryption, the file is stored in the cloud storage. When the user requires, decryption can be performed on the file and thus a key is generated for decrypting file from cloud. 6) PFP: Improving the Reliability of Deduplication-based Storage Systems with Per-File Parity [8]. During the process of deduplication, the reliability gets affected because of the removal of redundant chunks. These identified duplicate chunks are removed because they are common to other files too. At the end, there is just one chunk that is shared by all the files after the deduplication. This removal of chunks renders the sharing files unavailable.This paper gave an idea of Per File Parity(PFP) to improve the reliability. PFP performs the XOR parity within the parity groups of data chunks which are part of each file after the chunks are created but beforethesechunksarededuplicated. PFP performs XOR parity within the data chunks parity group. Therefore, PFP provides security from the parity redundancy. The PFP doesn't hinder the flow of data duplication.Thefingerprintcalculation,chunking,detectionof duplicate chunks and removal of these duplicate chunks, nothing changes.Parity generation is implemented on every file to prevent deduplication of the data. The resulting parity metadata is stored as the metadata of file. The incoming file data chunks arecreatedandthefingerprintsarechecked.PFP then divides these chunks into groups as per predefined group size of N. For N data chunk group the parity is calculated by usingXORoperation.Thisschemecantoleratea single chunk failures. 3. PROPOSED SYSTEM The proposed system comprises of components like the Client, Middleware and the Server. Middleware is the core component where all the processing takes place. Fig -1: System Architecture
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3802 A middleware is a software thatacts as a bridgebetweentwo endpoints, generally a client and a server. The middleware gives the users an interface to upload the files. The user has to create an account on the application in order to perform functions like upload or download. Once the user has an account, they can add files to their storage using the Middleware. The Middleware will be handling the process of data deduplication. Fig -2: Middleware and associated modules The middleware consists of the the following modules: 1. Authentication Module : It is used to authenticate the user so that the user can access the middleware dashboard. The user needs to supply their credentials in the login form to get access to their account. The authentication module also provided a service to sign up for the account of the middleware. 2. Chunking Module: The chunking module is invoked by an API call which chunks the file into small fixed size pieces in a synchronized manner which is then passed on tothehashing module. 3. Hashing Module: The hashing module takes the chunked file input from the chunking module and generates SHA1 hash of each chunk and stores it in an array until the the chunks of the current file being processed is completely supplied to the module. After the individual hashing of the chunks, the SHA1 of the wholearraywhichconsistsofhashof the chunk is hashed and returned to the API. 4. Drive Module: The drive module is used for uploading the user file to the user’s drive storage. It is also used by the cron job process to sync drive storage content information with the middleware by using the metadata of the files present on the cloud of the user and alsoto download the file from users cloud for deduplication processing. 5. API: The API Module is used by other modules to interact to other middleware or topassthedatathroughouttheentire middleware. It handles the incoming requests and data response forother modules using HTTP requests. Thedatais wrapped on the response body. 6. Cron Module: The cron module runs automatically everyday once to sync withtheuserscloudstorage.Thejobof the cron module is to check whether the user has uploaded any file directly to the cloud storage by fetching the file’s metadata and checking in the database. If the file has never been processed by the middleware, then it fetches the file from the cloud storage usingtheAPIanddrivemoduleforthe process of deduplication. The processing of the file can take place based on the two scenarios with respect to the method of file upload. In the first scenario, the user uploads the file to the middleware via the uploader. In the second scenario, if the user uploads the file directly to the cloud storage or for the old files on cloud storage need to go through the process of deduplication which is done using the cron job module. The file which is uploaded tothe middleware or downloaded from the user’s cloud storage will go through chunking and hashing. Based on the value of hash generated it is determined whether duplication is present or not by checking in the database withrespect to that individual user. If the hash of the file exists in the database then a pointer is added to the location where the originalfileexistsandthefile at hand is discarded. But, if the hash is not present in the database then the current fileinformationalongwithitshash value is added in the database. While syncing with the cloud storage, using file ID from the metadata of the file, we check whether the file has been ever processed or not. If not processed then it goes through the above process otherwise it is sent for batch processing. 3.1. Techniques Used The system focuses on removal of the duplicatedataoncethe file is uploaded by the user. The major techniques usedinthe system are chunking and hashing that is performed in the middleware. The chunking and hashing are used for the deduplication process. The deduplication is done in order to remove the redundancy. If the user uploads a file, and the same is present in the storage server then a pointer is assigned to the original file instead of saving the new file again. Deduplication reduces use of storage space. This is performed in two steps: 1. Chunking: This is a technique that breaks a large piece of information say file into multiple pieces or chunks. These chunks are helpful for further data deduplication. Chunking can be done by either using fixed size chunking or file level chunking.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3803 Fig -3: Chunking Process 2. Hashing: This is a method used to generate a value of the incoming data ofa file through a mathematical function. This is performed during the deduplication procedure in the middleware. The hash values can be generated by either using SHA1 or MD5 algorithm.SHA1ispreferredoverMD5as the collision experience is less. Also, SHA1 is faster with respect to it's successors. Fig -4: Hashing Process The chunks created from the chunking process are hashed and pushed into an array. Finally afterallthechunkshashare pushed into an array, the arrays' hash is generated that is used fordeduplication. Themiddleware runsaservicewhich is used to sync the data already present on the user's cloud storage. The sync checks whether the file present on the cloud storage is processed by the middleware or not using the fileID from metadata. The file is fetched and processed from the cloud server and same deduplication process is done if the file is not processed ever otherwise the files are sent for batch processing which is optional or not required if the cloud is only used for storage and not for editing files. 3.2 Proposed System Procedure Flow The user's file upload command is examined by the middleware. The middleware fetches the file from the user for chunking and hashing. The chunking process first takes place which gives small chunks of the file whose hash is generated. The number of chunks depend on the size of the file with respect to the chunk size. One by one the hash of the chunks are generated and pushed into an array. Once all the chunks of the file are created, the hashing of the complete array is generated. This generated hash of the file is now compared with the hash values present in the database. As shown in Fig 5, there are two scenarios as stated below: Fig -5: Procedural Flow Case I: Hash generated of the file is not present in the database. In this case the database is updated along with the new hash value. The file in this case is uploaded in the storage server. Case II: Hash generated of the file is present in the database. In this case the file is not saved again. Rather a pointer is added that will point to the original file presentinthestorage server. 3.3 Use Case Diagram There are three actors involved in the system as showninFig 6. Those can be described as: 1. User: The user is the one who will use the deduplication application. 2. Middleware: The middleware is the system which will check for deduplication and ensure that there is no duplicate file in the storage server. 3. Storage: Storage is a system where the file of the users is stored and retrieved. The system validates the user credentials and logs in them into the system. The middleware fetches the file from the storage server using the appropriate credentials. The middleware uploads the processed file to the storage server usingappropriatecredentials. Theuseruploadsthefiletothe middleware using the dashboard. The middleware create chunks of the file which was uploaded by the user. The chunks created will be hashed and stored in a hash digest. The middleware will now store the hash generated to the database.
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3804 Fig -6: Use Case Diagram 4. PERFORMANCE EVALUATION 4.1 Datasets Used The datasets used in this section are iso files. An ISO file is a disc image of an optical disc and contains everythingthatcan be written to the optical drive. In earlier days ISO was the format used to burn CDs and DVDs whereas, in the present days an ISO file is used to transfer a larger amount of files compressed into a single ISO file which is then shared over the internet using peer to peer or FTP servers. The name ISO is taken from the ISO 9660 file system.Thedatasetstakenare ISO image of Linux distributions that are most widely used. These datasets are relatively large, usually, the size rangesin Gigabytes. The large dataset will ensure harmonious and distinct results between fixed size hashing and file level hashing. Table -1: Datasets Involved Datasets Dataset Name Type Provide d By Size (GiB) Dataset 1 Antergos-18.12- x86_64.is ISO File Antergos .com 2.1 Dataset 2 openSUSE- Tubleweed-DVD- x86_64Snapshot2 0181128- Media.iso ISO File Suse Linux 4.1 Dataset 3 Popos_18.10_amd 64_intel_11.iso ISO File System 76 inc. 2 Dataset 4 Ubuntu-18.04.1- desktop- amd64.iso ISO File Canonica l inc. 1.8 4.2 Tools Involved The performance analysis done on the proposed system involves the use of following tools: 1. Time command linux: The time command runs the specified program with the provided arguments. When command finishes the execution timewrites back a message to standard error giving timing and processing statistics about the program that was runned. 2. htop for cpu usage: Htop is an interactive system viewer and process managing tool. Itis based on UNIX tool top. Htop allows the user to view process or tasks statistics like CPU usage, ram consumption, and the process uptime. 3. console.time() and console.timeEnd() for nodejs function: These inbuilt methods are used to benchmark javascript applications for response and average computation time required to complete the given task. console.time(label) is used to set the initial reference from where the timer would begin.console.timeEnd(label)isusedastheclosingstatement the tasks in between console.time() and console.timeEnd is bench marked for time. 4.3. Evaluation Metrics The metrics usedforbenchmarkingthesystemsperformance with respect to the defined constraints is the Evaluation metrics. 1. Hashing Time: The Hashing Time is the time required (in Seconds) to produce the hash of a given input. The hash function encodes each byte as two hexadecimal characters. The hash time should be lower in order to increase the efficiency of the application. Lower hash time also helps in scaling the application if needed. Table -2: Hashing Time in seconds Datasets File Level Fixed Size Dataset 1 67.3 24.19 Dataset 2 132.66 46.1 Dataset 3 82.59 22.6 Dataset 4 34.5 25
  • 6. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3805 Fig -7. Time consumption graph for hashing 2. Chunking Time: The time required to divide a file into chunks is termed as chunking time . It is usually measured in milliseconds (ms). The lesser the chunking time, the faster the file data will be processed and analyzed. It obtained results won't be better than file level processing, as the chunking time increase. Table -3: Chunking Time in milliseconds Datasets Fixed Size Dataset 1 1.7 Dataset 2 2.47 Dataset 3 2.93 Dataset 4 2.41 Fig -8. Time consumption graph for chunking 3. CPUload: CPU load measures theamountofcomputational work in percentage that the system performs in order chunk and hash the input data. The load on the CPU should be neutral and consistent. The more the load on the CPU the more the chances of the application getting halt. Also, less CPU load ensures less cost on dedicated hardware and increased scalability of the application. Table -4: CPU Load in Percentage Datasets File Level Fixed Size Dataset 1 40 23 Dataset 2 66 40 Dataset 3 38 37 Dataset 4 40 33 Fig -9: Tme consumption graph for CPU 5. CONCLUSION In effect of all the results that were obtained in our proposed system, the evaluation of the performance of our project, as well as the extensive Literature survey conducted by us in this project it can be concluded that the the middleware approach forperforming the deduplication is moreeffiecient and cost effective. The workload on the client and serverside isreduced.Theperformanceevaluationrevealsthatchunking is the most efficient form of file processing and hence it must be used in data dedupication. Thehashingisperformedusing SHA1 algorithm. We discovered that, SHA1 is significantly more securethan MD5 as it isprone to less collisionsand itis significantly faster than SHA256 and SHA512. Hence, it provides the bestbalancebetweensecurityandperformance. REFERENCES [1] W. Litwin, Thomas Schwarz,"Combining Chunk Boundary and Chunk Signature Calculations for Deduplication ", IEEE, 2012. [2] Jian Zhang, Xingyu Zhang, "Data Deduplication Cluster Based on Similarity-Locality Approach",IEEE, 2013.
  • 7. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 06 Issue: 04 | Apr 2019 www.irjet.net p-ISSN: 2395-0072 © 2019, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 3806 [3] Jin Li, Yan Kit Li, Xiaofeng Chen, Patrick P. C. Lee, Wenjing Lou, "A Hybrid Cloud Approach for Secure Authorized Deduplication", IEEE, 2014. [4] Yan Tang, Jianwei Yin, and Zhaohui Wu, "Try Managing Your Deduplication Fine-grained-ly: A Multi-tiered and Dynamic SLA-driven Deduplication Framework for Primary Storage", IEEE, 2016. [5] Nancy Digra1 , Sandeep Sharma, "A Noval Approach of Enhancing Security in Cloud Using Diffie Hellman Algorithm",IJSRM, 2017. [6] Vaishnavi Moorthy, Arpit Parwal and Udit Rout,"Deduplication in Cloud Storage Using Hashing Technique For Encrypted Data",IEEE, 2018. [7] Mr. Nishant N.Pachpor, Dr. Prakash S.Prasad, "Improving the Performance of System in Cloud by Using Selective Deduplication", IEEE, 2018. [8] Suzhen Wu, Bo Mao, Hong Jiang, Huagao Luan, Jindong Zhou,"PFP: Improving the Reliability of Deduplication- based Storage Systems with Per-File Parity",IEEE,2019.