Professional Documents
Culture Documents
i
4.6 Project Module Description 24
4.6.1 Admin Modules 24
4.6.2 User Modules 24
4.7. Technique and Algorithms 25
4.8 System Design 28
4.8.1 The Activities of Design Process 28
4.9 System Architecture 30
4.10 Use Case Diagram 31
4.10.1 Admin Use Case Diagram 31
4.10.2 User Use Case Diagram 32
4.11 Data Flow Diagram 33
4.11.1 DFD Admin Session 33
4.11.2 DFD User Session 34
4.11.3 DFD File Upload Process 35
4.11.4 DFD File Download Process 36
4.12 Sequence Diagram 37
4.13 Context Analysis 38
4.14 Flow Chart 39
CHAPTER 5 Testing 41-60
5.1. Definition 41
5.2. Validation and System Testing 41
5.3. Integration Testing 43
5.4. Unit Testing 44
5.5. Unit Test Cases 45
CHAPTER 6 Result 46-60
7. Conclusion 61
8. References 62-65
ii
LIST OF FIGURE AND GRAPH
iii
LIST OF TABLES
iv
LIST OF ABBREVIATION
v
CHAPTER 1
INTRODUCTION
Cloud storage is one of the digital storage solutions that utilize multiple servers (typically
spread across multiple locations) which ensures to safely store files such as site backups etc.
The data is stored on dedicated servers, which provides unlimited accessibility wherever an
internet connection is available, along with an increase in backup file security which ensures
our files not to be hacked. Distributed computing gives versatile, minimal effort, and area free
online administrations extending from basic reinforcement administrations to distributed
storage foundations. The quick development of information volumes put away in the
distributed storage has prompted an expanded interest for procedures for sparing circle space
and system transfer speed.
To decrease asset utilization, many distributed storage administrations, for example, Dropbox,
Wuala, Mozy, and Google Drive, utilize a deduplication method, where the cloud server stores
just a solitary duplicate of excess information and gives connects to the duplicate as opposed
to putting away other real duplicates of that information, paying little respect to what number
of customers make a request to store the information.
The investment funds are critical, and allegedly, business applications can accomplish plate
and transfer speed reserve funds of over 90. Be that as it may, from a security point of view,
the mutual utilization of clients' information raises another test. As clients are worried about
their private information, they may scramble their information before outsourcing with a
specific end goal to shield information security from unapproved outside enemies, and in
addition from the cloud specialist co-op.
This is advocated by current security patterns and various industry directions, for example,
PCI DSS. In any case, ordinary encryption makes deduplication outlandish for the
accompanying reason. Deduplication procedures exploit information similitude to recognize
similar information and diminish the storage room.
Conversely, encryption calculations randomize the encoded records with a specific end goal to
make cipher text indistinct from hypothetically arbitrary information. Encryptions of similar
Page 1
information by various clients with various encryption keys brings about various cipher texts,
which makes it troublesome for the cloud server to decide if the plain information are the
same and deduplicate them.
Say a client Alice encodes a document M under her mystery key skA and stores its comparing
cipher text CA. Weave would store CB, which is the encryption of M under his mystery key
skB. At that point, two issues emerge: (1) in what capacity can the cloud server recognize that
the hidden document M is the same, and (2) regardless of the possibility that it can identify
this, how might it enable both sides to recoup the put away information, in view of their
different mystery keys? Direct customer side encryption that is secure against a picked
plaintext assault with arbitrarily picked encryption keys avoids deduplication.
One credulous arrangement is to enable every customer to encode the information with
general society key of the distributed storage server. At that point, the server can deduplicate
the distinguished information by decoding it with its private key combine. Be that as it may,
this arrangement permits the distributed storage server to get the outsourced plain
information, which may abuse the protection of the information if the cloud server can't be
completely put stock in [13],[14].
Merged encryption [15] settle this issue adequately. A merged encryption calculation encodes
an info record with the hash estimation of the information document as an encryption key. The
cipher text is given to the server and the client holds the encryption key. Since united
encryption is deterministic1, indistinguishable documents are constantly scrambled into
indistinguishable cipher text , paying little heed to who encodes them.
In this manner, the distributed storage server can perform deduplication over the cipher text,
and all proprietors of the record can download the cipher text (after the confirmation of-
possession (PoW) handle alternatively) and decode it later since they have a similar
encryption key for the document.
United encryption has for some time been examined in business frameworks and has
distinctive encryption variations for secure deduplication [8],[16],[17],[18], which was
formalized as message blocked encryption later in [20]. Be that as it may, focalized
Page 2
encryption experiences security blemishes with respect to label consistency and proprietorship
repudiation.
Deduplication is a technique that ensures only one unique instance of data is retained on
storage media, such as disk, flash or tape. Redundant data blocks are replaced with a pointer
to the unique data copy. The technique is used to improve storage utilization and can also be
applied to network data transfers to reduce the number of bytes that must be sent. In the
deduplication process, unique chunks of data, or byte patterns, are identified and stored during
a process of analysis. Data deduplication is a specialized data compression technique for
eliminating duplicate copies of repeating data. Related and somewhat synonymous terms are
intelligent (data) compression and single-instance (data) storage.
At that point, the repudiated clients ought to be kept from getting to the information put away
in the distributed storage after the erasure or change ask for (forward mystery). Then again,
when a client transfers information that as of now exist in the distributed storage, the client
ought to be discouraged from getting to the information that were put away before he acquired
the proprietorship by transferring it (in reverse mystery).
These dynamic proprietorship changes may happen often in a handy cloud framework, and
hence, it ought to be appropriately overseen keeping in mind the end goal to maintain a
strategic distance from the security corruption of the cloud benefit. Be that as it may, the past
deduplication plans couldn't accomplish secure get to control under a dynamic possession
evolving condition, regardless of its significance to secure deduplication, in light of the fact
that the encryption key is inferred deterministically and once in a while refreshed after the
underlying key deduction.
In this way, for whatever length of time that denied clients keep the encryption key, they can
get to the comparing information in the distributed storage whenever, paying little respect to
the legitimacy of their proprietorship. This is the issue we endeavour to understand in this
investigation.
As of late, Li et al. [25] proposed a united key administration conspire in which clients
appropriate the concurrent key offers over various servers by misusing the Ramp mystery
Page 3
sharing plan [26]. Li et al. [27] additionally proposed an approved deduplication plot in which
differential benefits of clients, and in addition the information, are considered in the
deduplication technique in a half and half cloud condition. Jin et al. [31] proposed an
unknown deduplication conspire over scrambled information that adventures an intermediary
re-encryption calculation. Bellare et al. [28] proposed a server-helped MLE which is secure
against beast compel assault, which was as of late stretched out to intuitive MLE [29] to give
protection to messages that are both associated and reliant on people in general framework
parameters.
In any case, these plans don't deal with the dynamic possession administration issues required
in secure deduplication for shared outsourced information. Shin et al. [30] proposed a
deduplication plot over scrambled information that utilizations predicate encryption. This
approach permits deduplication just of documents that have a place with a similar client,
which extremely decreases the impact of deduplication. Therefore, in this paper, we
concentrate on deduplication crosswise over various clients to such an extent that
indistinguishable records from various clients are distinguished and deduplicated securely to
give more stockpiling investment funds.
1.2 Contribution
We propose a data deduplication scheme over mixed data. The proposed system ensures that
individual secure access to the common data is possible, which is believed to be the most
Page 4
basic test for successful and secure appropriated bigdata storage organizations in nature where
ownership changes effectively. At the point when diverged from the past deduplication
schemes over mixed data, the proposed system has the going with purposes of enthusiasm to
the extent security, dynamic data ownership and profitability. To begin with, dynamic data
ownership organization guarantees the retrogressive and forward secret of deduplicated data
upon any ownership change.
Page 5
CHAPTER 2
LITERATURE SURVEY
General
On the basis of extensive literature survey related to the data deduplication with dynamic
ownership management in cloud storage has been taken into consideration in this chapter.
D. T. Meyer, and W. J. Bolosky [1] has proposed that File frameworks regularly contain
repetitive duplicates of data: indistinguishable documents or sub-record locales, perhaps put
away on a solitary host, on a mutual stockpiling group, or moved down to optional capacity.
Deduplicating stock piling frameworks exploit this excess to decrease the fundamental space
expected to contain the record frameworks (or reinforcement pictures thereof).
Deduplication can work at either the sub-document or entire record level. All the more fine-
grained deduplication makes more open doors for space reserve funds, yet fundamentally
lessens the successive format of a few records, which may have critical execution impacts
when hard plates are utilized for capacity (and at times requires confused strategies to
enhance execution.
M. Dutch[2] has stated that with respect to the comprehension of information deduplication
proportions that information deduplication brings down business dangers, builds income
openings, and diminishes stockpiling level expenses, bringing about an ideal tempest for
organizations sending a versatile stockpiling foundation. Capacity flexibility advances, for
example, RAID or RAIN, shield the deduplicated information to guarantee high accessibility
of utilizations getting to the information.
Page 6
W. K. Ng et al. [3] has proposed about another idea which we call private information
deduplication convention, a deduplication method for private information stockpiling is
presented and formalized. Instinctively, a private information deduplication convention
permits a customer who holds a private information demonstrates to a server who holds an
outline string of the information that he/she is the proprietor of that information without
uncovering additional data to the server.
For people, recorded capacity is being called upon to save nostalgic and authentic ancient
rarities, for example, photographs, films and individual archives. Further, while few would
contend that business information calls for security, protection is similarly imperative for
people; information, for example, medicinal records and authoritative reports must be kept for
drawn out stretches of time however should not be openly available. Incomprehensibly, the
expanding estimation of authentic information is driving the requirement for cost-productive
capacity; reasonable capacity permits the conservation of all information that may in the long
run demonstrate helpful.
Page 7
client had worn out its distributed storage membership (in this paper we tend to suggest to that
as indweller off boarding).
C. Wang, Z. Qin, J. Peng, and J. Wang[8] has stated another information pressure innovation,
which is known as information deduplication, come in locate in pace with the gigantic
increment of electronic information. Information deduplication partitions information into
settled or variable size pieces and the cryptographic hash estimation of each lump is utilized
as the piece's worldwide one of a kind 10 with the end goal that excess information might be
distinguished.
These redundancies are utilized to either decrease stockpiling limit needs or to lessen organize
activity. Not at all like conventional pressure technique, deduplication recognizes normal
groupings of bytes both inside and amongst records, and just stores a solitary occasion of each
piece paying little mind to the quantity of times it happens.
P. Anderson, L. Zhang [12] had stated that a run of the mill cluster of moveable digital
computer shoppers share plenty of knowledge in like manner. This provides the chance to
essentially diminish reinforcement times, and capability conditions. In any case, we've got
incontestable that manual determination of the many info - as an example, moving down
simply home registries - may be a poor methodology; this neglects to reinforcement some
important records, within the meanwhile as superfluously repeating totally different
Page 8
documents. we've got exhibited a model reinforcement program that accomplishes a perfect
level of sharing within the mean while as taking care of secrecy.
This adventures a unique calculation to diminish the number of documents that ought to be
checked and later diminishes reinforcement times. we've got incontestable that run of the mill
cloud interfaces, as an example, Amazon S3 don't seem to be acceptable to the present
reasonably use, due to the time and price of run of the mill exchanges, and therefore the
absence of multi-client verification to shared info.
We have delineated a usage utilizing a neighbourhood server which might be from these
problems by storing and pre-preparing info before sending to the cloud. this can be looked as
if it would accomplish noteworthy price funds2.9: Safe backup of cloud system with
guaranteed deletion:
J. Xu, E. Chang, and J. Zhou[15] has expressed that Cloud reposition administration is
learning prominence as recently. To diminish quality utilization in each system transmission
capability and capability, several distributed storage administrations as well as Dropbox one
and Wuala a pair of utilizes client aspect deduplication. That is, the purpose at that a consumer
tries to transfer a document to the server, the server checks whether or not this specific record
is as of currently within the cloud (transferred by some consumer beforehand), and spares the
transferring procedure within the event that it's as of currently within the distributed storage.
Page 9
Along these lines, every and each record can have simply one duplicate within the cloud (i.e.
Single Instance Storage). SNIA study declared that the deduplication procedure will put aside
to ninetieth capability, dependent on applications. As per Halevi et al. what is a lot of, Drop
ship, a current usage of client aspect deduplication is as beneath: Cloud consumer Alice tries
to transfer a record F to the distributed storage.
Nowadays cloud computing gaining more popularity for sharing data through cloud. It runs a
lot of application to provide different features to users. It gives a platform for users to share
data through cloud storage. Cloud contains many schemes for data security and privacy.
To get these things distributed computing uses extensive gatherings in system of servers
which is running in ease buyer PC innovation to spread information preparing Virtualization
strategies for the most part used to build the force of distributed computing.
Conclusion
Page 10
This part for the most part talks about the papers that are alluded while making this
thesis report. Every one of these papers give data identified with learning of aggregate
conduct, their current arrangements, and strategies utilized furthermore their focal points &
constraints.
Page 11
Chapter 3
SOFTWARE REQUIREMENT SPECIFICATION
This part depicts about the prerequisites. It determines the equipment and
programming prerequisite that are needed for software to keeping in mind the end goal, to run
the application appropriately. The Software Requirement Specification (SRS) is clarified in
point of interest, which incorporates outline of this exposition and additionally the functional
and non-practical necessity of this thesis.
This can be utilized to relate the rising structure properties, for instance, immovable
quality, response time and store inhabitancies. Here again they ought to portray objectives on
the system, for instance, the capacity of the data yield devices and data representation used as
a piece of structure interfaces. In all probability all non helpful essentials can be relating to the
system as whole rather than to individual structure highlights. This suggests they are every
now and again essential appear differently in relation to the individual commonsense
Page 12
necessities. Non utilitarian necessity gets through the client needs, in view of spending plan
limitations, hierarchical approaches, and the requirement for interoperability with other
programming and equipment frameworks.
The going with non-valuable requirements are meriting thought.
Security: The framework ought to permit a secured correspondence between
information proprietor and beneficiary.
Reliability: The system should be trustworthy and ought not corrupt the execution of
the present structure and should not to provoke the hanging of the structure.
Page 13
3.4 Feasibility Study
Believability is the determination of paying little respect to whether an undertaking
justifies action. The framework followed in building their strength is called acceptability
Study, these kind of study if a task could and ought to be taken.
Three key thoughts included in the likelihood examination are:
Technical Feasibility
Economic Feasibility
Operational Feasibility
Page 14
What alterations will be carried through the framework?
What authoritative shapes are dispersed?
What new aptitudes will be needed?
Do the current framework employee’s individuals have these aptitudes?
If not, would they be able to be prepared over the span of time?
Page 15
Internationalization: Help for composing projects that could be restricted for clients
around the world. Projects can naturally adjust to particular local people and be shown
in the suitable dialect.
Security: Mutually low level and abnormal state, together with electronic marks, open
and private key administration, right of entry control and authentications.
Software components: Recognized as JavaBeans, could connect to existing parts
structural designs.
Object Serialization: Let’s Permits lightweight tirelessness and correspondence by
means of Remote Method Invocation (RMI).
Java Database Connectivity (JDBC): Give consistent entree to an extensive variety
of social databases.
Page 16
3.5.2 Java Server Page
Java Server Pages development allows you to put scraps of servlet coding particularly
into a substance based record. A java server page is a substance based record that holds two
sorts of substance: static format data, which should be imparted in several substance based
association, for instance, XML, WML, HTML and Java server page segments, which choose
how the page fabricates component content.
Java Server Page (JSP):An extensible Web innovation that make use of layout information,
custom components, scripting dialects, and server side Java programming language articles to
homecoming element substance to a customer. As indicated by JSP model1 we can build up
the application as:
For sweeping extent reaches out instead of using model1 it is perfect to use model2 Model
View Controller. Struts structure is in light of model 2.JSP allows you to specific the dynamic
bit of your pages with the static HTML. Here it just makes the steady HTML in the run of the
mill way, using any Web page building mechanical assemblies it frequently uses. Here again
encase the dynamic code for the parts in exceptional labels, the greater part of which begin
with "<%" and end with "%>". You ordinarily give your record a .jsp augmentation, and
regularly introduce it in wherever you could put a typical Web page.
Despite the fact that what you compose frequently seems more like a customary HTML
record compare to a servlet, in the background, the java server page just acquires change over
Page 17
to an average servlet, through the static HTML basically being printed to the yield stream
joined with the servlet's organization method.
3.5.3 JavaScript
Java Script is a one of the object oriented; lightweight, script based computer
programming language dialect tongue which was made by Netscape Communication
Corporation. JavaScript holds the change equally client and server parts of application of Web
program and the client side; this could be employ to form programs which are implemented
by a Web program inside the association of a Web page on the server side, this should be
employed to make a Web server programs which could deal with information set up together
by a Web project and thereafter overhaul the program's showcase moreover.
Despite the fact that JavaScript underpins client and server Web programming, this
lean toward JavaScript at customer side programming subsequent to a large portion of the
programs bolsters this one. JavaScript is as simple as to study as HTML, and JavaScript
explanations could be incorporated in HTML reports by encasing the announcements among a
couple of scripting labels.
JavaScript and Java are altogether diverse dialects. A couple of the obvious contrasts
are:
Page 18
Java applets are usually indicated for a situation inside the web chronicle; JavaScript
could impact every bit of the Web report itself.
JavaScript is most suitable to essential applications and inserting instinctive
components to Web pages and Java could be used for amazingly multifaceted
applications.
Here there are various distinctive complexities yet the fundamental object to review is that
JavaScript and Java are autonomous lingos. They are together useful for unmistakable objects;
really they could be used both to join their ideal circumstances.
3.5.4 JDBC
JDBC with a last goal to set a self administering database benchmark API for java made
database blend, or JDBC. This offers a non specific SQL database access portion that provides
an expected interface to a mixed pack of RDBMS. This trustworthy interface is refined in the
use of "unit" database framework units, or drivers. On the off chance that a database merchant
wishes to have JDBC strengthen, he or she can allot to the driver to every stage that the
database and java keep running on. To get a more expansive attestation of JDBC, sun builds
up JDBC's system concerning ODBC. Java database integration gives uniform access to an
extensive variety of social databases. MS Access databases are utilized for rapidly
overhauling the store table.
Plan objectives for JDBC are as per the following:
SQL Level API
The organizers experience that their essential objective was to portray a SQL interface
for java. Yet not the most diminished database interface level achievable, this is an
adequately low level for strange state devices and APIs to be made. Then again, it is at
an adequately abnormal state for application programming architect to use it surely.
Finishing this target looks into future mechanical assembly shippers to "make" JDBC
code and to cover countless difficulties from the end customer.
SQL Conformance
Page 19
SQL sentence structure changes as this movement from database shipper to database
vender. With an end target to support a wide grouping of traders; JDBC will authorize
every investigation articulation to be moved out through this to the basic database
driver. This authorizes the integration unit to grip non-standard usefulness in approach
that is appropriate for their clients.
CHAPTER 4
Page 20
SYSTEM ANALYSIS
4.3 DISADVANTAGES
Page 21
Client deduplication on the customer side can't create another label when they refresh
the document. In this circumstance, the dynamic Ownerships would come up short. As an
outline, existing dynamic Ownerships can't be reached out to the multi-client condition. At
whatever point information is changed, concerns emerge about potential loss of information.
By definition, information deduplication frameworks store information uniquely in contrast to
how it was composed.
Accordingly, clients are worried about the trustworthiness of their information. One strategy
for deduplicating information depends on the utilization of cryptographic hash capacities to
distinguish copy sections of information. In the event that two distinct snippets of data
produce similar hash esteem, this is known as an impact. The likelihood of a crash relies on
the hash work utilized, and in spite of the fact that the probabilities are little, they are
dependably non zero.
Page 22
• volume of data
• Integrity auditing and secure deduplication directly on encrypted data.
Download a File
View details of all the uploaded file
User has to select the file to download and initiate the download process
Retrieve the Block details of the selected file.
Get the block storage details in cloud
Download all the blocks
Page 23
Integrity Check
Generate the Hash Tag for all the Blocks.
Compare the Hash Tags of the block with Hash Tags in Table.
If Hash Tags Comparison pass for all the blocks then merge the blocks and
download the file to the user system Else show Integrity Check file message
to the user.
Transaction
Sign out
5: EmitIntermediate(w, "1")
key: word
result += ParseInt(v);
Emit(AsString(result));
Page 24
Secure Auditing Multi Level Block Signature Verification: Pseudocode
Begin
Get Block Signature (BS)
Convert block signature in to three parts
First Bit (FB), Second Bit (SB), Third Bit (TB) to Last Bit (LB)
Get the Index of Code(IC) FB from zero level BS
Indentifying the corresponding next level BS index using IC
If (presence of SB in second level BS index)
If (get (TB to LB) and check the presence of byte in 3rd level))
Block exist
Else
Else
Block not exist
End
Page 25
of navigation mechanisms. Interface Control mechanisms- to implement navigation
options, the designer selects form one of a number of interaction mechanism;
a. Navigation menus
b. Graphic icons
c. Graphic images
Interface Design work flow- the work flow begins with the identification of user, task, and
environmental requirements. Once user tasks have been identified, user scenarios are
created and analyzed to define a set of interface objects and actions.
2. Aesthetic design-also called graphic design, describes the “look and feel” of the WebApp.
Includes colour schemes, geometric layout. Text size, font and placement, the use of
graphics, and related aesthetic decisions.
3. Content design-defines the layout, structure, and outline for all content that is presented as
part of the WebApp. Establishes the relationships between content objects.
4. Navigation design-represents the navigational flow between contents objects and for all
WebApp functions.
5. Architecture design-identifies the overall hypermedia structure for the WebApp.
Architecture design is tied to the goals establish for a WebApp, the content to be
presented, the users who will visit, and the navigation philosophy that has been
established.
a. Content architecture, focuses on the manner in which content objects and structured
for presentation and navigation.
b. WebApp architecture, addresses the manner in which the application is structure to
manage user interaction, handle internal processing tasks, effect navigation, and
present content. WebApp architecture is defined within the context of the development
environment in which the application is to be implemented.
Page 26
J2EE uses MVC Architecture
Page 27
Figure 4.9.1 System Architecture
Page 28
4.10.1 Admin Use Case Diagram
Page 29
4.10.2User Use Case Diagram
Page 30
4.11 Data Flow Diagram
Page 31
4.11.2DFD User Session
Page 32
4.11.3DFD File Upload Process
Page 33
4.11.4DFD File Download Process
Page 34
4.12 Sequence Diagram
Page 35
4.13 Context Analysis
Page 36
4.14 Flow Chart Draw the flow chart with proper arrow pointing correct flow
Page 37
Fig. 4.14.2 Flow Chart for Block
Page 38
CHAPTER 5
TESTING
5.1 Definition
Unit testing is an improvement system where software engineers make tests as they create
programming. The tests are basic short tests that test practically of a specific unit or module of
their code, for example, a class or capacity.
Utilizing open source libraries like cunit, oppunit and religious recluse it (for C, C++ and
Java) these tests can be consequently run and any issues discovered rapidly. As the tests are
created in parallel with the source unit test exhibits its rightness.
Security Testing: Where unauthorized attempts to operate the software, or parts of it,
attempted it might also include attempts to obtain access the data, or harm the software
installation or even the system software. As with all types of security determined will
be able to obtain unauthorized access and the best that can be achieved is to make this
process as difficult as possible.
Page 39
Stress Testing: Where abnormal demands are made upon the software by increasing
the rate at which it is asked to accept, or the rate t which it is asked to produce
information. More complex tests may attempt to crate very large data sets or cause the
soft wares to make excessive demands on the operating system.
Alpha and beta testing: This is where the software is released to the actual end users.
An initial release, the alpha release, might be made to selected users who be expected
to report bugs and other detailed observations back to the production team. Once the
application changes necessitated by the alpha phase can be made to larger more
representative set users, before the final release is made to all users.
The final process should be a Software audit where the complete software project is
checked to ensure that it meets production management requirements. This ensures
that all required documentation has been produced, is in the correct format and is of
acceptable quality. The purpose of this review is: firstly to assure the quality of the
production process and by implication construction phase commences. A formal hand
over from the development team at the end of the audit will mark the transition
between the two phases.
Page 40
exhibit just as stubs. Subprogram stubs were introduced in section2 as fragmented
subprograms which are just present to permit the higher. Level control schedules to be tried.
Top down testing can continue in a profundity first or an expansiveness first way. For
profundity first reconciliation every module is tried in expanding subtle element, supplanting
an ever increasing number of levels of detail with real code as opposed to stubs. On the other
hand expansiveness initially would handle by refining every one of the modules at a similar
level of control all through the application .by and by a blend of the two methods would be
utilized. At the underlying stages every one of the modules may be just halfway useful,
potentially being actualized just to manage non-mistaken information. These eventual tried in
expansiveness first way, yet over some stretch of time each future supplanted with progressive
refinements which were nearer to the full usefulness. This permits profundity initially testing
of a module to be performed at the same time with expansiveness initially testing of the
considerable number of modules.
The other significant classification of mix testing is Bottom Up Integration Testing where an
individual module is tried frame a test outfit. Once an arrangement of individual module have
been tried they are then consolidated into an accumulation of modules ,known as assembles,
which are then tried by a moment test saddle. This procedure can proceed until the point that
the fabricate comprises of the whole application. By and by a blend of best down and base up
testing would be utilized. In a substantial programming venture being produced by various
sub-groups, or a littler venture where diverse modules were worked by people. The sub
groups or people would lead base up testing of the modules which they were developing
before discharging them to a coordination group which would gather them together for top-
down testing.
Page 41
additionally called discovery testing on the grounds that exclusive the points of interest of the
interface are unmistakable to the test. Limits that are worldwide to a unit are tried here.
In the development business, platform is a transitory, simple to amass and dismantle, outline
set around a working to encourage the development of the building. The development
specialists initially construct the platform and afterward the building. Later the framework is
expelled, uncovering the finished building likewise, in programming testing, one specific test
may require some supporting programming. This product builds up can a right assessment of
the test occur. The platform programming may build up state and qualities for information
structures and giving sham outside capacities to the test. Distinctive platform programming
might be required frame one test to another test. Platform programming seldom is considered
piece of the framework.
A few times the platform programming winds up noticeably bigger than the framework
programming being tried. Typically the platform programming is not of an indistinguishable
quality from the framework programming and oftentimes is very delicate. A little change in
test may prompt considerably bigger changes in the framework.
Inner and unit testing can be computerized with the assistance of scope apparatuses. Dissects
the source code and created a test that will execute each option string of execution. Regularly,
the scope device is utilized as a part of a marginally extraordinary way. To begin with the
scope instrument is utilized to enlarge the source by putting data prints after each line of code.
At that point the testing suite is executed creating a review trail. This review trail is broke
down and reports the percent of the aggregate framework code executed amid the test suite.
On the off chance that the scope is high and the untested source lines are of low effect to
theframework's general quality, at that point not any more extra tests are required.
Page 42
5.5Unit Test cases
Page 43
Sl. No TEST CASE INPUT EXPECTED RESULT STATUS
RESULT OBTAINED
1 Upload File File Uploaded Uploaded Pass
Successfully Successfully
2 Download Selected File Downloaded Downloaded Pass
File Successfully Successfully
Conclusion
This segment deals with a couple sorts of testing, for instance, module testing that is a
framework for testing the exact working of a particular unit of the source code. This is
similarly suggested as unit testing. It also provides a brief knowledge regarding different sorts
of blend testing in that solitary programming units are joined and attempted as a social event.
This part similarly focuses on ensuring nature of the item.
Chapter 6
RESULT
The accompanying depictions layout the outcomes or yields that we are going to get
once regulated execution of the considerable number of modules of the framework.
Upload Process
Page 44
While uploading the file, first step is break the file in small blocks based on given block size
after that hash code get generated for all blocks, while generating hash code it will check
whether it is new block of data or duplicate block of data based on hash code if hash code
matched with existing hash code means it is duplicate block of data and if it is not matching
means it is new data, all new block of data we will encrypt using AES encryption then we will
upload to the cloud drive. As graph showing the result if file size is less it will take less time
to upload and if file size is big it will take more time to execute.
Download Process
Page 45
make a single file. So if file size is less it will take less time to download and file size is big it
will take more time to download.
Screenshot Results
Page 46
Screenshot-1 Admin Login Page
The screenshot-1 shows the Admin Login page.
Page 47
Screenshot-2 Admin Home Page
The screenshot-2 shows the admin home page.
Page 48
Screenshot-3 Admin Profile
The screenshot-3 shows the admin profile details.
Page 49
Screenshot-4 User Details page
The screenshot-4 shows the user details page which has the option to add new user, edit user
details and delete user account.
Page 50
Screenshot-5 Cloud Details
The screenshot-5 shows the cloud details page which consist cloud drive URL, folder name,
Login id and password.
Page 51
Screenshot-6 Hash Details
The screenshot-6 shows the Hash details. Hash is the unique code generated by hash
algorithm for every new block of data.
Page 52
Screenshot-7 Transaction Details
The screenshot-7shows the Transaction details.
Page 53
Screenshot-8 User Login Page
The screenshot-8 shows the User Login Page.
Page 54
Screenshot-9User Home Page
The screenshot-9 shows the User Home page.
Page 55
Screenshot-10 User Details
The screenshot-10 shows the User Details.
Page 56
Screenshot-11 File Upload
The screenshot-11 shows Upload File. Here we have to browse the file from system and click
on submit button. After submitting the file for upload file breaks in small block and each
block hash code get generated, that code first we will match with existing code in data base if
it is new code in that case we will update that hash code in database and encrypt that block
and upload to the cloud drive.
Page 57
Screenshot-12 File Download
The screenshot-11 shows the file download page. Here we have to select the file which one
we want to download then click on option download after that download process get started in
that first it will check how many block is there in that file and that blocks it will download
from cloud drive and merge all small block and make it single file.
Page 58
Screenshot-12 User Transaction Page
The screenshot-11 shows the user Transaction Page.
Page 59
CONCLUSION
Dynamic data ownership services is a basic and testing issue in secure deduplication over
mixed data in dispersed capacity. In this proposed system, we proposed a new secure data
deduplication intend to enhance procedure for the cloud data organization structure. In this
way, the proposed system enhances data assurance and security in conveyed capacity for
every customer who doesn’t have access regarding data. Name consistency is in like manner
guaranteed, while the arrangement empowers full ideal position to be taken of capable data
deduplication over encoded data. To the extent the correspondence cost, the proposed system
is more suitable than the existing system, while hash tag comparison our approach taking less
time compare to existing system. In this way, the proposed system achieves more secure and
well performing approach in dynamic storage for secure and powerful data deduplication.
Future Enhancement
We should reduce comparison time while deduplication of data, because when data
block will increase it will take more time to compare with new block hash tag.
For data security we should use Advanced Encryption Techniques. Like AES and
DES.
Page 60
REFERENCES
Page 61
[14] A. Rahumed, H. C. H. Chen, Y. Tang, P. P. C. Lee, J. C. S. Lui, “A secure cloud
backup system with assured deletion and version control,” Proc. International
Workshop on Security in Cloud Computing, 2011.
[15] J. Xu, E. Chang, and J. Zhou, “Leakage-resilient client-side deduplication of encrypted
data in cloud storage,” ePrint, IACR, http://eprint.iacr.org/2011/538.
[16] M. Bellare, S. Keelveedhi, and T. Ristenpart, “Message-locked encryption and secure
deduplication,” Proc. Eurocrypt 2013, LNCS 7881, pp. 296–312, 2013.Cryptology
ePrint Archive, Report 2012/631, 2012.
[17] S. Halevi, D, Harnik, B. Pinkas, and A. Shulman-Peleg, “Proofs of ownership in
remote storage systems,” Proc. ACM Conference on Computer and Communications
Security, pp. 491–500, 2011.
[18] M. Mulazzani, S. Schrittwieser, M. Leithner, and M. Huber, “Dark clouds on the
horizon: using cloud storage as attack vector and online slack space,” Proc. USENIX
Conference on Security, 2011.
[19] A. Juels, and B. S. Kaliski, “PORs: Proofs of retrievability for large files,” Proc. ACM
Conference on Computer and Communications Security, pp. 584–597, 2007.
[20] G. Ateniese, R. C. Burns, R. Curtmola, J. Herring, L. Kissner, Z. Peterson, and D.
Song, “Provable data possession at untrusted stores,” Proc. ACM Conference on
Computer and Communications Security, pp. 598–609, 2007.
[21] J. Li, X. Chen, M. Li, J. Li, P. Lee, and W. Lou, “Secure deduplication with efficient
and reliable convergent key management,” IEEE Transactions on Parallel and
Distributed Sytems, Vol. 25, No. 6, 2014.
[22] G.R. Blakley, and C. Meadows, “Security of Ramp schemes,” Proc. CRYPTO 1985,
pp. 242–268, 1985.
[23] J. Li, Y. K. Li, X. Chen, P. Lee, and W. Lou, “A hybrid cloud approach for secure
authorized deduplication,” IEEE Transactions on Parallel and Distributed Systems,
Vol. 26, No. 5, pp. 1206–1216, 2015.
[24] M. Bellare, S. Keelveedhi, T. Ristenpart, “DupLESS: Serveraided encryption for
deduplicated storage,” Proc. USENIX Security Symposium, 2013.
[25] M. Bellare, S. Keelveedhi, “Interactive message-locked encryption and secure
deduplication,” Proc. PKC 2015, pp. 516–538, 2015.
Page 62
[26] Y. Shin and K. Kim, “Equality predicate encryption for secure data deduplication,”
Proc. Conference on Information Security and Cryptology (CISC-W), pp. 64–70,
2012.
[27] X. Jin, L. Wei, M. Yu, N. Yu and J. Sun, “Anonymous deduplication of encrypted data
with proof of ownership in cloud storage,” Proc. IEEE Conf. Communications in
China (ICCC), pp.224-229, 2013.
[28] D. Naor, M. Naor, and J. Lotspiech, ”Revocation and tracing schemes for stateless
receivers,” Proc. CRYPTO 2001, Lecture Notes in Computer Science, vol. 2139, pp.
41–62, 2001.
[29] K. C. Almeroth, and M. H. Ammar, “Multicast group behavior in the Internet’s
multicast backbone (MBone),” IEEE Communication Magazine, vol. 35, pp. 124–129,
1997.
[30] Crypto++ Library 5.6.2, http://www.cryptopp.com/.
[31] S. Rafaeli, D. Hutchison, “A Survey of Key Management for Secure Group
Communication,” ACM Computing Surveys, Vol. 35, No. 3, pp. 309-329, 2003.
[32] L. Mingqiang, C. Qin, P.P.C. Lee, and J. Li, “Convergent Dispersal: Toward Storage-
Efficient Security in a Cloud-of- Clouds,” Proc. USENIX Conference on Hot Topics in
Storage and File Systems, 2014.
[33] L. Mingqiang, C. Qin, P.P.C. Lee, “CDStore: Toward Reliable, Secure, and Cost-
Efficient Cloud Storage via Convergent Dispersal Dispersal,” Proc. USENIX Annual
Technical Conference, pp. 120, 2015.
[34] J. Li, X. Chen, X. Huang, S. Tang, Y. Xiang, M. Hassan, and A. Alelaiwi, “Secure
Distributed Deduplication Systems with Improved Reliability,” IEEE Transactions on
Computer, Vol. 64, No. 2, pp. 3569–3579, 2015.
Page 63