View
161
Download
0
Category
Tags:
Preview:
Citation preview
Exploiting Dynamic Resource Allocation for Efficient
Parallel Data Processing in the Cloud
A PROJECT REPORT
Submitted by
KARTHIKA.RN 21708205038
LAKSHMIPRIYA ANBU 21708205042
NAGA BHARGAVI.CH 21708205050
in partial fulfillment for the award of the degree
of
BACHELOR OF TECHNOLOGYIn
INFORMATION TECHNOLOGY
SREE SASTHA INSTITUTE OF ENGINEERING
AND TECHNOLOGY
CHEMBARAMBAKKAM ,CHENNAI-600123
ANNA UNIVERSITY: CHENNAI-600 025 APRIL-2012
ANNA UNIVERSITY: CHENNAI 600 025BONAFIDE CERTIFICATE
Certified that this project report “Exploiting Dynamic Resource
Allocation for Efficient Parallel Data Processing in the
Cloud” is the bonafide work of
“RN.KARTHIKA” ,“LAKSHMI PRIYA
ANBU” ,“CH.NAGA BHARGAVI” who carried out the project
work under my supervision.
SIGNATURE SIGNATURE
S.MUHAMAD IBRAHIM A.RENGARAJAN
BADHUSHA
SUPERVISOR HEAD OF THE DEPARTMENT
Department of Information Department of Information
Technology Technology
Sree Sastha Institute Of Sree Sastha Institute Of
Engineering and technology Engineering and technology
Chembarambakkam, Chembarambakkam,
Chennai-600 123 Chennai-600 123
ACKNOWLEDGEMENT
Authentic thanks to Our Ethical Founder & Chairman
Thiru Dr. “A.M.K. JAMBULINGA MUDALLIAR”, for his genuine venture
in educating us in his esteemed institution.
I would like to express my heartfelt thanks to
,Principal of our college for his kind endorsement and inspiration.
We are very obliged to thank Prof Mr. “A.RENGARAJAN”, Assistant
Professor and HOD Department Of Information And Technology ,for his
valuable counseling and evocation during the course. We are profound to him
for having reassuring us.
We are copiously beholden to our Project Co-ordinator
Mr“S.MUHAMAD IBRAHIM BADHUSHA”,LECTURER,Department Of
Information And Technology for his sustained mentorship and benefaction to
take up this project and for his zeal shown towards this project
We are gratified to express our devout owing to our Project Guide
Mr.“S.MUHAMAD IBRAHIM BADHUSHA”,Lecturer in Department Of
Information And Technology for his valuable guidance and extensive efforts
offered in making the project success.
We affirm our thanks to all our department staff members for the
unceasing benefaction and inducement for the efficacious accomplishment of
our project.
We oblige our thanks to our library staff and management for their
extensive support by providing the information and resources that helped us to
complete the project successfully
TABLE OF CONTENTS
CHAPTER NO. TITLE PAGE NO.
List of Tables
List of Figures
List of Abbreviations
1 Introduction
1.1.1 Objective
2 System Analysis
2.1 Existing System
2.1.1 Drawbacks
2.2 Proposed System
2.2.1 Advantages
2.3 Feasibility Study
2.3.1 Economical Feasibility
2.3.2 Operational Feasibility
2.3.3 Technical Feasibility
3 System Specification
3.1 Hardware Requirements
3.2 Software Requirements
4 Software Description
4.1 Front End
4.2 Features
5 Project Description
5.1 Problem Definition
5.2 Overview of the Project
5.3 Module Description
5.3.1 Modules
5.4 Data Flow Diagram
5.5 E-R Diagram
5.6 Database Design
5.6.1 Table 1
5.6.2 Table 2
5.7 Input Design
5.8 Output Design
6 System Testing
6.1 Unit Testing
6.2 Acceptance Testing
6.3 Test Cases
7 System Implementation
8 Conclusion & Future Enhancements
8.1 Conclusion
8.2 Future Enhancements
9 Appendix
9.1 Source Code
9.2 Screen Shots
10 References
LIST OF FIGURES
FIGURE NO FIGURE NAME PAGENO
1.0 FEATURES OF .NET
1.1 THE .NET FRAMEWORK
1.2 LANGUAGES SUPPORTED BY .NET
1.3 OBJECTIVES OF. NET FRAMEWORK
1.4 FEATURES OF SQL-SERVER
1.0 FEATURES OF .NET:
Microsoft .NET is a set of Microsoft software technologies for
rapidly building and integrating XML Web services, Microsoft Windows-based
applications, and Web solutions. The .NET Framework is a language-neutral
platform for writing programs that can easily and securely interoperate. There’s
no language barrier with .NET: there are numerous languages available to the
developer including Managed C++, C#, Visual Basic and Java Script. The .NET
framework provides the foundation for components to interact seamlessly,
whether locally or remotely on different platforms. It standardizes common data
types and communications protocols so that components created in different
languages can easily interoperate. “.NET” is also the collective name given to
various software components built upon the .NET platform. These will be both
products (Visual Studio.NET and Windows.NET Server, for instance) and
services (like Passport, .NET My Services, and so on).
1.1.THE .NET FRAME WORK :
The .NET Framework has two main parts:
1. The Common Language Runtime (CLR).
2. A hierarchical set of class libraries.
The CLR is described as the “execution engine” of .NET. It provides the
environment within which programs run. The most important features are
Conversion from a low-level assembler-style language, called
Intermediate Language (IL), into code native to the platform being
executed on.
Memory management, notably including garbage collection.
Checking and enforcing security restrictions on the running code.
Loading and executing programs, with version control and other
such features.
The following features of the .NET framework are also worth
description:
Managed Code
The code that targets .NET, and which contains certain extra Information
- “metadata” - to describe itself. Whilst both managed and unmanaged code can
run in the runtime, only managed code contains the information that allows the
CLR to guarantee, for instance, safe execution and interoperability.
Managed Data
With Managed Code comes Managed Data. CLR provides
memory allocation and Deal location facilities, and garbage collection.
Some .NET languages use Managed Data by default, such as C#, Visual
Basic.NET and JScript.NET, whereas others, namely C++, do not. Targeting
CLR can, depending on the language you’re using, impose certain constraints
on the features available. As with managed and unmanaged code, one can have
both managed and unmanaged data in .NET applications - data that doesn’t get
garbage collected but instead is looked after by unmanaged code.
Common Type System
The CLR uses something called the Common Type System (CTS) to
strictly enforce type-safety. This ensures that all classes are compatible with
each other, by describing types in a common way. CTS define how types work
within the runtime, which enables types in one language to interoperate with
types in another language, including cross-language exception handling. As
well as ensuring that types are only used in appropriate ways, the runtime also
ensures that code doesn’t attempt to access memory that hasn’t been allocated to
it.
Common Language Specification :
The CLR provides built-in support for language interoperability. To
ensure that you can develop managed code that can be fully used by developers
using any programming language, a set of language features and rules for using
them called the Common Language Specification (CLS) has been defined.
Components that follow these rules and expose only CLS features are
considered CLS-compliant.
The class library
.NET provides a single-rooted hierarchy of classes, containing over 7000
types. The root of the namespace is called System; this contains basic types like
Byte, Double, Boolean, and String, as well as Object. All objects derive from
System. Object. As well as objects, there are value types. Value types can be
allocated on the stack, which can provide useful flexibility. There are also
efficient means of converting value types to object types if and when necessary.
The set of classes is pretty comprehensive, providing collections, file,
screen, and network I/O, threading, and so on, as well as XML and database
connectivity.
The class library is subdivided into a number of sets (or
namespaces), each providing distinct areas of functionality, with dependencies
between the namespaces kept to a minimum.
1.2 LANGUAGES SUPPORTED BY .NET
The multi-language capability of the .NET Framework and Visual
Studio .NET enables developers to use their existing programming skills to
build all types of applications and XML Web services. The .NET framework
supports new versions of Microsoft’s old favorites Visual Basic and C++ (as
VB.NET and Managed C++), but there are also a number of new additions to
the family.
Visual Basic .NET has been updated to include many new and
improved language features that make it a powerful object-oriented
programming language. These features include inheritance, interfaces, and
overloading, among others. Visual Basic also now supports structured exception
handling, custom attributes and also supports multi-threading.
Visual Basic .NET is also CLS compliant, which means that any
CLS-compliant language can use the classes, objects, and components you
create in Visual Basic .NET.
Managed Extensions for C++ and attributed programming are just
some of the enhancements made to the C++ language. Managed Extensions
simplify the task of migrating existing C++ applications to the new .NET
Framework.
C# is Microsoft’s new language. It’s a C-style language that is
essentially “C++ for Rapid Application Development”. Unlike other languages,
its specification is just the grammar of the language. It has no standard library
of its own, and instead has been designed with the intention of using the .NET
libraries as its own.
Microsoft Visual J# .NET provides the easiest transition for Java-
language developers into the world of XML Web Services and dramatically
improves the interoperability of Java-language programs with existing software
written in a variety of other programming languages.
Active State has created Visual Perl and Visual Python, which
enable .NET-aware applications to be built in either Perl or Python. Both
products can be integrated into the Visual Studio .NET environment. Visual
Perl includes support for Active State’s Perl Dev Kit.
Other languages for which .NET compilers are available include
FORTRAN
COBOL
Eiffel
Fig1 .Net Framework
ASP.NET
XML WEB SERVICES
Windows Forms
Base
Class Libraries
Common
Language Runtime
Operating System
C#.NET is also compliant with CLS (Common Language Specification) and
supports structured exception handling. CLS is set of rules and constructs that
are supported by the CLR (Common Language Runtime). CLR is the runtime
environment provided by the .NET Framework; it manages the execution of the
code and also makes the development process easier by providing services.
C#.NET is a CLS-compliant language. Any objects, classes, or components that
created in C#.NET can be used in any other CLS-compliant language. In
addition, we can use objects, classes, and components created in other CLS-
compliant languages in C#.NET .The use of CLS ensures complete
interoperability among applications, regardless of the languages used to create
the application.
Constructors and destructors:
Constructors are used to initialize objects, whereas destructors are used
to destroy them. In other words, destructors are used to release the resources
allocated to the object. In C#.NET the sub finalize procedure is available. The
sub finalize procedure is used to complete the tasks that must be performed
when an object is destroyed. The sub finalize procedure is called automatically
when an object is destroyed. In addition, the sub finalize procedure can be
called only from the class it belongs to or from derived classes.
Garbage collection
Garbage Collection is another new feature in C#.NET. The .NET Framework
monitors allocated resources, such as objects and variables. In addition,
the .NET Framework automatically releases memory for reuse by destroying
objects that are no longer in use.
In C#.NET, the garbage collector checks for the objects that are not currently in
use by applications. When the garbage collector comes across an object that is
marked for garbage collection, it releases the memory occupied by the object.
Overloading
Overloading is another feature in C#. Overloading enables us to define multiple
procedures with the same name, where each procedure has a different set of
arguments. Besides using overloading for procedures, we can use it for
constructors and properties in a class.
Multi threading:
C#.NET also supports multithreading. An application that supports
multithreading can handle multiple tasks simultaneously, we can use
multithreading to decrease the time taken by an application to respond to user
interaction.
Structured exception handling
C#.NET supports structured handling, which enables us to detect
and remove errors at runtime. In C#.NET, we need to use Try…Catch…Finally
statements to create exception handlers. Using Try…Catch…Finally statements,
we can create robust and effective exception handlers to improve the
performance of our application.
1.3 OBJECTIVES OF. NET FRAMEWORK
1. To provide a consistent object-oriented programming environment whether
object codes is stored and executed locally on Internet-distributed, or executed
remotely.
2. To provide a code-execution environment to minimizes software deployment
and guarantees safe execution of code.
3. Eliminates the performance problems.
There are different types of application, such as Windows-based applications
and Web-based applications.
1.4 FEATURES OF SQL-SERVER
The OLAP Services feature available in SQL Server version 7.0 is
now called SQL Server 2000 Analysis Services. The term OLAP Services has
been replaced with the term Analysis Services. Analysis Services also includes
a new data mining component. The Repository component available in SQL
Server version 7.0 is now called Microsoft SQL Server 2000 Meta Data
Services. References to the component now use the term Meta Data Services.
The term repository is used only in reference to the repository engine within
Meta Data Services
SQL-SERVER database consist of six type of objects,
They are,
1. TABLE
2. QUERY
3. FORM
4. REPORT
5. MACRO
Table:
A database is a collection of data about a specific topic.
Views of table:
We can work with a table in two types,
1. Design View
2. Datasheet View
Design view
To build or modify the structure of a table we work in the
table design view. We can specify what kind of data will be hold.
Datasheet view
To add, edit or analyses the data itself we work in tables
datasheet view mode.
Query:
A query is a question that has to be asked the data. Access gathers data
that answers the question from one or more table. The data that make up the
answer is either dynaset (if you edit it) or a snapshot (it cannot be edited).Each
time we run query, we get latest information in the dynaset. Access either
displays the dynaset or snapshot for us to view or perform an action on it, such
as deleting or updating.
LIST OF ABBREVIATIONS
BIR Big Integer Reader
BIS Big Integer Sorter
DAG Directed Acyclic Graph.
EI Execution Instance
EV Execution vertices
HDFS Hadoop Distributed File System
HTC High-Throughput Computing .
IAAS Infrastructure-as-a-Service
JM Job Manager
JMX Java Management Extensions
MTC Many Task Computing
PACT Parallel Architectures and Compilation Techniques
TSM TeraSortMap
TM Task Manager
TSR TeraSortReduce
VMs Virtual Machines
CHAPTER 1
INTRODUCTION
The Cloud computing paradigm is sometimes viewed as the
commercial successor of the academic Grid computing paradigm.Cloud
computing as offered by companies such as Amazon is easier to use than the
typical Grid and offers more user rights and tighter security.This has led to
much greater commercial adoption of Cloud computing than of Grid
computing.However,unlike aGrid,a Cloud doesn’t not assist its users with the
selection and use of different remote Cloud sites.Due to commercial interests,a
Cloud provider usually has a proprietary system where a user creates an image
that only runs on the provider’s site.This Represents a significant step back
compared to grid computing.While many cross-site technologies applied in grid
computing such as virtual organization management and single sign-on can
quite easily be adopted by Cloud computing,the acyual configuration and
management of user software in the cloud is significantly different compared to
the traditional Grid.
In The Cloud,Users expect to be able to install complex software
With super-user privileges on-demand.If the cloud paradigm is to be extended
to facilitate cross-provider utilization of resources,several challenges need to be
solved.One of the requirements for Cross-Cloud computing discussed in this
paper is the capability to create and efficiently distribute interoperable virtual
machines.First,We resent a Cross-Cloud virtual machine creation solution that
is based on a layered approach to allow a single user image to be deployed to
multiple cloud sites,including desktop Cloud sites.Next,we analyse methods for
the bulk transefer of these layered images to the cloud sites.Finally,We present a
study of the performance of our system with and without security.
1.1 OBJECTIVE:
For efficient parallel data processing in cloud environments
and presented Nephele framework, the first data processing framework to
exploit the dynamic resource provisioning offered by today’s IaaS clouds.
ABSTRACT:
In recent years ad-hoc parallel data processing has emerged to be one of the
killer applications for Infrastructure-as-a-Service (IaaS) clouds. Major Cloud
computing companies have started to integrate frameworks for parallel data
processing in their product portfolio, making it easy for customers to access
these services and to deploy their programs. However, the processing
frameworks which are currently used have been designed for static,
homogeneous cluster setups and disregard the particular nature of a cloud.
Consequently, the allocated compute resources may be inadequate for big parts
of the submitted job and unnecessarily increase processing time and cost. In this
paper we discuss the opportunities and challenges for efficient parallel data
processing in clouds and present our research project. It is the first data
processing framework to explicitly exploit the dynamic resource allocation
offered by today’s IaaS clouds for both, task scheduling and execution.
Particular tasks of a processing job can be assigned to different types of virtual
machines which are automatically instantiated and terminated during the job
execution.
CHAPTER 2
SYSTEM ANALYSIS
2.1 Existing System
A growing number of companies have to process huge amounts of
data in a cost-efficient manner. Classic representatives for these companies are
operators of Internet search engines. The vast amount of data they have to deal
with every day has made traditional database solutions prohibitively
expensive .Instead, these companies have popularized an architectural paradigm
based on a large number of commodity servers. Problems like processing
crawled documents or regenerating a web index are split into several
independent subtasks, distributed among the available nodes, and computed in
parallel.
2.1.1 Disadvantages of Existing System:
The cloud’s virtualized nature helps to enable promising new use
cases for efficient parallel data processing. ,However, it also imposes
new challenges compared to classic cluster setups.
The major challenge we see is the cloud’s opaqueness with prospect
to exploiting data locality.
2.2 Proposed System:
In recent years a variety of systems to facilitate MTC has been developed.
Although these systems typically share common goals (e.g. to hide issues of
parallelism or fault tolerance), they aim at different fields of application. Map
Reduce is designed to run data analysis jobs on a large amount of data, which is
expected to be stored across a large set of share-nothing commodity servers.
Once a user has fit his program into the required map and reduce pattern, the
execution framework takes care of splitting the job into subtasks, distributing
and executing them. A single Map Reduce job always consists of a distinct map
and reduce program.
2.2.1 Advantages of Proposed System:
The first data processing framework to exploit the dynamic resource
provisioning offered by today’s IaaS clouds .The performance evaluation gives
a first impression on how the ability to assign specific virtual machine types to
specific tasks of a processing job, as well as the possibility to automatically
allocate/deallocate virtual machines in the course of a job execution, can help to
improve the overall resource utilization and, consequently, reduce the
processing cost.
2.3 Feasibility Study:
Feasibility studies aim to objectively and rationally uncover the
strengths and weaknesses of the existing business or proposed venture,
opportunities and threats as presented by the environment, the resources
required to carry through, and ultimately the prospects forsuccess.In its simplest
term, the two criteria to judge feasibility are cost required and value to be
attained. As such, a well-designed feasibility study should provide a historical
background of the business or project, description of the product or service,
accounting statements, details of the operations and management, marketing
research and policies, financial data, legal requirements and tax
obligations.Generally, feasibility studies precede technical development and
project implementation
2.3.1 Economical Feasibility:
Economic analysis is the most frequently used method for evaluating the
effectiveness of a new system. More commonly known as cost/benefit analysis,
the procedure is to determine the benefits and savings that are expected from a
candidate system and compare them with costs. If benefits outweigh costs, then
the decision is made to design and implement the system. An entrepreneur must
accurately weigh the cost versus benefits before taking an action.
Cost-based study: It is important to identify cost and benefit factors, which can
be categorized as follows: 1. Development costs; and 2. Operating costs. This is
an analysis of the costs to be incurred in the system and the benefits derivable
out of the system.
Time-based study: This is an analysis of the time required to achieve a return on
investments. The future value of a project is also a factor.
2.3.2 Operational Feasibility:
Operational feasibility is a measure of how well a proposed system solves
the problems, and takes advantage of the opportunities identified during scope
definition and how it satisfies the requirements identified in the requirements
analysis phase of system development
2.3.3 Technical Feasibility:
The assessment is based on an outline design of system requirements in
terms of Input, Processes, Output, Fields, Programs, and Procedures. This can
be quantified in terms of volumes of data, trends, frequency of updating, etc. in
order to estimate whether the new system will perform adequately or not.
Technological feasibility is carried out to determine whether the company has
the capability, in terms of software, hardware, personnel and expertise, to
handle the completion of the project. When writing a feasibility report the
following should be taken to consideration:
A brief description of the business to assess more possible factor/s which
could affect the study
The part of the business being examined
The human and economic factor
The possible solutions to the problems
At this level, the concern is whether the proposal is
both technically and legally feasible (assuming moderate cost).
CHAPTER 3
SYSTEM SPECIFICATION
HARDWARE REQUIREMENTS:
System : Pentium IV 2.4 GHz.
Hard Disk : 40 GB.
Floppy Drive : 1.44 Mb.
Monitor : 15 VGA Colour.
Mouse : Logitech.
Ram : 512 MB.
SOFTWARE REQUIREMENTS:
Operating system : Windows XP Professional.
Coding Language : ASP .Net,C#
Database : Sql Server 2005.
CHAPTER 4
SOFTWARE DESCRIPTION
4.1 Front End:
ASP.NET is a Web application framework developed and marketed by
Microsoft to allow programmers to build dynamic Web sites, Web applications
and Web services. It was first released in January 2002 with version 1.0 of
the .NET Framework, and is the successor to Microsoft's Active Server Pages
(ASP) technology. ASP.NET is built on the Common Language Runtime
(CLR), allowing programmers to write ASP.NET code using any
supported .NET language. The ASP.NET SOAP extension framework allows
ASP.NET components to process SOAP messages.
4.2 Features Of .Net :
Microsoft .NET is a set of Microsoft software technologies for rapidly
building and integrating XML Web services, Microsoft Windows-based
applications, and Web solutions. The .NET Framework is a language-neutral
platform for writing programs that can easily and securely interoperate. There’s
no language barrier with .NET: there are numerous languages available to the
developer including Managed C++, C#, Visual Basic and Java Script. The .NET
framework provides the foundation for components to interact seamlessly,
whether locally or remotely on different platforms. It standardizes common data
types and communications protocols so that components created in different
languages can easily interoperate.“.NET” is also the collective name given to
various software components built upon the .NET platform. These will be both
products (Visual Studio.NET and Windows.NET Server, for instance) and
services (like Passport, .NET My Services, and so on).
CHAPTER 5
PROJECT DESCRIPTION
5.1 Problem Definition:
5.2 Overview of the Project:
Abstract
Existing System
Proposed System
Hardware Requirements
Software Requirements
The Nephele Execution Engine
Nephele Architecture
Demonstration of an Example Task
5.3 Module Description :
5.3.1 . NETWORK MODULE:
Server - Client computing or networking is a distributed application
architecture that partitions tasks or workloads between service providers
(servers) and service requesters, called clients. Often clients and servers operate
over a computer network on separate hardware. A server machine is a high-
performance host that is running one or more server programs which share its
resources with clients. A client also shares any of its resources; Clients therefore
initiate communication sessions with servers which await (listen to) incoming
requests.
5.3.2. LBS SERVICES:
In particular, users are reluctant to use LBSs, since revealing their
position may link to their identity. Even though a user may create a fake ID to
access the service, her location alone may disclose her actual identity. Linking a
position to an individual is possible by various means, such as publicly
available information city maps. When a user u wishes to pose a query, she
sends her location to a trusted server, the anonymizer through a secure
connection (SSL). The latter obfuscates her location, replacing it with an
anonymizing spatial region (ASR) that encloses u. The ASR is then forwarded
to the LS. Ignoring where exactly u is, the LS retrieves (and reports to the AZ) a
candidate set (CS) that is guaranteed to contain the query results for any
possible user location inside the ASR. The AZ receives the CS and reports to u
the subset of candidates that corresponds to her original query.
5.3.3. SYSTEM MODEL:
The ASR construction at the anonymization process abides by the
user’s privacy requirements. Particularly, specified an anonymity degree K by
u, the ASR satisfies two properties: (i) it contains u and at least another K * 1
users, and (ii) even if the LS knew the exact locations of all users in the system.
We propose an edge ordering anonymization approach for users in road
networks, which guarantees K-anonymity under the strict reciprocity
requirement (described later).
We identify the crucial concept of border nodes, an important indicator of
the CS size and of the query processing cost at the LS.
We consider various edge orderings, and qualitatively assess their query
performance based on border nodes.
We design efficient query processing mechanisms that exploit existing
network database infrastructure, and guarantee CS inclusiveness and
minimality. Furthermore, they apply to various network storage schemes.
We devise batch execution techniques for anonymous queries that
significantly reduce the overhead of the LS by computation sharing.
5.3. 4. SCHEDULED TASK:
Recently, considerable research interest has focused on preventing
identity inference in location-based services. Proposing spatial cloaking
techniques. In the following, we describe existing techniques for ASR
computation (at the AZ) and query processing (at the LS). At the end, we cover
alternative location privacy approaches and discuss why they are inappropriate
to our problem setting. This offers privacy protection in the sense that the actual
user position u cannot be distinguished from others in the ASR, even when
malicious LS is equipped/advanced enough to possess all user locations. This
spatial K-anonymity model is most widely used in location privacy
research/applications, even though alternative models are emerging.
5.3.5. QUERY PROCESSING:
Processing is based on implementation of the theorem uses
(network-based) search operations as off the shelf building blocks. Thus, the
NAP query evaluation methodology is readily deployable on existing systems,
and can be easily adapted to different network storage schemes. In this case, the
queries are evaluated in a batch. we propose the network-based anonymization
and processing (NAP) framework, the first system for K- anonymous query
processing in road networks. NAP relies on a global user ordering and
bucketization that satisfies reciprocity and guarantees K-anonymity. We
identify the ordering characteristics that affect subsequent processing, and
qualitatively compare alternatives. Then, we propose query evaluation
techniques that exploit these characteristics. In addition to user privacy, NAP
achieves low computational and communication costs, and quick responses
overall. It is readily deployable, requiring only basic network operations.
5.4 Data Flow Diagram:
Join Request Nodes
Address &routing Un addressed dataTable
Network Tree
Address
Nodes
Cache
Routing cache Network Path
New Node
ReOrganizer
Address Resolver
Routing system
5.5 E-R Diagram:
5.6 Database Design:
5.6.1 Table 15.6.2 Table 2
User Mode of transport
Personal information
Location
Gives
Pin Area
Distance
Flight
Train
License id
Bus Personal travel
Name Phone no
Email id
5.7 Input Design5.8 Output Design
CHAPTER 6
SYSTEM TESTING
6.1 Unit Testing:
Unit testing is conducted to verify the functional performance of
each modular component of the software.Unit testing focuses on the smallest
unit of the software design (i.e), the module. The white-box testing techniques
were heavily employed for unit testing.
6.2 Acceptance Testing:
Acceptance testing is a term used in agile software development
methodologies, particularly Extreme Programming, referring to the functional
testing of a user story by the software development team during the
implementation phase.
The customer specifies scenarios to test when a user story has been correctly
implemented. A story can have one or many acceptance tests, whatever it takes
to ensure the functionality works. Acceptance tests are black box system tests.
Each acceptance test represents some expected result from the system.
Customers are responsible for verifying the correctness of the acceptance tests
and reviewing test scores to decide which failed tests are of highest priority.
Acceptance tests are also used as regression tests prior to a production release.
A user story is not considered complete until it has passed its acceptance tests.
This means that new acceptance tests must be created for each iteration or the
development team will report zero progress
6.3 Test Cases:
A test case in software engineering is a set of conditions or
variables under which a tester will determine whether an application
or software system is working correctly or not. The mechanism for determining
whether a software program or system has passed or failed such a test is known
as a test oracle. In some settings, an oracle could be a requirement or use case,
while in others it could be a heuristic. It may take many test cases to determine
that a software program or system is considered sufficiently scrutinized to be
released. Test cases are often referred to as test scripts, particularly when
written. Written test cases are usually collected into test suites.
CHAPTER 7
SYSTEM IMPLEMENTATION
Implementation is the stage of the project when the theortical
design is turned out into a working system.Thus it can be considered to be the
most critical stage in achieving a successful new system and in giving the
user,confidence thtat the new system will work and be effective.
The Implementation stage involves carefull planning,investigation
of the existing system and it’s constraints on implementation,designing of
methods to achieve change over and evaluation of change over methods.
A hierarchial structuring of relations may result in more classes and a
more complicated structure to implement.Therefore it is advisable to transform
the hierarchial relation structure to a simpler structure such as a classical flat
one.It is ratherstraight forward to transform the developed hierarchial model
into a bipartite,flat model,consisting of classes on the one hand and flat relations
on the other.
Flat relations are preferred at the design level for reasons of simplicity
and implementation ease.there is no identity or functionality associated with a
flat relation.A flat relation corresponds with the relation concept of entity-
relationship modeling and many object oriented methods
CHAPTER 8
CONCLUSION AND FUTURE ENHANCEMENT
8.1 CONCLUSIONS:
For efficient parallel data processing incloud environments and
presented Nephele, the first data processing framework to exploit the dynamic
resource provisioning offered by today’s IaaS clouds. Wehave described
Nephele’s basic architecture and presented a performance comparison to the
well-established data processing framework Hadoop. The performance
evaluation gives a first impression on how the ability to assign specific virtual
machine types to specific tasks of a processing job, as well as the possibility to
automatically allocate/deallocate virtual machines in the course of a job
execution, can help to improve the overall resource utilization
and,consequently, reduce the processing cost.
8.2 FUTURE ENHANCEMENT:
With a framework like Nephele at hand, there are a variety
of open research issues, which we plan to address for future work. In particular,
we are interested in improving Nephele’s ability to adapt to resource overload
or underutilization during the job execution automatically. Our current profiling
approach builds a valuable basis for this, however, at the moment the system
still requires a reasonable amount of user annotations.
CHAPTER 9
APPENDIX
9.1 Source Code:
9.2 Screen Shots:
Recommended