You are on page 1of 53

WAREHOUSE MANAGEMENT SYSTEM IMPLEMENTATION

WITH ZERO INVENTORY


ABSTRACT:
Big sensing data is prevalent in both industry and scientific research applications where
the data is generated with high volume and velocity. Cloud computing provides a promising
platform for big sensing data processing and storage as it provides a flexible stack of massive
computing, storage, and software services in a scalable manner. Current big sensing data
processing on Cloud have adopted some data compression techniques. However, due to the high
volume and velocity of big sensing data, traditional data compression techniques lack sufficient
efficiency and scalability for data processing. Based on specific on-Cloud data compression
requirements, we propose a novel scalable data compression approach based on calculating
similarity among the partitioned data chunks. Instead of compressing basic data units, the
compressions will be conducted over partitioned data chunks. To restore original data sets, some
restoration functions and predictions will be designed. MapReduce is used for algorithm
implementation to achieve extra scalability on Cloud. With real world meteor logical big sensing
data experiments on U-Cloud platform, we demonstrate that the proposed scalable compression
approach based on data chunk similarity can significantly improve data compression efficiency
with affordable data accuracy loss.
CHAPTER 1

INTRODUCTION
OVER VIEW

In line with the numerous electronics manufacturers producing new mobile devices such
as smart phones and smart tablets, various mobile services are being provided as applications for
these devices.there are more than 200,000 Android and 300,000 iPhone applications available as
of March 2011 and these numbers are increasing rapidly. One recent trend for mobile services is
their change to cloud-based mobile services. Cloud-based mobile services benefit users by richer
communications and higher flexibility. Richer communications mean advanced techniques
supporting such as enhanced phonebooks, messaging with push notification, and enriched call
with multi-media content sharing.Massive computational processing is performed through cloud
computing infrastructure instead of low-speed mobile devices.The data stored in cloud
infrastructure can be accessed at any time and from anywhere through mobile devices.

As a result, richer communications and higher flexibility can be provided to mobile


device users through cloud computing.

Through the convergence of mobile devices and cloud services, we expect that new
mobile cloud services will be provided with the virtualization of mobile devices in cloud
infrastructure. Virtual smartphone over IP is one example of provisioning virtual mobile
instances to users. Each virtual instance in cloud infrastructure represents a mobile device, and
users can connect to and use this instance.Here, we present a mobile cloud infrastructure as an
infrastructure that provides virtual mobile instances, and those instances are managed in cloud
computing architecture with massive computational processing power and storage.
However, service providers should be aware of security problems that may arise when
they adopt and launch new cloud services. According to an IDC report , when questioned, 74.6%
of service providers answered that the most importantissue for cloud services is security.
In addition, recent cloud computing attacks make it difficult to guarantee the trust and
safety of cloud services. For mobile cloud services, malicious mobile applications can be run on
virtual mobile instances and therefore any security problems may be much severe if those
applications target on the virtualization of mobile cloud infrastructure.

Here we focuses on the abnormal behavior detection in mobile cloud infrastructure.


Although signature-based vaccine applications can target on virtual mobile instances to detect
malware, it makes additional overhead on instances, and it is difficult for users to install vaccine
software by force when those instances are provided as a service. Behavior-based abnormal
detection can address those problems by observing activities in the cloud infrastructure. To
achieve this, we design a monitoring architecture using both the host and network data. Using
monitored data, abnormal behavior is detected by applying a machine learning algorithm. To
validate our methodology, we built a test bed for mobile cloud infrastructure, intentionally
installed malicious mobile programs onto several virtual mobile instances, and then successfully
detected the abnormal behavior that arose from those malicious programs.

Cloud Computing refers to the infrastructure in which applications are delivered as services over
the Internet. These infrastructures are supported by very large networked distributed machines.
Users typically can pay for the time they would like to use these resources, e.g. CPU usage per
hour or storage costs per day. This mode of computation is beneficial to both the user and
provider for several reasons:

by allowing pay-per-use model, the users can run their jobs with less cost investment compared
to owning the machines themselvessince there is a cost associated with the loan of the resources,
users will always have an incentive to return them, when no longer neededan easy way for the
cloud provider to add resources when the demands are not met anymore
Aim & Objective
To improve the data compression efficiency by avoiding traditional compression based on
each data unit, which is space and time costly due to low level data traverse and manipulation. At
the same time, because the compression happens at a higher data chunk level, it reduces the
chance for introducing too much usage of iteration and recursion which prove to be main trouble
in processing big graph data.

Security a major Concern:

1. Security concerns arising because both customer data and program are residing in
Provider Premises.

2. Security is always a major concern in Open System Architectures

Data centre Security?

1. Professional Security staff utilizing video surveillance, state of the art intrusion detection
systems, and other electronic means.

2. When an employee no longer has a business need to access datacenter his privileges to
access datacenter should be immediately revoked.

3. All physical and electronic access to data centers by employees should be logged and
audited routinely.

Data Location:

1. When user uses the cloud, user probably won't know exactly where your data is hosted,
what country it will be stored in?
2. Data should be stored and processed only in specific jurisdictions as define by user.

3. Provider should also make a contractual commitment to obey local privacy requirements
on behalf of their customers,

4. Data-centered policies that are generated when a user provides personal or sensitive
information, that travels with that information throughout its lifetime to ensure that the
information is used only in accordance with the policy

Backups of Data:

1. Data store in database of provider should be redundantly store in multiple physical


location.

2. Data that is generated during running of program on instances is all customer data and
therefore provider should not perform backups.

3. Control of Administrator on Databases.

Data Sanitization:

1. Sanitization is the process of removing sensitive information from a storage device.

2. What happens to data stored in a cloud computing environment once it has passed its
user’s “use by date”

Network Security:

1. Denial of Service: where servers and networks are brought down by a huge amount of
network traffic and users are denied the access to a certain Internet based service.

2. Like DNS Hacking, Routing Table “Poisoning”, XDoS attacks


3. QoS Violation : through congestion, delaying or dropping packets, or through resource
hacking.

4. Man in the Middle Attack: To overcome it always use SSL

5. IP Spoofing: Spoofing is the creation of TCP/IP packets using somebody else's IP


address.

6. Solution: Infrastructure will not permit an instance to send traffic with a source IP or
MAC address other than its own.

How secure is encryption Scheme:

1. Is it possible for all of my data to be fully encrypted?

2. What algorithms are used?

3. Who holds, maintains and issues the keys? Problem:

4. Encryption accidents can make data totally unusable.

5. Encryption can complicate availability Solution

6. The cloud provider should provide evidence that encryption schemes were designed and
tested by experienced specialists.

Information Security:

1. Security related to the information exchanged between different hosts or between hosts
and users.

2. This issues pertaining to secure communication, authentication, and issues concerning


single sign on and delegation.

3. Secure communication issues include those security concerns that arise during the
communication between two entities.
4. These include confidentiality and integrity issues. Confidentiality indicates that all data
sent by users should be accessible to only “legitimate” receivers, and integrity indicates
that all data received should only be sent/modified by “legitimate” senders.

5. Solution: public key encryption, X.509 certificates, and the Secure Sockets Layer (SSL)
enables secure authentication and communication over computer networks.
CHAPTER 2

SYSTEM STUDY

FEASIBILITY STUDY

When complex problem and opportunities are to be defined, it is generally desirable to


conduct a preliminary investigation called a feasibility study. A feasibility study is conduct to
obtain an overview of the problem and to roughly assess whether feasible solution exists prior to
committing substantial resources to a project. During a feasibility study, the system analyst
usually works with representatives from the departments(s) expected to benefit from the solution.

Every project is feasible if given unlimited resource and infinite time. Unfortunately, the
development of computer based systems is more likely to be plagued by scarcity of resources
and difficult delivery of data it is both necessary and prudent to evaluate the feasibility of a
project at the earliest possible time. Precious time and money can be saved and untold
professional embarrassment can be averted if an ill conceived system is recognized early in the
definition phase. So a detailed study is carried out to check the workability of the system.

Feasibility study is undertaken to evaluate its workability, impact on the organization,


ability to meet user needs, and effective se of resources. The main objective of feasibility study is
to test the technical, operational and economical feasibility of developing the computer system.
Thus, during feasibility analysis for this project the following three primary areas of interest
were considered very carefully. The feasibility of a project can be ascertained in terms of
technical factors, economic factors, or both. A feasibility study is documented with a report
showing all the ramification of the project.

The primary objective of a feasibility study is to assess three types of feasibility.

1) Technical feasibility: can a solution be supported with existing technology?


2) Economical feasibility: is existing technology cost effective?

3) Operational feasibility: will the solution work in the organization if implemented?

3.1 TECHNICAL FEASIBILITY

A systems development project may be regarded as technically feasibility or ‘practical’ if


the organization has the necessary expertise and infrastructure to develop, install, operate and
maintain the proposed system. Organizations will need to make this assessment based on:

• Knowledge of current and emerging technological solutions.

• Availability of technically qualified staff in house for the duration of the project and
subsequent maintenance phase.

• Availability of infrastructure in house to support the development and maintenance of the


proposed system.

• Where necessary, the financial and/or technical capacity to procure appropriate


infrastructure and expertise from outside.

• Capacity of the proposed system to accommodate increasing levels of use over the
medium term and capacity of the proposed system to meet initial performance expectations and
accommodate new functionality over the medium term.

The existing computer system has a good hardware configuration and good software facilities in
such a way that any alteration can be implemented with slight modifications of the existing
process. Hence this project is technically feasible.

3.2 ECONOMIC FEASIBILITY

A systems development project may be regarded as economically feasible or ‘good


value’ to the organization if its anticipated benefits outweigh its estimated costs. However, many
of the organizational benefits arising from record keeping projects are intangible and may be
hard to quantify. In contrasts, many development costs are easier to identify.
These costs may include the time, budget and staff resources invested during the
design and implementation phase as well as infrastructure, support, training and maintenance
costs incurred after implementation. In these high risk situations it may be appropriate
assessments of financial feasibility.

3.3. OPERATIONAL FEASIBILITY

A systems development project is likely to be operationally feasible if it meets the


‘needs’ and expectations of the organization. User acceptance is an important determinant of
operational feasibility.

FEASIBILITY STUDY OF THE PROPOSED SYSTEM

The feasibility study of the proposed system has been carried out in all the three areas.

Technical Feasibility

The proposed system can be easily developed using resources available in the
organization. Hence it is technically feasible.

3.4 ECONOMIC FEASIBILITY

The proposed system can be easily developed using the resources available in the
organization and they do not invest in procurement of additional hardware or software. The cost
of developing the system, including all the phases have been taken into account and it is strict
minimum. Hence the system is economically feasible.

3.5 OPERATIONAL FEASIBILITY

The system has been developed after extensive discussion with the end user and all the
operational requirements has been taken into account during the planning and implementation
stages. Hence the system is operationally feasible.
Related Work

In this section we present some work related to this area of research.Arshad et al. [2] presents a
framework for intrusion detection and diagnosis for clouds. The goal of the paper was to map the
input call sequences to one of the five severity levels: "minimal", "medium", "serious", "critical",
and "urgent". The authors have used decision trees for this task. The tree learns rules which can
perform predictions on unseen instances. Experiments with publicly available system call
sequences from the University of New Mexico (UNM) show that the algorithm exhibits good
performance. A similar approach was also developed by Zheng et al. [3] and [4]. The last paper
uses canonical correlation analysis (CCA) for tracking maximally correlated subspaces over
time. one problem with both these techniques is that they both need labeled examples for training
which are difficult to acquire.

Most of the existing techniques for failure detection are rule-based [5] which defines a set of
watchdogs. The method comprises of monitoring a single sensor using some hard thresholds.
Whenever, the sensor value crosses the threshold, an alarm is raised. However, this threshold
needs to be changed for different types of jobs to prevent missed detections and false alarms.

Bodik et al. [6] develop a method for identifying time cycles in machine performance which
fall below a certain threshold. They use quantiles of the measured data to statistically quantify
faults. They optimize the false positive rate and provide the user to directly control it. This
method was evaluated on a real datacenter running enterprise level services giving around 80%
detection accuracy. However, as with some of the previous techniques, this method too requires
labeled examples. An overview article on this topic is available at [7].
Pelleg et al. [8] explore failure detection in virtual machines. They use decision trees to
monitor counters of the system. First of all, this method requires labeled instances for training
and. Moreover, the counters which are monitored are manually detected which reduces the scope
of its general applicability. It is only suitable for well managed settings that include predictable
workloads and previously seen failures.

Some data mining techniques have also been applied for monitoring distributed systems e.g. the
Grid Monitoring System (GMS) by Palatin et al. [9] and the fast outlier detection by Bhaduri et
al. [10]. GMS uses a distributed distance-based outlier detection algorithm, which detects
outliers using the average distance to k nearest neighbors. Similar to our method, GMS is based
on outlier detection and is unsupervised and requires no domain knowledge. But the detection
rate of GMS can be very slow due to the quadratic time complexity of k-nn computation. The
authors in [10] propose to speed up this computation using fast database indexing and distributed
computation

FAULT DETECTION IN CLOUD SYSTEMS (FDCS)

In this section we describe our Fault Detection in Cloud Systems (FDCS) framework in which
the participating machines in a cloud computing environment can collaboratively track the
performance of other machines in the system and raise an alarm in case of faults. our algorithm
relies on in-network processing of messages, thereby making it faster than the brute force
alternative approach of data centralization. Moreover, as we discuss in this section, it also allows
fault isolation — determining which features are most faulty — which is valuable to take
remedial actions.

In our distributed setup, we assume that there is a central machine in the cloud
infrastructure called reporter which does the final reporting of all the outliers. We also assume
that all computational entities P1 , . . . , Ppform a unidirectional communication ring (except the
leader machine Po) i.e. any machine Pican communicate with the machine with the higher id Pi
+ 1, 1 <i < p. Furthermore, each machine holds its own data partition Diwhile the test points are
either sent by Po or read from the disk.
PROBLEM DEFINITION

In this section we present some notations which are necessary for discussing our FDCS
framework.

Let P 1 , . . . , P p be p machines in the cloud infrastructure connected to each other via a


communication infrastructure such that the set of (one-hop) neighbors of Pi, 1^ is known to Pj.
Each Pi holds a dataset D i (e.g. its status or log file) containing n vectors each in R d . We assume

• Disjoint property: Din Dj = 0, Vi = j

• Global property: D = |J p=1D i

In real applications, it is not feasible to compute D due to massive data sizes, changing datasets
or both. In this paperwe have only introduced this notation to formally define our global fault
detection task via distributed processing.
Given two user-defined parameters t, k > 0, let Nk(x, D) denote the set of k nearest neighbors
from {D \ {x}} to x (with respect to Euclidean distance with ties broken according to some
arbitrary but fixed total ordering -<). Let Sk(x, D) denote the maximum distance between x and
all the points in Nk(x, D) i.e. the distance between x and its k-nearest neighbors in D. Sk(x, D)
can be viewed as an outlier ranking function. Let Ot,k(D) denote the top t points (outliers) in D
according to Sk(.,D). In the rest of the description, for simplicity, we rewrite Nk(b,D), Sk(b,D)
and Ot,k(D) as Nk(b), Sk(b) and Ok.
Definition 3.1 (Distributed fault detection): Given integers t, k > 0, and dataset Di at each
machine Pi, the goal of distributed fault detection algorithm is to compute the outliers Ok (in D =
Di).

CHAPTER-3

SYSTEM ANALYSIS

EXISTING SYSTEM:
 In an existing system, an anomaly detection technique was used for through-wall human
detection to demonstrate the big sensing data processing effectiveness. This technique is
totally based on compressive sensing. The results showed that the proposed anomaly
detection algorithm could effectively detect the existence of a human being through
compressed signals and uncompressed data.
 An adaptive data gathering scheme by compressive sensing for wireless sensor networks
was developed. By introducing autoregressive (AR) model into the reconstruction of the
sensed data, the local correlation in sensed data is exploited and thus local adaptive
sparsity is achieved. Up to about 8dB SNR gain can be achieved over conventional CS
based method.
 There is also technique focusing on parallel data storing over large-scale distributed
storage stock of Cloud platform. The stored big graph data or stream data sets will be
queried and evaluated as the model of distributed data-base in Cloud, such as “Hydoop”
and its related “Hive”, “HBase”, “Zookeeper” and so on.

EXISTING ALGORITHM
 Iterations and recursive algorithms

DISADVANTAGES OF EXISTING SYSTEM:


 Due to the size and speed of big sensing data in real world, the current data compression
and reduction techniques still need to be improved.
 Iterations and recursive algorithms may cause computation problems such as parallel
memory bottlenecks, deadlocks on data accessing, algorithm inefficiency.
 In other words, under some circumstances, even with Cloud platform, the task of big data
processing may introduce unacceptable time cost, or even lead to processing failures.
 Due to the huge volume of the big sensing data, the only data size reduction at that level
is not enough.

PROPOSED SYSTEM

PROPOSED SYSTEM:
 To further improve the data size reduction, reduce the processing time cost and release
the iterations in processing big sensing data, in this paper, we propose a novel technique
based on data chunk partitioning for effectively processing big data, especially streaming
big sensing data on Cloud. With this novel technique, big sensing data stream will be
filtered to form standard data chunk sat first based on our predefined similarity model.
Then, the coming sensing data stream will be compressed according to the generated
standard data chunks.
 With the above data compression, we aim to improve the data compression efficiency by
avoiding traditional compression based on each data unit, which is space and time costly
due to low level data traverse and manipulation.
 At the same time, because the compression happens at a higher data chunk level, it
reduces the chance for introducing too much usage of iteration and recursion which prove
to be main trouble in processing big graph data.

PROPOSED ALGORITHM
 Data chunk generation algorithm and scalable compression algorithm

ADVANTAGES OF PROPOSED SYSTEM:


 Our compression algorithm can guarantee the acceptable data accuracy when make
significant performance gains in data compression.
 It was demonstrated that our proposed scalable compression based on data chunk
similarity significantly improved data compression performance gains with affordable
data accuracy loss.
 The significant compression ratio brought dramatic space and time cost savings.

DESIGN DEVELOPED

1. Designing computer output should proceed in an organized, well thought out manner;
the right output must be developed while ensuring that each output element is designed so that
people will find the system can use easily and effectively. When analysis design computer
output, they should Identify the specific output that is needed to meet the requirements.

2.Select methods for presenting information.

3.Create document, report, or other formats that contain information produced by the
system.
For our experiments, we monitored the cluster perfor-mance in a controlled environment
by submitting a fixed set of 64 jobs to run on 8 machines for 3 days. Our job consists of reading
200 MB numerical data followed by a kernel and SVD computation, and finally writing the
solution on disk files. The code written in MATLAB is shown in Figure 3. The FDCS algorithm
in our experiment uses the last 27 parameters.

CHAPTER 4

SOFTWARE REQUIREMENTS

4.1 HARDWARE REQUIREMENTS:

The minimum hardware requirements are as follows:

Processor : Pentium series

RAM(Primary Processor) : 512 MB RAM

Hard Disk Capacity : 20GB

Monitor : Colors/SVGA(800 x 600)

Printer : Laser Compatible

4.2 SOFTWARE REQUIREMENTS:


The software used is as follows:

Operating System Windows XP

Platform Java 2 Enterprise Edition

Front End Java Swing

Back End Microsoft SQL Server 2005

4.2.1 SOFTWARE METHODOLOGY USED

Problem & Assign requirements to Design System


LiteratureSurvey Modules Architecture

Develop Testing of Integrate Validate


Modules Modules System System

Fig 1.1: Addition of extra modules (Future Enhancements)


SYSTEM SPECIFICATION

4.1 HARDWARE REQUIREMENTS:

 System : Pentium IV 2.4 GHz.


 Hard Disk : 80 GB.
 Monitor : 15 VGA Colour.
 Mouse : Logitech.
 Ram : 2 GB.

4.2 SOFTWARE REQUIREMENTS:

• Operating system : Windows XP/7.


• Platform : Visual Studio.NET 2010
• Front-end : ASP. Net and Window Forms
• Coding language : ASP.Net with C#
• Data Base : SQL Server 2008.
• Web Browser : Mozilla Firefox

4.3 SOFTWARE REQUIREMENT SPECIFICATIONS

4.3.1 Introduction to .NET Framework


 The Microsoft .NET Framework is a software technology that is available with several
software windows operating systems.

 It includes a large library of pre-coded solutions to common programming problems and


a virtual machine that manages the execution of the programs written specifically for the
framework.

 The pre-coded solutions that form the framework’s Base Class Library cover a large
range of programming needs in a number of areas, including user, interface, data access,
database connectivity, cryptography, web application development, numeric algorithms
and network communications.

 The class library is used by programmers, who combine if with their own code to produce
applications.

 Programmers written for the .NET framework executes in the software environment that
manages the program’s runtime requirements. Also part of the .NET framework, this
runtime environment is known as the common Language Runtime (CLR)..

Architecture of .NET Framework

Figure 3.1 .NET Framework


Principal design features
 Interoperability
 Common Language Runtime

 Base Class Library

 Portability

 Common Language Infrastructure

 Assemblies

 Metadata

 Security

Interoperability
Because interaction between new and older applications is commonly required, the .NET
framework provides means to access functionality that is implemented in programs that execute
outside the .NET environment. Access to COM components is provided in the
System.Runtime.Interopservices and System.EnterpricesServices namespace of the framework,
access to other functionality is provided using the p/Invoke function.

Common Language Runtime


The Common Language Runtime (CLR) is the virtual machine component of the .NET
framework. All .NET programs execute under the supervision of the CLR, guaranteeing certain
properties and behavior in the area of memory management, security and executing handling.

Base Class Library


The Base Class Library (BCL), part of the FrameworkClassLibrary (FCL), is the library
of functionality available to all languages using the .NET framework. The BCL provides classes
which encapsulate a number of common functions, including file reading and writing, graphics
rendering, database interaction and XML document manipulation.

Portability
The design of the .NET framework allows it to theoretically be platform agnostic, and
thus cross-platform compatible. That is, a program written to use the framework should run
without change on any type operating system for which the framework is implemented.
Microsoft’s commercial implementations of the framework cover windows, windows CE and
xbox 360.

Common Language Infrastructure


The core aspect of the .Net framework infrastructure lies within the Common Language
Infrastructure, or CLI. The purpose of CLI is to provide the language neutral platform for
application development and execution, including functions for execution handling, garbage
collection, security and interoperability. Microsoft’s implementation of the CLI is called the
Common Language Runtime or CLR.

Assemblies
The intermediate CLI code is hosted in .NET assemblies. As mandated by specification,
assemblies are stored in the Portable Executable (PE) format, common on the windows platform
for all dll and exe files

Architecture
Figure 3.2 Visual overview of Common Language Infrastructure

Metadata
All CLI is self describing through .NET metadata. The CLR checks the metadata to
ensure that the correct method is called. Metadata is usually generated by language compilers but
developers can create their own metadata through custom attributes. Metadata contains
information about the assembly, and is also used to implement the relative programming
capabilities of .NET framework.

Security
.NET has to own security mechanism with two general features: Code Access Security
(CAS), and validation and verification.
Code Access Security is based on evidence that is associated with a specific assembly.
When assembly is loaded the CLR performs various two tasks namely validation and
verification.

Memory management
The .NET Garbage collector (GC) is a non-deterministic, compacting, mark-and-sweep
garbage collector. The GC runs only when a certain amount of memory has been used or there is
enough pressure for memory on the system.

Figure 3.3 .Net framework stack


4.3.2 ASP.NET - Introduction

ASP.NET is a web development platform, which provides a programming model, a


comprehensive software infrastructure and various services required to build up robust web
applications for PC, as well as mobile devices.ASP.NET works on top of the HTTP protocol, and
uses the HTTP commands and policies to set a browser-to-server bilateral communication and
cooperation.ASP.NET is a part of Microsoft .Net platform. ASP.NET applications are compiled
codes, written using the extensible and reusable components or objects present in .Net
framework. These codes can use the entire hierarchy of classes in .Net framework.The ASP.NET
application codes can be written in any of the following languages:

 C#
 Visual Basic.Net
 Jscript
 J#
ASP.NET is used to produce interactive, data-driven web applications over the internet. It
consists of a large number of controls such as text boxes, buttons, and labels for assembling,
configuring, and manipulating code to create HTML pages.

ASP.NET Web Forms Model

ASP.NET web forms extend the event-driven model of interaction to the web applications. The
browser submits a web form to the web server and the server returns a full markup page or
HTML page in response.All client side user activities are forwarded to the server for stateful
processing. The server processes the output of the client actions and triggers the reactions.Now,
HTTP is a stateless protocol. ASP.NET framework helps in storing the information regarding the
state of the application, which consists of:

 Page state
 Session state

The page state is the state of the client, i.e., the content of various input fields in the web
form. The session state is the collective information obtained from various pages the user visited
and worked with, i.e., the overall session state. To clear the concept, let us take an example of a
shopping cart.User adds items to a shopping cart. Items are selected from a page, say the items
page, and the total collected items and price are shown on a different page, say the cart page.
Only HTTP cannot keep track of all the information coming from various pages. ASP.NET
session state and server side infrastructure keeps track of the information collected globally over
a session.

The ASP.NET runtime carries the page state to and from the server across page requests
while generating ASP.NET runtime codes, and incorporates the state of the server side
components in hidden fields.

The ASP.NET Component Model

The ASP.NET component model provides various building blocks of ASP.NET pages.
Basically it is an object model, which describes:
 Server side counterparts of almost all HTML elements or tags, such as <form> and
<input>.
 Server controls, which help in developing complex user-interface. For example, the
Calendar control or the Gridview control.

ASP.NET is a technology, which works on the .Net framework that contains all web-related
functionalities. The .Net framework is made of an object-oriented hierarchy. An ASP.NET web
application is made of pages. When a user requests an ASP.NET page, the IIS delegates the
processing of the page to the ASP.NET runtime system.The ASP.NET runtime transforms the
.aspx page into an instance of a class, which inherits from the base class page of the .Net
framework. Therefore, each ASP.NET page is an object and all its components i.e., the server-
side controls are also objects.

4.4 SQL Server 2008

Microsoft SQL Server is a relational database management system developed by


Microsoft. As a database server, it is a software product with the primary function of storing and
retrieving data as requested by other software applications-which may run either on the same
computer or on another computer across a network (including the Internet).

SQL is Structured Query Language, which is a computer language for storing, manipulating
and retrieving data stored in relational database.SQL is the standard language for Relation
Database System. All relational database management systems like MySQL, MS Access, Oracle,
Sybase, Informix, postgres and SQL Server use SQL as standard database language.Also, they
are using different dialects, such as:

 MS SQL Server using T-SQL,


 Oracle using PL/SQL,
 MS Access version of SQL is called JET SQL (native format) etc.

History

The history of Microsoft SQL Server begins with the first Microsoft SQL Server product
- SQL Server 1.0, a 16-bit server for the OS/2 operating system in 1989 - and extends to the
current day.As of December 2016 the following versions are supported by Microsoft:
 SQL Server 2008
 SQL Server 2008 R2
 SQL Server 2012
 SQL Server 2014
 SQL Server 2016

The current version is Microsoft SQL Server 2016, released June 1, 2016. The RTM
version is 13.0.1601.5.SQL Server 2016 is supported on x64 processors only.

SQL Process

When you are executing an SQL command for any RDBMS, the system determines the
best way to carry out your request and SQL engine figures out how to interpret the task.There are
various components included in the process. These components are Query Dispatcher,
Optimization Engines, Classic Query Engine and SQL Query Engine, etc. Classic query engine
handles all non-SQL queries but SQL query engine won't handle logical files.

Data storage

Data storage is a database, which is a collection of tables with typed columns. SQL
Server supports different data types, including primary types such as Integer, Float, Decimal,
Char (including character strings), Varchar (variable length character strings), binary (for
unstructured blobs of data), Text (for textual data) among others. The rounding of floats to
integers uses either Symmetric Arithmetic Rounding or Symmetric Round Down (fix) depending
on arguments: SELECT Round(2.5, 0) gives 3.

Microsoft SQL Server also allows user-defined composite types (UDTs) to be defined
and used. It also makes server statistics available as virtual tables and views (called Dynamic
Management Views or DMVs). In addition to tables, a database can also contain other objects
including views, stored procedures, indexes and constraints, along with a transaction log. A SQL
Server database can contain a maximum of 231 objects, and can span multiple OS-level files
with a maximum file size of 260 bytes (1 exabyte). The data in the database are stored in primary
data files with an extension .mdf. Secondary data files, identified with a .ndf extension, are used
to allow the data of a single database to be spread across more than one file, and optionally
across more than one file system. Log files are identified with the .ldf extension

Storage space allocated to a database is divided into sequentially numbered pages, each 8
KB in size. A page is the basic unit of I/O for SQL Server operations. A page is marked with a
96-byte header which stores metadata about the page including the page number, page type, free
space on the page and the ID of the object that owns it. Page type defines the data contained in
the page: data stored in the database, index, allocation map which holds information about how
pages are allocated to tables and indexes, change map which holds information about the
changes made to other pages since last backup or logging, or contain large data types such as
image or text.

Buffer management

SQL Server buffers pages in RAM to minimize disk I/O. Any 8 KB page can be buffered
in-memory, and the set of all pages currently buffered is called the buffer cache. The amount of
memory available to SQL Server decides how many pages will be cached in memory. The buffer
cache is managed by the Buffer Manager. Either reading from or writing to any page copies it to
the buffer cache. Subsequent reads or writes are redirected to the in-memory copy, rather than the
on-disc version. The page is updated on the disc by the Buffer Manager only if the in-memory
cache has not been referenced for some time. While writing pages back to disc, asynchronous I/O
is used whereby the I/O operation is done in a background thread so that other operations do not
have to wait for the I/O operation to complete. Each page is written along with its checksum
when it is written.

Concurrency and locking

SQL Server allows multiple clients to use the same database concurrently. As such, it
needs to control concurrent access to shared data, to ensure data integrity-when multiple clients
update the same data, or clients attempt to read data that is in the process of being changed by
another client. SQL Server provides two modes of concurrency control: pessimistic concurrency
and optimistic concurrency. When pessimistic concurrency control is being used, SQL Server
controls concurrent access by using locks. Locks can be either shared or exclusive. Exclusive
lock grants the user exclusive access to the data-no other user can access the data as long as the
lock is held. Shared locks are used when some data is being read-multiple users can read from
data locked with a shared lock, but not acquire an exclusive lock. The latter would have to wait
for all shared locks to be released.

SQLCMD

SQLCMD is a command line application that comes with Microsoft SQL Server, and
exposes the management features of SQL Server. It allows SQL queries to be written and
executed from the command prompt. It can also act as a scripting language to create and run a set
of SQL statements as a script. Such scripts are stored as a .sql file, and are used either for
management of databases or to create the database schema during the deployment of a database.

SQLCMD was introduced with SQL Server 2005 and this continues with SQL Server
2012 and 2014. Its predecessor for earlier versions was OSQL and ISQL, which is functionally
equivalent as it pertains to TSQL execution, and many of the command line parameters are
identical, although SQLCMD adds extra versatility.
CHAPTER-5

SYSTEM DESIGN

SYSTEM ARCHITECTURE
DATAFLOW DIAGRAM
SERVER ROUTER
CLIENT

IP Address IP Address
Connecting..

Socket
Select a file to send Connection

NO
no link
check probe
Via Hidden yes packets
YES
Link yes

Start Finding
nearest path File Recieving
no File Recieving Failed
Send probe packets
Connecting.. to all node
no link
failure?

File Received
YES

File sending to client


failed

File Transfered

File Received

End

CLASS DIAGRAM
SERVER CLIENT
FILENAME RECPATH
IPADDRESS RECSTATUS
FILEBYTES FILEBYTES
LENGTH LENGTH
PORTNUMBER ACK
ACK

SOCKETCON() SOCKETCON()
SEND() RECEIVE()

ROUTER
FILENAME
IPADDRESS
FILEBYTES
LENGTH
PROBENODE
ROUTER
PATH
ACK

PROBE SEND()
SHORTESTPATH()
SEND()

ACTIVITY DIAGRAM
CLIENT SERVER ROUTER

IP Address
Browse IP Address

Client socket
connection
Select a Browse a
Receiving Path File

NO check probe
packet
Connecting.. NO
socket Yes
connection
Find Shortest
Path
Send Probe
FILE RECEIVE Yes packets

NO
Connecting.. link failure

Yes

Connecting..
FILE TRANSFER

TRANSACTION
FAILED

Sending Packets

FILE receive

SEQUENCE DIAGRAM
SERVER ROUTER
CLIENT

Start File Transfer

Socket Connection

Socket Connection
Routing
Sending Probe
Packets
Bytes Transferred

Bytes transfer

Routing Finish File transfer failed

Finding shortest path


File transfered

Acknowledgement

File Received

USE CASE DIAGRAM


Receiving
Path

IP Address

Sending Probe
packets

CLIENT
SERVER
Finding
Shortest Path

ROUTER

socket connection

Browse a File

Receive a File

File Sending

ALGORITHM DESCRIPTION
CHAPTER 6
SYSTEM IMPLEMENTATION

IMPLEMENTATION

Implementation is the stage of the project when the theoretical design is turned out into a
working system. Thus it can be considered to be the most critical stage in achieving a successful
new system and in giving the user, confidence that the new system will work and be effective.

The implementation stage involves careful planning, investigation of the existing system
and it’s constraints on implementation, designing of methods to achieve changeover and
evaluation of changeover methods.

Module Description:
 Data chunk similarity and compression
 Data chunk similarity based compression on cloud
 MapReduce for Big Data Analysis
 Jaccard coefficient

Data chunk similarity and compression:


 The similarity models for our compression and clustering will be developed. The
similarity model is critical and fundamental for deploying the data chunk based data
compression because the similarity model is used for generating the standard data
chunks.
 There are five types of models are commonly used including common element approach,
template models, geometric models, feature models and Geon theory.
 Similarity models work on two types of data sets, multi-dimensional numerical data and
text data.

Data chunk similarity based compression on cloud:


 Big sensing data compression on Cloud, two important stages, standard chunk
generation and chunk based compression are essential. So the algorithms are
developed respectively to conduct the related data processing.
 To guarantee the scalability of the proposed data compression algorithm based on
data chunks, MapReduce programming model and Hadoop platform are adopted for
implementation.

MapReduce for Big Data Analysis:


 There is a growing trend of applications that should handle big data. However,
analyzing big data is a very challenging problem today.
 For such applications, the MapReduce framework has recently attracted a lot of
attention. Google’s MapReduce or its open source equivalent Hadoop is a powerful
tool for building such applications
 Effective management and analysis of largescale data poses an interesting but critical
challenge.
 Recently, big data has attracted a lot of attention from academia, industry as well as
government.
Jaccard Coefficient:
 It is used as an information similarity measure. In addition of the similarity measure
Jaccard.
 It has the advantage which protects the privacy of the data.
 It proposed the SJCM protocol (Secure computation of the Jaccard Coefficient for
Multisets)using the existing dot product method.
 Jaccard similarity algorithm is used to finding thebetween the data chunks and the data
stream
Increase the accuracy of the data. To show that the significant storage saving is achieved due
to compressed data blocks
USER MODULE:

In this module, Users are having authentication and security to access the detail which is
presented in the ontology system. Before accessing or searching the details user should have the
account in that otherwise they should register first.

CLOUD SERVICE:
Here new mobile cloud service through the virtualization of mobile devices in cloud
infrastructure. We describe two main service scenarios to explain how this mobile cloud service
can be used. Service scenarios are useful to discuss security threats on mobile cloud
infrastructure, because they include users, places, mobile devices, and network types, and user’s
interesting contents.We define mobile cloud computing as processing jobs for mobile devices in
cloud computing infrastructure and delivering job results to mobile devices. we propose a new
mobile cloud service as providing virtual mobile instances through mobile cloud computing. The
proposed mobile cloud service provides virtual mobile instances through the combination of a
mobile environment and cloud computing. Virtual mobile instances are available on mobile
devices by accessing the mobile cloud infrastructure. This means that users connect to virtual
mobile instances with their mobile devices and then use computing resources such as CPU,
memory, and network resources on mobile cloud infrastructure. In this case, such mobile devices
will have smaller roles to play than current mobile devices.

MALWARE DATA :
We chose ‘GoldMiner’ malware applications to obtain abnormal data in our mobile cloud
infrastructure. We installed the malware onto two hosts and ran it. It gathers location coordinate
and device identifiers (IMEI and IMSI), and sends the information to its server. The malware
target affecting each mobile instance as zombie, and there are many other malware which have
the same purpose although their functionality and behavior are little different from each other.
This kind of malware is more threatening to mobile cloud infrastructure because there are lots of
similar virtual mobile instances and they are closely connected to each other.
Entered data are not same, compare the database data that is called malwaredata. when If
some abnormal behavior’s help to modify the date in External object.
ABNORMAL BEHAVIOR DETECTION:
We used the Random Forest (RF) machine learning algorithm to train abnormal behavior
with our collected data set. The RF algorithm is a combination of decision trees that each tree
depends on the values of a random vector sampled independently and with the same distribution
for all trees in the forest. We represented the collected features as a vector with the data
subsequently used to train our collected data set.
CHAPTER 7

SOFWARE DESIGNING

The most creative and challenging phase of system lifecycle is the system design. System design
is the process that states the details of how a system will meet the requirements identified during
system analysis. When the analyst prepares logical system design, they specify the user needs at
level of detail that virtually determines information flow into and out of the system and the
required data source. First step in the design is to determine how the output is to be produced and
in what format. Secondly, input data and master files have to be designed to meet the
requirement of proposed output. Finally, the end of the design phase, the system flow chart will
be ready which is used as the base of coding phase.

SYSTEM FLOW CHART

A high level view of the organization of a system can be shown by system flowchart. The
rectangular boxes represent original aggregates of computer software. It is often very useful
diagram for the system designers who must develop overall system architecture of hardware and
software to implement the user requirements. There is one situation where the system flowcharts
could be a useful modeling tool: At the end of the system analyst’s activity, when the user
implementation model is being developed. At this point, the user, the system analyst, and the
implementation team discuss the implementation constraints that must be imposed upon the
system: these include such things as the determination of the automation boundary and the
human interface. Input design is creative as well as technical activity, which includes the
following tasks.

 Approaching the terms of reference.


 Designing the format medium of files.
 Determining methods of data capture and data input.
 Designing forms.

INPUT DESIGN
The input of a system can be defined as the information that is provided to the system. This is
used for future processing by the system to obtain meaningful information, which helps in
decision-making. Input design is the process of converting user-oriented inputs to a computer-
based format. Input is a part of overall system design, which requires special attention.
Inaccurate input data are the most common cause of errors in error processing. Input design can
control errors entered by data entry operators. Entered data have to be checked for their accuracy
and direction of errors. Appropriate error message have to be displayed. When an invalid data is
entered, the user should not be allowed to type it.

OUTPUT DESIGN

The computer output is the most important and direct source of information to the user. Efficient
and intelligible output design improves the system’s relationship with the user and helps in
decision-making. Output design was studied going actively during the study phase. The objective
of the output design is defined the contents and format of all documents and reports in an
attractive and useful format.

Types of Outputs

 External output
 Internal output
 Operational outputs
 Interactive outputs

Output generally refers to the results and information that are generated by the system. It can be
in the form of operational documents and reports. The major form of output is a hard copy from
the printer. Once the output requirements are identified, the output devices used also should be
determined. Factors like compatibility of the output device with the system and response time
requirement should be considered while descending the output device to be utilized.

DEVELOPMENT MODEL
The development phase focuses on how the engineer attempts to develop the system. It also deals
with how data are to be structured, how the function is to be implemented within software
architecture, how procedural details are to be implemented, how interfaces are characterized,
how the design will be translated into a programming language, and how testing will be
performed. The methods applied during the development phase will vary but three specific
technical tasks should always occur.

 Software design
 Code generation
 Software testing

The classical life cycle model or the linear sequential model is a systematic and sequential
approach to software development. It progress through analysis, design, coding, testing and
support.

ANALYSIS

The requirements gathering process is intensified and focused on software. The nature of the
programs to be built, the software engineer must understand the information domain for the
software, as well as required function, behavior, performance and interface. Requirements for
both the system and software are documented and reviewed with the customer.

DESIGN

Software design actually a multistep process that focuses on four attributes of a program: data
structure, software architecture, interface representations and procedural detail. The design
process translates the requirements into a representation of the software that can be assessed for
the quality before coding begins.

CODE GENERATION

The design must be translated into a machine-readable form. The code generation step performs
this task. If the design is performed in a detailed manner, code generation can be accomplished
mechanistically. The requirement was converted into code.

TESTING
Once code has been generated, program testing begins. The testing process focuses on the logical
internals of the software, ensuring that all statements have been tested, and on the functional
CHAPTER 8

SYSTEM TESTING

The purpose of testing is to discover errors. Testing is the process of trying to discover
every conceivable fault or weakness in a work product. It provides a way to check the
functionality of components, sub assemblies, assemblies and/or a finished product. It is the
process of exercising software with the intent of ensuring that the Software system meets its
requirements and user expectations and does not fail in an unacceptable manner. There are
various types of test. Each test type addresses a specific testing requirement.

TYPES OF TESTING

UNIT TESTING

Unit testing involves the design of test cases that validate that the internal program logic
is functioning properly, and that program input produce valid outputs. All decision branches and
internal code flow should be validated. It is the testing of individual software units of the
application .it is done after the completion of an individual unit before integration. This is a
structural testing, that relies on knowledge of its construction and is invasive. Unit tests perform
basic tests at component level and test a specific business process, application, and/or system
configuration. Unit tests ensure that each unique path of a business process performs accurately
to the documented specifications and contains clearly defined inputs and expected results.

Unit testing is usually conducted as part of a combined code and unit test phase of the
software lifecycle, although it is not uncommon for coding and unit testing to be conducted as
two distinct phases.

Test Strategy and approach:

Field testing will be performed manually and functional tests will be written in detail.
Test objectives:

 All field entries must work properly.


 Pages must be activated from the identified link.
 The entry screen, messages and responses must not be delayed.
 Features to be tested

INTEGRATION TESTING

Integration tests are designed to test integrated software components to determine if they
actually run as one program. Testing is event driven and is more concerned with the basic
outcome of screens or fields. Integration tests demonstrate that although the components were
individually satisfaction, as shown by successfully unit testing, the combination of components is
correct and consistent. Integration testing is specifically aimed at exposing the problems that
arise from the combination of components.

Software integration testing is the incremental integration testing of two or more


integrated software components on a single platform to produce failures caused by interface
defects.

The task of the integration test is to check that components or software applications, e.g.
components in a software system or – one step up – software applications at the company level –
interact without error.

FUNCTIONAL TESTING

Functional tests provide systematic demonstrations that functions tested are available as
specified by the business and technical requirements, system documentation and user manuals.
Functional testing is centered on the following items:

Valid Input : identified classes of valid input must be


accepted.

Invalid Input : identified classes of invalid input must be


rejected.

Functions : identified functions must be exercised.

Output : identified classes of application outputs


must be exercised.

Systems/Procedures : interfacing systems or procedures must be


invoked.

Organization and preparation of functional tests is focused on requirements, key functions, or


special test cases. In addition, systematic coverage pertaining to identify

Business process flows; data fields, predefined processes, and successive processes must
be considered for testing. Before functional testing is complete, additional tests are identified and
the effective value of current tests is determined.

SYSTEM TESTING

System testing ensures that the entire integrated software system meets requirements. It
tests a configuration to ensure known and predictable results. An example of system testing is the
configuration oriented system integration test. System testing is based on process descriptions
and flows, emphasizing pre-driven process links and integration points.

WHITE BOX TESTING


White Box Testing is a testing in which in which the software tester has knowledge of the
inner workings, structure and language of the software, or at least its purpose. It is purpose. It is
used to test areas that cannot be reached from a black box level.

BLACK BOX TESTING

Black Box Testing is testing the software without any knowledge of the inner workings,
structure or language of the module being tested . Black box tests, as most other kinds of tests,
must be written from a definitive source document, such as specification or requirements
document, such as specification or requirements document. It is a testing in which the software
under test is treated, as a black box .you cannot “see” into it. The test provides inputs and
responds to outputs without considering how the software works.

TESTING

User Acceptance Testing is a critical phase of any project and requires significant
participation by the end user. It also ensures that the system meets the functional requirements.

TEST RESULTS

All the test cases mentioned above passed successfully. No defects encountered.
CHAPTER 8

CONCLUSION

To present a new mobile cloud service with the virtualization of mobile devices and
discussed some possible scenarios for individual users and office workers. To address security
issues in mobile cloud infrastructure, we proposed abnormal behavior monitoring methodology
and architecture to detect malware. These were then tested by deploying our mobile cloud test
bed. Host and network data are used together to detect abnormal behavior. Our abnormal
behavior detection using the RF machine learning algorithm shows that our proposed
methodology and architecture successfully detect abnormal behavior.

FUTURE WORK

For future work, we will investigate on the service feasibility of this new mobile cloud
service. In addition to the monitoring of mobile cloud infrastructure focusing on security issues,
other monitoring metrics should be considered for the provisioning and configuration, of
services, and for the charging of users. We will also measure the performance of our proposed
monitoring architecture. To deal with security aspects on this service, we will gather various
additional types of sample malware for training in order to improve the accuracy of using various
machine learning algorithms. Further, we will consider other monitoring features to improve the
accuracy of detecting abnormal behavior. But there is an overhead issue such as time complexity
and battery consumption if we gather lots of features. So we should also consider this aspect
together.

CHAPTER 9
BIBLIOGRAPHY

Good Teachers are worth more than thousand books, we have them in Our Department

References Made From:

1. Birget, J.C., D. Hong, and N. Memon. Graphical Passwords Based on Robust


Discretization. IEEE Trans. Info. Forensics and Security, 1(3), September 2006.
2. Blonder, G.E. Graphical Passwords. United States Patent 5,559,961, 1996.
3. Chiasson, S., R. Biddle, R., and P.C. van Oorschot. A Second Look at the Usability of
Click-based Graphical Passwords. ACM SOUPS, 2007.
4. Cranor, L.F., S. Garfinkel. Security and Usability. O’Reilly Media, 2005.
5. Davis, D., F. Monrose, and M.K. Reiter. On User Choice in Graphical Password
Schemes. 13th USENIX Security Symposium, 2004.
6. R. N. Shepard, "Recognition memory for words, sentences, and pictures," Journal of
Verbal Learning and Verbal Behavior, vol. 6, pp. 156-163, 1967.
7. A. Perrig and D. Song, "Hash Visualization: A New Technique to Improve Real-World
Security," in Proceedings of the 1999 International Workshop on Cryptographic

Techniques and E-Commerce, 1999.

SAMPLE SCREENS AND SAMPLE CODE


CHAPTER 10
BOOKS & WEBSITES

Books:

• Pro ASP.NET 3.5 in C# 2008 (Apress)

• Unleashed ASP.NET 2.0 (SAMS Publication)

• ASP.NET Bible

• The Complete Reference Asp.net

• Professional VB.NET 3.5 with Visual Basic (Wrox)

Sites Referred:

http://www.sourcefordgde.com

http://www.networkcomputing.com/

http://www.ieee.org

http://www.almaden.ibm.com/software/quest/Resources/

http://www.computer.org/publications/dlib

http://www.ceur-ws.org/Vol-90/

You might also like