collage report on online trading
Post on 22-May-2015
1.094 Views
Preview:
DESCRIPTION
TRANSCRIPT
INTRODUCTION
1. INTRODUCTION
The internet’s arrival and its subsequent popularity in India have made online trading inIndia, which is about the online purchase and sales of shares, one of the extremely popular
means of trading. Both beginner and experienced traders and investors in India are milking this
opportunity by trading online in futures and options, stocks and currencies worldwide. Such
opportunities are in the form of reduced brokerage and commissions, better broking services
etc.
Constructive uses of new technologies have always contributed positively towardsimproving human life standards and the economy of a country .Such as online trading, in equity
markets it increased trade volumes and number of investors trading in stock markets. Online
trading was started in India in the year 1995, where a new system is formed which allows the
investor to trade through an internet site where the major financial products and services like
equities, mutual funds, life insurance, general insurance, loans, share trading, commodities
trading, portfolio management and financial planning etc. are directly available for the
customer.
There is also stock option trading in India. An option is a financial agreement, with apredetermined maturity period and price, for the purchase or sales of the underlying products.
Stock options enable the protection of dealers and control of their stocks, in addition to
generation of higher earnings.
For carrying out online trading in India, you have to open an online demat and tradingaccount, followed with online trading software. For this purpose, you would require a
Depository Participant (DP), selection of which should be preceded by extensive research on
various determinants.
This is time consuming and inefficient. This imposed limits on trading volumes andefficiency. In order to provide efficiency, liquidity, and transparency NSE and BSE introduced
nationwide online fully automated “SCREEN BASED TRADING SYSTEM”.
Online trading in India is the internet based investment activity that involves no directinvolvement of the broker. There are many leading online trading portals in India along with the
online trading platforms of the biggest stock houses like the National stock exchange and the
Bombay stock exchange. The total portion of online share trading India has been found to have
grown from just 3 per cent of the total turnover in 2003-04 to 16 per cent in 2006-07.
1
The investor has to register with an online trading portal and get into an agreement with
the firm to trade in different securities following the terms and conditions listed down on the
agreement. The order processing is done in correct timings as the servers of the online trading
portal are connected to the stock exchanges and designated banks all around the clock. They
can also get updates on the trading and check the current status of their orders either through
e-mail or through the interface. Brokerages also provide research content on their websites,
such that the clients can take their own decisions on stocks before investing.
In spite of many private stock houses at present involved in online trading in India, theNSE (National Stock Exchange) and BSE (Bombay Stock Exchange) are among the largest
exchanges. They handle huge daily trading volumes, supporting large amounts of data traffic
and processing a countrywide network. The automated online systems used for trading by the
national stock exchange and the Bombay stock exchange are the NIBIS or NSE’s Internet based
System and NEAT for the national stock exchange and the BSE online trading system or BOLT for
the Bombay stock exchange.
Thanks to the ever-rising number of people owning computers along with a readilyavailable internet access, online stock trading in India is simplified manifolds. This is because
investments can now be easily controlled by traders themselves as a result of extensive
availability of all types of information on the web.
2
HISTORY
2. HISTORY
As early as 1969, digital trading systems called electronic communications networks (ECNs) were being used by brokerages to display in-house the bid and ask prices for stocks. By the late 1980s, as the financial industry began to realize the potential of a public internet and ownership of personal computers was growing in popularity, some of the leading brokerages began to look more closely at ECNs. Several industrious brokerages either developed software or bought companies that had developed software to link up stock traders with current stock price information, thereby matching up buyers with sellers easily and efficiently and with significant cost savings.
It was 1981 at my local Radio Shack. But just one year later in 1982, the first full service electronic consumer equity trading system for buying and selling stocks, mutual funds and commodities using a PC came online. It was called NAICO-NET and it was offered by a company called North American Holding Corp. located in East Hartford, CT. The system was ANSI based, meaning it was terminal based but IBM PCs could connect to it via a simple application and it connected traders from all over the world. Trades were sent directly to Pershing Corp. (Donaldson Lufkin & Jenrette) for high speed clearing and the trader subscription base quickly rose to over 5,000. And the rest, as they say, is history – the history of online stock trading.
Over the next 10 years, the history of online stock trading developed slowly, temporarily crippled by the stock market crash of 1987, a recession and the Gulf War. Additionally, the costs were still generally high for online access for retail stock traders. It was in the 1990s that things shifted and online stock trading really took off.
Trade*Plus was another pioneering company in the history of online stock trading and in 1985 offered some of the very first retail trading platforms on America Online and CompuServe. Trade*Plus continued to offer its services to brokerages, but in 1991 one of the founders of Trade*Plus, William Porter, created a new subsidiary company called E*TRADE Securities, Inc.
For retail traders, the coming together of better accessibility to information, the ease of an electronic trade and decreased cost to process a trade enhanced the popularity of online stock trading and it began to take off. By the mid-1990s, more than 20 percent of the nation’s population was investing in stock, compared with less than 5 percent the decade before. To illustrate the impact this had on E*Trades’ bottom line, its revenues increased from $850,000 in 1992 to $11 million in 1994 and was rated as the fastest growing company in America that year. The rush to become an online stock trading broker was on!
3
Another well-known company, TD Ameritrade, originally started out as a company called First Omaha Securities and in 1995 the company acquired K. Aufhauser & Company, Inc. and its WealthWeb, which had begun offering online stock trading in August 1994. The company that was to eventually become TD Ameritrade continued to merge and acquire other companies, quickly becoming one of the largest online brokerages.
A conversation about the history of online stock trading must include the fact that Charles Schwab, for a long time, had been cornering a huge portion of the financial industry’s market share through branch offices and discount fees. They continued to build market share at a strong pace all through the 1990s. Charles Schwab was only slightly late to the online stock trading party and became another strong leader as an online stock trading provider in the late 1990s, combining their market share with strong promotional activity. They started by providing a web presence in 1995 and at the time had $181.7 billion in total client assets. The very next year their online stock trading went live and in 1996, they ended the year with $253 billion in total client assets. By the next year, 1997, Charles Schwab registered it’s one millionth online account, reached total client assets of $437 billion and ranked as the top online broker in the U.S.
SYSTEM ANALYSIS
3. SYSTEM ANALYSIS
3.1 DEFINATION
System Analysis is the detailed study of the various operations performed by the system
and their relationships within and outside the system. Analysis is the process of breaking
something into its parts so that the whole may be understood. System analysis is concerned with
becoming aware of the problem, identifying the relevant and most decisional variables, analyzing
and synthesizing the various factors and determining an optimal or at least a satisfactory
solution. During this a problem is identified, alternate system solutions are studied and
recommendations are made about committing the resources used to design the system.
3.2 DESCRIPTION OF PRESENT SYSTEM
Existing system refers to the system that is being followed till now. Presently all the
registrations are done manually. If a person wants to buy computer and accessories, domestic
used things, mobile and accessories and many different goods etc., he should directly contact to
the various websites such as flip kart, Naaptol, Snap deal etc. which provides the online trading
facilities. The main disadvantage is that there will be lot of difficulties for the citizens. So, all
these procedures will be a time consuming one.
3.3 LIMITATIONS OF PRESENT SYSTEM
4
Difficult for persons.
Time consuming.
3.4 PROPOSED SYSTEM
Online Trading is aimed at developing a web-based system. In this system the person can
sales online and do many things. The details of all the things are made available to them through
the website.
ADVANTAGES:
This website provides online help for legal queries.
This website helps all the users to view the registration.
The user can post thread in the forum.
The system is user friendly.
3.5 FEASIBILITY STUDY
A feasibility analysis usually involves a thorough assessment of the operational (need),
financial and technical aspects of a proposal. Feasibility study is the test of the system proposal
made to identify whether the user needs may be satisfied using the current software and
hardware technologies, whether the system will be cost effective from a business point of view
and whether it can be developed with the given budgetary constraints. A feasibility study should
be relatively cheap and done at the earliest possible time. Depending on the study, the decision is
made whether to go ahead with a more detailed analysis.
5
When a new project is proposed, it normally goes through feasibility assessment.
Feasibility study is carried out to determine whether the proposed system is possible to develop
with available resources and what should be the cost consideration. Facts considered in the
feasibility analysis were.
Technical Feasibility
Economic Feasibility
Behavioral Feasibility
3.5.1 Technical Feasibility
Technical Feasibility deals with the hardware as well as software requirements.
Technology is not a constraint to type system development. We have to find out whether the
necessary technology, the proposed equipment have the capacity to hold the data, which is used
in the project, should be checked to carry out this technical feasibility.
The technical feasibility issues usually raised during the feasibility stage of investigation
includes these
This software is running in windows 7 Operating System, which can be easily installed.
The hardware required is Pentium based server.
The system can be expanded.
3.5.2 Economical Feasibility
This feasibility study present tangible and intangible benefits from the prefect by
comparing the development and operational cost. The technique of cost benefit analysis is often
used as a basis for assessing economic feasibility. This system needs some more initial
investment than the existing system, but it can be justifiable that it will improve quality of
service.
Thus feasibility study should center along the following points:
Improvement resulting over the existing method in terms of accuracy, timeliness.
Cost comparison
Estimate on the life expectancy of the hardware
Overall objective
6
Our project is economically feasible. It does not require much cost to be involved in the overall
process. The overall objectives are in easing out the requirement processes.
3.5.3 Behavioral/ Operational Feasibility
This analysis involves how it will work when it is installed and the assessment of
political and managerial environment in which it is implemented. People are inherently resistant
to change and computers have been known to facilitate change. The new proposed system is very
much useful to the useful to the users and there for it will accept broad audience from around the
world.
3.6 PROJRCT PLANNING & SCHEDULING
7
Waterfall Model
The waterfall model derives its name due to the cascading effect from one phase to theother as is illustrated in above figure. In this model each phase well define starting and endingpoint, with identifiable deliveries to the next phase. Note that this model is sometime referred toas the linear sequential model or the software life cycle model. The water fall diagram isbasically divided into following 5 models.
Requirement Design Implementation Verification Maintenance
Requirement:-
In the requirement phase the need to create the application is specified. What isthe need of the system is defined. What information to be feeder to create the applicationwill come under the requirement phase?
Design:
After the requirement phase the next phase is the Design phase where theapplication is designed according to the forms and other modules created. This phase ismuch important phase because it will structure the layout of your application.
Implementation:
Implementation is the process of having a system personnel phase check out andput new equipment into use, train users, install new application and construct any file ofdata need to use it.
Verification:
8
After the whole application is being the developed the main phase is theverification phase where the whole application tested and verified to check the wholeapplication.
Maintenance:
After the successful verification of the application the main phase is themaintenance phase where the application needs to be maintained for its successfuloperation in future.
SYSTEM SPECIFICATION
4. SYSTEM SPECIFICATIONS
4.1 HARDWARE DESCRIPTION
The selection of hardware is very important in the existence and proper working of any
software. When selecting hardware, the size and requirements are also important.
Minimum Requirements:
Processor : Pentium II class, 450MHz
RAM : 128MB
Hard Disk Drive : 3GB
Video : 800X600, 256 colors
CD-ROM : Required
The proposed System is developed on:
Processor : INTEL Pentium 4
RAM : 512MB
Hard Disk Drive : 40GB
Key Board : Standard 101/102 or Digi Sync Family
Monitor : Display Panel (1024 X 764)
Display Adapter : Trident Super VGA
9
Network Adapter : SMC Ethernet Card Elite 16 Ultra
Mouse : Logitech Serial Mouse
4.2 SOFTWARE DESCRIPTION
Operating System : Windows XP
Front- End : C#. NET with ASP. NET
Back- End : MS SQL SERVER 2005 EXPRESS
project will be done in ASP. NET with C# 2.0 as front end and SQL Server 2005 as back
end. Microsoft .NET is software that connects information, people, systems and devices. It spans
clients, servers and developer tools and consists of:
The .NET Framework programming model that enables developers to build Web-
based applications which expose their functionality programmatically over a
network using standard protocols such as SOAP and HTTP.
Developer tools such as Microsoft Visual Studio .NET, which provide a rapid
application integrated development environment for programming with the .NET
Framework.
A set of servers including Microsoft Windows 2000, Microsoft SQL, Server and
Microsoft BizTalk Server that integrates, runs, operates and manages XML Web
services and applications.
Client software such as Windows XP, Windows CE and Microsoft Office XP that
helps developers deliver a deep and compelling user experience across a family of
devices and existing products.
The .NET Framework is the programming model of the .NET environment for building,
deploying and running Web- based applications, smart client applications and XML Web
services. It manages much of the plumbing, enabling developers to focus on writing the business
logic code for their applications. The .NET Framework includes the common language runtime
and class libraries.
10
ORGANIZATIONAL PROFILE
5. ORGANIZATIONAL PROFILE
KOMOLINE established in 1990 is a subsidiary of Astra Microwave Products Limited. It
has in-house design, development, test and manufacturing facility of hi-reliability automated
weather monitoring apparatus, Satellite Communication Equipment and Satellite Payload
subsystems.
AWS product range comprises precision Sensors, Data-loggers, Transmitters, Software,
including integrated in-situ measurement equipment and solutions. Complete system applications
cater to parametric sensing requirements of Weather forecasting, Hydrometeorology, Agro-
meteorology, Maritime, Aviation meteorology and oceanography.
Remote Automated Weather Stations, Agro-met Towers, MBLM-met Towers, 50meters
FLUX-met Towers, Tide Gauge Systems, High Speed data acquisition systems are the core
expertise with over a decade experience.
Satellite Communication Equipment comprises UHF, L-Band and S-Band portable
Voice-Data-Video Modems. Applications of these systems are Disaster management, Data-
buoys, Tsunami Early warning system, Border and Coastal Security services.
Business commitment with State-of-the-art microelectronics technology is the key to
successful implementation of these wide range projects.
SPECIFICATION OF LANGUAGE
6. OVERVIEW OF THE LANGUAGE USED
6.1 MICROSOFT VISUAL STUDIO
Microsoft Visual Studio is Microsoft’s flagship software development product for
computer programmers. It centers on an integrated distribution environment which has
11
programmers create stand alone, and web services that run on any platforms supported by
Microsoft’s .Net Framework (for all versions after 6). Supported platforms include Microsoft
windows, servers and workstations, Pocket PC, Smart Phones and World Wide Web browsers
not the Java Virtual Machine that all other java tools target.
VISUAL STUDIO 2005
The most important language added in this version was the introduction of generics,
which are similar in many aspects to C++ templates. This potentially increases the number of
bugs caught at compile- time instead of run- time by encouraging the use of strict type checking
in areas where it was not possible before, C++ also got a similar upgrade with the addition of
C++/CLI which is slated to replace the use of Managed C++. Other new features of Visual
Studio 2005 include the “Development Designer” which allows application designs to be
validated before deployments, an improved environment for web publishing when combined
with ASP.NET 2.0 and load testing to see application performance under various sorts of user
loads. Visual Studio 2005 also added extensive 64- bit support. While the development
environment itself only available as a 32- bit application, visual C++ 2005 supports compiling
for (x64AMD64 and EM64T) as well as IA- 64 (Itanium). The platforms SDK include 64- bit
and 64-bit versions of the libraries.
6.2 NET FRAMEWORK SDK
The .NET framework is an integral windows component that supports building and
running the next generation of applications and XML web services. The key component of the
.NET frame work are the common language run time and the. NET frame work class library,
which includes ADO.NET, ASP.NET and windows forms. The .NET framework provides a
wide variety of programming languages.
This framework is made up of the following parts:
12
The common language runtime(CLR)
The base class libraries.
Object oriented internet development with ASP.NET
Rich client user interface using windows forms
RAD for the internet using web forms
6.3 OVERVIEW OF THE .NET FRAME WORK
The .NET framework is a new computing platform that simplifies application
development in the highly distributed environment of the internet. The .NET framework is
designed to fulfill following objectives:
To provide a consistent object oriented programming environment whether object code
is stored and executed locally but internet- distributed or executed remotely.
To provide a code execution environment that minimizes software deployment and
versioning conflicts.
To provide a code execution environment that guarantees safe execution of code,
including code created by an unknown or semi trusted third party.
To provide a code execution environment that eliminates the performance problem of
scripted or interpreted environments.
To make the developer experience consistent across widely types of application, such as
windows based applications and web based applications.
To build all communication on industry standards to ensure that code based on the .NET
framework can integrate with any other code.
The .NET framework has two main components: the common language runtime and
the .Net framework class library. The common language runtime is the foundation of the .NET
framework. You can think of the runtime as an agent that manages code at execution time, and
remoting while also enforcing strict type safely and other forms of code accuracy that ensure
13
security and robustness in fact the concept of code management is a fundamental principle of the
runtime.
Code that targets the runtime is known as managed code, while code that does not
target the runtime is known as unmanaged code. The class library, the other main component of
the .NET frameworks is a comprehensive, object-oriented collection reusable types that you can
use to develop applications ranging from traditional command line or graphical user interface
(FGUI) applications to application base d on the latest innovations provided by ASP.NET, such
as web forms and XML web services.
The .NET framework can be hosted by unmanaged component that load the common
language runtime into their processes and initiate the execution of managed code. ASP.NET
works directly with the runtime to enable ASP.NET application and XML web services, both of
which are discussed later in this topic, Internet explorer is an example of unmanaged application
that hosts the runtime (in the form of a MIME type extension). Using internet explorer to the
host runtime enables you to embed managed components or windows forms controls in HTML
documents. Hosting the runtime in this way makes mobile code 9similar to Microsoft Active Xr
controls) possible, but with significant improvement that only managed code can offer, such as
semi-trusted execution and secure isolated file storage.
The following illustration shows the relationship of the common language runtime and
the class library to your application and to the overall system. The illustration also shows how
managed code operated with in a larger architecture.
We can use the .NET framework to develop the following types of application and services:
Console applications
Window GUI application (Windows Forms) ASP.NET applications
XML Web services
Windows services
14
6.4 COMMON LANGUAGE RUNTIME (CLR)
The common language runtime (CLR) is responsible for runt-time services such as
language integration; security enforcement; and memory, process and thread management. In
addition, it has a roll at development time when features such as life cycle management strong
type naming, cross-language exception handling, dynamic binding and so on, reduce the amount
of code that a developer must write to turn the business logic the reusable component. The
runtime can be hosted by high performance, server-side applications, such as Microsoft Internet
Information Services (IIS) for building web applications with ASP.NE and the next release of
Microsoft SQL Server. This infrastructure enables you to use code “managed “ by the .NET
framework to write your business logic, while still enjoying the superior performance of the
industry’s best enterprises servers that support runtime hosting.
6.5 ASP.NET
ASP.NET is a set of Microsoft.NET framework technologies used for building web
applications and XML Web services. ASP.NET page execute on the server and generate markup
such as HTML, WML or XML that is sent to a desktop or mobile browser. ASP.NET pages use
a compiled, event-driven programming model that improves performance and enables the
separation of application logic and user interface. Both ASP.NET pages and ASP.NET web
services files contain server-side (as opposed to client side logic) written in Visual basic .NET,
C#.NET or any .NET compatible language, Web applications and XML Web Services take
advantage of the features of the common language runtime, such as type safety, inheritance,
language, interoperability, versioning, and integrated security.
6.6 IIS
Microsoft Internet Information S4ervices (IIS; formerly called Server is a set of
Internet-based services for servers using Microsoft Windows. It is the world’s second most
popular web server in terms of overall websites. As of September 2007 it served 34.94% of all
websites and 36.63% of all active websites according to Net craft. The servers currently include
FTP, SMTP, NNTOP, and HTTP/HTTPS.
6.7 ADO.NET
15
ADO.NET provides consistent access to data sources such as Microsoft SQL Server and
XML, as well as to data sources exposed through OLE DB and ODBC. Data sharing consumer
applications can use ADO.NET to connect to these data sources and retrieve, manipulate and
update the data that they contain.
ADO.NET separates data access from data manipulation into discrete components that
can be used separately or in tandem. ADO.NET includes .NET Frame work data providers for
connecting to a database, executing commands and retrieving results. Those results are either
processed directly, placed in and ADO.NET Dataset objects in order to be exposed to the used in
an ad hoc manner, combined with data from multiple sources or remoted between tiers. The
ADO.NET Dataset object can also be used independently of a .NET Framework data provider to
manage data local to the application or sourced from XML.
The ADO.NET classes are found in System.Data.dll and are integrated with the XML
classes found in System.Xml.dll. When compiling code that uses the System. Data, namespace
reference both System.Data.dll and System.Xml.dll.
ADO.NET provided functionality to developers writing managed code similar to the
functionality provided to native component object model (COM) developers by ActiveX Data
Objects (ADO).
6.8 ADO.NET COMPONENTS
There are two components of ADO.NET that you can use to access and manipulate
data:
.NET Framework data providers.
The Dataset
.NET FRAMEWORK DATA PROVIDERS
The .NET Framework Data providers are components that have been explicitly
16
designed for data manipulation and fast, forward-only, read-only access to data. The connection
object provides connectivity to a data source. The command object enables access to database
commands to return data, modify data, run stored procedures and send or retrieve parameter
information. The Data Adapter provides a high-performance stream of data from the data
source. Finally, the Data Adapter provides the bridge between the Data Set object and the data
source. The Data Adapter uses command object to execute SQL commands at the data source to
both load the Data Set with data and reconcile changes made to the data in the Data Set back to
the data source.
THE DATASET
The ADO.NET Data Set is explicitly designed for data access independent of any data
source. As a result, it can be used with multiple and differing data sources used with XML data
or used to manage data local to the application. The Data Set contains a collection n of one or
more Data Table objects made up to rows and columns of data as well as primary key, foreign
key, constraint and relation information about the data in the Data Table objects.
6.9 BENEFITS OF ADO.NET
ADO.NET offers several advantages over previous versions of ADO and over other
data access components. These benefits fall into the following categories:
1. Interoperability
2. Maintainability
3. Programmability
4. Salability
6.10 MICROSOFT DATA ACCSS COMPONENTS (MDAC)
Microsoft Data Access Components (MDAC) is a collection of core files provided to
help applications by providing a means of accessing data. MDAC includes core files for Open
Database Connectivity (ODBC), ActiveX Data Objects (ADO), OLEDB, Network libraries and
client configuration tool for SQL Server. Depending on your data access strategy, you may need
to ensure that MDAC is installed on the client computers, the business servers, the Web servers
or the database servers. MDAC 2.6 or later is required by the .NET Framework and at least
17
MDAC 2.7 SP1 is recommended.
SYSTEM DESIGN
7. SYSTEM DESIGN
DEFINATION
The most creative and challenging face of the system development is System Design. It
provides the understanding and procedural details necessary for implementing the system
recommended in the feasibility study. Design goes through the logical and physical stages of
development.
In designing a new system, the system analyst must have a clear understanding of the
objectives, which the design is aiming to fulfill. The first step is to determine how the output is to
be produced and in what format. Second, input data and master files have to be designed to meet
the requirements of the proposed output. The operational phases are handled through program
construction and testing.
Design of a system can be defined as a process of applying various techniques and
principles for the purpose of defining a device, a process or a system in sufficient detail to permit
its physical realization. Thus system design is a solution to “how to” approach to the creation of
a new system. Thus important phase provides the understanding and the procedural details
necessary for implementing the system recommended in the feasibility study. The design step
provides a data design, architectural design, and a procedural design.
7.1 OUTPUT DESIGN
In the output design, the emphasis is on producing a hard copy of the information
requested or displaying the output on the CRT screen in a predetermined format. Two of the
18
most output media today are printers and the screen. Most users now access their reports from a
hard copy or screen display. Computer’s output is the most important and direct source of
information to the user, efficient, logical, output design should improve the systems relations
with the user and help in decision-making.
As the outputs are the most important source of information to the user, better design
should improve the system’s relation and also should help in decision-making. The output
device’s capability, print capability, print capability, response time requirements etc. should also
be considered form design elaborates the way output is presented and layout available for
capturing information. It’s very helpful to produce the clear, accurate and speedy information for
end users.
7.2 INPUT DESIGN
In the input design, user-oriented inputs are converted into a computer based system
format. It also includes determining the record media, method of input, speed of capture and
entry on to the screen. Online data entry accepts commands and data through a keyboard. The
major approach to input design is the menu and the prompt design. In each alternative, the user’s
options are predefined. The data flow diagram indicates logical data flow, data stores, source and
destination. Input data are collected and organized into a group of similar data. Once identified
input media are selected for processing.
In this software, importance is given to develop Graphical User Interface (GUI), which is
an important factor in developing efficient and user-friendly software. For inputting user data,
attractive forms are designed. User can also select desired options from the menu, which
provides all possible facilities.
Also the important input format is designed in such a way that accidental errors are
avoided. The user has to input only just the minimum data required, which also helps in avoiding
the errors that the users may make. Accurate designing of the input format is very important in
developing efficient software. The goal or input design is to make entry as easy, logical and free
from errors.
7.3 LOGICAL DESIGN
Logical data design is about the logically implied data. Each and every data in the form
can be designed in such a manner to understand the meaning. Logical data designing should give
a clear understanding and idea about the related data used to construct a form.
19
7.4 DATA FLOW DIAGRAM
A Data Flow Diagram (DFD) is a diagram that describes the flow of data and the
processes that change data throughout a system. It’s a structured analysis and design tool that can
be used for flowcharting in place of or in association with information. Oriented and process
oriented system flowcharts. When analysts prepare the Data Flow Diagram, they specify the user
needs at a level of detail that virtually determines the information flow into and out of the system
and the required data resources. This network is constructed by using a set of symbols that do not
imply physical implementations. The Data Flow Diagram reviews the current physical system,
prepares input and output specification, specifies the implementation plan etc.
Four basic symbols are used to construct data flow diagrams. They are symbols that represent data source, data flows, and data transformations and data storage. The points at which data are transformed are represented by enclosed figures, usually circles, which are called nodes.
DATA FLOW DIAGRAM SYMBOLS:-
Source or Destination of data
- Data Flow
- Process
20
- Storage
7.5 DATA DIRECTORY
Table Name: Registration
Table Name: Trading
Table Name: Online Trading
Field Name Data Type SizeProduct Varchar 50Quality Varchar 50Amount Varchar 50
21
Field Name Data Type Size
Email ID Varchar 50
User Name Varchar 50
Password Varchar 50
Name Varchar 50
Address Varchar 50
Mobile No Varchar 50
Phone No Varchar 50
Designation Varchar 50
Bank Name Date/Time 50
Account No Varchar 50
Field Name Data Type SizeProduct Varchar 50Quality Varchar 50Quantity Varchar 10Bank Name Varchar 50Account No Varchar 50Amount Varchar 50
Table Name: Online Price
Table Name: Project
Table Name: Annual Report
Field Name Data Type SizeYear Varchar 50Product Varchar 50Quality Varchar 100Quantity Varchar 50Total Amount Varchar 50Income Tax Varchar 50
Table Name: Feedback
22
Field Name Data Type SizeDate Date/Time 50Product Varchar 50Total Sales Varchar 50Amount Varchar 100Quality Varchar 50
Field Name Data Type SizeProduct Varchar 50Down Rate Varchar 50High Rate Varchar 50Today Rate Varchar 50
Field Name Data Type SizeName Varchar 50Address Varchar 50Email Id Varchar 100
Quantity Varchar50
Total Amount Varchar 50
Income Tax Varchar 50
Table Name: Login
7.6 ER – Diagram
23
Field Name Data Type SizeUsername Varchar 50Password Varchar 50
7.7 Process Model
Spiral Model
DEFINITION
The spiral model, also known as the spiral lifecycle model, is a systemsdevelopment method (SDM) used in information technology (IT). This model ofdevelopment combines the features of the prototyping model and the waterfall model .The spiral model is intended for large, expensive, and complicated projects.
24
The steps in the spiral model can be generalized as follows:
1. A preliminary design is created for the new system.
2. A first prototype of the new system is constructed from the preliminary design. Thisis usually a scaled-down system, and represents an approximation of thecharacteristics of the final product.
3. A second prototype is evolved by a fourfold procedure:(I) evaluating the first prototype in terms of its strengths, weaknesses and
risks;(ii) Defining the requirements of the second prototype;(iii) Planning and designing the second prototype;(iv) Constructing and testing the second prototype.
4. At the customer's option, the entire project can be aborted if the risk is deemed toogreat. Risk factors might involve development cost overruns, operating-costmiscalculation, or any other factor that could, in the customer's judgment, result in aless-than-satisfactory final product.
5. The existing prototype is evaluated in the same manner as was the previousprototype, and, if necessary, another prototype is developed from it according to thefourfold procedure outlined above.
6. The preceding steps are iterated until the customer is satisfied that the refinedprototype represents the final product desired.
7. The final system is constructed, based on the refined prototype.
8. The final system is thoroughly evaluated and tested. Routine maintenance is carriedout on a continuing basis to prevent large-scale failures and to minimize downtime.
Advantages Estimates (i.e. budget, schedule, etc.) get more realistic as work progresses,
because important issues are discovered earlier. It is more able to cope with the (nearly inevitable) changes that software
development generally entails. Software engineers (who can get restless with protracted design processes) can
get their hands in and start working on a project earlier.
25
CODING
8. SOURCE CODE
Home.apsx.cs
using System;using System.Data;using System.Configuration;using System.Web;using System.Web.Security;using System.Web.UI;using System.Web.UI.WebControls;using System.Web.UI.WebControls.WebParts;using System.Web.UI.HtmlControls;using System.Data.SqlClient;
public partial class Home : System.Web.UI.Page{
SqlConnection con;SqlDataReader dr;
SqlCommand cmd;string conn;protected void Page_Load(object sender, EventArgs e){
display();
//con = new SqlConnection("server=SUTHAHAR-F92F17\\TEST;database=sample;user id=sa;pwd=sukitha");
}public void display(){
conn = ConfigurationSettings.AppSettings["ConnectionString"];con = new SqlConnection(conn);
}
//USER LOGIN
protected void ImageButton1_Click(object sender, ImageClickEventArgs e){
if ((TextBox1.Text == "") &&(TextBox2.Text == "")){
Label1.Text = "Enter Username and Password to Login";}else
{
26
cmd = new SqlCommand("select * from login", con);con.Open();
dr = cmd.ExecuteReader();while (dr.Read ()){
TextBox1.Text =dr["username"].ToString();//{
// Session["username"] = TextBox1.Text;// Server.Transfer("media selection.aspx");
//TextBox1.Text = "";//TextBox2.Text = "";
//}
// else//{
//// Label1.Text = "Invalid Username and Password";
// }
}}
}}
Adminlogin.aspx.cs
using System;using System.Data;using System.Configuration;using System.Collections;using System.Web;using System.Web.Security;using System.Web.UI;using System.Web.UI.WebControls;using System.Web.UI.WebControls.WebParts;using System.Web.UI.HtmlControls;
public partial class Default : System.Web.UI.Page{
protected void Page_Load(object sender, EventArgs e){
}protected void ImageButton1_Click(object sender, ImageClickEventArgs e){
if ((TextBox1.Text == "") || (TextBox2.Text == "")){
Label1.Text = "Enter Login Details";}else if ((TextBox1.Text == "zecco") && (TextBox2.Text == "456") ||
(TextBox1.Text == "admin") && (TextBox2.Text == "ADMIN")){
27
Response.Redirect("Online Trading.aspx");}
else{
Label1.Text = "Invalid Username & Password";}
}}
TESTING
9. TESTINGTesting is a process to show the correctness of the program. Testing is needed to
show completeness, it improve the quality of the software and to provide the maintenance aid.
Some testing standards are therefore necessary reduce the testing costs and operation time.
Testing software extends throughout the coding phase and it represents the ultimate review of
configurations, design and coding. Based on the way the software reacts to these testing, we can
decide whether the configuration that has been built is study or not. All components of an
application are tested, as the failure to do so many results in a series of bugs after the software is
put to use.
9.1 Black box Testing
Black box testing, also called behavioral testing, focuses on the functional requirements
of software. This testing approach enables the software engineer to derive the input conditions
that will fully exercise all requirements for a program. Black box testing attempts to find the
errors like
Incorrect or missing functions
Interface errors
Errors in data structures or external database access
Behavior or performance errors
Initialization and termination errors
In Black box testing software is exercised over a full range of inputs and outputs are
observed for correctness.
9.2 White box Testing
White box testing is also called Glass box testing is a test case design control; structure of
the procedural design to derive test cases using White box testing method, the software engineer
can derive the test cases that guarantee that all independent paths within the module have been
28
exercised at least once. Exercise all logic decisions on their true or false sides. Execute all loops
at their boundaries and within their operational bounds. Exercise internal data structure to ensure
their validity.
9.3 Software Testing Strategies
Testing involves
Unit testing
Integration testing
Acceptance testing
The first level of test is unit testing. The purpose of unit testing is to ensure that each
program is fully tested.
The second step is integration testing. In this individual program units or programs are
integrated and tested as a complete system to ensure that the software requirements are met.
Acceptance Testing involves planning and the execution of various types of tests in order
to demonstrate that the implemented software system satisfies the requirements. Finally our
project meets the requirements after going through all the levels of testing.
29
SECURITY
10. SECURITY
The system security problem can be divided into four related issues: security, integrity,
privacy and confidentiality. They determine the file structure, data structure and access
procedures.
System security refers to the technical innovations and procedures applied to the
hardware and operating systems to protect against deliberate or accidental damage from a
defined threat. In contrast, data security is the protection of data from loss, disclosure,
modifications and destruction.
System integrity refers to the proper functioning of programs, appropriate physical
security and safety against external threats such as eavesdropping and wiretapping. In
comparison, data integrity makes sure that do not differ from original from others and how the
organization can be protected against unwelcome, unfair or excessive dissemination of
information about it.
The term confidentiality is a special status given to sensitive information in a data base to
minimize the possible invasion of privacy. It is an attribute of information that characterizes its
need for protection. System security is the technical means of providing such protection. In
contrast privacy is largely a procedural matter of how information is used.
30
SCREEN SHOTS
11. SCREEN SHOTS
10.1. Home
31
11.2 Registration
32
11.3 About Us
33
11.4 Admin Login
11.5 Online Price34
11.6 Online Trading
35
CONCLUSION
36
12. CONCLUSION
Now a day’s manual process for the citizens to sales for their product such like mobile,
computer, laptop, stock share etc… has become a huge task. The main object of the website is to
reduce the effort by the candidate and save his time and avoid unwanted rushes at the ZECCO
and assure a smooth working schedule at ZECCO for Trading.
The main features of this site includes flexibility, reduce manual work in an efficient
manner, a quick, convenient, reliable and effective way to apply for their online trading market
committee records. The project could very well be enhanced further as per the requirements.
Online trading is the new concept in the stock market. In India, online trading is still at
its infancy stage. Online trading has made it easy to trade in the stock market as now people
can trade while sitting at their home. Now stock market is easily accessible by the people. There
are some problems while doing the trade through the internet. Major problem faced by online
trader is that the investors are loyal to their traditional brokers, they rely upon the suggestions
given by their brokers. Another major problem is that the people don't have full knowledge
regarding online trading. They find it difficult to trade themselves, as a wrong entry made by
them, can bring them huge losses.
BIBLIOGRAPHY
37
13. BIBLIOGRAPHY
Windows programming Charles Petzoid
ASP. NET Unleashed Stephen Walther
Visual Basic .NET Black Book Steven Holzer
Professional ASP. NET Web services Alex Homer, Dave Sussman
SQL Server Essential Reference Sharon Dooley
www.nseindia.com
www.bseindia.com
www.on-linetrading.com
www.sebi.gov.in
www.zecco.com
38
top related