Sie sind auf Seite 1von 43

INDEX

1. Abstract

2 . Organizational Profile

3. Sof tware Development Methodology

4. In troduction 4.1 Purpose 4.2 Scope 4.2.1 Existing System with Limitations 4.2.2 Proposed System Features

5. Feasibility Report 5.1 Technical Feasibility 5.2 Operational Feasibility 5.3 Economical Feasibility

6. Requir ements Specification 6.1 Hardware Requirements 6.2 Software Requiremen ts

7. Beh avioral Description 7.1 Data Flow. 7.1.1 Data Flow Diagrams

7.1.2 Use-Case Documentation 7.2 Process Flow 7.2.1 System Flow Chart 7.2.2 Activity Diagrams

8. Functional Description

9. Perf ormance Requirements

10. Design 10.1 Sequence and Collaboration Diagrams 10.2 Class Diagrams 10.3 E-R Diagrams 10.4 Database Desi gn 10.4.1 Data Dictionary 10.4.2 Normalization 11. Samp le Code

12 . System Testing and Implementation

13. Samp le Output Screens

14. Con clusion and Future Enhancements

15. Bibliography

ABSTRACT

ORGANIZATION PROFILE TRY LOGIC Soft Solutions AP Private Limited, having its Registered Office at: # 765, 3rd Floor, Habsiguda X Roads, Nacharam Road, adj: Bharat Petroleum, HYDERABAD 500 007. A.P. (India), home of a world-class professional software services company. We specialize in bespoke software development for business applications and outsourced commercial product development (R&D). Our domain of expertise ranges from B2B eCommerce, ERP, POS, Portal solutions, Human Resource & Payroll and Customer Relationship Management. TRY LOGIC is a software development, web design & development Services Company that specializes in creating tailor made custom applications for enterprises. We use innovative engagement models, balancing a sophisticated understanding of business with a

close attention to detail, to custom-design our process models and meet our client's aomation development needs. The company has expertise to develop new solutions and products right from the requirement definition stage to implementation and support. The company has specific skills Client/Server, databases, intranets, data warehousing, web technologies, WAN, LAN, networking, software design technologies. The company is capable of creating software solutions covering the entire spectrum of typical Software Project Life-Cycle consisting of system studies, re-engineering, system design, software development, testing, data migration, performance tuning and support, this is possible only because of our core technology expertise, project management skills, software engineering methodologies, talented, motivated team of software engineers and business domain knowledge consultants. With a view to identify and improve the latent talent in the students, primarily to cater their ACADEMIC requirements and mainly to groom them as INDUSTRY READY PROFESSIONAL, TRY LOGIC provides an opportunity for the much talented yet untapped students who were studying professional academic courses of various stream(s) on processes and methodologies under our unique program called TAC. TAC (TRYLOGIC ACADEMIC CONCEPT) is a program from TSSPL, under which we shape future of many students as techies who were supposed to undergo industry relevant technical education program. The program consist of Technology updating as of the current industry Requirements along with soft skills improvement and attitude enrichment, more importantly hands on experience expertise during the project work of the participants under the observation of our senior programmers on different platforms into our client / real-time projects. The concept of Knowledge for Knowledge is replaced by 'Knowledge for Application'. In such a situation, TAC try to accomplish everything possible to make learning application oriented. The advantage and idea is to overcome time, energy, and money in the recruitment process and to provide reciprocal benefit to the industries as well as the students who roll out of academic life. The company constantly focuses on learning, upgrading and disseminating state of art technologies among its engineers. Keeping abreast with the changes is an on going process

with regular in house training programs. Besides technologies the engineers are trained on quality standards and personality development Our VISION is to be globally recognized as a provider of a trusted, reliable resource to quality Information Technology (IT) service / solutions. TRY LOGIC has applied for the membership of STPI Hyderabad unit for software import & export authorization and support as well for D & O trade license. TRY LOGIC is much interested to be part of HYSEA & NASSCOM in near future as our applications are in progress. We are a learning organization and we continually invest time and effort in the area of developing standards & process documentation in line with our objective of getting ISO Quality Management System certification and preparation for level 3 CMM-I certification. Why Us Our Corporate Philosophy is Give a solution which is an amalgamation of expertise and the clients needs.

We began with a vision of meeting the massive requirements for IT talent and services in a world moving into an information-based economy. We are a team of software professionals with experience in providing robust and optimally designed IT-Solutions. We have successfully executed projects in various industry segments and have a wealth of experience and expertise in project and resource management. By partnering with TRYLOGIC, you are dealing with an Indian Company with the Quality and Cost benefit advantages of an international corporation. Our objective is to anticipate and respond to the changing needs of our customers and fulfill their expectations better than our competitors. We build lasting, high quality customer relationships in diverse domains - from ecommerce to corporate companies. We enhance the growth and profits of our clients by delivering effective and economical IT-Solutions on schedule. Background: It is without doubt, that growth of any nation is affected by information technology more than any other and also Indians have proved to be the best analytical minds world over. To

maintain this status an urgent need is to keep pace with the state of art technologies and generate world-class professionals. The industry has been seized with the problem of shortage of computer professionals in various fields, while several individuals with high computer skill-set and analytical outlook are unable to be profitably employed. This is partly due to the non-standardized methodologies, models being learnt and used by these individuals. For the well informed and well trained, the availability of new technologies represents a real push for enhancing the quality in all departments improving the productivity and hence profitability. Considering the need of the hour TRYLOGIC has formulated a unique program for assisting thousands of such aspiring software professionals to prove their mettle and be a part of winning team. Thus the fitting answer for the situation can be TRYLOGIC. SOFTWARE DEVELOPMENT METHODOLOGY After analyzing the requirements of the task to be performed, the next step is to analyze the problem and understand its context. The first activity in the phase is studying the existing system and other is to understand the requirements and domain of the new system. Both the activities are equally important, but the first activity serves as a basis of giving the functional specifications and then successful design of the proposed system. Understanding the properties and requirements of a new system is more difficult and requires creative thinking and understanding of existing running system is also difficult, improper understanding of present system can lead diversion from solution. ANALYSIS MODEL Mainly there are four phases in the "Spiral Model": Planning Evolutions Risk Analysis Engineering Software Development India

Planning: In this phase, the aims, option and constraints of the project are determined and are documented. The aims and other specifications are fixed so as to determine the strategies/approaches to go after during the project life cycle. Risk Analysis: It is the most significant phase of "Spiral Model". In this phase the entire possible option that are available and helpful in developing a cost efficient project are analyzed and strategies are determined to employ the available resources. This phase has been added particularly so as to recognize and resolve all the possible risks in the project Orphan Foundation Development. If any indication shows some uncertainty in needs, prototyping may be utilized to continue with the obtainable data and discover out possible software development solution so as to deal with the potential modification in the needs. Engineering: In this phase, the specific software development of the project is worked out. The output of developed of modules by modules is passed through all the phases iteratively so as to obtain development in the same. Customer Evaluation: In this phase, before releasing the developed product, the product is passed on to the customer so as to obtain customers views and suggestions and if some is left or the desire result is not achieved then all the needs will be identified and resolve all the possible problems/errors in the Farmers Buddy. One can compare it from the TESTING phase. The spiral model, illustrated in below figure, combines the iterative nature of prototyping with the controlled and systematic aspects of the waterfall model, therein providing the potential for rapid development of incremental versions of the software. In this model the software is developed in a series of incremental releases with the early stages being either paper models or prototypes. Later iterations become increasingly more complete versions of the product. Depending on the model it may have 3-6 task regions our case will consider a 6-task region model.

These regions are: 1. The User communication task to establish effective communication between developer and User.

2. The planning task to define resources, time lines and other project related information.. 3. The risk analysis task to assess both technical and management risks. 4. The engineering task to build one or more representations of the application. 5. The construction and release task to construct, test, install and provide user support (e.g., documentation and training). 6. The User evaluation task to obtain customer feedback based on the evaluation of the software representation created during the engineering stage and implemented during the install stage. The evolutionary process begins at the centre position and moves in a clockwise direction. Each traversal of the spiral typically results in a deliverable. For example, the first and second spiral traversals may result in the production of a product specification and a prototype, respectively. Subsequent traversals may then produce more sophisticated versions of the software. An important distinction between the spiral model and other software models is the explicit consideration of risk. There are no fixed phases such as specification or design phases in the model and it encompasses other process models. For example, prototyping may be used in one spiral to resolve requirement uncertainties and hence reduce risks. This may then be followed by a conventional waterfall development.

Note that each passage through the planning stage results in an adjustment to the project plan. Each of the regions is populated by a set of work tasks called a task set that are adapted to characteristics of the project to be undertaken. For small projects the number of tasks and their formality is low. Conversely, for large projects the reverse is true. Advantages of the Spiral Model The spiral model is a realistic approach to the development of large-scale software products because the software evolves as the process progresses. In addition, the

developer and the client better understand and react to risks at each evolutionary level. The model uses prototyping as a risk reduction mechanism and allows for the development of prototypes at any stage of the evolutionary development. It maintains a systematic stepwise approach, like the classic life cycle model, but incorporates it into an iterative framework that more reflect the real world. If employed correctly, this model should reduce risks before they become problematic, as consideration of technical risks are considered at all stages. Disadvantages of the Spiral Model Demands considerable risk-assessment expertise It has not been employed as much proven models (e.g. the WF model) and hence may prove difficult to sell to the client that this model is controllable and efficient.

INTRODUCTION:

4.1 PURPOSE: 4.2 SCOPE: 4.2.1 Existing system with limitations 4.2.2 Proposed system with features: SPECIFICATIONS REQUIREMENTS TOOLS Software Requirements: Language Specification Databases Framework Operating Systems Hardware Requirements: Processor RAM Hard Disk Functional Requirements: In software engineering, a functional requirement defines a function of a software system or its component. A function is described as a set of inputs, the behavior, and outputs (see also software). Functional requirements may be calculations, technical details, data manipulation and processing and other specific functionality that define what a system is supposed to accomplish. Intel Pentium IV & LATER 512MB RAM & above 20GB Min C#.Net SQLSERVER 2005 .Net 3.5 Windows 2000, windows NT 4.0,Xp sp2

Behavioral requirements describing all the cases where the system uses the functional requirements are captured in use cases. Functional requirements are supported by non-functional requirements (also known as quality requirements), which impose constraints on the design or implementation (such as performance requirements, security, or reliability). How a system implements functional requirements is detailed in the system design. As defined in requirements engineering, functional requirements specify particular results of a system. This should be contrasted with non-functional requirements which specify overall characteristics such as cost and reliability. Functional requirements drive the application architecture of a system, while non-functional requirements drive the technical architecture of a system. Non-Functional Requirements: In systems engineering and requirements engineering, a non-functional requirement is a requirement that specifies criteria that can be used to judge the operation of a system, rather than specific behaviors. This should be contrasted with functional requirements that define specific behavior or functions. In general, functional requirements define what a system is supposed to do whereas non-functional requirements define how a system is supposed to be. Non-functional requirements are often called qualities of a system. Other terms for non-functional requirements are "constraints", "quality attributes", "quality goals" and "quality of service requirements". Qualities, that is, non-functional requirements, can be divided into two main categories: 1. Execution qualities, such as security and usability, which are observable at run time. 2. Evolution qualities, such as testability, maintainability, extensibility and scalability, which are embodied in the static structure of the software system. .NET Framework: The .NET Framework is a new computing platform that simplifies application development in the highly distributed environment of the Internet. The .NET Framework is designed to fulfill the following objectives:

To provide a consistent object-oriented programming environment whether object code is stored and executed locally, executed locally but Internet-distributed, or executed remotely.

To provide a code-execution environment that minimizes software deployment and versioning conflicts. To provide a code-execution environment that guarantees safe execution of code, including code created by an unknown or semi-trusted third party. To provide a code-execution environment that eliminates the performance problems of scripted or interpreted environments. To make the developer experience consistent across widely varying types of applications, such as Windows-based applications and Web-based applications. To build all communication on industry standards to ensure that code based on the .NET Framework can integrate with any other code. The .NET Framework has two main components: the common language runtime and

the .NET Framework class library. The common language runtime is the foundation of the .NET Framework. You can think of the runtime as an agent that manages code at execution time, providing core services such as memory management, thread management, and remoting, while also enforcing strict type safety and other forms of code accuracy that ensure security and robustness. In fact, the concept of code management is a fundamental principle of the runtime. Code that targets the runtime is known as managed code, while code that does not target the runtime is known as unmanaged code. The class library, the other main component of the .NET Framework, is a comprehensive, object-oriented collection of reusable types that you can use to develop applications ranging from traditional command-line or graphical user interface (GUI) applications to applications based on the latest innovations provided by ASP.NET, such as Web Forms and XML Web services. The .NET Framework can be hosted by unmanaged components that load the common language runtime into their processes and initiate the execution of managed code, thereby creating a software environment that can exploit both managed and unmanaged features. The .NET Framework not only provides several runtime hosts, but also supports the development of third-party runtime hosts.

For example, ASP.NET hosts the runtime to provide a scalable, server-side environment for managed code. ASP.NET works directly with the runtime to enable Web Forms applications and XML Web services, both of which are discussed later in this topic. Internet Explorer is an example of an unmanaged application that hosts the runtime (in the form of a MIME type extension). Using Internet Explorer to host the runtime enables you to embed managed components or Web Forms controls in HTML documents. Hosting the runtime in this way makes managed mobile code (similar to Microsoft ActiveX controls) possible, but with significant improvements that only managed code can offer, such as semi-trusted execution and secure isolated file storage. The following illustration shows the relationship of the common language runtime and the class library to your applications and to the overall system. The illustration also shows how managed code operates within a larger architecture. Features of the Common Language Runtime The common language runtime manages memory, thread execution, code execution, code safety verification, compilation, and other system services. These features are intrinsic to the managed code that runs on the common language runtime. With regards to security, managed components are awarded varying degrees of trust, depending on a number of factors that include their origin (such as the Internet, enterprise network, or local computer). This means that a managed component might or might not be able to perform file-access operations, registry-access operations, or other sensitive functions, even if it is being used in the same active application. The runtime enforces code access security. For example, users can trust that an executable embedded in a Web page can play an animation on screen or sing a song, but cannot access their personal data, file system, or network. The security features of the runtime thus enable legitimate Internet-deployed software to be exceptionally feature rich. The runtime also enforces code robustness by implementing a strict type- and codeverification infrastructure called the common type system (CTS). The CTS ensures that all managed code is self-describing. The various Microsoft and third-party language compilers

generate managed code that conforms to the CTS. This means that managed code can consume other managed types and instances, while strictly enforcing type fidelity and type safety. In addition, the managed environment of the runtime eliminates many common software issues. For example, the runtime automatically handles object layout and manages references to objects, releasing them when they are no longer being used. This automatic memory management resolves the two most common application errors, memory leaks and invalid memory references. The runtime also accelerates developer productivity. For example, programmers can write applications in their development language of choice, yet take full advantage of the runtime, the class library, and components written in other languages by other developers. Any compiler vendor who chooses to target the runtime can do so. Language compilers that target the .NET Framework make the features of the .NET Framework available to existing code written in that language, greatly easing the migration process for existing applications. While the runtime is designed for the software of the future, it also supports software of today and yesterday. Interoperability between managed and unmanaged code enables developers to continue to use necessary COM components and DLLs. The runtime is designed to enhance performance. Although the common language runtime provides many standard runtime services, managed code is never interpreted. A feature called just-in-time (JIT) compiling enables all managed code to run in the native machine language of the system on which it is executing. Meanwhile, the memory manager removes the possibilities of fragmented memory and increases memory locality-of-reference to further increase performance. Finally, the runtime can be hosted by high-performance, server-side applications, such as Microsoft SQL Server and Internet Information Services (IIS). This infrastructure enables you to use managed code to write your business logic, while still enjoying the superior performance of the industry's best enterprise servers that support runtime hosting. .NET Framework Class Library The .NET Framework class library is a collection of reusable types that tightly integrate with the common language runtime. The class library is object oriented, providing types from

which your own managed code can derive functionality. This not only makes the .NET Framework types easy to use, but also reduces the time associated with learning new Features of the .NET Framework. In addition, third-party components can integrate seamlessly with classes in the .NET Framework. For example, the .NET Framework collection classes implement a set of interfaces that you can use to develop your own collection classes. Your collection classes will blend seamlessly with the classes in the .NET Framework. As you would expect from an object-oriented class library, the .NET Framework types enable you to accomplish a range of common programming tasks, including tasks such as string management, data collection, database connectivity, and file access. In addition to these common tasks, the class library includes types that support a variety of specialized development scenarios. For example, you can use the .NET Framework to develop the following types of applications and services:

Console applications. Scripted or hosted applications. Windows GUI applications (Windows Forms). ASP.NET applications. XML Web services. Windows services.

For example, the Web Forms classes are a comprehensive set of reusable types that vastly simplify Web GUI development. If you write an ASP.NET Web Form application, you can use the Web Forms classes. C#.NET Introduction to Web Forms (C#.NET) Web Forms is the new platform for Microsoft Web application development, based on the .NET Framework. This framework provides a clear, object-oriented, extensible set of classes that enable you to develop rich Web applications. Additionally, Web Forms can act as the local user interface in a multi-tier distributed solution. Web Forms is a framework for building

Windows client applications that utilize the common language runtime. Web Forms applications can be written in any language that the common language runtime supports. ACTIVE X DATA OBJECTS.NET ADO.NET Overview ADO.NET is an evolution of the ADO data access model that directly addresses user requirements for developing scalable applications. It was designed specifically for the web with scalability, statelessness, and XML in mind. ADO.NET uses some ADO objects, such as the Connection and Command objects, and also introduces new objects. Key new ADO.NET objects include the DataSet, DataReader, and DataAdapter. The important distinction between this evolved stage of ADO.NET and previous data architectures is that there exists an object -- the DataSet -- that is separate and distinct from any data stores. Because of that, the DataSet functions as a standalone entity. You can think of the DataSet as an always disconnected recordset that knows nothing about the source or destination of the data it contains. Inside a DataSet, much like in a database, there are tables, columns, relationships, constraints, views, and so forth. A DataAdapter is the object that connects to the database to fill the DataSet. Then, it connects back to the database to update the data there, based on operations performed while the DataSet held the data. In the past, data processing has been primarily connection-based. Now, in an effort to make multi-tiered apps more efficient, data processing is turning to a messagebased approach that revolves around chunks of information. At the center of this approach is the DataAdapter, which provides a bridge to retrieve and save data between a DataSet and its source data store. It accomplishes this by means of requests to the appropriate SQL commands made against the data store. The XML-based DataSet object provides a consistent programming model that works with all models of data storage: flat, relational, and hierarchical. It does this by having no 'knowledge' of the source of its data, and by representing the data that it holds as collections and data types. No matter what the source of the data within the DataSet is, it is manipulated through the same set of standard APIs exposed through the DataSet and its subordinate objects.

While the DataSet has no knowledge of the source of its data, the managed provider has detailed and specific information. The role of the managed provider is to connect, fill, and persist the DataSet to and from data stores. The OLE DB and MSACCESS .NET Data Providers (System.Data.OleDb and System.Data.SqlClient) that are part of the .Net Framework provide four basic objects: the Command, Connection, DataReader and DataAdapter. In the remaining sections of this document, we'll walk through each part of the DataSet and the OLE DB/MSACCESS .NET Data Providers explaining what they are, and how to program against them. The following sections will introduce you to some objects that have evolved, and some that are new. These objects are: Connections. For connection to and managing transactions against a database. Commands. For issuing SQL commands against a database. DataReaders. For reading a forward-only stream of data records from a MSACCESS data source. DataSets. For storing, remoting and programming against flat data, XML data and relational data. DataAdapters. For pushing data into a DataSet, and reconciling data against a database. When dealing with connections to a database, there are two different options: MSACCESS .NET Data Provider (System.Data.SqlClient) and OLE DB .NET Data Provider (System.Data.OleDb). In these samples we will use the MSACCESS .NET Data Provider. These are written to talk directly to Microsoft MSACCESS. The OLE DB .NET Data Provider is used to talk to any OLE DB provider (as it uses OLE DB underneath). Connections Connections are used to 'talk to' databases, and are represented by provider-specific classes such as SQLConnection. Commands travel over connections and result sets are returned in the form of streams which can be read by a DataReader object, or pushed into a DataSet object. Commands

Commands contain the information that is submitted to a database, and are represented by provider-specific classes such as SQLCommand. A command can be a stored procedure call, an UPDATE statement, or a statement that returns results. You can also use input and output parameters, and return values as part of your command syntax. The example below shows how to issue an INSERT statement against the Northwind database. DataReaders The DataReader object is somewhat synonymous with a read-only/forward-only cursor over data. The DataReader API supports flat as well as hierarchical data. A DataReader object is returned after executing a command against a database. The format of the returned DataReader object is different from a recordset. For example, you might use the DataReader to show the results of a search list in a web page. DataSets and DataAdapters DataSets The DataSet object is similar to the ADO Recordset object, but more powerful, and with one other important distinction: the DataSet is always disconnected. The DataSet object represents a cache of data, with database-like structures such as tables, columns, relationships, and constraints. However, though a DataSet can and does behave much like a database, it is important to remember that DataSet objects do not interact directly with databases, or other source data. This allows the developer to work with a programming model that is always consistent, regardless of where the source data resides. Data coming from a database, an XML file, from code, or user input can all be placed into DataSet objects. Then, as changes are made to the DataSet they can be tracked and verified before updating the source data. The GetChanges method of the DataSet object actually creates a second DatSet that contains only the changes to the data. This DataSet is then used by a DataAdapter (or other objects) to update the original data source. The Dataset has many XML characteristics, including the ability to produce and consume XML data and XML schemas. XML schemas can be used to describe schemas interchanged via WebServices. In fact, a DataSet with a schema can actually be compiled for type safety and statement completion.

DataAdapters (OLEDB/SQL) The DataAdapter object works as a bridge between the DataSet and the source data. Using the provider-specific SqlDataAdapter (along with its associated SqlCommand and SqlConnection) can increase overall performance when working with a Microsoft MSACCESS databases. For other OLE DB-supported databases, you would use the OleDbDataAdapter object and its associated OleDbCommand and OleDbConnection objects. The DataAdapter object uses commands to update the data source after changes have been made to the DataSet. Using the Fill method of the DataAdapter calls the SELECT command; using the Update method calls the INSERT, UPDATE or DELETE command for each changed row. You can explicitly set these commands in order to control the statements used at runtime to resolve changes, including the use of stored procedures. For ad-hoc scenarios, a CommandBuilder object can generate these at run-time based upon a select statement. However, this run-time generation requires an extra round-trip to the server in order to gather required metadata, so explicitly providing the INSERT, UPDATE, and DELETE commands at design time will result in better run-time performance. 1. 2. 3. 4. 5. ADO.NET is the next evolution of ADO for the .Net Framework. ADO.NET was created with n-Tier, statelessness and XML in the forefront. ADO.NET can be used to get data from a stream, or to store data in a cache for There is a lot more information about ADO.NET in the documentation. Remember, you can execute a command directly against the database in order to

Two new objects, the DataSet and DataAdapter, are provided for these scenarios. updates.

do inserts, updates, and deletes. You don't need to first put data into a DataSet in order to insert, update, or delete it. 6. Also, you can use a DataSet to bind to the data, move through the data, and navigate data relationships ASP.Net

Server Application Development Server-side applications in the managed world are implemented through runtime hosts. Unmanaged applications host the common language runtime, which allows your custom managed code to control the behavior of the server. This model provides you with all the features of the common language runtime and class library while gaining the performance and scalability of the host server. The following illustration shows a basic network schema with managed code running in different server environments. Servers such as IIS and MSACCESS can perform standard operations while your application logic executes through the managed code. Server-side managed code ASP.NET is the hosting environment that enables developers to use the .NET Framework to target Web-based applications. However, ASP.NET is more than just a runtime host; it is a complete architecture for developing Web sites and Internet-distributed objects using managed code. Both Web Forms and XML Web services use IIS and ASP.NET as the publishing mechanism for applications, and both have a collection of supporting classes in the .NE XML Web services, an important evolution in Web-based technology, are distributed, server-side application components similar to common Web sites. However, unlike Web-based applications, XML Web services components have no UI and are not targeted for browsers such as Internet Explorer and Netscape Navigator. Instead, XML Web services consist of reusable software components designed to be consumed by other applications, such as traditional client applications, Web-based applications, or even other XML Web services. As a result, XML Web services technology is rapidly moving application development and deployment into the highly distributed environment of the Internet. If you have used earlier versions of ASP technology, you will immediately notice the improvements that ASP.NET and Web Forms offers. For example, you can develop Web Forms pages in any language that supports the .NET Framework. In addition, your code no longer needs to share the same file with your HTTP text (although it can continue to do so if you prefer). Web Forms pages execute in native machine language because, like any other

managed application, they take full advantage of the runtime. In contrast, unmanaged ASP pages are always scripted and interpreted. ASP.NET pages are faster, more functional, and easier to develop than unmanaged ASP pages because they interact with the runtime like any managed application. The .NET Framework also provides a collection of classes and tools to aid in development and consumption of XML Web services applications. XML Web services are built on standards such as SOAP (a remote procedure-call protocol), XML (an extensible data format), and WSDL (the Web Services Description Language). The .NET Framework is built on these standards to promote interoperability with non-Microsoft solutions. For example, the Web Services Description Language tool included with the .NET Framework SDK can query an XML Web service published on the Web, parse its WSDL description, and produce C# or Visual Basic source code that your application can use to become a client of the XML Web service. The source code can create classes derived from classes in the class library that handle all the underlying communication using SOAP and XML parsing. Although you can use the class library to consume XML Web services directly, the Web Services Description Language tool and the other tools contained in the SDK facilitate your development efforts with the .NET Framework. If you develop and publish your own XML Web service, the .NET Framework provides a set of classes that conform to all the underlying communication standards, such as SOAP, WSDL, and XML. Using those classes enables you to focus on the logic of your service, without concerning yourself with the communications infrastructure required by distributed software development. Finally, like Web Forms pages in the managed environment, your XML Web service will run with the speed of native machine language using the scalable communication of IIS. SQL SERVER DATABASE A database management, or DBMS, gives the user access to their data and helps them transform the data into information. Such database management systems include dBase,

paradox, IMS, Sql Server and SQL Server. These systems allow users to create, update and extract information from their database. A database is a structured collection of data. Data refers to the characteristics of people, things and events. SQL Server stores each data item in its own fields. In SQL Server, the fields relating to a particular person, thing or event are bundled together to form a single complete unit of data, called a record (it can also be referred to as raw or an occurrence). Each record is made up of a number of fields. No two fields in a record can have the same field name. During an SQL Server Database design project, the analysis of your business needs identifies all the fields or attributes of interest. If your business needs change over time, you define any additional fields or change the definition of existing fields. SQL Server Tables SQL Server stores records relating to each other in a table. Different tables are created for the various groups of information. Related tables are grouped together to form a database.

Primary Key Every table in SQL Server has a field or a combination of fields that uniquely identifies each record in the table. The Unique identifier is called the Primary Key, or simply the Key. The primary key provides the means to distinguish one record from all other in a table. It allows the user and the database system to identify, locate and refer to one particular record in the database. Relational Database Sometimes all the information of interest to a business operation can be stored in one table. SQL Server makes it very easy to link the data in multiple tables. Matching an employee to the department in which they work is one example. This is what makes SQL Server a relational database management system, or RDBMS. It stores data in two or more tables and

enables you to define relationships between the table and enables you to define relationships between the tables. Foreign Key When a field is one table matches the primary key of another field is referred to as a foreign key. A foreign key is a field or a group of fields in one table whose values match those of the primary key of another table. Referential Integrity Not only does SQL Server allow you to link multiple tables, it also maintains consistency between them. Ensuring that the data among related tables is correctly matched is referred to as maintaining referential integrity. Data Abstraction A major purpose of a database system is to provide users with an abstract view of the data. This system hides certain details of how the data is stored and maintained. Data abstraction is divided into three levels. Physical level: This is the lowest level of abstraction at which one describes how the data are actually stored. Conceptual Level: At this level of database abstraction all the attributed and what data are actually stored is described and entries and relationship among them. View level: This is the highest level of abstraction at which one describes only part of the database. Advantages of RDBMS Redundancy can be avoided Inconsistency can be eliminated Data can be Shared Standards can be enforced Security restrictions can be applied

Integrity can be maintained Conflicting requirements can be balanced.

Disadvantages of DBMS A significant disadvantage of the DBMS system is cost. In addition to the cost of purchasing of developing the software, the hardware has to be upgraded to allow for the extensive programs and the workspace required for their execution and storage. adequately backed up so that in case of failure the data can be recovered. FEATURES OF SQL SERVER (RDBMS) SQL SERVER is one of the leading database management systems (DBMS) because it is the only Database that meets the uncompromising requirements of todays most demanding information systems. From complex decision support systems (DSS) to the most rigorous online transaction processing (OLTP) application, even application that require simultaneous DSS and OLTP access to the same critical data, SQL Server leads the industry in both performance and capability SQL SERVER is a truly portable, distributed, and open DBMS that delivers unmatched performance, continuous operation and support for every database. SQL SERVER RDBMS is high performance fault tolerant DBMS which is specially designed for online transactions processing and for handling large database application. SQL SERVER with transactions processing option offers two features which contribute to very high level of transaction processing throughput, which are The row level lock manager While centralization reduces duplication, the lack of duplication requires that the database be

Enterprise wide Data Sharing The unrivaled portability and connectivity of the SQL SERVER DBMS enables all the systems in the organization to be linked into a singular, integrated computing resource. Portability

SQL SERVER is fully portable to more than 80 distinct hardware and operating systems platforms, including UNIX, MSDOS, OS/2, Macintosh and dozens of proprietary platforms. This portability gives complete freedom to choose the database sever platform that meets the system requirements. Open Systems SQL SERVER offers a leading implementation of industry standard SQL. SQL Servers open architecture integrates SQL SERVER and non SQL SERVER DBMS with industries most comprehensive collection of tools, application, and third party software products SQL Servers Open architecture provides transparent access to data from other relational database and even non-relational database. Distributed Data Sharing SQL Servers networking and distributed database capabilities to access data stored on remote server with the same ease as if the information was stored on a single local computer. A single SQL statement can access data at multiple sites. You can store data where system requirements such as performance, security or availability dictate. Unmatched Performance The most advanced architecture in the industry allows the SQL SERVER DBMS to deliver unmatched performance. Sophisticated Concurrency Control Real World applications demand access to critical data. With most database Systems application becomes contention bound which performance is limited not by the CPU power or by disk I/O, but user waiting on one another for data access. SQL Server employs full, unrestricted row-level locking and contention free queries to minimize and in many cases entirely eliminates contention wait times. DATA TABLES

DATA FLOW DIAGRAMS 3.4 Data flow diagrams DATAFLOW DIAGRAMS Data flow diagrams (bubble charts) are directed graphs in which the nodes specify processing activities and the arcs specify data items transmitted between processing nodes. A data flow diagram might represent data flow between individual statements or blocks of statements in a routine, data flow between sequential routines, data flow between concurrent process, or data flow in a distributed computing system, where each node represents a geographically remote processing unit. Unlike flow charts, data flow diagrams do not indicate decision logic or condition under which various nodes in the diagram might be activated.Data flow diagrams can be expressed using special symbols. These symbols can be used to denote processing nodes, data nodes, data sources and data stores. A Data Flow Diagram (DFD) is a graphical technique that depicts information flow and transforms that are applied as data move to represent a system or software at any level of abstraction. Data Flow Diagram may be partitioned into levels that represent increasing information flow and functional detail. The notation used to create a DFD is illustrated in the given as: 1. Process:A Process transforms incoming data flow(s) into outgoing data flow(s). 2. External Entity:An External Entity is a system element (hardware a person) that produces information for transformation by the software or received information produced by the software. 3. Data Store:A repository of data for the use of one or more process.

4.Arrow ( ):A data item or collection of data items, the arrowhead the direction of data flow.

indicates

Data Flow

Processing Node

Source or Destination

Data Store

DATA FLOW DIAGRAMS

Context Level DFD/LEVEL 0:

TopLevelDFD/LEVEL1/HIGHLEVEL:

3.6

UML Diagrams:

UML DATA The UMLs four structural diagram exists to visualize, specify, construct and document the static aspects of system. Just as the static aspects of a house encompass the existence and placement of such things as walls, doors, windows, pipes, wires and vents. So to do the static aspects of a software system encompass the existence and placement of such things classes, interfaces, collaborations, components and nodes. The UMLs structural diagrams are roughly organized around the major groups of things you will find when modeling a system. 1. 2. 3. 4. CLASS DIAGRAM: classes, interfaces and collaborations OBJECT DIAGRAM: objects COMPONENT DIAGRAM: components DEPLOYMENT DIAGRAM: nodes.

CLASS DIAGRAM: A class diagram shows a set of classes, interfaces, collaborations and their relationships. Class diagrams are the most common diagrams found in modeling object oriented system. Class diagram that includes active classes are used to address the static process view of a system. COMPONENT DIAGRAM: A component diagram shows a set of components and their relationships. Component diagrams are related to class diagrams in that component typically maps to one or more classes, interfaces or collaborations. It is used to illustrate the static implementation view of a system. DEPLOYMENT DIAGRAM: A deployment diagrams shows a set of nodes and their relationships deployment diagrams are related to component diagrams in that node typically encloses one or more components. USECASE DIAGRAMS:

A use case diagram shows as set of use cases and actors and their relationships. Use case diagrams are especially important in organizing and modeling behavior of a system.

SEQUENCE DIAGRAM: A sequence diagram is an interaction diagram that emphasizes the time ordering of messages. A sequence diagram shows a set of objects and messages sent and receive by those objects. The objects are typically named or anonymous instances of other things, such as collaborations, components and nodes. We can use sequence diagrams to illustrate the dynamic view of a system. COLLABORATION DIAGRAM: A collaboration diagram is an inter action diagram that emphasizes the structural organization of the objects that send and receive messages. It shows a set of objects, links among those objects and messages sent and receive by the objects. The objects are typically named or anonymous instances of classes but may also represent instance of other things such as collaborations, components and nodes. RELATIONSHIPS 1. Dependency: Dependency is a semantic relationship between two things in which change to one thing may affect the semantics of other things. Graphically a dependency rendered as a dashed-line, possibly directed occasionally including a label.

Fig: Dependency 2. Association:

Association is a structural relationship that describes set of links.A link being a connection among objects. Aggregation is a special kind of association a structural relationship between a whole and its parts.It is graphically rendered as a solid-line. Such as multiplicity and role names. _________________ Employer Fig: Association 3. Generalization: Generalization is a specialization/generalization relationship in which objects of specialized element (child) are substitutable for object of the generalized element (parent).in this way the child shares the structure and behavior of the parent.Graphically rendered as a solid-line with a hollow over-head pointing to the parent. employee

Fig: Generalization 4. Realization: A realization is a semantic relationship between classifiers. Where in one classifier specifies a contract that another classifier guarantees to carry-out you will encounter relationships in two places between interfaces and classes or components between use case collaborations.Graphically rendered as cross between generalization and dependency relationships.

Fig: Realization 5. Class:

A class is a description of set of objects and the same attributes, operations relationships and semantics. A class implements one or more interfaces. Graphically rendered as a rectangle, usually including a name, attributes and operations.

Fig: Classes 6. Use case: Use case is a description of set of sequence of actions that a system performs that yields an observable result of value to a particular actor. Graphically rendered as an ellipse with solid-lines, usually including only its names.

Fig: Use case 7. Collaborations: Collaboration defines an interaction is a society of roles and other elements that work together to provide some co-operative behavior thats bigger than the some of elements. Therefore collaborations have structural as well as behavioral dimensions. Graphically rendered as an ellipse with dashed-lines usually including only its names.

Fig: collaboration 8. Component: A component is a physical and replicable part of a system that confirms to provides the relation of set of interfaces. In a system we can find and encounter different kinds of deployment components such as com+ components and java-beans. A component typically represents the physical package logical elements such as classes, interfaces and collaborations. It is graphically rendered as rectangle with tabs usually including only its names.

Fig: Component 9. Node: A node is a physical element that exists at run-time and represents a computational resources generally having at least some memory and often processing capability. Graphically rendered as a cube, usually including only its names.

Fig: Node UML DIAGRAMS

Usecase analysis: The use case diagram is used to identify the primary element and processes. That form the system the primary element are termed as actors and the processes are called use cases.The use case diagram show which actors interacts with each use case. Use case diagram of UML are used to view system from this perspective as a set of discrete Activities or transactions.A ellipse represents a use case and a Stick figure represents an actor operating with in the use case. An actor can be user, system or other entity. The initiating actor appears on the left of the use case and the receiving actor appears on the right. The actor is identified by a name below the stick figure The name of the use case appears either in the side or below the ellipse. An association line is a that connects an actor and to the use case and represents communication between the actor and use cases The association line is solid similar to the line that connect associated classes.

USECASE DIAGRAM:

CLASS DIAGRAM:

SEQUENCE DIAGRAM

COLLABORATION DIAGRAM

DEPLOYMENT DIAGRAM

Component diagram: We map the available classes into respective component here component are nothing physical entities. COMPONENT DIAGRAM

Activity diagram:

ER DIAGRAM

SAMPLE CODE

TESTING Testing Concepts Testing Testing Methodologies Black box Testing: White box Testing. Gray Box Testing. Levels of Testing Unit Testing. Module Testing. Integration Testing. System Testing. User Acceptance Testing.

Types Of Testing Smoke Testing. Sanitary Testing. Regression Testing. Re-Testing. Static Testing. Dynamic Testing. Alpha-Testing.

Beta-Testing. Monkey Testing. Compatibility Testing. Installation Testing. Adhoc Testing. Ext.

TCD (Test Case Documentation) STLC Test Planning. Test Development. Test Execution. Result Analysis. Bug-Tracing. Reporting.

Microsoft Windows Standards Manual Testing Automation Testing (Tools) Win Runner. Test Director.

Testing: The process of executing a system with the intent of finding an error. Testing is defined as the process in which defects are identified, isolated, subjected for rectification and ensured that product is defect free in order to produce the quality product and hence customer satisfaction. Quality is defined as justification of the requirements Defect is nothing but deviation from the requirements Defect is nothing but bug. Testing --- The presence of bugs Testing can demonstrate the presence of bugs, but not their absence Debugging and Testing are not the same thing! Testing is a systematic attempt to break a program or the AUT Debugging is the art or method of uncovering why the script /program did not execute properly.

Testing Methodologies: Black box Testing: is the testing process in which tester can perform testing on an application without having any internal structural knowledge of application. Usually Test Engineers are involved in the black box testing.

White box Testing: is the testing process in which tester can perform testing on an application with having internal structural knowledge. Usually The Developers are involved in white box testing.

Gray Box Testing: is the process in which the combination of black box and white box tecnics are used.

Levels of Testing:

Module1 Units

Module2 Units

Module3 Units

i/p

Integration o/p i/p

Integration o/p

System Testing: Presentation + business +Databases

UAT: user acceptance testing STLC (SOFTWARE TESTING LIFE CYCLE)


Test Planning: 1.Test Plan is defined as a strategic document which describes the procedure how to perform various testing on the total application in the most efficient way. 2.This document involves the scope of testing, 3. Objective of testing, 4. Areas that need to be tested, 5.Areas that should not be tested, 6. Scheduling Resource Planning, 7. Areas to be automated, various testing tools Used. Test Development:

1. Test case Development (check list) 2. Test Procedure preparation. (Description of the Test cases) Test Execution: Result Analysis: 1. Implementation of test cases. Observing the result. 1. Expected value: is nothing but expected behavior Of application.

3. Actual value: is nothing but actual behavior of application Bug Tracing: Reporting: Types Of Testing: > Smoke Testing: is the process of initial testing in which tester looks for the availability of all the functionality of the application in order to perform detailed testing on them. (Main check is for available forms) > Sanity Testing: is a type of testing that is conducted on an application initially to check for the proper behavior of an application that is to check all the functionality are available before the detailed testing is conducted by on them. > Regression Testing: is one of the best and important testing. Regression testing is the process in which the functionality, which is already tested before, is once again tested whenever some new change is added in order to check whether the existing functionality remains same. >Re-Testing: is the process in which testing is performed on some functionality which is already tested before to make sure that the defects are reproducible and to rule out the environments issues if at all any defects are there. Static Testing: is the testing, which is performed on an application when it is not been executed.ex: GUI, Document Testing Dynamic Testing: is the testing which is performed on an application when it is being executed.ex: Functional testing. Alpha Testing: it is a type of user acceptance testing, which is conducted on an application when it is just before released to the customer. Beta-Testing: it is a type of UAT that is conducted on an application when it is released to the customer, when deployed in to the real time environment and being accessed by the real time users. Monkey Testing: is the process in which abnormal operations, beyond capacity operations are done on the application to check the stability of it in spite of the users abnormal behavior. Collect all the failed cases, prepare documents. Prepare document (status of the application)

Compatibility testing: it is the testing process in which usually the products are tested on the environments with different combinations of databases (application servers, browsersetc) In order to check how far the product is compatible with all these environments platform combination. Installation Testing: it is the process of testing in which the tester try to install or try to deploy the module into the corresponding environment by following the guidelines produced in the deployment document and check whether the installation is successful or not. Adhoc Testing: Adhoc Testing is the process of testing in which unlike the formal testing where in test case document is used, with out that test case document testing can be done of an application, to cover that testing of the future which are not covered in that test case document. Also it is intended to perform GUI testing which may involve the cosmotic issues. TCD (Test Case Document: Test Case Document Contains Test Scope (or) Test objective Test Scenario Test Procedure Test case This is the sample test case document for the Acadamic details of student project: Test scope: Test coverage is provided for the screen Acadamic status entry form of a student module of university management system application Areas of the application to be tested

Test Scenario: When the office personals use this screen for the marks entry, calculate the status details, saving the information on students basis and quit the form.

Test Procedure: The procedure for testing this screen is planned in such a way that the data entry, status calculation functionality, saving and quitting operations are tested in terms of Gui testing, Positive testing, Negative testing using the corresponding Gui test cases, Positive test cases, Negative test cases respectively

Test Cases:

Template for Test Case Description Exp Act Result

T.C.No

Guidelines for Test Cases: 1. GUI Test Cases: Total no of features that need to be check Look & Feel Look for Default values if at all any (date & Time, if at all any require) Look for spell check Expected value Actual Result value

Example for Gui Test cases: T.C.No Description

Check for all the The screen must contain all the features in the screen features 1 Check for the alignment The alignment should be in of the objects as per the proper way 2 validations 2. Positive Test Cases: The positive flow of the functionality must be considered Valid inputs must be used for testing Must have the positive perception to verify whether the requirements are justified.

Example for Positive Test cases: T.C.No 1 Description Expected value Actual value Check for the date The date and Time Auto Display time system 2 of the must Result

be displayed Enter the valid Roll It should no into the student accept roll no field 3. Negative Test Cases:

Must have negative perception. Invalid inputs must be used for test.

Example for Negative Test cases: T.C.No 1 Description Expected value Actual value Try to modify The Modification information 2 in should not be not save not date and time allow Enter invalid data It should in to the student accept details form, click data, on save should allow Result

invalid

SCREEN SHOTS

CONCLUSION: This project is used to maintain the details and marks of the students belonging to NALANDA INSTITUTE OF ENGINEERING AND TECHNOLOGY.Individual marks report of a student can be generated.Semister wise marks of the students can also be generated.

Das könnte Ihnen auch gefallen