Ieee

  • November 2019
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Ieee as PDF for free.

More details

  • Words: 9,503
  • Pages: 50
SOFTWARE QULAITY ASSURANCE & MANAGEMENT A CASE STUDY ON WEB METRICS FRAME WORK FOR IMPROVING QUALITY ASSURANCE

(Phase 3)

Presented To: Dr. Qaiser S. Durrani

Presented By: Muhammad Irfan Khan

(07L-0858)

Saqib Aziz

(07L-0861)

Ahmad Mohsin

(07L-0852)

Case Study Phase III

Table of Contents Objectives of Study (Revised)...................................................................................................4 To define a framework for quality assurance.............................................................................5 To identify Web Metrics that will boost the quality of a product..............................................5 To understand the failure causes for web application................................................................5 To develop a model that ensures adherence to software product standards, processes and procedures. To assure that standards and procedures are established and are followed throughout the software acquisition life cycle. .........................................................................5 To capture standards that will directly affect a business function.............................................5 To investigate the issues, and present a distinguishable model.................................................5 To measure the subject adherence of web pages.......................................................................5 To measure the content quality of web pages............................................................................5 To measure the popularity of web pages....................................................................................5 To measure the linkage between web pages..............................................................................5 To measure the similarity between web pages...........................................................................5 Problem Statement (Revised)....................................................................................................6 Problem Elaboration (Revised)..................................................................................................7 Scope..........................................................................................................................................8 Prior Literature (Revised)........................................................................................................11 Research Methodology (Revised)............................................................................................13 Develop Web Metrics Framework.....................................................................................14 Detailed metrics of performance, security and usability were developed to help us measure the CRMs from different quality aspects. This detailed framework is explained in the later section................................................................................................................................14 Selection of CRMs.............................................................................................................14 Data Collection and Analysis.............................................................................................19 Size...................................................................................................................................20 No. of Projects under Study.............................................................................................20 Interpretation of Data.........................................................................................................20 Research Report for Findings............................................................................................20 Development of Hypotheses (Revised)...................................................................................21 Hypotheses:..............................................................................................................................21 Hypothesis I.......................................................................................................................21 Null Hypothesis H0 Perf .................................................................................................21 Alternate Hypothesis H1 Perf..........................................................................................22 Hypothesis II......................................................................................................................22 Null Hypothesis H0 Sec...................................................................................................22 Alternate Hypothesis H1 Sec...........................................................................................22 Hypothesis III....................................................................................................................22 Null Hypothesis H0 Usab................................................................................................22 Alternate Hypothesis H1 Usab.........................................................................................22 Mathematical Description of Hypotheses (Revised)...............................................................23 Web Metrics Framework .........................................................................................................23 Formula for Calculating Page Performance......................................................................27

Page 2 of 50

Case Study Phase III Research Results and Data Analysis........................................................................................27 Discussion of Findings with Statistical Analysis.....................................................................40 Conclusion...............................................................................................................................46 Future Work.............................................................................................................................47 References (Revised)...............................................................................................................48 [4]. A Quality Framework For Web Site Quality ..............................................................48 [5] R.S. Pressman & Associates Inc..................................................................................48

Page 3 of 50

Case Study Phase III

Objectives of Study (Revised) The objective of our study is to use web metrics for Web Applications to improve overall Quality. We will be studying previous and current trends for web applications related to quality and hence Web metrics can be used to improve the Quality of CRM Applications. Within a short period, the Internet and World Wide Web have become ubiquitous, surpassing all other technological developments in our history. They have also grown rapidly in their scope and extent of use, significantly affecting all aspects of our lives. Industries such as manufacturing, travel and hospitality, banking, education, and government are Web-enabled to improve and enhance their operations. E-commerce has expanded quickly, cutting across national boundaries. Even traditional legacy information and database systems have migrated to the Web[1]. Advances in wireless technologies and Web-enabled appliances are triggering a new wave of mobile Web applications. As a result, we increasingly depend on a range of Web applications. Now that many of us rely on Web based systems and applications, they need to be reliable and perform well. To build these systems and applications, Web developers need a sound methodology, a disciplined and repeatable process, better development tools, and a set of good guidelines. The emerging field of Web engineering fulfils these needs. It uses scientific, engineering, and management principles and systematic approaches to successfully develop, deploy, and maintain high-quality Web systems and applications. It aims to bring the current chaos in Web based system development under control, minimize risks, and enhance Web site maintainability and quality. The objective of our study is to have a thorough understanding of web applications their trends and technologies being used. When we talk about Software Engineering the word Web Engineering comes into our minds. Web Engineering is evolving every day. Our focus is to analyze current web applications critically, define Quality Assurance Frame work for Web Applications, defining Web Metrics and creating a relationship of metrics regarding Quality Assurance of Web Applications.

Page 4 of 50

Case Study Phase III The main objectives of this study are: •

To define a framework for quality assurance.



To identify Web Metrics that will boost the quality of a product.



To understand the failure causes for web application.



To develop a model that ensures adherence to software product standards, processes and procedures. To assure that standards and procedures are established and are followed throughout the software acquisition life cycle.



To capture standards that will directly affect a business function.



To investigate the issues, and present a distinguishable model.



To measure the subject adherence of web pages.



To measure the content quality of web pages.



To measure the popularity of web pages.



To measure the linkage between web pages.



To measure the similarity between web pages.

Page 5 of 50

Case Study Phase III

Problem Statement (Revised) Businesses these days are employing web as an integral part of their strategies to make their products available worldwide. This led to an increase in the number of web applications deployed for business expansion. With this increase, quality has become the most overlooked aspect of web products. Due to global reach, a small compromise in quality has serious implications on the reputation of the business. This demands formulation of a standard framework to ensure the quality of online applications. We intend to study and analyze the Quality Assurance related issues with regard to web as quality is always not given a too much importance in the whole SDLC. But now industry is grooming and understanding of Quality Assurance throughout the life cycle of a product under construction has improved a lot. When we talk about Web Quality, Web Metrics come into play. Web Metrics and Quality assurance are closely related to each other. In our problem statement we will be defining a framework for Quality of web Applications. Metrics, as we know, refer to standards of measurement. Therefore, web metrics are standardized ways of measuring something that relates to the Web. Here in our case study we are focusing on Web Applications and to more precise we will be using CRM Applications and mapping of these metrics to evaluate Quality. We have decided to get main data from a renowned Software House Mindshare Solutions as it specializes in CRM applications. The problem is that they have introduced three demo versions of their CRM application named as Unify CRM. But results reported are that it bears a low quality as for as overall performance of the product is concerned. We decided to have a comprehensive comparisons of Unify CRM with other renowned open Source CRMs and evaluate it on the basis of our Web Metrics Frame work in which we will be focusing more on the Performance, Usability and Security and see the Quality of the products in association with these Quality metrics for Web Applications. When considering the Web, it becomes clear that there is an abundance of different things we can measure. As an example, consider web traffic, while it certainly is

Page 6 of 50

Case Study Phase III possible to review the entire Web and how "busy" it is. [2], consider web page similarity such as executed by Google and other search engines if not performed with a wide (meaning we only consider more than a small amount of pages, such as those of a particular site), its practical usefulness is likely to be limited. In other words different metrics can be applied on different views.

Problem Elaboration (Revised) In our case study we are trying to frame a web metric for Quality Assurance for web applications. For this purpose we have devised a Web metric Frame work for Quality Assurance. To conduct our studies on the basis of empirical data we visited one of the most leading Software Houses in industry. The name of the Software House is Mindshare. It has the specialization in the CRM applications. It produces 3 different levels of CRMs. The problem is that produced three demos for their CRM but they failed badly. As suggested that we will be conducting an empirical analysis on how to improve Quality of Web Applications based on Web Metrics Framework devised. The organization wanted to know how to improve the Quality of their web products (CRM). To make our studies more useful and to conduct the analysis we will be choosing four projects and will be comparing how to improve the Quality of Web applications especially related to Web Applications on the basis of our Web Applications and examine how Quality is closely related to these Web Metrics. We will then compare our results on the basis Web metrics to how much extent they are improving the Quality of the products more efficiently. Quality Assurance has been a debatable topic in the Web development industry. There are views which say that quality is only a value added service, not an integral part of the software delivered. Another approach is that Quality can be assured if the cost and time permits. As mostly the schedules are tight so all the saving is done on Quality Assurance phase. However, most of the sophisticated IT Companies realize the importance of the Quality Assurance and therefore have evolved whole department for the very purpose. Quality Assurance has different dimensions which vary according to the nature of the projects. For example a Web based application

Page 7 of 50

Case Study Phase III must incorporate proper performance, security, usability and other measures and should perform efficiently. Similarly critical systems need to be quality assured against breakdown and should have an efficient backup flows. We will be defining a framework for Quality of web Applications. Metrics refer to standards of measurement. Therefore, web metrics are standardized ways of measuring something that relates to the Web Applications Quality. Web metrics do play a very important role in determining the actual characteristics related to the web applications functionalities. Here in our study we have identified key web metrics that actually effect the Quality of Web Applications. To support our study in a more specific way we decided to select few of the Web metrics and to check them on Web Applications to see the impact of these on Quality Assurance. Being more realistic we chose different CRM Applications to check the impact of these metrics related to Quality Assurance. Following metrics will be used in order to determine the extent to which CRM Applications adhere to Quality: -

Performance

-

Security

-

Usability

Although at our initial study phase we had identified few other metrics as well but we considered above mentioned metrics to be more precise to our case study.

Scope This study encompasses key web QA metrics that can be used to ensure the quality of Web Applications and specifically to CRM Applications. Following are the important key areas:

1- Web Traffic

Page 8 of 50

Case Study Phase III Web traffic can be about number of hits a site has during any given interval, but can also be measured in the amount of data that is transferred during a similar interval. It can be useful in determining popularity, and through trend analysis also in estimating future needs. 2- Web Page Significance Significance can be seen as a formalization of quality and relevance of web pages with regard to their information content. (Here, quality does indeed refer to information content, and not to design quality, as discussed above). 3- Accessibility Accessible design principles often results in a significant overall increase in the usability of a Web site - in terms of faster completion of tasks, with lower error rates, and more effective retention of knowledge of the site by repeat users. 4- Relevance Relevance is a direct measure of how well a particular page satisfies the information need of some user, typically expressed as a set of query words. As the Web grows exponentially, it seems logical to assume that the number of documents that contain the same query words is typically also increasing. The need for metrics that can order all such documents so that those that are most relevant can be examined first is thus greater than ever. 5- Web Page Similarity Web search engines such as Google have for some time, in addition to relevance, allowed users to retrieve similar pages. There are three ways on which similarity is typically measured content-based, link-based and usagebased similarity.

Page 9 of 50

Case Study Phase III

6- Search Engine Optimization Website visibility and its ranking in the search engines can be very important in the web-commerce industry. A website optimized for search engine can yield more profit. 7- Performance A fast website will increase the user experience very much and this brings returning visitors. 8- Security Online security is perhaps overlooked most often in local software houses, websites with poor security implementations will invariably damage users and the business. 9- Ease of Use Quality issues regarding the ease of use of a web application are important in that they help a business to retain their clientele. Also, such applications are easier to maintain and change. 10- Portability With a growing range of computer hardware and software platforms, it is important for ecommerce applications to be able to perform consistently and provide similar functionality in different computing environments. 11- Reliability As with traditional software, reliability is always an important quality issue for users. A system application should always produce consistent results and outputs for a given fixed input. Otherwise the application can not be trusted for high quality service.

Page 10 of 50

Case Study Phase III

Prior Literature (Revised) Quality Assurance is an important step in the website development process and, by all means, should not be skipped. A broken link or a misspelled word may seem like trivial mistakes, but they can greatly undermine the credibility of your website. You want people who visit your site to feel assured of the quality of the information they find. As an emerging discipline, Web engineering actively promotes systematic, disciplined and quantifiable approaches towards successful development of highquality, ubiquitously usable Web-based systems and applications [3] A simple definition in the context of quality for Web sites is that ‘quality is meeting requirements’. This definition works because by creating technical specifications and requirements that describe various attributes of a Web site as well as how it should function you have set yourself goals to achieve and have determined specific indicators of quality. Quality can then be measured by testing various aspects of your Web site and the complex relationships between all areas of the site at intervals. The current World Wide Web has many flaws, with a great many resources failing to comply with. As we move towards a richer, more structure Web, it will be essential that quality assurance is built into development processes – unlike HTML, XML applications formally require string adherence with the standards and may fail to render if this is not the case. However, even when a resource does comply with standards it does not mean that the user experience will necessarily be a happy one. Thus, a combination of supplier QA and user satisfaction assessment are needed. However, linking the subjective perceptions of users with the QA practices of suppliers is not a simple task. The next stage of work is to model the relationships between user satisfaction and supplier initiatives (such as QA procedures). One way in which this might be done is through quality function deployment (QFD): “a structured and disciplined process that provides a means to identify and carry the voice of the customer through each stage of product and or service development and implementation” [4].

Page 11 of 50

Case Study Phase III Software process and product metrics are quantitative measures that enable software people to gain insight into the efficacy of the software process and the projects that are conducted using the process as a framework. Basic quality and productivity data are collected. These data are then analyzed, compared against past averages, and assessed to determine whether quality and productivity improvements have occurred. [5] The Internet and the world wide web (WWW or simply the web) are some specific example of general heterogeneous systems. QA for these systems is gaining importance.[6] Web applications have become very complex and crucial, especially when combined with areas such as CRM (Customer Relationship Management) and BPR (Business Process Reengineering). The scientific community has focused attention to Web applications design, development, analysis, and testing, by studying and proposing methodologies and tools [7] Given the organic growth of the Web, we require new metrics that provide deeper insight on the Web as a whole and also on individual sites from different perspectives. Arguably, the most important motivation for deriving such metrics is the role they can play in improving the quality of information available on the Web. [8] When you can measure what you are speaking about, and express it in numbers, you know something about it; but when you cannot express it in numbers, your knowledge is of a meager and unsatisfactory kind;

it may be the beginning of

knowledge, but you have scarcely in your thoughts advanced to the state of science.[9] Metrics help organizations generate more effective Web sites and provide measures that understand and that academics can replicate and analyze. To provide practical value, metrics should identify frequency of measurement, frequency of review, source of data, rationale for introducing the measure, who will act on the data, and the purpose of the measure (Neely 1998). For scientific, quantitative rigor, metrics should exhibit, at a minimum, construct validity and reliability (Straub 1989, Cook and Campbell 1979).[10]

Page 12 of 50

Case Study Phase III For effective measurement, the measurement activity should have clear objectives and identify types of attributes that can be measured and appropriate scales. Empirical relations for an ttribute should be identified in advance and the measurement scale must be meaningful. [10]

Research Methodology (Revised) We worked with a software house, Mindshare Solutions, to find quality issues in their CRM system “UnifyCRM”, which they sell as a product to different international clients. We used our web metrics framework to assess its level of quality. Same quality metrics framework was run on the CRM systems of the company’s competitors. The results generated by our framework were used to find out the differences in quality levels between Unify CRM and competitors’ CRMs. We were asked to evaluate the quality of CRMs by focusing on three parameters: 1) Performance 2) Security 3) Usability These metrics will be elaborated in detail in the coming sessions for understandings. To develop a detailed metrics framework for the above mentioned parameters of quality, we explored the existing web engineering practices, techniques and core web metrics used in web development industry. Our research methodology tried to identify the successful and problematic areas in the existing approaches used during the development life cycle of web-based systems. We also identified different aspects that are being used in local industry to evaluate the quality of a website. In the case study we define, how to measure these aspects and how to improve and how these metrics can be used to improve the quality of Web Applications, especially CRMs. Now we describe the steps we used to formulate the structured research methodology for the case study. For our case study, we used some real world CRM

Page 13 of 50

Case Study Phase III projects. Our main aim is to identify a set of key quality aspects and then formulate a workable framework for the quality metrics thus identified. Our research methodology is a systematic process to achieve the objectives of carrying out this study. The methodology that was formulated for case study is briefly explained. Develop Web Metrics Framework Detailed metrics of performance, security and usability were developed to help us measure the CRMs from different quality aspects. This detailed framework is explained in the later section. Selection of CRMs We have selected certain Customer Relationship Management (CRM) systems with the focus on QA metrics. Our selection consists of a mix of open-source and proprietary CRMs. We have selected four systems: 1. Unify

CRM

a

product

of

Mindshare

Solutions

Pvt.

Ltd.

(http://unifycrm.com/UnifyCRM/Login.aspx) 2. Salesforce CRM (http://www.salesforce.com/) 3. Enterprise

CRM

and

Groupware

System

(http://sourceforge.net/projects/egs/) 4. SugarCRM, an open-source CRM ( http://www.sugarcrm.com/crm/)

Salesforce CRM

Page 14 of 50

Case Study Phase III The proven leader in on-demand customer relationship management (CRM), salesforce.com empowers customers to stand out from the crowd. We do so by delivering the most innovative technology and making it as easy as possible to share and

manage

business

information.

Our

solutions

combine

award-winning

functionality, proven integration, point-and-click customization, global capabilities, and the best user experience and the result is CRM success. That's why Salesforce has earned the trust of its customers and a customer success rate of 95%.Salesforce SFA enables companies to drive sales productivity, increase visibility, and expand revenues with an affordable, easy-to-deploy service that delivers success to companies of all sizes. Following are key features of Salesforce CRM:-

Service & Support The Salesforce solution for customer service gets companies up and running in a matter of weeks with a call center application that is loved by agents and a customer self-service application— powered by Web 2.0—that generates new levels of customer loyalty.

Partner Relationship Management Salesforce Partners makes it easy for partners to access leads, collaborate on deals, and locate all the information they need in order to be successful. The Salesforce Partners is seamlessly integrated with Salesforce SFA to deliver unparalleled visibility to your company's entire sales pipeline for direct and indirect channels.

Marketing

Page 15 of 50

Case Study Phase III Salesforce Marketing enables closed-loop marketing to execute, manage, and analyze the results of multichannel campaigns. Marketing executives can measure the ROI of their budgets, tie revenue back to specific marketing programs, and make adjustments in real time.

Content Salesforce Content brings Web 2.0 usability to your business content so you can share it more effectively and enhance collaboration within your organization. Empower employees to find the exact documents they need, right from the business applications they use on a daily basis.

Analytics Salesforce Analytics empowers business users at every level to gain relevant insight and analysis. With real-time reporting, calculations, and dashboards, businesses can optimize performance, decision making, and resource allocation.

Custom Applications Build enterprise-class applications on salesforce.com's powerful on-demand platform. Deliver all your company's business applications in a single environment with one data model, one sharing model, and one user interface.

Page 16 of 50

Case Study Phase III Industry Applications Meet all of your industry-specific needs with salesforce.com's award-winning CRM, a broad variety of on-demand apps from the AppExchange, and the Force.com platform. Our industry applications are built on the successes of hundreds of companies in your industry. And because no two companies are exactly alike, all industry apps are fully and easily customizable.

AppExchange Applications The AppExchange is your one-stop marketplace for on-demand business applications. The AppExchange makes it easy to find, sample, and select from hundreds of apps for your business, all preintegrated with Salesforce.

SOURCE FORGE CRM KEY FEATRUES Sourceforge CRM application is a highly flexible Sales Force Automation (SFA) tool that meets both the needs of sales managers and the sales rep. In addition to standard SFA functionality such as lead, account and opportunity management, Sales Management provides a powerful sales management system to improve a sales organization's productivity, allowing management to plan ahead of economic changes in order to effectively manage any market condition. It provides sales reps the capability to develop accurate forecasts, seamlessly share information across sales teams, and configure products and services to meet the unique needs of each customer. Its simple user interface is designed to improve sales rep productivity, yet support best practices across the entire sales organization. Sales Management's key features and capabilities include: Improve forecast accuracy - The sales pipeline is continually updated in real time so that everyone in your organization is provided with a clear view, allowing resources to be focused accordingly.

Page 17 of 50

Case Study Phase III Close more deals - Sales process visibility enables each member of your sales team to know precisely what the other teammates are working on, allowing them to collaborate to transform prospects into profitable customers. Shorten and standardize unique sales cycles - Because different channels require different sales processes, Sales Manager allows an unlimited number of unique sales methodologies or sales process to be created. Teams, both inside and outside your organization can effectively work together to close accounts by scheduling events, assigning tasks, coordinating meetings, flagging new opportunities, and updating client files on every account. Enable collaborative and consistent customer management - Real-time, secure access to detailed account data enables you and your channel partners to collaborate with sales, customer service & support, and marketing personnel. With instant access to all communication, including email, notes, calls, resolutions, and more, you can collectively manage customer relationships across your entire extended enterprise. Recognize "big picture" market trends - With Sourceforge flexible reporting system, it's easy to review and analyze sales data, both current and historical, allowing sales management to spot changes in customer behavior or shifts in key market indicators. Armed with a comprehensive contextual view of both past and current events, your sales organization can respond to evolving customer needs and economic condition Project Admins :

jstride, nsuk

Developers :

6

Database Environment :

ADOdb,PostgreSQL (pgsql)

Development Status :

5 - Production/Stable

License :

GNU General Public License (GPL) Operating

System :

OS Independent (Written in an interpreted

language) Programming Language : Translations :

PHP English

Page 18 of 50

Case Study Phase III User Interface :

Web-based

Project UNIX name :

egs

Registered :

2003-05-22 06:18

Sugar 5.0 Key Features SugarCRM is becoming a disruptive force in the small enterprise Customer Relationship Management (CRM) market. Its commercial open source model, CRM appliance option, low price, and strong set of CRM features are impacting more traditional methods of CRM delivery and the perception of CRM value among small enterprises. Despite its small size compared to other vendors, SugarCRM scored highest in the Product Index of our evaluation, due to its broad features and extremely flexible deployment option. New Module Builder allows users to build custom modules from scratch or combine existing or custom objects into a brand new CRM module. New Metadata Driven User Interface (UI) stores customizations in a metadata repository and combines the benefits of custom CRM with the ability to incorporate new features in future releases. Improved Access Control offers better support for team hierarchies and access control functions that manage and protect information at the field level. New AJAX Email Client delivers the functionality of a desktop email client with the portability of a web-based email application. Improved Dashboards with new charting capabilities, including support for funnel, pie charts, line and bar graphs, and performance gauges. Multiple Dashboards allows users to access any number of pre-built or custom dashboards from their homepage. Data Collection and Analysis

All CRMs will be passed through metrics framework and data will be collected against different parameters of quality. Data would be entered into statistical software. We will use SPSS 15.0 for statistical analysis of our data. We intend to find

Page 19 of 50

Case Study Phase III out if there is a significant difference in the quality of the selected CRMs in terms of performance, security and usability. It can be accomplished with the application of a statistical test. As the total number of CRMs being used are four(4), we will use ANOVA test. [11]

Organization Information 53 Size Expertise Enterprise Solutions Provider Project Information No.

of

Projects

4

under Study Domain of Projects

Customer

Communication

Management System Project Manager of corresponding

Person Table – 1

Relationship

Organization’s Quality Team.

Interpretation of Data After the statistical analysis of the data, we will describe the results of draw inferences and recommend a course of action about the quality areas where Unify CRM needs improvements.

Research Report for Findings We will be making a comprehensive report to conclude our study and present a web metric framework that can be employed as a tool to assess the quality of any CRM. Pictorial Representation of Research Methodology (REVISED)

Page 20 of 50

Case Study Phase III

Fig – 1 (Representation of Research Methodology)

Development of Hypotheses (Revised) Hypotheses: For the purpose of analytical study, we have assumed following hypothesis Hypothesis I This hypothesis enables us to evaluate the impact of our framework’s performance metrics on the quality of a web application. Null Hypothesis H0 Perf Performance metrics in our framework have no effect on the quality of a web application.

Page 21 of 50

Case Study Phase III Alternate Hypothesis H1 Perf Performance metrics in our framework have a significant effect on the quality of a web application. Hypothesis II Through this hypothesis we can evaluate the impact of our framework’s security metrics on the quality of a web application. Null Hypothesis H0 Sec Security metrics in our framework have no effect on the quality of a web application. Alternate Hypothesis H1 Sec Security metrics in our framework have a significant effect on the quality of a web application.

Hypothesis III This hypothesis helps us evaluate the impact of our framework’s usability metrics on the quality of a web application. Null Hypothesis H0 Usab Usability metrics in our framework have no effect on the quality of a web application. Alternate Hypothesis H1 Usab Usability metrics in our framework have a significant effect on the quality of a web application.

Page 22 of 50

Case Study Phase III

Mathematical Description of Hypotheses (Revised) For this study, we have selected the value of significance as α = 0.05. This is the probability of rejecting the Null hypothesis, when the Null hypothesis is true i.e. there are 5 in 100 chances that the Null hypothesis is rejected Hypotheses that were formulated with the hope that they be rejected led to the use of the term null hypothesis. Today this term is applied to any hypothesis we wish to test and is denoted by Ho. The rejection of Ho leads to the acceptance of an alternative hypothesis, denoted by H1. In the hypotheses, the performance, security and usability are the independent variables, and quality is the dependent variable. This is because quality depends on metrics in the framework.

Web Metrics Framework Security Metrics Following are the security parameters to evaluate quality of the CRM application that to how much extent security is affecting the overall Quality of the proudct. 1. SQL injection 2. Cross-Site Scripting Attacks 3. Session Hijacking 4. Denial of Service 5. Buffer Overflows

SQL injection: SQL injection is a type of security exploit in which the attacker adds Structured Query Language (SQL) code to a Web form input box to gain access to resources or

Page 23 of 50

Case Study Phase III make changes to data. An SQL query is a request for some action to be performed on a database. [12] Cross-Site Scripting Attacks Cross-site scripting (XSS) is a security exploit in which the attacker inserts malicious coding into a link that appears to be from a trustworthy source. When someone clicks on the link, the embedded programming is submitted as part of the client's Web request and can execute on the user's computer, typically allowing the attacker to steal information. Denial of Service A denial of service (DoS) attack is an incident in which a user or organization is deprived of the services of a resource they would normally expect to have. In a distributed denial-of-service, large numbers of compromised systems (sometimes called a botnet) attack a single target. Common forms of denial of service attacks are: Services SYN Attack

Detail When a session is initiated between the Transport Control Program (TCP) client and server in a network, a very small buffer space exists to handle the usually rapid "hand-shaking"

Teardrop

exchange of messages that sets up the session. This type of denial of service attack exploits the way that the

Attack

Internet Protocol (IP) requires a packet that is too large for the next router to handle be divided into fragments.

Smurf Attack

In this attack, the perpetrator sends an IP ping (or "echo my message back to me") request to a receiving site The ping packet specifies that it be broadcast to a number of hosts within

Viruses

the receiving site's local network. Computer viruses, which replicate across a network in various ways, can be viewed as denial-of-service attacks where the Page 24 of 50

Case Study Phase III victim is not usually specifically targetted but simply a host Physical

unlucky enough to get the virus. Here, someone may simply snip a fiber optic cable. This kind of

Infrastructure

attack is usually mitigated by the fact that traffic can sometimes

Attacks quickly be rerouted. Table – 2 (Security Metric Measures)

Buffer Overflows A buffer overflow occurs when a program or process tries to store more data in a buffer (temporary data storage area) than it was intended to hold. Since buffers are created to contain a finite amount of data, the extra information - which has to go somewhere - can overflow into adjacent buffers, corrupting or overwriting the valid data held in them. Although it may occur accidentally through programming error, buffer overflow is an increasingly common type of security attack on data integrity. Session Hijacking Session hijacking, also known as TCP session hijacking, is a method of taking over a Web user session by surreptitiously obtaining the session ID and masquerading as the authorized user. Once the user's session ID has been accessed (through session prediction), the attacker can masquerade as that user and do anything the user is authorized to do on the network.

Usability Metrics Here we discuss the aspects / parameters to measure the usability of CRM applications to see to how much extent they make an impact on the quality of the product.

Page 25 of 50

Case Study Phase III To determine if a system meets user/human-centered design goals of effectiveness, efficiency, and user satisfaction, it is necessary to collect objective, quantifiable data.

The Human Factors Specialist develops experimental plans and data

collection procedures to obtain and statistically analyze measures of usability. Some of the most frequently used metrics for this purpose are listed by goal as follows: [13] •





Effectiveness o

Training time

o

Time to reach proficiency

o

Number of commands/actions per task

o

Number of commands/features that are never used

o

Number of times "help" is accessed

Efficiency o

Time to complete a task

o

Error rate

o

Number of tasks completed within a given time

o

Error recovery time

o

Decision time/delay

User satisfaction o

Positive statements recorded during observations

o

Negative statements recorded during observations

Performance Metrics Following are the parameters to measure the performance of a CRM application to evaluate the Quality of the product.[14] Parameters Simultaneous

Detail browser Number of users using the simultaneous connections

Page 26 of 50

Case Study Phase III connections Warm up time (secs)

to the server. The time required to open the first page, first time are

called warp up time. Total number of requests Total Number of requests sent during one iteration. Total number of Total Number of active connections to the server connections Average requests

during one iteration. per Average request in one second during one iteration.

second Average time to first byte Average time required to download first byte in one (msecs) millisecond, after sending request. Average time to last byte Average time required to download last byte in one (msecs) millisecond, after sending request. Number of bytes sent Number of bytes sent in one iteration. (bytes) Http Errors Numeric value for the http error. Table – 3 (Performance Metric Measures)

Formula for Calculating Page Performance The following formula will quantify the performance of a Web application by measuring the number of processor cycles needed for each request. The formula requires dividing the amount of cycles spent by the number of requests that were handled. (processor speed x processor use) __________________________________ = cost, in cycles/request (or Hz/rps) number of requests per second

Research Results and Data Analysis We took an approach in which statistical analysis will be covered in Discussion section of the study where we will be discussing Statistical Test ANOVA and its results in terms of the Significance on the Quality of the CRM applications using our devised Quality Frame work of metrics. In order to carry out our study in a more successful way we did various Page 27 of 50

Case Study Phase III testing for the metrics defined in our frame work for the four (4) applications or CRM. For performance, security and usability various testes were conducting using different testing tools in order to see the results in more proper way. On the basis of these tests conducted one can foresee the Quality of the CRM applications in terms of Web Metrics formulated. Following tests were conducted for our analysis of the test. PERFORMANCE TESTING SECURITY TESTING USABILITY TESING For these tests to be conducted various testing techniques were used. Now we briefly focus on these test results and show them in tabular form.

Performance Testing We did conduct performance testing on the four applications using MICROSOFT APPLICATION TEST CENTER. Application test center is designed to focus on stress testing of Web servers and analyze performance and scalability problems with Web applications, including Active Server Pages (ASP) and the components they use. Application Center Test simulates a large group of users by opening multiple connections to the server and rapidly sending HTTP requests. Application test center supports several different authentication schemes and the SSL protocol, making it ideal for testing personalized and secure sites. Although long-duration and high-load stress testing is Application Center Test's main purpose, the programmable dynamic tests will also be useful for functional testing.Application Center Test is compatible with all Web servers and Web applications that adhere to the HTTP protocol.

Page 28 of 50

Case Study Phase III To check the quality of the CRM applications we did stress testing as it is supported in Test center. Stress testing is a form of testing that is used to determine the stability of a given system or entity. It involves testing beyond normal operational capacity, often to a breaking point, in order to observe the results. Stress testing may have a more specific meaning in certain industries.[16] As we did use Microsoft Application Test center for conducting our tests so here we discuss some of the features of test center. It uses the features of dynamic testing techniques. Dynamic tests consist of a script that sends requests to the Web server during a test run. These tests are called "dynamic" because the request order, the URL being requested, and many other properties, are determined at the time the test runs. The code in the test can examine the server's previous response before creating the properties of the next request. Because most of the request's behavior is under the control of the script, an understanding of the HTTP protocol is necessary. If your test will be emulating the behavior of a particular HTTP user agent, particularly something other than a common Web browser running on a personal computer, you will need to understand how that user agent behaves. Note that dynamic tests must run within the Application Center Test UI. This is required because the program, not the test script, is responsible for tracking and managing the request properties, such as the HTTP headers and user cookies. Here we used five categories with respect to concurrent users with the range of 20,40,60,80,100 respectively. We did create lots of tests for performance metric to assess the quality of the product. We also considered other aspects related to Quality of the product like Dot Net frame work but as our CRMs had different platforms so this could not be extended. Tabular Representation of the tested CRM in terms of Performance All the results have been shown in tabular form show exact measures of the each parameter used for the performance.

Page 29 of 50

Case Study Phase III These are the performance Tables to check the Quality of 4 CRMs. PERFORMANCE TEST-1

Performance

Unify

Sales

Sugar

Enterprise

CRM

Force

CRM

CRM

Simultaneous browser connections Warm up time (secs)

20

20

20

20

1.45

0.5

0.5

0.9

9800 1315

13625

10200

116932

5

12727

10900

12689

66

29.5

35

42.9

1.5

0.6

1.325

0.92

5.32 7E+0

2.43 976065

3.5 866632

4.33

6

7

5

6738890

1

0

0

0

Total number of requests Total number of connections Average requests per second Average time to first byte (msecs) Average time to last byte (msecs) Number of bytes sent/Sec (bytes) HTTP Errors

Table – 4 (Performance of CRMs for 20 concurrent users) PERFORMANCE TEST-2

Performance

Unify

Sales

Sugar

Enterprise

CRM

Force

CRM

CRM

Simultaneous browser connections 40

40

40

40

1.5

0.5

0.7

0.75

Warm up time (secs)

Page 30 of 50

Case Study Phase III Total number of requests 19571

25255

20317

23843

26133

25255

21995

25578

63.57

30.05

37.03

47.11

1.28

0.85

1.03

1.11

5.32

2.43

3.5

4.33

655084

976065

866632

3

7

5

7436890

5

0

0

2

Total number of connections

Average requests per second

Average time to first byte (msecs) Average time to last byte (msecs) Number of bytes sent/Sec (bytes) HTTP Errors

Table – 5 (Performance of CRMs for 40 concurrent users)

PERFORMANCE TEST-3

Performance Simultaneous browser

Unify

Sales

Sugar

Enterpri

CRM

Force

CRM

se CRM

connections 60

60

60

60

Page 31 of 50

Case Study Phase III Warm up time (secs) 1.5

0.72

0.79

0.8

29443

20265

22425

28342

39465

32721

35432

33825

63.57

30.05

37.03

47.11

1.28

0.85

1.03

1.11

5.32

2.43

3.5

4.33

3

9760657

8666325

7436890

10

4

2

7

Total number of requests

Total number of connections

Average requests per second

Average time to first byte (msecs) Average time to last byte (msecs) Number of bytes sent/Sec (bytes)

655084

HTTP Errors

Table – 6 (Performance of CRMs for 60 concurrent users)

PERFORMANCE TEST-4

Performance Simultaneous browser

Unify

Sales

Sugar

Enterpris

CRM

Force

CRM

e CRM

connections 80

80

80

80

Page 32 of 50

Case Study Phase III Warm up time (secs) 1.45

0.5

0.5

0.9

38980

30453

36522

37675

54620

40822

45987

50465

66

29.5

35

42.9

1.5

0.6

1.325

0.92

5.32

2.43

3.5

4.33

685485

976065

866632

3

7

5

6738890

22

9

14

17

Total number of requests

Total number of connections

Average requests per second

Average time to first byte (msecs)

Average time to last byte (msecs)

Number of bytes sent/Sec (bytes)

HTTP Errors

Table – 7 (Performance of CRMs for 80 concurrent users)

PERFORMANCE TEST-5

Performance

Unify

Sales

Sugar

Enterpri

CRM

Force

CRM

se CRM

Page 33 of 50

Case Study Phase III Simultaneous browser connections

100

100

100

100

1.5

0.8

0.9

1.3

49825

37568

40918

48431

65775

12727

10900

67293

66

29.5

35

42.9

1.5

0.6

1.325

0.92

5.32

2.43

3.5

4.33

685485

976065

866632

3

7

5

6738890

52

18

25

47

Warm up time (secs)

Total number of requests

Total number of connections

Average requests per second

Average time to first byte (msecs) Average time to last byte (msecs) Number of bytes sent/Sec (bytes) HTTP Errors

Table – 8 (Performance of CRMs for 100 concurrent users)

PERFORMANCE TEST-6

Page 34 of 50

Case Study Phase III Suga Performance Warm up time (secs)

Unify

Sales

r

Enterpri

Standard

CRM

Force

CRM

se CRM

Value

72.5

25

25

45

2

65.3

90.8

68

77.9

15000

87.7

84.8

72.6

84.5

15000

66

29.5

35

42.9

100

75

30

66.2

46

2

53.2

24.3

35

43.3

10

68.5

97.6

86.6

67.3

10,000,000

52

18

25

47

100

Total number of requests

Total number of connections

Average requests per second

Average time to first byte (msecs) Average time to last byte (msecs) Number of bytes sent/Sec (bytes) HTTP Errors

Table – 9 (Performance Test with random No. of Users)

Graphical Representation of performance measures

Page 35 of 50

Case Study Phase III As performance parameters have different measuring units. They are in milli seconds to seconds

from bits to bytes so in this order is quite hectic to present

them in such a way. For this reason we assigned all parameters a percentile of hundred according to which their measures have been shown. We have used specific standardized values for all the CRM applications to show the accurate results on the graph.

100% 80% Enterprise CRM 60%

Sugar CRM

40%

Sales Force Unify CRM

20%

Number of bytes

Average time to first

Total number of

Warm up time (secs)

0%

Fig – 2 (Graphical Representation of Performance Measures)

Security Testing To conduct security testing we used IBM Rational AppScan Standard Edition is an industry-leading Web application security testing suite that scans and tests for all common web application vulnerabilities - including those identified in the WASC threat classification - such as SQL-Injection, Cross-Site Scripting and Buffer Overflow.

Page 36 of 50

Case Study Phase III •

Provides broad application coverage, including Web 2.0/Ajax applications



Generates advanced remediation capabilities including a comprehensive task list to ease vulnerability remediation



Simplifies security testing for non-security professionals by building scanning intelligence directly into the application



Features over 40 out-of-the-box compliance reports including PCI Data Security Standards, ISO 17799, ISO 27001, Basel II, SB 1386 and PABP (Payment Application Best Practices [17]

It scans and tests for all common Web application containing vulnerabilities including SQL-Injection, Cross-Site Scripting and Buffer Overflow.

Tabular Representation of Security Measures in terms of Quality Here we present a tabulated representation of our security metric parameters.

Security Parameters

Unify

Sales

Sugar

Enterprise

CRM

Force

CRM

CRM

SQL injection

15

0

2

7

Cross-Site Scripting Attacks

42

5

7

Session Hijacking

3

0

0

1

Denial of Service

50

38

42

43

Buffer Overflows

87

35

59

58

Table – 10 (Representation of Security Parameters)

Page 37 of 50

Case Study Phase III Graphical Representation of Security Measures in terms of Quality

Unify CRM Sales Force Sugar CRM

Overflows

Buffer

Hijacking

Session

Enterprise CRM

SQL injection

100 90 80 70 60 50 40 30 20 10 0

Fig – 3 (Graphical Representation of Security Measures)

Usability Testing We did conduct usability testing in terms of Effectiveness, Efficiency and User Satisfaction. Usability is defined by five quality components: •

Learnability: How easy is it for users to accomplish basic tasks the first time they encounter the design?



Efficiency: Once users have learned the design, how quickly can they perform tasks?



Memorability: When users return to the design after a period of not using it, how easily can they reestablish proficiency?



Errors: How many errors do users make, how severe are these errors, and how easily can they recover from the errors?



Satisfaction: How pleasant is it to use the design? Page 38 of 50

Case Study Phase III There are many other important quality attributes. A key one is utility, which refers to the design's functionality: Does it do what users need? Usability and utility are equally important: It matters little that something is easy if it's not what you want. It's also no good if the system can hypothetically do what you want, but you can't make it happen because the user interface is too difficult. To study a design's utility, you can use the same user research methods that improve usability.

Usability Testing Tool To conduct our usability testing we devised a usability Questionnaire which aided us in determining usability of the CRM applications. Usability Questionnaire has already been discussed in the prior section of Web Metrics . Tabular Representation of Usability Measures in terms of Quality Sale

Security Parameters Training time Number of commands/actions per task Number of commands/features that are never used Number of times "help" is accessed Time to complete a task Error rate Number of tasks completed within a given time Error recovery time

Unif

s

S

y

Forc

ugar

Enterpri

CRM

e

CRM

se CRM

80 10

70 55

60 49

65 70

75 89 10 90 95 10

62 64 59 5 70 67

53 47 50 20 50 52

10 20 64 56 62 20

Table – 11 (Usability Metrics)

Graphical Representation of Usability Measures in terms of Quality

Page 39 of 50

Case Study Phase III 100% 80% 60%

Enterprise CRM Sugar CRM

40%

Sales Force 20% 0% Training time

Unify CRM Time to complete a task

Fig – 4 (Graphical Representation of Usability Measures)

Discussion of Findings with Statistical Analysis Since we used four(4) different CRM applications to verify our metrics framework, ANNOVA Statistical Analysis Test was undertaken to test all hypotheses.

Hypothesis 1 From below table we can see that the Sig. (significant) value against all performance metrics is less than 0.05 level of significance. This leads us to reject null hypothesis H0

Perf

and accept alternate hypothesis H1

Perf

Thus, we can infer that the

performance metrics in the framework have significant impact on the quality of a web application.

ANOVA - TEST Sum of

Mean

Squares df

Square

F

Page 40 of 50

Sig.

Case Study Phase III SimultaneousBrows Between

1767.29

erConnections

Groups Within

4 887289

Groups Total

5.206 887466

Between

2.500 647130

Groups

221571

WarmUpTime

493000 Within

0.000 5874811

Groups

100031 970000

Total

347000

TotalConnections

883.647

471

18838.419

.047

.024

25.886

.000

473 323565110 2

78574660 00.000 12499598

470

08517442 00.000

0.000 652194 132160

TotalRequests

2

472

Between

00.000 1622.98

Groups Within

9 2313.47

Groups Total

7 3936.46

Between

6 894384

Groups

83925.9 2

Within

43 484780

Groups

11510.3

Total

97 137916

2

811.495

471

4.912

165.212 .000

473

471

44719241 962.971

434.481 .02

10292571 4.459

495436. 473 340

Page 41 of 50

Case Study Phase III AvgRequestsPerSe

Between

179255

c

Groups

44532.2 2

Within

97 113753

Groups

60433.1 471

Total

56 293009

89627722 66.149

371.106 .005

24151508. 351

04965.4 473 54 AvgTimeFirstByte

Between Groups Within

AvgTimeLastByte

6.227 47872.0

Groups Total

68 47878.2

Between

95 117490

Groups Within

6.874 399889

Groups Total

9.936 517380 6.810

NumberOfBytesSen Between t

HttpErrors

Groups Within Groups Total Between Groups Within Groups Total

Table - 12

2

3.114

471

101.639

.031

.014

69.192

.007

13.763

.003

13.763

.011

473 2 471

587453.43 7 8490.233

473

4.482

2

2.241

76.699

471

.163

81.181

473

4.482

2

5.114

76.699

471

125.163

81.181 473 First ANOVA Test

Hypothesis II

Page 42 of 50

Case Study Phase III We can clearly see below that the Sig. (significant) value against all security metrics is less than 0.05 level of significance which means we can reject hypothesis H0 Sec and accept alternate hypothesis H1 Sec Thus, it is inferred that the security metrics in the framework have significant impact on the quality of a web application. ANOVA - TEST Sum of Square SQLinjection

Between Groups Within Groups Total

CrossSiteScripting

Between Groups Within Groups Total

SessionHijacking

Between Groups Within Groups Total

DenialOfService

Between Groups Within Groups Total

Mean

s

df

Square

F

Sig.

.604

2

.302

.377

.032

15.214

19

.801

15.818

21

.286

2

.143

.173

.004

15.714

19

.827

16.000

21

.073

2

.037

.138

.002

5.018

19

.264

5.091

21

.006

2

.003

.014

.021

4.357

19

.229

4.364

21

Page 43 of 50

Case Study Phase III BufferOverflows

Between Groups Within Groups Total

Table – 13

63.955 807.50 0 871.45

5 Second ANOVA Test

2

31.977

19

42.500

.752

.014

21

Hypothesis III As from below table, it is clear that Sig. (significant) value against all usability metrics is less than 0.05 level of significance which means we can reject hypothesis H0 Usab and accept alternate hypothesis H1 Usab Thus, we can infer that the usability metrics in the framework have significant impact on the quality of a web application.

ANOVA - TEST

TrainingTime

Between Groups Within

PorifciencyTime

Groups Total Between Groups Within Groups Total

Sum of

Mean

Squares df

Square

F

Sig.

.286

2

.143

.173

.031

15.714

19

.827

16.000

21

63.955

2

.752

.009

807.500 19

31.977 42.500

871.455 21

Page 44 of 50

Case Study Phase III ErrorRate

Between Groups Within

ErrorRecoveryTime

Groups Total Between Groups Within

DecisionTime

Groups Total Between Groups Within

PositieStatements

Groups Total Between Groups Within

Groups Total NegativeStatements Between Groups Within

ActionsPerTask

Groups Total Between Groups Within

FeaturesNeverUsed

Groups Total Between Groups Within Groups Total

9.657

2

134.161 19

4.829

.684

.006

.931

.002

1.578

.232

.367

.000

.479

.004

.377

.090

.138

.000

7.061

143.818 21 2.255

2

1.127

23.018

19

1.211

25.273

21

16.578

2

8.289

99.786

19

5.252

116.364

21

3.448

2

1.724

89.143

19

4.692

92.591

21

4.294

2

2.147

85.161

19

4.482

89.455

21

.604

2

.302

15.214

19

.801

15.818

21

.073

2

.037

5.018

19

.264

5.091

21

Page 45 of 50

Case Study Phase III TaskCompletionTim

Between

e

Groups Within

Table – 14

.006

4.357 Groups Total 4.364 Third ANOVA Test

2

.003

19

.229

.014

.002

21

Conclusion Based on the study conducted for the web applications in general and CRM applications in particular driven from the Web Metrics Framework , we have drawn following conclusions about the improvement of CRM application for the organization under study as compared to open source CRM applications. 1. It is quite evident that having more adequate measures about performance, security and usability we have more progressive standard towards achieving high Quality of the product. 2. Results show that in the presence of more empirical measures of performance the overall quality of the product is increased the same is the case with usability and security measures. 3. From our studies we come into conclude that Salesforce CRM is more effective in terms of performance, security and usability. Second is the Sourceforge CRM which also shows good results when these measures are tested. Enterprise CRM is at third level in terms of overall Quality and forth one is the Unify CRM from a local software House. 4. Quality Manager of the Software House has been provided with a set of Metrics to improve their overall productivity at individual levels as well as at team performance levels. 5. The metrics framework is not specific to only CRM applications this can also be applied to other web based applications for improving the Quality.

Page 46 of 50

Case Study Phase III Based on the statistical analysis of data collected from the organization, we have concluded that: 1. There is no documentation being done as for as Metrics which are related to the Quality of the product. 2. It is now become clear that why their first three versions of Unify CRM had been failed just because in the absence of proper measures. 3. If these measures are used at appropriate stages of deliverables the Quality can be met at heights in order to remove the defects of the application. 4. The post-releases issues have no relationship with the software defects after release of the software.

Future Work In this study, we presented three hypotheses, which were related to Web metrics frame work model. Next step is to apply the framework on the application, and finally, we have to study the effect of proposed Web Metrics framework on the software overall Quality. This test is the limitation in this study, and in our future studies, we will provide the quality framework analysis on other applications. More work can be done on the other metrics related to the Quality of web applications. We intend to improve upon this frame work and test it on other web applications as well.

Page 47 of 50

Case Study Phase III

References (Revised) [1]

Athula Ginige and San Murugesan, "Web Engineering: An Introduction," IEEE Multimedia, Vol. 8, No. 1, January 2001, pp 14-18

[2]

http://www.abo.fi/~kaisa/ Web Metrics a research Paper by (c) Jukka Heinonen, Marcus Hägert 2004

[3].

Institutional Web Management Workshop 2002: The Pervasive Web http://www.ukoln.ac.uk/web-focus/events/workshops/webmaster-2002/

[4].

A Quality Framework For Web Site Quality www.ukoln.ac.uk/web-focus/papers/www2005

[5]

R.S. Pressman & Associates Inc.

[6]

A Survey of Web Metrics [ Bibliography of Web metrics] http://www.nsdl.comm.org/ National Science Digital Library DEVANSHU DHYANI and NG WEE KEONG & SOURAV S BHOWMICK Nanyang Technological University

[7]

A Concerns-based Metrics Suite for Web Applications Dipartimento di Informatica e Comunicazione Università degli Studi di Milano Via Comelico 39, 20135 Milano, Italy Alessandro.Marchetto@unimi Page 48 of 50

Case Study Phase III

[8]

A Survey of Web Metrics DEVANSHU DHYANI and NG WEE KEONG and SOURAV S BHOWMICK Nanyang Technological University

[9]

http://www.clickz.com/showPage.html?page=992351

[10]

Web Site Usability, Design, and Performance Metrics Jonathan W. Palmer University of Maryland, R. H. Smith School of Business, Decision and Information Technologies, 4348Van Munching Hall, College Park, Maryland 20742-1871 [email protected]

[11] ccnmtl.columbia.edu/projects/qmss/anova_about.html [12] searchsoftwarequality.techtarget.com/generic/0,295582,sid92.html [13] http://shiflett.org/articles/the-truth-about-sessions [14] http://msdn2.microsoft.com/en-us/library/ms998581.aspx [15] http://www.useit.com/jakob/ [16] http://www-306.ibm.com/software/awdtools/appscan/standard/features/?S_CMP=wspace

Web Metrics Related References 1. Website Performance http://www.hurolinan.com/resources/resource.asp?LocatorCode=416 http://searchsoftwarequality.techtarget.com/originalContent/0,289142,sid92_gci12 60130,00.html

Page 49 of 50

Case Study Phase III 2.

Web Security http://www.webopedia.com/TERM/S/security.html http://www.websense.com/global/en/ResourceCenter/Glossary/websecurity.php http://www.arctecgroup.net/pdf/0703-OWASPMetrics.pdf http://www.securitymetrics.org/content/attach/Metricon2.0/Grossman_Metrico n_2.pdf

3

Usability http://www.useit.com/alertbox/20030825.html www.ccs.neu.edu/home/tarase/vita.htm http://sigchi.org/chi97/proceedings/sig/jms.htm

Page 50 of 50

Related Documents

Ieee
May 2020 64
Ieee
July 2020 52
Ieee
November 2019 70
Ieee-via
November 2019 63
Ieee 802
October 2019 55
Ieee Si.docx
April 2020 32