Integrated Product and Process Attribute - Quantitative Model for Software Quality

Similar documents
Software Quality Engineering Courses Offered by The Westfall Team

Software Quality Engineering Courses Offered by The Westfall Team

CONTRIBUTORY AND INFLUENCING FACTORS ON THE LABOUR WELFARE PRACTICES IN SELECT COMPANIES IN TIRUNELVELI DISTRICT AN ANALYSIS

5 CHAPTER: DATA COLLECTION AND ANALYSIS

Enhancing Cost Estimation Models with Task Assignment Information

BUSINESS STRATEGY: USING SHIFT LEFT PRINCIPLES TO MANAGE IT PROJECTS EFFECTIVELY

A Study on Software Metrics and Phase based Defect Removal Pattern Technique for Project Management

CMMI SM Model Measurement and Analysis

The Effect of Managerial Competencies on Employee Engagement in Multinational IT Industries

GUIDE TO THE CHANGES IN PMP simpl learn i

Building a Mathematical Model for Predicting the Cost of the Communication Towers Projects Using Multifactor Linear Regression Technique

Chapter Six- Selecting the Best Innovation Model by Using Multiple Regression

STATISTICS PART Instructor: Dr. Samir Safi Name:

Introduction to Business Research 3

Fuzzy Logic for Software Metric Models throughout the Development Life-Cycle

B.H. Far

SE Effectiveness Leading Indicators. Garry Roedler

Building quality into the software from the. Keeping and. the software. software life cycle

Leveraging Smart Meter Data & Expanding Services BY ELLEN FRANCONI, PH.D., BEMP, MEMBER ASHRAE; DAVID JUMP, PH.D., P.E.

McKinsey BPR Approach

FACTORS AFFECTING JOB STRESS AMONG IT PROFESSIONALS IN APPAREL INDUSTRY: A CASE STUDY IN SRI LANKA

BaiGuo ML4 Journey. Tim Kasse Kasse Initiatives LLC USA +45 (0) Europe NREL

PMP. Processexam.com. PMI Project Management Professional. Exam Summary Syllabus Questions

Book Outline. Software Testing and Analysis: Process, Principles, and Techniques

M. Zhao, C. Wohlin, N. Ohlsson and M. Xie, "A Comparison between Software Design and Code Metrics for the Prediction of Software Fault Content",

Fuzzy Logic based Short-Term Electricity Demand Forecast

SOFTWARE REQUIREMENT SCALING USING FUZZY LOGIC

Comparative study on demand forecasting by using Autoregressive Integrated Moving Average (ARIMA) and Response Surface Methodology (RSM)

A Case Study of Strategic Metrics Use in a CMM- Based Outsourcing Environment

Leading Indicators for Systems Engineering Effectiveness Presentation for NDIA SE Conference October 28, 2009

Personal Software Process SM for Engineers: Part I

Planning for Quality using Defect Prediction Model

A Study of Service Quality of Banks in Mumbai

Chapter 6. Software Quality Management & Estimation

DESIGN AND DEVELOPMENT OF SOFTWARE FAULT PREDICTION MODEL TO ENHANCE THE SOFTWARE QUALITY LEVEL

Applying Regression Techniques For Predictive Analytics Paviya George Chemparathy

APPLICATION OF TIME-SERIES DEMAND FORECASTING MODELS WITH SEASONALITY AND TREND COMPONENTS FOR INDUSTRIAL PRODUCTS

Models in Engineering Glossary

ITIL: Planning, Protection & Optimization

Examining the Test Process: Predicting the Return on Investment of a Process Change

SHRI ANGALAMMAN COLLEGE OF ENGINEERING & TECHNOLOGY (An ISO 9001:2008 Certified Institution) SIRUGANOOR,TRICHY

Service Quality of BRAC Bank in Bangladesh: A Case Study

TEST CASE PRIORITIZATION USING FUZZY LOGIC BASED ON REQUIREMENT PRIORITIZING

ADVANCE: Implementing a Defect Model for Performance Prediction

An Overview of Software Process

Measuring and Improving Process Capabilities Best Practices

KINGS COLLEGE OF ENGINEERING DEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK

CHAPTER 5 RESULTS AND ANALYSIS

New Customer Acquisition Strategy

In Chapter 3, we discussed the two broad classes of quantitative. Quantitative Forecasting Methods Using Time Series Data CHAPTER 5

A Study on Employee Engagement and its importance for Employee Retention in IT industry in India

This resource is associated with the following paper: Assessing the maturity of software testing services using CMMI-SVC: an industrial case study

Comparison of Different Independent Component Analysis Algorithms for Sales Forecasting

Collaborative Project Management A Practical and Effective Mechanism For On time Delivery

Software Project & Risk Management Courses Offered by The Westfall Team

PMBOK Guide Fifth Edition Pre Release Version October 10, 2012

A Comparative Study on the existing methods of Software Size Estimation

PRES The Effects of Software Process Maturity on Software Development Effort


An Automated Approach to Requirement Elicitation Using Stakeholder Recommendation and Prediction Analysis

The CMM Level for Reducing Defects and Increasing Quality and Productivity

Apply Lean Methodology-VSM to Improve IT Regression Testing

Quantitative Code Review based on Weighted Checklist

Selection of a Forecasting Technique for Beverage Production: A Case Study

A Realistic Approach: RTST to Reduce Cost & Time

CMMI and FPA. the link and benefit of using FPA when rolling out CMMI. Christine Green IFPUG - Certified Function Point Specialist EDS

Quantitative Analysis Using Statistics for Forecasting and Validity Testing. Course #6300/QAS6300 Course Material

How mature is my test organization: STDM, an assessment tool

Specification and Prediction of Net Income using by Generalized Regression Neural Network (A Case Study)

TDWI Analytics Principles and Practices

World Journal of Engineering Research and Technology WJERT

Comparing Scrum And CMMI

ANALYSIS OF FACTORS CONTRIBUTING TO EFFICIENCY OF SOFTWARE DEVELOPMENT

Abstract. Keywords. 1. Introduction. Rashmi N 1, Suma V 2. Where, i = 1 requirement phase, n = maintenance phase of software development process [9].

Examples of Statistical Methods at CMMI Levels 4 and 5

USING GEOSTATISTIC ANALYSIS FOR PREDICTION OF SAR IN SOUTH OF IRAN

Informatics solutions for decision support regarding electricity consumption optimizing within smart grids

Study On Suitability And Role Of Business Process Reengineering In Objective Attainment Of Small And Medium Sized Enterprises.

Solutions Manual. Object-Oriented Software Engineering. An Agile Unified Methodology. David Kung

Software Data Analytics. Nevena Lazarević

PREDICTION OF DEFECT DENSITY FOR OPEN SOURCE SOFTWARE USING REPOSITORY METRICS

Breast Cancer Diagnostic Factors Elimination via Evolutionary Neural Network Pruning

Logistic and production Models

Keywords COCOMO model, cost estimation, genetic algorithm, ant colony optimization.

IS VALUE CREATION BE SEGMENTED IN LIFE WELLNESS PROGRAM OF SME S: A COMPARATIVE ANALYSIS

Analysis of Factors that Affect Productivity of Enterprise Software Projects

7. Model based software architecture

Flood Forecasting using Adaptive Neuro-Fuzzy Inference System (ANFIS)

DORNERWORKS QUALITY SYSTEM

Analysis of Customer Satisfaction during Online Purchase

Predicting and Explaining Price-Spikes in Real-Time Electricity Markets


Simulation Analytics

Project Management CTC-ITC 310 Spring 2018 Howard Rosenthal

Financial Forecasting:

Metric systems for executive overview

INTRODUCTION. It is the process used to identify the correctness, completeness and quality of developed computer software.

Comparison of sales forecasting models for an innovative agro-industrial product: Bass model versus logistic function

Define 6. Improve 80. Measure 20. Control 100. Analyse 46. Analyse cont... Hypothesis Testing Correlation & Regression 74-79

Given the competitive importance of

Transcription:

International Journal of Performability Engineering, Vol. 2, No. 3, July 2006, pp. 265-276 RAMS Consultants Printed in India Integrated Product and Process Attribute - Quantitative Model for Software Quality PIYUSH MEHTA, A. K. VERMA and A. SRIVIDYA Interdisciplinary Programme in Reliability Engineering, Indian Institute of Technology Bombay, Mumbai, India (Received on March 10, 2006) Abstract: The software size and complexities are increasing and software project management has become very crucial. The industry demands for consistent software quality within the stipulated cost and time frame. In order to achieve quality, cost and schedule target, quantification and prediction of these attributes early in the life cycle of development have become an important problem. Software does not manifest its quality properties directly; instead it is exhibited through certain contributory measures of the process steps and intermediate work products. Attempts have been made in past by various researchers to correlate the product and process attributes, however these type of modeling is done only for subset of the attributes spanning to one and two software development phase[1]. In the present paper Integrated Product and Process Attribute - Quantitative Model (IPPA-QM) is proposed which is based on the relationship of elemental product and process attributes. Prediction equations are developed for the software development phases. Requirements phase equations are instantiated using quantitative techniques. IPPA-QM provides the holistic view of the product and process attributes throughout the software development life cycle by applying various quantitative techniques. IPPA- QM enables prediction based planning and corrective action early in the development lifecycle thereby improving the execution capability of the IT organization and achieving quality, cost and schedule targets. Key Words: software quality, product attribute, process attribute, software metrics taxonomy, IPPA-QM 1. Introduction The software size and complexities are increasing and software project management has become very crucial. The industry demands for consistent software quality within the stipulated cost and time frame. In order to achieve quality, cost and schedule target, quantification and prediction of these attributes early in the life cycle of development have become an increasingly important problem. A methodology for building model, which relate product and process attributes of the software is needed. Benefits of quantitative prediction of the product attributes and process attributes are two fold: first it will improve the process, second it will lead to the better project Communicating author s email: piyush.mehta@iitb.ac.in 265

266 Piyush Mehta, A. K. Verma and A. Srividya management. Improved software process and better project management is needed in order to deliver high quality product, at low cost, at agreed time and schedule with high customer satisfaction. IPPA-QM is devised to present the holistic view of process and product relationship and is based on the quantitative techniques. Section 3 describes the Software Metrics Taxonomy (SMT) which provides foundation for the IPPA-QM. Section 4 describes the typical lifecycle model, model premise and graphical representation of model. Section 5 represents the phase wise prediction equations for quantitative modeling. Section 6 explains the instantiation of the model for the requirements phase; Section 7 summarizes the work and describes the direction for future research. 2. Notation S Normalized Phase Phase TS Phase D Phase S Phase DD Phase Project Size Normalized Phase Effort Phase Duration Phase Team Size Phase Defects Phase Product Size Phase Defect Density Where Phase equals to Requirements, Design, Construction and Testing 3. Software Metrics Taxonomy A very large number of software metrics are suggested by various researchers in different contexts [2] [3] [4] [5]. Practitioners/new project managers are often confused by various software metrics proposed. They are not able to understand the importance/benefits of these metrics and hence are not able to choose the appropriate metrics set for the project. The classification scheme of the software metrics is needed in order to understand what practices are prevalent in the industry. After classifying the metrics in the various categories and creating the precise definition, the data collection can be done. This promotes the understanding of the software measures and will aid in the data analysis and interpretation. The metrics reported by the researchers are used to find the generic classification attributes as shown in Table 1. These generic classifiers form the basic building blocks for the software metrics. Generic classifiers provide the clear understanding of the underlying elements of the proposed software metrics and their interrelationship. Software Metrics Taxonomy (SMT) [6] is generated by using the generic classifier attribute and representing their relationship in the form of logical diagram. This logical diagram is the abstraction of the software metrics classification. It clearly distinguishes the product and process attribute in the top two tables from the rest of segmentation factors. The elemental attributes will be used for predicting the downstream product and process characteristics. Thus SMT provides the definitive foundation for the IPPA-QM model.

Integrated Product and Process Attribute - Quantitative Model for Software Quality 267 Table 1: Generic Classification of the Software Attributes Type of Measure Product Process Base Measures Size Effort Complexity Duration Defects Review Competency Derived Segmentation Factors Productivity Cost Technology Type 3GL 4GL Product Type Project Type Project Management Tools (Debugging, IDE) Process Type Software Engineering Process Management Process Support Process Phase wise classification Project Type Project Management Tools (CASE, Process Workbench) PRODUCT BASE MEASURE DTL PRODUCT BASE NM SIZE COMPLEXITY DEFECTS PROCESS BASE MEASURE DTL PROCESS BASE NM EFFORT DURATION REVIEW COMPETENCY TECHNOLOGY TYPE SEG FCT DTL TECHNOLOGY TYPE NM 3GL 4GL PRODUCT TYPE SEC FCT DTL PRODUCT TYPE NM SRS HLD LLD CODE USER MANUAL OTHER DOCUMENT REUSE PHASE SEG FCT DTL PHANSE NM REQUIREMENT ANALYSIS IMPACT ANALYSIS DESIGN REVIEW EFFORT SYSTEM TESTING DEPLOYMENT PROJECT MANAGEMENT PRODUCT TYPE - BASE MEASURE DTL PROCESS TYPE SEC FCT DTL PROJECT MANAGEMENT NM PRODUCT BASE NM (FK) PROCESS TYPE NM ATTRIBUTE NM (FK) PLANNED VALUES PRODUCT TYPE NM (FK) ACTUAL VALUES ENGINEERING SUPPORT VARIANCE PROCESS MANAGEMENT PROJECT MANAGEMENT NM (FK) TECHNOLOGY TYPE NM (FK) PROJECT MANAGEMENT ATTRIBUTES DTL TOOLS DTL ATTRIBUTE NM PROJECT TYPE SEC FCT DTL TOOLS NM NO OF DEFECTS NO OF HRS PROJECT TYPE NM DEBUGGING DEVELOPMENT IDE MAINTENANCE. CASE PROCESS WORKBENCH PHASE-PROCESS BASE DTL PHANSE NM (FK) ATTRIBUTE NM (FK) PROCESS BASE NM (FK) PROJECT MANAGEMENT NM (FK) PROCESS TYPE-BASE MEASURE DTL TECHNOLOGY TYPE NM (FK) PROCESS TYPE NM (FK) ATTRIBUTE NM (FK) TOOLS - EFFORT DTL PROCESS BASE NM (FK) TOOLS NM (FK) ATTRIBUTE NM (FK) PROCESS BASE NM (FK) PHASE-PROCESS BASE - PROJECT TYPE DTL TECHNOLOGY TYPE NM (FK) PHANSE NM (FK) PROJECT MANAGEMENT NM (FK) PROCESS BASE NM (FK) ATTRIBUTE NM (FK) PROJECT TYPE NM (FK) PROJECT MANAGEMENT NM (FK) 4. IPPA-QM Premise Fig.1: Software Metrics Taxonomy (SMT) The transformation of high level business requirements to the fully functional software is done in various phases. Work products are created out of life cycle phase are subjected to verification and validations. Measures of the intermediate products and process can explain the attribute of next phases. Figure 2 represents the typical lifecycle model followed during the software development. It also depicts the output of life cycle phases, verification and validation steps and associated product and process measures.

268 Piyush Mehta, A. K. Verma and A. Srividya Life Cycle Phases Outputs V&V Steps Typical Attribute/Measures Requirement Analysis Software Requirements Specification Review Product Attribute: Specification Size Specification Complexity Review Defects Process Measure: Effort, Duration, Review Effectiveness,Review Competency Design (High Level Design & Low Level Design) High Level Design Document Low Level Design Document Relationship Data Model Design Flow Diagram Class Diagrams Review Product Measure: Size Complexity Review Defects Process Measure: Effort, Duration, Review Effectiveness,Review Competency Construction Code Test Cases Review Product Measure: Code Size (LOC,FP, UCP) Code Review Defects Test Defects Process Measure: Effort, Duration, Review Effectiveness,Review Competency Testing Unit Integration User Acceptance Tesing Fixed Codes Test Product Measure: Code Size Test Coverage Process Measure: Test Effectiveness Test Effort & Duration Testing Competency Product Measure: Code Size Test Defects / Size Implementation Roll Out Implemented Code (Live) User Feedback Process Measure: Effort, Duration, Review Effectiveness (Final), Test Effectiveness (Final) Typical Life Cycle Model with phases, outputs, v&v steps and measures * This is a sample model & varies within and outside organization Fig. 2: Typical Life Cycle Model The product and process attributes in the life cycle are correlated. The early prediction of product and process attributes leads to corrective actions and is less costly to implement. The downstream predictions are costlier to implement. Correlation between product and process attributes leads to dependable prediction model. Based on above premise, IPPA-QM (shown in Figure 3) is proposed in this paper. IPPA-QM is an end to end model which correlates the attributes in various software development phases. It uses various quantitative techniques to determine the relationship between the product and process attributes. This early prediction of the attributes leads to

Integrated Product and Process Attribute - Quantitative Model for Software Quality 269 better planning, control and management of the software project. It also increases the execution capability of the organization. P1- Product Attribute Goals* P1 P2 P3 Delivered Product Predictio n Model Pred- P1 Act- P1 Predictio n Model Pred - P2 Act- P2 Prediction Based Planning and Corrective Action Predictio n Model Pred- P3 Act- P3 Predictio n Model Prediction Model Predictio n Model Actual Product Attributes Fine tuning (leanings) for future projects Predictio n Model Fig. 3: Integrated Product and Process Attributes Quantitative Model (IPPA-QM) 5. IPPA-QM Details IPPA-QM model takes into account the various product and process attributes available at the start of life cycle phase. These attributes are used to estimate the attributes for the later life cycle phases. Following section represents the phase wise prediction equations. These equations are of skeleton type and exact selection of the parameters will vary with organizational environment, policies, processes and prevailing practices. 5.1 Project Start Software project starts with the feasibility analysis. The management authorization for the approval of the project is given based on the high level project cost estimates. The cost estimates are based on high level estimate of the size of the project. Thus available attributes at the start of the project is project size ball park estimate. Requirements Phase Effort Estimation: The predicted size can be used to estimate the requirements phase effort. The size can be in terms of function points or number of components. The requirements effort is in person days. The prediction equation is represented as Requirements = f (S Normalized ) (1)

270 Piyush Mehta, A. K. Verma and A. Srividya Requirements Phase Duration Estimation: Duration of the requirements phase is dependent on the size of the product to be developed and team size. Size of the product is directly proportional to the duration of requirements phase and team size is inversely proportional to the duration. It is important to note that the above relationship is heterostedistic in nature. For example increasing the team size to a threshold value will certainly decrease the duration to a certain extent, after that the duration will increase. This behavior can explained in terms of the overheads associated with the communications, project management, work decomposition, integration effort. The following equation is used to determine the requirements duration based on the predicted size and the team size of requirements phase. Requirements = f (S Normalized, TS Requirements ) (2) Requirements Team Size Estimation: The duration of requirements phase is usually a constraint for the project. Team size need to be estimated based on the duration and size of the project. The following equation is used to determine the team size. TS Requirements = f (S Normalized, Requirements ) (3) Requirements Defect Prediction: The number of defects found in requirements inspection depends on S Requirements, Requirements, TS Requirements, Requirements Thus the prediction equation for requirements defects is as follows D Requirements = f (S Requirements, Requirements, TS Requirements, Requirements ) (4) 5.2 Requirements Analysis End Feasibility phase is followed by the requirements analysis phase. In requirements analysis phase the requirements are elicited from the various stakeholders and requirements specification is developed. The requirement specification is reviewed in order to detect any defects. Defects found in requirement reviews are corrected and is followed by verification of defect closure activity. Requirement phase is followed by design phase. The attributes available at requirement end can be used to predict design phase attributes [7]. Design Effort Prediction Model: Design effort can be predicted by project size, requirements specification size, requirements effort, requirements defects. Thus the prediction equation for design effort is Design = f (S Normalized, S Specification, Requirements, D Requirements ) (5) Design Duration Prediction Model: Design duration is predicted based on the S Specification, Requirements, Requirements, TS Requirements, TS Design. Thus design duration prediction equation is Design = f (S Specification, Requirements, Requirements, TS Requirements, TS Design ) (6) Assumption: Design Team Size is constraint in the organization and depends on the availability of the resources in the organization Design Team Size Estimation:The duration of design phase can be a constraint for the project. Design team size can be estimated based on the duration and size of the project. TS Design = f (S Specification, Requirements, Requirements, TS Requirements, Design ) (7)

Integrated Product and Process Attribute - Quantitative Model for Software Quality 271 Design Size Prediction Model: Design size is dependent primarily on the specification size and the specification effort. Thus design size can be predicted as S Design = f (S Specification, Requirements ) (8) Design Defect Prediction: The number of defects found in design inspection depends on DD Requirements, S Design, Design, TS Design, Design Thus the prediction equation for design defect is D Design = f (S Design, Design, TS Design, Design, DD Requirements ) (9) The best set parameters which can predict the design defect need to be identified and will vary based on environmental parameters, organizational policies and processes. 5.3 Design End Software construction follows the design phase. The attributes available at design end are used to predict the construction phase attributes. Construction phase prediction equations are as follows. Construction Effort Prediction Model: Construction effort is dependent on the attribute like S Specification, S Design, Requirements, Design Construction Effort = f (S Specification, S Design, Requirements, Design ) (10) Construction Duration Prediction Model: Duration of the construction phase is dependent on S Specification, Requirements, Requirements, TS Requirements, S Design, Design, Design, TS Design, TS Construction Assumption: Design = f (S Specification, Requirements, Requirements, TS Requirements, S Design, Design, Design, TS Design, TS Construction ) (11) 1. Construction Team Size is constraint in the organization and depends on the availability of the resources in the organization 2. Parameter pruning need to be done to improve the dependability of the model 3. Derived measures can be used in the equation for example, Specification Size and Specification Effort can be replaced by specification productivity Construction Team Size Estimation:The duration of construction phase can be a constraint for the project. The construction team size is estimated based on the following equation TS Construction = f (S Specification, Requirements, Requirements, TS Requirements, S Design, Design, Design, TS Design, Design ) (12) Construction Size Prediction Model: Construction or code size is dependent on S Specification, S Design Thus the code size prediction is as follows S Code = f (S Specification, S Design ) (13) Code Review Defect Prediction: The number of defects found in code review depends on DD Requirements, DD Design, S Code, Code, TS Code, Code Thus the code review defects is predicted as

272 Piyush Mehta, A. K. Verma and A. Srividya D Code Review = f (S Code, Code, TS Code, Code, Code Review, Code Review, TS Code Review, DD Requirements, DD Design) (14) 6. IPPA-QM Instantiation The proposed IPPA-QM was instantiated for all the phases. The model was instantiated based on the dataset collected from CMMI level 5 maturity organizations. The data was collected from the same organization representing the homogeneous development environment. The projects were commercial business systems projects. The data was randomly divided into training set comprising 100 project data and validation set comprising 19 projects. Regression and Artificial Neural network were two principal techniques used to develop the organization specific prediction equations. Different regression equation variants like enter, backward selection, forward selection and stepwise were fitted in order to find the best fit regression equation. Neural network techniques were used as alternatives to the regression fit. Regression equation was applied to the 19 projects of the validation set. The predicted attributes were compared with the actual attribute values. The accuracy of the prediction equation was gauged by calculating standard error [8] as shown in equation no. 15-18. Neural network model built is used to predict the same 19 projects of the validation set. The errors are calculated for the neural network predictions and compared with regression equations for determining the accuracy of method in the organizational context. Mean Absolute Error (MAE): 1 n MAE actualvaluei forecastedvaluei n i 1 Where the N is the number of the forecasted values. Mean Squared Error (MSE): 1 n MSE ( actualvalue forecastedvalue ) 2 n i i i 1 Root Mean Square Error (RMSE): (15) (16) n 1 RMSE ( actualvalue forecastedvalue ) 2 i i n i 1 Mean Absolute Percentage Error (MAPE): (17) 1 n actualvalue i forecastedvalue MAPE i n i 1 forecastedvaluei Requirements phase effort estimation: The normalized size is used to estimate the requirements phase effort. The normalized size is measured in terms of components. Linear regression model is used to predict the requirements phase effort. The model summary, ANOVA and coefficients of predictors are as follows. (18)

Integrated Product and Process Attribute - Quantitative Model for Software Quality 273 Table 2: Model Summary Requirements Phase Effort Model R R Square Adjusted R Square Std. Error of the Estimate 1 0.925 0.855 0.854 27.86604 Table 3: ANOVA Requirements Phase Effort Sum of Df Mean Square F Sig Squares Regression 449973.6 1 449973.565 579.477 0.000 Residual 76098.595 98 776.516 Total 526072.2 99 Table 4: Regression Coefficients Model Unstandardized Coefficients Standardized Coefficients t Sig B Std. Error Beta Constant 43.163 6.192 6.970 0.000 Size 1.181 0.049 0.925 24.072 0.000 The above model is statistically significance and the regression equation for requirements phase effort estimation is as follows Requirements = 43.163 + 1.181 (S Normalized ) (19) The regression equation is applied to the validation dataset to predict the requirements phase effort. Generalized Regression Neural Network (GRNN) is applied to the 100 project data set. The trained network is applied on the validation dataset to predict the requirements phase effort. Error of prediction by both the method is calculated. Table 5: Error Comparison of Requirements Effort Prediction Model Method Name MAE MSE RMSE MAPE Linear Regression Generalized Regression Neural Network 20.65938 579.0549515 24.06356 0.1349 18.63677 554.7918675 23.55402 0.1175 Requirements Team Size Estimation: The duration of requirements phase is constraint for the projects. The requirements team size is estimated based on the duration and size of the project. Best fit model summary, ANOVA and regression coefficients are shown in Table 6 through 8. Table 6: Model Summary Requirements Team Size Model R R Square Adjusted R Square Std. Error of the Estimate 1 0.732 0.535 0.526 0.44854

274 Piyush Mehta, A. K. Verma and A. Srividya Table 7: ANOVA Requirements Team Size Sum of Df Mean Square F Sig Squares Regression 22.484 2 11.242 55.878 0.000 Residual 19.516 97 0.201 Total 42.000 99 Table 8: Regression Coefficients Model Unstandardized Coefficients Standardized Coefficients t Sig B Std. Error Beta Constant 2.280 0.148 15.372 0.000 Size 0.12 0.001 1.069 10.258 0.000 Duration_Req -0.17 0.003-0.623-5.975 0.000 TS Requirements = 2.280 + 0.12 * S Normalized -0.17 * Requirements (20) Table 9: Error Comparison of Requirements Team Size Prediction Model Method Name MAE MSE RMSE MAPE Multiple Regression 3.476824 20.33709016 4.509666 0.6606 Generalized Regression Neural Network 0.224492 0.066087054 0.257074 0.0973 Requirements Defect Prediction: Multiple regression models with method enter, stepwise, remove, backward and forward is used to predict the requirements defects. The following models were statistically significant out of various models generated. Table 10: Model Summary Requirements Defects Model R R Square Adjusted R Square Std. Error of the Estimate 1 0.871 0.758 0.755 8.184579 2 0.876 0.768 0.763 8.05074 1. Independent Variable: Size_req 2. Independent Variable: Size_req, Team_Size_req Table 11: ANOVA Requirements Defects Model Sum of Df Mean Square F Sig Squares 1 Regression 20546.298 1 20546.298 306.704 0.000 Residual 6565.092 98 66.991 Total 27111.390 99 2 Regression 20824.394 1 10412.197 160.646 0.000 Residual 6286.996 98 64.814 Total 27111.390 99

Integrated Product and Process Attribute - Quantitative Model for Software Quality 275 Table 12: Regression Coefficients Model Parameters Unstandardized Coefficients Standardized Coefficients t Sig B Std. Error Beta 1 Constant 7.095 2.729 2.600 0.011 Size_Req 0.359 0.021 0.871 17.513 0.000 2 Constant 3.106 3.304 0.940 0.349 Size_Req 0.334 0.024 0.808 14.085 0.000 Team_Size_Req 3.020 1.458 0.119 2.071 0.041 D Requirements = 7.095 + 0.359 * S Requirements (21) D Requirements = 3.106 + 0.334 * S Requirements + 3.020 TS Requirements (22) The GRNN model was trained on all the above parameters. The above two regression models and GRNN model was applied to the validation dataset. The error comparison for the is mentioned in table Table 13: Error Comparison of Requirements Defect Prediction Models Method Name MAE MSE RMSE MAPE Multiple Regression 5.858185 54.28470049 7.367815 0.135 Model I Multiple Regression Model II Generalized Regression Neural Network 5.863852 52.43206981 7.240999 0.1335 5.43487 41.8059925 6.465755 0.1269 Thus it is observed that predictions of the various requirements phase attributes can be done by means of regression or neural network model. The prediction accuracy of the neural network model is better than the traditional regression methods in the current organizational data context. 7. Conclusions The research work has collated the various metrics suites proposed by the researchers and industry practitioners and has arrived at Software Metrics Taxonomy which is pictorial representation of product and process attributes. SMT simplifies the understanding of the state of the art measurement in Software Engineering and forms the foundation for developing the IPPA-QM. Proposed IPPA-QM demonstrates the prediction equation for Effort, Size, Duration, Team Size and Defects for phase starting from Requirements through Construction phase. These prediction equations can be instantiated based on the organizational data. Multiple regression and Neural Network techniques were used to demonstrate the development and application of prediction equations for a high maturity software development organization. Organization specific prediction equations are useful for prediction based planning and corrective actions early in the development phases.

276 Piyush Mehta, A. K. Verma and A. Srividya IPPA-QM provides prediction of software attributes based on the past performance trend and current software project state. These attribute prediction can be used to develop integrated software reliability model. IPPA-QM technique set can be extended to include emerging techniques like genetic algorithm and fuzzy set. This sets the direction for future research in this area. References [1]. Pregibon, D., H. Chernoff, et al., Chapter 4: Critique of Some Current Applications of Statistics in Software Engineering, Report on Statistical Software Engineering, National Academy of Sciences, 1996. [2]. Burr A and M. Owen M., Statistical Methods for Software Quality: Using Metrics for Software Process Improvement, Boston, MA: International Thomson publishing, 1996. [3]. Daskalantonakis M. K, A Practical View of Software Measurement and Implementation Experiences within Motorola, IEEE Transactions on Software engineering, Vol. 18, No. 11, pp. 998-1010, Nov 1992. [4]. Florac W. A, A. D. Carleton, Measuring the Software Process: Statistical Process Control for Software Process Improvement, MA: Addison Wesley, 1999. [5]. Kan S. H, Metrics and Models in Software Quality Engineering, Addison Wesley, 1995. [6]. Mehta Piyush, A. Srividya, A. K. Verma, Software Metrics Taxonomy for Metrics Guided Software Development, Proceedings of International Conference on Reliability, Safety and Hazard, Narosa Publications, pp. 209-213, 2005. [7]. Verma A. K, Piyush Mehta, A. Srividya, A Framework for Design Phase Prediction using Integrated Product and Process Attribute Approach, International Workshop on Future Software Technology 2005, China, 2005. [8]. Walsh, D. M., G. Y, -G Tsou, Forecasting Index Volatility: Sampling Interval and Non-Trading Effects, Applied Financial Economics, Vol. 8, No. 5, 477-486, 1998. Piyush Mehta received his Master of Engineering degree in Production from Government Engineering College Pune in 2000. He is currently pursuing his research in the area of software quality prediction at IIT, Bombay. He performed as project leader for numerous software development projects in India and Abroad. He has also executed several quality consulting engagements in the field of IT. He has published papers in international conferences and his research interests include software quality prediction, estimation, productivity and process improvement. A. K. Verma is a Professor in Reliability Engineering in the department of Electrical Engineering, Indian Institute of Technology, Mumbai. He has published about 100 research papers in journals and conferences. He is also on the editorial board of various international journals. He is a senior member of IEEE and Fellow (life) of IETE. A. Srividya is an Associate Professor in Reliability engineering. She has published many papers in the areas of quality and reliability and her research interests include application of quality tools and techniques in software and the service sector. She has been a co-editor (guest) for the special issue of IETE technical review on quality management. She has been the chairperson and co-chairperson of various International Conferences.