Thomas Michael Lintner, Steven D. Smith, Scott Smurthwaite / Paper for the 2009 ISAHP

Similar documents
The Aerospace Performance Factor (APF)

The goal of the Flight Operations Risk Assessment System (FORAS)

Proposal for using AHP method to evaluate the quality of services provided by outsourced companies

Use of AHP Method in Efficiency Analysis of Existing Water Treatment Plants

Cloud adoption decision support for SMEs Using Analytical Hierarchy Process (AHP)

CANSO Global Approach to Safety. Sam Espig NATS

An Automated System for Prioritizing Highway Improvement Locations and Analyzing Project Alternatives

Project Scope Management

A Decision Support System for Performance Evaluation

A RANKING AND PRIORITIZING METHOD FOR BRIDGE MANAGEMENT

Analytic Hierarchy Process, Basic Introduction

VENDOR RATING AND SELECTION IN MANUFACTURING INDUSTRY

SESAR The European ATM Improvement Programme. Regional ANC 2012 Preparatory Symposium Michael STANDAR Moscow March 2012

ANALYSIS OF THE EFFICIENCY OF BANKS IN MONTENEGRO USING THE AHP

SELECTING A PROFOUND TECHNICAL SERVICE PROVIDER TO PERFORM A TECHNICAL FIELD DEVELOPMENT STUDY FROM GIVEN MULTIPLE CRITERIA

6.4 ACCEPTABLE LEVEL OF SAFETY (ALoS) (SMM Chapter 6.4, 2nd edition 2009)

Supplier Selection Using Analytic Hierarchy Process: An Application From Turkey

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 9, March 2014

Collaborative Decision Making (CDM) Update

Topic areas for the PEGASAS COE Summer Experience for 2018

OJT INSTRUCTOR. Learning outcomes. Why study this course? Aim. ICAO Code 212

Preventing, reporting, evaluating, and mitigating Airside accidents and incidents at airports.

ATC BASIC. Learning outcomes. Why study this course? Aim. ICAO Code 051

ERP Reference models and module implementation

Leader Culling Using AHP - PROMETHEE Methodology

Airport Collaborative Decision Making Enhancing Airport Efficiency

Score. Developing a Lean Performance PERFORMANCE MEASUREMENT. Here s a way to track the progress of your lean journey.

NAS-Wide Performance: Impact of Select Uncertainty Factors and Implications for Experimental Design

SESAR: a European initiative for an enhanced ATM system

First European Air traffic controller Selection Test (FEAST) package. Information to Candidates applying as Student Air Traffic Controllers

The Multi criterion Decision-Making (MCDM) are gaining importance as potential tools

Determining and ranking essential criteria of Construction Project Selection in Telecommunication of North Khorasan-Iran

WebShipCost - Quantifying Risk in Intermodal Transportation

Evaluation method for climate change mitigation instruments

Application of AHP for Lean Implementation Analysis in 6 MSMEs

Modeling ATM day to day operations

Commission hearing on the preparation for RP3

Airport Collaborative Decision Making (A-CDM) Safety Case Guidance Material

TITLE -SUPPLIER SELECTION: FUZZY-AHP APPROACH

Joint JAA/EUROCONTROL Task- Force on UAVs

EUROPEAN COMMISSION DIRECTORATE-GENERAL FOR ENERGY AND TRANSPORT MANDATE TO CEN/CENELEC/ETSI FOR THE DEVELOPMENT OF

Training Crewmembers in the Soft Skills

ESTABLISHING RELATIVE WEIGHTS FOR CONTRACTOR PREQUALIFICATION CRITERIA IN A PRE-QUALIFICATION EVALUATION MODEL

Quizzes for 1 st Study Group Session

Selecting an appropriate forestry extension model for the Zagros area in Iran through the AnalyticHierarchy Process (AHP)

START Saratoga Taking Action and Report Tool Help Document

RISK MANAGEMENT IN PRODUCT DEVELOPMENT CURRENT METHODS

INTERNATIONAL JOURNAL OF INDUSTRIAL ENGINEERING RESEARCH AND DEVELOPMENT (IJIERD)

WORKGROUP-LEVEL OVERVIEW. What You Will Learn. What You Will Apply To Your Workgroup

METHODOLOGY FOR SAFETY RISK ASSESSMENT IN FUTURE AIR TRAFFIC MANAGEMENT CONCEPT OF OPERATIONS

Over the last fifty years or so,

APPLICATION OF ANALYTIC HIERARCHY PROCESS FOR STRATEGIC PLANNING AND IMPLEMENTATION AT NEPALESE UNIVERSITIES AND COLLEGES

Measuring En Route Separation Assurance Performance. Ella Pinska, Brian Hickling

Federal Aviation Administration Safety Management System Manual

Presentation of the main functionalities. SafeGround. July Safety at airside ground movements

Big Data Example of SUR Data Application

Guidance for Modeling Noise for Non-Standard Aircraft Profiles

Group Decision Support System for Multicriterion Analysis A Case Study in India

Safety Risks in an Airworthiness Organisation

Measuring Safety Performance

RELEVANCE OF STRATEGIC MANAGEMENT IN ICT BASED SMALL AND MEDIUM ENTERPRISES

Master Plan edition 2

MAINTENANCE COST OPTIMIZATION FOR THE PROCESS INDUSTRY

ANALYTIC HIERARCHY PROCESS IN SELECTING BEST GROUNDWATER POND

Engagement Portal. Employee Engagement User Guide Press Ganey Associates, Inc.

IHST Global Safety Survey Results

A Stochastic AHP Method for Bid Evaluation Plans of Military Systems In-Service Support Contracts

The Search for Shared Situational Awareness for Enhanced Collaboration in Airport Systems

Best Practices in Dashboard Design

Gulfstream SMS. Safety Management International Collaboration Group Meeting Seattle - October 25, Fred Etheridge / Rick Trusis / Carmen Schooley

20. Alternative Ranking Using Criterium Decision Plus (Multi-attribute Utility Analysis)

CAHIER DE RECHERCHE n E4

PRIORITIZATION OF INFRASTRUCTURE INVESTMENT FOR RAIL SAFETY PROJECTS: A CORRIDOR-LEVEL APPROACH

RESEARCH ON DECISION MAKING REGARDING HIGH-BUSINESS-STRATEGY CAFÉ MENU SELECTION

AN AHP APPLICATION IN VENDOR SELECTION

Analytic Hierarchy Process (AHP) in Ranking Non-Parametric Stochastic Rainfall and Streamflow Models

7 9 May 2012, Dubrovnik, Croatia 2 nd International Conference on Road and Rail Infrastructure

Risk Management for Remotely Piloted Aircraft Systems

ISE480 Sequencing and Scheduling

Criteria For Selection of Software Development Environment For Construction Robotic Systems

Air Traffic Management Capacity-Driven Operational Concept Through 2015

Gulfstream Flight Test Safety Management System. Evaluation Tool Guidance

Enhancing the role of Human Factors in Safety Investigation and Analysis

Multi-criteria decision making for supplier selection using AHP and TOPSIS method

Clean Sky 2 Joint Undertaking. 5 th Call for Proposals (CFP05): List and full description of Topics. Call Text. - November

Decision Analysis Part 4: Multi-Attribute Models. Motivation for Multi-Attribute Modeling. ID s and Multiple Objectives. Questions.

A study on Supply Chain issues of Indian Railway in Samastipur division using AHP technique

Capacity Planning and Assessment Additional considerations

Chapter 3. Integrating AHP, clustering and association rule mining

Curbfront Case Studies

The International Pilot Training Consortium (IPTC) IPTC Case Study: Pilot Competencies

Multi Criteria Decision Analysis for Optimal Selection of Supplier Selection in Construction Industry A Case Study

APPLY THE ANALYTIC HIERARCHY PROCESS METHOD FOR SELECTING SUPPLIER IN AUTOMOTIVE BUSINESS

Transforming Projectile System Combining Lethality and Intelligence

ANALYTICAL HIERARCHY PROCESS BASED ON DATA FLOW DIAGRAM

Combining AHP and TOPSIS Approaches to Support Site Selection for a Lead Pollution Study

Evaluation Models. Multi-Attribute Models. Multi-Attribute Modelling: Why? Multi-Attribute Model Structure. Decision Support: Multi-Attribute Methods

Simulation Design Approach for the Selection of Alternative Commercial Passenger Aircraft Seating Configurations

Human Reliability Assessment In ATM: the CARA tool

NextGen Implementation Plan An Update

Transcription:

THE AEROSPACE PERFORMANCE FACTOR: UTILIZATION OF THE ANALYTICAL HIERARCHY PROCESS TO DEVELOP A BALANCED PERFORMANCE AND SAFETY INDICATOR OF THE NATIONAL AIRSPACE SYSTEM FOR THE FEDERAL AVIATION ADMINISTRATION Thomas Michael Lintner Office of Safety Federal Aviation Administration Washington, D.C., USA E-mail: thomas.lintner@faa.gov Steven D. Smith Office of Safety Federal Aviation Administration Washington, D.C., USA E-mail: steven.d.smith@faa.gov Scott Smurthwaite, PhD Office of Safety Federal Aviation Administration Washington, D.C., USA E-mail: scott.smurthwaite@faa.gov ABSTRACT The central mission of the Federal Aviation Administration (FAA) is the safety of the national airspace system (NAS). Historically, basic metrics such as traffic counts, delays, flight cancellations, incidents, and accidents were used to gauge performance. This method of analysis was effective at looking at isolated events but failed to take into account an overall system-wide assessment of the national airspace system. Currently, the Office of Safety is developing the Aerospace Performance Factor (APF) in order to better understand the overall system-wide events and their relative contribution to the overall safety of the national airspace. In this process, the APF utilizes aspects of the Analytical Hierarchy Process (AHP) method to establish structure and create a measurement of historical FAA aerospace incidents. It presents the results of the measurement as a graphical representation of the system s performance over time. The APF is an effective tool for assessing the relative performance of the FAA s airspace and has been adopted by other groups within the aviation arena including the European-based commercial airline easyjet, the Irish Aviation Authority (IAA), the United Kingdom's National Air Traffic Services (NATS), and the European Organisation for the Safety of Air Navigation (EUROCONTROL). Keywords: Aerospace Performance Factor, Analytical Hierarchy Process, national airspace system, Office of Safety, Federal Aviation Administration Corresponding author 1

1. Traditional methods of assessing performance Generally, air navigation service providers (ANSP s) have relied on isolated metrics to assess the performance of their systems. Information such as annual traffic counts and annual flight delays have traditionally been used as a method of quantifying and reporting on the system s performance. These reports might state that in 2008 Airport-X supported x-amount of flights and had y-amount of total delays, resulting in z-amount of total minutes of delays. Then when compared to 2007, the results for the same airport might show a 13% increase in overall traffic with a 10% increase in delays for an overall increase of z-amount of minutes in delays. Although these reports were effective the overall efficiency and general performance trend was unclear. Other views of reporting performance were also effective for individual lines of business but were still less able to view system-wide impact. A sample report might state that in 2008 there were x-amount of mishaps and when compared to 2007 there was a 3% increase while traffic increased by 12%. These snapshot reports, which have traditionally been used by the FAA and other international ANSP s, provide a quick view of important portions of air traffic control systems but do not fully take into account other factors that contribute to a better system-wide view of the overall program. Despite the narrow scope of the reporting, individual incident or mishap reports contain substantial details. For example, a report of a specific event might contain extensive details on the circumstances of the incident. A severity rating associated with assessing the risk of a mishap would be generated using vector mathematics. Similar details were gathered on other incident types. If a truck was inadvertently driven onto a runway, a runway incursion (RI) incident report would be filed and as part of the details of the incident a rating of A, B, C, or D would be assigned to the report with a classification of A being most significant and D being the least significant. These individual reports on the performance aspects of the air traffic control (ATC) system were effective for specific aspects of the system but were not aggregated into a comprehensive report that would highlight a system-wide view of the performance of the NAS. 2. The need for multiple safety indicators In order to better assess the overall system-wide performance, it was recognized that all relevant safety reports would need to be integrated into a single unified measurement. By including all reports (including subject matter expert s assessments) all data could be included in the measurement. By following this approach, where all data are placed on the table for all to see, a comprehensive multi-factor indicator could be developed which might identify current performance, long-term trends, and ultimately allow for the better management of the overall system. Recognizing there would be numerous issues associated with creating new reporting criteria, and knowing the importance of historical data to support trending, a decision was made to utilize all of the traditional safety reports. The spectrum of traditional data collected, and used within the APF, included the following: Operational Errors (OE), where less than the prescribed separation exists between two aircraft. (OE s are subdivided as Category A, B, C, D, PE, MVA, and Flight Formations, depending on certain criteria.) Operational Deviations (OD), where an aircraft infringed on an area of airspace. Pilot Deviations (PD), where the pilot deviated from a rule or ATC instruction. Runway Incursions (RI), where there is an unauthorized aircraft, vehicle, or person on a runway. (RI s are subdivided as Category A, B, C, and D depending on certain criteria.) Near Midair Collisions (NMAC) where a pilot has reported an encounter with another aircraft that the pilot considered potentially unsafe. 2

Vehicle and Pedestrian Deviations (VPD), a type of RI involving the unauthorized presence of a vehicle or person on a runway. By capitalizing on existing safety reports which spanned various areas of ATC operations it was believed the total impact that each of these indicators had on the total performance of the system could be established. What was less clear was how these specific reports would be included into a comprehensive measure. For example, what was the effect of a NMAC and a PD? Similarly, what was the relative contribution of the RI with rating of C to the overall performance of the system? What impact do PD s have on the overall assessment of the performance and safety of the national airspace system? 3. The Analytic Hierarchy Process 3.1 Background of AHP The Analytic Hierarchy Process (AHP) is based on the experience gained by its developer, T.L. Saaty, while directing research projects in the US Arms Control and Disarmament Agency. It was developed as a reaction to the finding that there is a lack of common, easily understood, and easy-to-implement methodology to enable the making of complex decisions (Bhushan and Ria, 2004). Since its development, the simplicity and power of the AHP has lead to its rapid and widespread use across multiple domains beyond the process of making complex decisions. 3.2 The AHP process Given its simplicity and ease of use, and relying on the three primary functions of AHP (structuring complexity, measurement, and synthesis), decision-maker s thoughts on the problem can be organized in a manner that is simple to follow and analyze. These characteristics make AHP an effective methodology to a wide range of applications and allow any complex situation that requires structuring, measurement, and synthesis to be better understood. In addition to the three primary functions, there are three related basic principles of AHP (decomposition, comparative judgments, and hierarchical composition or synthesis of priorities) which are utilized in the AHP methodology. The decomposition principle is applied to structure a complex problem into a hierarchy of clusters and sub-clusters. The principle of comparative judgments is applied to construct pairwise comparisons of all combinations of elements in a cluster with respect to the parent of the cluster. These pairwise comparisons are used to derive local priorities of the elements in a cluster with respect to their parent. The principle of hierarchical composition or synthesis is applied to multiply the local priorities of the elements in a cluster by the global priority of the parent element, producing global priorities throughout the hierarchy and then adding the global priorities for the lowest level elements. 3.3 Steps in AHP Saaty (2008) describes the steps in the AHP in the following way: 1. Define the problem and determine the kind of knowledge sought. 2. Structure the decision hierarchy from the top with the goal of the decision, then the objectives from a broad perspective, through the intermediate levels (criteria on which subsequent elements depend) to the lowest level (which usually is a set of the alternatives). 3. Construct a set of pairwise comparison matrices. Each element in an upper level is used to compare the elements in the level immediately below with respect to it. 4. Use the priorities obtained from the comparisons to weigh the priorities in the level immediately below. Do this for every element. Then for each element in the level below add its weighted 3

values and obtain its overall or global priority. Continue this process of weighting and adding until the final priorities of the alternatives in the bottom most level are obtained. In this process users of the AHP decompose their problem into a hierarchy of more easily comprehended sub-problems, each of which can be analyzed independently. Once the hierarchy is built, subject matter experts systematically evaluate the elements by comparing them to one another two at a time. The AHP then converts these evaluations to numerical values that can be processed and compared over the entire range of the problem. From this process, a numerical weight is derived for each element of the hierarchy. 4. The Aerospace Performance Factor The APF does not use the AHP as a multi-criteria decision tool. Since no decision is being sought, no alternatives are used in the APF AHP. Utilizing the AHP the APF looked at all available contributing factors which could be used to measure the overall safety performance of the system. All available reported factors that involved safety issues were included to ensure a complete system-wide view. 4.1 Building the hierarchy All factors were obtained through the use of historical mishap reports of events and incidents within the system. The elements chosen were comprised of Operational Errors (OE), Pilot Deviations (PD), Near Midair Collisions (NMAC), Operational Deviations (OD), Runway Incursions (RI), and Vehicle and Pedestrian Deviations (VPD). The elements were then placed in a hierarchy, using a mindmap, that represents the logical relationships of the elements to each other. The completed structures of the elements are displayed in Figure 1. 4

4.2 Judging the pairwise comparisons Figure 1. The national airspace hierarchy structure. Once the hierarchy was built, subject matter experts systematically evaluated the various elements by comparing them to one another two at a time (Table 1). The subject matter experts were presented with the paired comparison and a rating scale (Table 2) and asked Between these 2 choices, how strongly do you feel that this item represents a risk to the safety of the system? By making the comparison of the various elements the experts used their judgments about the elements' relative significance and a ratio scale was created (Table 3). Table 1. Sample Airborne events pairwise comparison matrix. Operational Errors (OE) Pilot Deviations (PD) Near Midair Collision (NMAC) Operational Errors (OE) ------------------------- Pilot Deviations (PD) ------------------------- ------------------------ Near Midair Collision (NMAC) ------------------------- ------------------------ -------------------------- 5

Table 2. Rating scale. Equal Between Equal and Moderately Moderately Between Moderately and Between and Very Very Between Very and Extreme Extreme Table 3. Sample Airborne events pairwise comparison matrix [note: these are sample data]. Operational Errors (OE) Pilot Deviations (PD) Near Midair Collision (NMAC) Operational Errors (OE) Pilot Deviations (PD) Near Midair Collision (NMAC) ------------------------- 1/5 1/2 ------------------------- ------------------------ 5 ------------------------- ------------------------ ------------------------- In this example of Airborne events (Table 3) a judgment was made that an Operational Error represents a greater risk than a Pilot Deviation and that Pilot Deviation represented a greater risk than a Near Midair Collision. 4.3 Deriving the weighted score from the matrix The AHP then converts these pairwise evaluations created by the subject matter experts and the following weights are created (Table 4). Table 4. Sample weighted score for Airborne events [note: these are sample data]. Weighted Score Operational Errors (OE) 0.65 Pilot Deviations (PD) 0.20 Near Midair Collision (NMAC) 0.15 TOTAL 1.00 4.4 Calculating the element score Once the judgment evaluations are converted to numerical values the data is applied to each of the branches of the model for a particular time period. For any element the total count would be divided by the total count for traffic for that particular element and then multiplied by its calculated weight. X = (incident count x weight) / traffic count 4.5 Summing the element scores The same calculations were then performed on each element of the hierarchy to calculate each element s relative contribution to the overall model. By calculating each element of the branches comprising a collection of the hierarchy and summing the score, the relative contribution of the element and branch to 6

the overall model could be rolled-up and overall scores could be determined. In so doing, all safety elements are rolled up to create a cumulative value on the overall system. 4.6 Summing the sums In the same way that individual elements within an incident-type were rolled-up to give a branch its relative score; branches were rolled-up and relative scores were attached to them. By summing up the calculated values of the branches of the hierarchy and continuing this roll-up process until the root of the hierarchy is reached a comprehensive system-wide performance measurement tool was complete. Total score for an event = Σ (sub-events) 5. Putting it all together The goal of the APF was to create a visual representation of the performance of the national airspace system. To achieve this, the final step in this process was to apply real data based on the existing historical safety indicators. 5.1 Applying the model to real data By defining the elements of the model (and having subject matter experts quantify the relative contribution of the elements in the model) the APF tool could be applied to series of real data. A comparison of the model for a specified period of time would reveal the relative contribution of each of the elements and branches of the model. 5.2 Performance over time The relative contribution of the elements (and branches) of the model to the performance of the airspace could also be applied to time-series data collected over many years. By examining the performance of the system over time a view of the historical performance could lead to an expectation of the future direction (Figure 2). 7

Figure 2. System-wide view of performance of the national airspace system. Figure 2 shows the performance of the national airspace during 2008. The dark line represents the total performance of the national airspace and the lightly shaded lines represent the sub-categories of the hierarchy (Airborne and Surface). The lightly shaded line directly below the dark top line represents the contribution of Airborne incidents while the lightly shaded line nearer the x-axis represents the relative contribution of Surface incidents. By summing these two sub-categories the total is shown in black. 5.3 Establishing a baseline The performance of the national airspace analyzed across time allowed for a direct comparison to other time periods. For example, the performance of the system could be compared to an agreed upon baseline. The baseline in Figure 2 is indicated by the horizontal line at the zero point of the y-axis. The baseline is calculated by taking the mean of the 36 monthly values of the performance for the years 2004 and 2006. This time period was selected because it represented real data over an extended period. By establishing this baseline a visual comparison can be made between a particular point in time and the baseline. 8

5.4 An interactive APF Figure 3. An interactive APF. In order for the APF to be interactive the figures allow the user to click on the sub-category lines of the figure and display the underlying data associated with this sub-category. For example, Figure 2 displays the two sub-categories that comprise the black line (i.e., performance of the system). By clicking on the sub-category lines (in this example the Airborne line) a new figure is displayed with the details for this sub-category (Figure 3). When the user navigates to a new figure all the functionalities of the figure remain the same but are calculated for the chosen category. For example, the baseline is calculated for the Airborne incidents calculated by taking the mean of the 36 monthly Airborne values for the years 2004 and 2006. The zero point on the y-axis then represents the mean value for Airborne on dates between Jan 2004 to Dec 2006. By establishing this baseline the comparisons of any given point in time is always being measured against an established baseline particular to that category. In addition to displaying a baseline particular to the user-chosen category, the APF application also displays the sub-categories that comprised that category (Figure 3). In the example of Airborne incidents, the categories making up Airborne are displayed. The top grey line is the Pilot Deviations (PD), the middle grey line is the Operational Errors (OE), and the bottom line is the Near Midair Collision (NMAC). In the same manner that the other lines on the figures allow the user to drill-down the lines on Figure 3 also allow the user to make a selection and viewing the details of the sub-categories that comprise the category (Figure 4). 9

Figure 4. An interactive APF: ALL > Airborne > Airborne OE. 6. Conclusion By adhering to the three primary functions (structuring complexity, measurement, and synthesis) and applying three related basic principles of AHP (decomposition, comparative judgments, and hierarchical composition or synthesis of priorities) a visual representation of the performance of the system over time was created. The APF was created by defining the elements of the model and having subject matter experts judge the relative importance of the elements in the model. Once this was completed the APF tool was applied to a series of real data and a picture of the performance of the entire landscape of the national airspace emerged. Based on the success of this effort other groups within the arena of aviation are working to develop their own APF s. Recently the European-based commercial airline easyjet created an APF based on their industry-specific categorization structure and data. In addition to easyjet, who is also being supported by the Imperial College of London, several European ANSP s are assessing the incorporation of APF methodology into their operations. This paper describes the process of establishing an indicator of performance within the arena of aviation, however, given the flexibility of the AHP framework, the process of providing a visual representation of performance can be applied to other industries. 10

REFERENCES Bhushan, N. and Ria, K. (2004). Strategic Decision Making: Applying the Analytic Hierarchy Process, London: Springer-Verlag London Limited. Saaty, T.L. (2008). Decision making with the analytic hierarchy process. Int., J. Services Sciences, 1(1), 83-98. 11