This is the accessible text file for GAO report number GAO-05-728 entitled 'Aviation Safety: FAA Management Practices for Technical Training Mostly Effective; Further Actions Could Enhance Results' which was released on September 7, 2005. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: September 2005: Aviation Safety: FAA Management Practices for Technical Training Mostly Effective; Further Actions Could Enhance Results: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-728]: GAO Highlights: Highlights of GAO-05-728, a report to congressional requesters: Why GAO Did This Study: One key way that the Federal Aviation Administration (FAA) makes air travel safer is to inspect the manufacture, operation, and maintenance of aircraft that fly in the United States. To better direct its resources, FAA is shifting from an inspection process that relied on spot-checks of compliance with regulations to one that evaluates operating procedures and analyzes inspection data to identify areas that pose the most risk to safety (called system safety). While FAA believes the new approach requires some technical knowledge of aircraft, Congress and GAO have long-standing concerns over whether FAA inspectors have enough technical knowledge to effectively identify risks. GAO reviewed the extent that FAA follows effective management practices in ensuring that inspectors receive up-to-date technical training. In addition, GAO is reporting on technical training that the aviation industry provides to FAA. What GAO Found: For its technical training, FAA follows many of the effective management practices for training that GAO has advocated and is improving its efforts in others. (See below.) In planning, FAA has linked technical training efforts to its goal of safer air travel and has identified technical proficiencies needed to improve safety inspectors’ performance in meeting this goal. It plans to better relate training to job tasks and is in the early stages of developing an approach to set priorities for new courses and course revisions. FAA Mostly Follows Effective Management Practices for Its Technical Training: Element: Practices in planning training efforts; Extent followed: Mostly. Element: Practices in developing training curriculum and courses; Extent followed: Mostly. Element: Practices in delivering training; Extent followed: Partially. Element: Practices in evaluating training efforts; Extent followed: Mostly. Source: GAO. [End of table] In developing technical courses, FAA has a structured process aimed at ensuring that courses meet performance objectives. It allows inspectors and others to identify the need for new training courses and to aid in developing courses. FAA is developing an initiative to systematically identify specific technical competencies and training requirements for inspectors. In delivering courses, FAA offers a wide array of technical courses from which inspectors can select to meet job needs. From GAO’s survey of FAA’s inspectors, we estimate that only about half think that they have the technical knowledge needed for their jobs. FAA officials told us that inspectors’ negative views stem from their wanting to acquire proficiencies that are not as crucial in a system safety environment. GAO also estimates that 28 percent of inspectors believe that they get the technical training that they request. However, FAA’s records show that FAA approves about 90 percent of these requests, and inspectors are making good progress in receiving training. Over half of the inspectors have completed at least 75 percent of technical training that FAA considers essential. In evaluating courses, FAA continuously assesses technical training through end-of-course evaluations and surveys of inspectors and supervisors. FAA is developing an approach to measure the impact of training on FAA’s mission goals, such as reducing accidents. This is a difficult task. Technical and Other Training Enables FAA to Inspect a Wide Variety of Aircraft: [See PDF for image] [End of figure] What GAO Recommends: Within the context of an overall system safety approach, GAO recommends that FAA take several actions, including systematically assessing inspectors’ technical training needs. FAA officials generally agreed with the contents of this report and agreed to consider GAO’s recommendations. www.gao.gov/cgi-bin/getrpt?GAO-05-728. To view the full product, including the scope and methodology, click on the link above. For more information, contact Gerald Dillingham at (202) 512-2834 or DillinghamG@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: Strategic Planning Activities Generally Reflect Effective Practices and Focus on Reducing a Large Gap in System Safety Knowledge: FAA Follows Effective Management Practices in Developing Individual Courses but Recognizes the Need to Develop a Unified Curriculum: FAA Provides Extensive Support for Delivering Training; However, Many Inspectors Believe Improvements Could Help Them Do Their Jobs More Effectively: Although FAA Uses Several Approaches to Evaluate Technical Training Provided, Assessing Impact on Performance Remains to Be Done: Industry Provides Much of FAA's Technical Training; Additional Safeguards Needed to Prevent Real or Appearances of Conflicts of Interest: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendixes: Appendix I: Inspector-Reported Travel for Technical Training: Appendix II: Additional Details on Training Data and Selected Inspector Survey Responses: Appendix III: Scope and Methodology: Appendix IV: GAO Contact and Staff Acknowledgments: Related GAO Products: FAA Safety Inspector Training: Human Capital: Related FAA Training: Tables: Table 1: Types of Inspectors, Responsibilities, and Numbers, as of April 2005: Table 2: Extent That FAA Followed Effective Management Practices in Planning for Training: Table 3: Extent That FAA Followed Effective Management Practices in Developing Courses: Table 4: Extent That FAA Follows Effective Management Practices in Delivering Technical Training: Table 5: Percent of Inspectors Completing Essential Technical Courses: Table 6: Average Number of Technical and Nontechnical Training Courses Taken, Fiscal Years 2002 through 2004: Table 7: Extent That FAA Followed Effective Management Practices in Evaluating Its Training Program: Table 8: Number of Memoranda of Understanding and Fleets Enrolled as Part of the Aircrew Designated Examiner Program and Agreements with Training Centers: Table 9: Numbers of Inspectors Trained under Aircrew Designated Examiner Program and Agreements with Training Centers, Fiscal Years 2002 through 2004: Table 10: Percent of Essential Courses That Are Technical in Nature: Table 11: Percent of Inspectors Completing Essential Courses: Table 12: Average Number of Technical Training Courses Taken Outside of Requirements, Fiscal Years 2002 through 2004: Table 13: Inspectors' Views on Extent to Which They Currently Have Enough Technical Knowledge to Do Their Jobs: Table 14: Inspectors' Views on Extent to Which Requested Technical Training Is Approved: Table 15: Inspectors' Views on Whether Availability of Courses Helped or Hindered Their Ability to Take Requested Technical Training: Table 16: Inspectors' Views on Whether Availability of Funds Helped or Hindered Their Ability to Take Requested Technical Training: Table 17: Inspectors' Views on Whether Management's Determination of Need Helped or Hindered Their Ability to Take Requested Technical Training: Table 18: Inspectors' Views on Whether Inspection Workload Helped or Hindered Their Ability to Take Requested Technical Training: Table 19: Inspectors' Views on the Degree to Which Technical Training Is Delivered in a Timely Manner: Table 20: Inspectors' Views on the Extent That They Receive Technical Training Prior to Scheduled Oversight Activities: Table 21: Percent of Technical Training Provided by Industry as Reported by FAA, Fiscal Years 2002 through 2004: Table 22: Inspectors' Views on the Extent to Which Technical Training Opportunities Exist Closer to Their Work Location: Table 23: Experts Consulted for Our Work: Figures: Figure 1: FAA's Safety Inspections Cover a Wide Range of Activities: Figure 2: FAA Safety Inspector Training Roles and Responsibilities: Figure 3: FAA's Structured Approach for Course Development: Figure 4: FAA Inspectors Receiving Training in a Classroom Setting: Figure 5: Inspectors Responding that to a Great or Very Great Extent They Currently Have Enough Technical Knowledge to Do Their Jobs: Figure 6: Extent to Which Requested Technical Training Is Approved: Figure 7: Inspectors' Views on Factors Hindering Their Ability to Take Requested Technical Training: Figure 8: Inspectors' Views on the Extent to Which Technical Training Is Delivered in a Timely Manner: Figure 9: Inspectors' Views on the Extent to Which They Received Technical Training Prior to Scheduled Oversight Activities: Figure 10: Percent of Technical Training Provided by Industry as Reported by FAA, Fiscal Years 2002 through 2004: Figure 11: Number of Weeks Inspectors Reported Spending on Travel for Technical Training within the Past 12 Months: Figure 12: Inspectors' Views on the Extent to Which Technical Training Opportunities Exist Closer to Their Work Location: Abbreviations: FAA: Federal Aviation Administration: ATOS: Air Transportation Oversight System: Letter September 7, 2005: The Honorable Ted Stevens, Chairman: The Honorable Daniel K. Inouye, Co-Chairman: Committee on Commerce, Science and Transportation: U.S. Senate: The Honorable Don Young, Chairman: The Honorable James L. Oberstar, Ranking Member: Committee on Transportation and Infrastructure: House of Representatives: FAA's overarching goal for technical training is to improve aviation safety. One key way that the Federal Aviation Administration (FAA) makes air travel safe for the public and the movement of goods is to inspect the manufacture, operation, and maintenance of aircraft that fly in the United States. To do so, about 3,700 FAA inspectors perform hundreds of thousands of inspections annually.[Footnote 1] Carrying out these inspections has become more challenging with the rapid growth in the number and type of aircraft in use and their increasing technical sophistication. Concerns about the quality of inspections heightened after the investigation of the 1996 crash of ValuJet flight 592 revealed deficiencies in FAA's inspection system. In response, FAA began to make fundamental changes in its approach to inspections. Traditionally, FAA aviation safety inspectors relied on their expertise to conduct inspections that spot-checked manufacturing processes, aircraft operations, and aircraft maintenance for compliance with regulations. FAA is transitioning to a risk-based system safety approach to inspections that requires inspectors to apply data analysis and auditing skills to identify, analyze, assess, and control the potential hazards and risks of flying and to prevent accidents.[Footnote 2] While we have endorsed FAA's move toward a system safety approach to inspections, congressional oversight committees and we have had long- standing concerns over whether FAA inspectors have sufficient knowledge of increasingly complex aircraft, aircraft parts, and systems to effectively identify safety risks. The Vision 100-Century of Aviation Reauthorization Act, enacted in December 2003, requires that we report on FAA's actions to ensure that inspectors receive up-to-date training on the latest technologies. We call this technical training, although this use of the term "technical" differs somewhat from FAA's use of the term.[Footnote 3] Consistent with the act, this report focuses on the extent to which FAA follows effective management practices for (1) planning, (2) developing, and (3) delivering up-to-date technical training, and (4) ensuring that technical training for inspectors contributes to improved performance and results. It also discusses the degree to which the aviation industry provides technical training to FAA safety inspectors and discusses the safeguards in place to help preclude the appearance of or an actual conflict of interest when inspectors receive certain kinds of training from a regulated entity. Finally, as required by the act, the report provides information on the amount of travel required of inspectors in receiving technical training. (See app. I.) This report focuses on how FAA ensures that its inspectors possess the technical proficiency they need to do their jobs through following effective management practices and whether inspectors are receiving the technical training that FAA has determined is essential for its inspectors.[Footnote 4] We did not attempt to assess the technical proficiency that FAA's workforce requires (and will require in the near future) and compare it with the proficiency that currently exists. Because of the diversity and size of the inspector workforce and the wide variety of aircraft technologies that FAA is responsible for overseeing, this type of assessment would have been a massive undertaking and would be more properly done by FAA. We also did not attempt to compare the technical training received by inspectors with the tasks and activities that inspectors perform. FAA's inspector activity database contains tens of thousands of task and activity records, and the manner in which these records are stored did not allow us to electronically sort and analyze the data. However, to provide some insight into these two issues, we did discuss these issues with FAA officials and surveyed FAA's inspectors on their views, as described below. To assess whether FAA follows effective management practices regarding technical training, we compared FAA's management of its inspector technical training efforts with effective management practices outlined in our 2004 guide for assessing strategic training activities in the federal government and determined the extent to which FAA followed the relevant elements of this guidance.[Footnote 5] In addition, we analyzed FAA documents pertaining to planning, developing, delivering, and evaluating inspector training and discussed these activities with FAA officials involved in inspector training and the management of inspection programs at FAA headquarters in Washington, D.C., and the FAA Training Academy in Oklahoma City. To examine the training provided, including technical training, we analyzed FAA data on training courses taken by inspectors from 2002 through 2004 and FAA's evaluation of technical training courses during that period. We discussed technical training with safety inspectors and their supervisors at 7 of FAA's approximately 130 field locations. The locations were chosen to represent the range of FAA inspection responsibilities. We also conducted a self-administered electronic survey posted on the World Wide Web to a stratified random sample of FAA safety inspectors to obtain their views about their technical proficiency and the technical training they receive. We received useable responses from 79 percent of the inspectors surveyed. This report does not contain all the results from the survey. The survey and a complete tabulation of the overall results (excluding results by type of inspector, which are too voluminous to present) can be viewed at [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-704SP]. Finally, we obtained further perspective on FAA's training curriculum through semistructured interviews with 16 experts from the aviation industry and the field of aviation education who were selected on the basis of having extensive background and knowledge of the technical areas covered by FAA inspections. As part of our review, we assessed internal controls and the reliability of FAA's data on the amounts and types of training received that are pertinent to this effort. We determined that the data elements were sufficiently reliable for our purposes. We conducted our work from March 2004 through July 2005 in accordance with generally accepted government auditing standards. (See app. III for additional information on our scope and methodology.) Results in Brief: FAA has made training an integral part of its safety inspection system, which in recent years has emphasized risk analysis techniques over individual inspector technical knowledge of aircraft, aircraft parts, and systems. FAA has generally followed several effective management practices for planning, developing, delivering, and assessing the impact of its technical training for its aviation safety inspectors, although some practices have yet to be fully implemented. Regarding planning for technical training, for example, FAA's training efforts for the most part follow effective management practices and are intended to support its goals for improving aviation safety, and they largely focus on effectively implementing a system safety approach to inspections. According to FAA, it has identified gaps in several of the competencies required to conduct system safety inspection, including risk assessment, data analysis, auditing, and systems thinking, and the agency is currently working to address these gaps. In FAA's view-- although it recognizes the importance of inspectors staying up to date with changes in aviation technology--the competencies needed for system safety inspections are the most critical for inspectors, and the gaps in these competencies are much larger than gaps in technical skills and competencies relating to the production, operation, and maintenance of aircraft. In addition, FAA Office of Aviation Safety officials said that inspectors do not need a substantial amount of technical training because inspectors are hired with a high degree of technical knowledge of aircraft and aircraft systems, and they can sufficiently keep abreast of many of the changes in aviation technology through FAA and industry training courses and on-the-job training. Nevertheless, FAA plans to identify specific technical competencies and training requirements as part of a process intended to better relate training to the job tasks of each inspector specialty.[Footnote 6] FAA also for the most part follows effective management practices for developing its inspector technical training curriculum. For example, FAA integrates the development of courses with overall strategies to improve performance and to meet emerging demands. In this regard, FAA develops courses that support changes in inspection procedures resulting from regulatory changes or agency initiatives, such as the implementation of the system safety approach to inspections. FAA will also consider developing training courses that are requested by inspectors and managers. FAA also works to match the training delivery approach with the nature of the material presented to best meet inspector and agency needs--such as delivery at a central location in FAA's Training Academy in Oklahoma City, Oklahoma; in multiple locations closer to field offices; or through computer-based instruction. While following many effective practices in this area, FAA has not systematically identified the technical skills and competencies each type of inspector needs to effectively perform inspections. As a result, technical courses are developed on an ad hoc basis rather than as part of an overall curriculum for each inspector specialty. FAA has recognized this problem and is developing an initiative that will systematically assess whether the complete array of training for each inspector specialty meets performance requirements. In delivering technical courses, FAA has followed effective management practices to differing degrees. For example, FAA has established clear accountability for ensuring that inspectors have access to technical training, developed a way for inspectors to choose courses that meet job needs and further professional development, and offers a wide array of technical and other courses. However, inspectors are for the most part dissatisfied with the technical training they receive. From an analysis of the survey, we estimate that only about half of FAA's inspectors think that they have the technical knowledge needed to do their jobs,[Footnote 7] only about one-third are satisfied with the technical training they have recently received, and less than half believe that they get to take the technical training that they request. However, our analysis of FAA training data indicates that FAA has approved about 90 percent of the technical courses requested by inspectors, and inspectors in general are making good progress in completing the technical training essential for their positions (77 percent of the inspectors have completed at least half of their essential courses, and 46 percent have completed at least 80 percent of their essential courses). In addition, according to the survey, we estimate that only 23 percent of FAA's inspectors think that they receive technical training in time to do their current job. FAA's records do not allow us to assess the timeliness of training. FAA officials told us that inspectors' negative views on their technical knowledge and the training they received stem from their not accepting FAA's move to a system safety approach. That is, the inspectors are concerned about acquiring individual technical proficiency that is not as crucial in a system safety/risk management environment. Given that it has not completed assessing whether training for each inspector specialty meets performance requirements, FAA is not in a position to make definitive conclusions concerning the adequacy of inspector technical training. FAA for the most part follows several effective management practices in evaluating individual technical training courses. For example, it continuously assesses technical training through participant end-of- course evaluations and surveys of inspectors and supervisors that focus on the application of skills and knowledge to the job. FAA also requires that each training course receive a systematic evaluation every 3 years to determine if the course is up to date and relevant to inspectors' jobs, although training officials noted that many courses have yet to undergo such an evaluation. However, FAA has taken limited action to evaluate the overall impact of its technical training on inspector performance in achieving mission goals, such as reducing accidents. Although FAA surveys its employees on their attitudes regarding many aspects of their employment, including the extent to which they were able to apply agency training to their jobs and perform their jobs effectively, it is not able to isolate inspectors' responses from those of its other employees. Moreover, the survey does not ask employees to differentiate between the types of training they receive, such as technical and nontechnical training. Experts on training in government agencies emphasize the importance of using an approach to evaluating training that goes beyond individual course evaluations and includes such indicators as the amount of learning that occurs from training programs and their organizational impact. However, training experts acknowledge that isolating performance improvements resulting from training programs is difficult for any organization. FAA has increasingly relied on the aviation industry to provide technical training in fiscal years 2002 through 2004. In fiscal year 2004 (latest data available), industry delivered nearly half of FAA's technical training. Although FAA pays for most of the technical training that industry provides, from fiscal years 2002 through 2004 about, 17 percent of industry-provided technical training was supplied to FAA in exchange for an in-kind service, such as delegating authority to conduct inspections, (called quid pro quo arrangements) with some apparently limited additional training supplied at no cost to the agency. To a large degree, FAA has established safeguards to help preclude actual or appearances of a conflict of interest, such as executing agreements with aviation industry training providers it regulates outlining the conditions under which it will accept training for in-kind service or at no cost. However, FAA has not included provisions covering its enforcement and oversight authority in all agreements with aviation industry training providers. In addition, two regional officials said that their regions accept free training on a limited basis outside the formal agreements with the training providers; one of these officials identified 57 instances over the past 5 years in which inspectors received free training from aircraft manufacturers or operators. Because these opportunities generally arise at the local office level, whether such an offer is reviewed by legal counsel is dependent on the office manager, the manager's understanding of the FAA policy, and a judgment about whether a specific training opportunity raises any concern that should be reviewed by legal counsel. Although FAA has followed effective management practices in many areas in providing technical training to its safety inspectors, we are making several recommendations aimed at, among other things, improving FAA's identification of gaps in inspectors' technical knowledge that relate to their jobs, better aligning the timeliness of training to when inspectors need the training to do their jobs, gaining inspectors' acceptance for changes made or planned to their training, and ensuring that the acceptance of training from aviation industry providers does not limit FAA's enforcement authority or pose a real or potential conflict of interest. In commenting on a draft of this report, the Department of Transportation generally agreed with the information that we presented and agreed to consider our recommendations. However, the department expressed the view that we should have considered, as positive responses, the views of inspectors who responded to survey questions as "moderate extent," along with those who responded to a "great extent" or "very great extent." The extent scale that we used in our survey represents a unidirectional scale. As such, it is possible to interpret any point along that scale, other than "no extent," as positive, depending upon how a question is worded. Generally, we presented information in the report with both "very great extent" and "great extent" combined to represent the clearly most positive responses. Background: Ensuring the safety of the nation's aviation system is the shared responsibility of FAA and the aviation industry. Aircraft manufacturers are responsible for building safe aircraft. Aircraft operators are responsible for the safe maintenance and operation of aircraft. FAA is responsible for, among other things, certifying that the manufacture of aircraft and aircraft parts meets FAA standards, encouraging the development of new aviation technologies, and conducting periodic inspections to ensure continued compliance with safety regulations. Within FAA, the Office of Aviation Safety (1) directs and manages aviation safety through inspection (called surveillance by FAA) and oversight programs; (2) creates and amends standards and policies; and (3) certifies that aircraft, manufacturers, maintenance services, and individuals who operate aircraft meet FAA safety standards before they carry out their activities (called certification). FAA's 3,700 inspectors are located in more than 130 offices throughout the world. About 3,000 of these are front-line inspectors. These inspectors specialize in conducting inspections of various aspects of the aviation system, such as aircraft and parts manufacturing, aircraft operation, aircraft airworthiness, and cabin safety. (See table 1.) Table 1: Types of Inspectors, Responsibilities, and Numbers, as of April 2005: Inspector type: Air carrier operations; Areas of responsibility: Responsible for evaluating airmen (pilots, aviators, or aviation technicians) for initial and continuing qualifications, airmen training programs, equipment, and facilities, and aircraft operations for adequacy of facilities, equipment, and procedures to ensure the safe operation of the aircraft. Air carrier inspectors are responsible for evaluating pilots, dispatchers, air carriers, and similar operators; Number: 908. Inspector type: Air carrier maintenance; Areas of responsibility: Focuses on evaluating mechanics and repair stations for initial and continuing certification and mechanic training programs. Examines the overall aircraft maintenance program, including the development of maintenance manuals and the procedures for repairing aircraft and their components. Inspects aircraft and related equipment for airworthiness. Air carrier inspectors evaluate maintenance programs of air carriers and similar operators; Number: 831. Inspector type: General aviation operations; Areas of responsibility: Duties are similar to air carrier operations inspectors, with the exception that general aviation operations inspectors are responsible for evaluating pilots, flight instructors, air taxis, and similar operators; Number: 636. Inspector type: General aviation maintenance; Areas of responsibility: Duties are similar to air carrier maintenance inspectors, with the exception that general aviation maintenance inspectors evaluate maintenance programs of air taxis and similar operators; Number: 571. Inspector type: Air carrier avionics; Areas of responsibility: Responsible for inspecting aircraft electronics and related systems for airworthiness, evaluates avionics technicians, repair stations, and technician training programs. Air carrier inspectors conduct surveillance and oversight of air carriers and similar operators; Number: 341. Inspector type: Aircraft certification; Areas of responsibility: Administers and enforces safety regulations and standards for the production and/or modification of aircraft. Evaluates and oversees the plants that build or assemble aircraft. Inspects prototype or modified aircraft, aircraft parts, and avionics for conformity with design specifications and safety standards. Issues certificates for all civil aircraft; Number: 187. Inspector type: General aviation avionics; Areas of responsibility: Duties are similar to air carrier avionics inspectors, with the exception that general aviation avionics inspectors conduct surveillance and oversight of air taxis, travel clubs, and similar operators; Number: 180. Inspector type: Cabin safety; Areas of responsibility: Serves as a resource and technical authority on cabin safety requirements, such as verifying that emergency equipment is onboard the aircraft, as they relate to activities affecting civil aviation; Number: 60. Total; Number: 3,714. Source: GAO summary of FAA information. [End of table] Some inspectors, such as operations and airworthiness inspectors, further specialize according to the type of aircraft and aircraft operators they oversee. Other inspectors, such as general aviation inspectors, are responsible for inspecting a wide range of aircraft, such as those used for agriculture, air taxi service, industry, and pleasure. (See fig. 1.) In addition, they inspect flight instructors. Some air carrier inspectors are assigned to one of the 16 carriers that are currently subject to the Air Transportation Oversight System (ATOS) program, which is intended to identify safety problems through risk analysis;[Footnote 8] while other air carrier inspectors are responsible for overseeing the operations of several smaller carriers in a geographic area. Figure 1: FAA's Safety Inspections Cover a Wide Range of Activities: [See PDF for image] Note: As a workforce, FAA inspectors conduct a wide variety of inspections, including ensuring that pilots are qualified to operate air carrier and general aviation aircraft and inspecting air carrier and general aviation aircraft for safety. [End of figure] FAA requires that candidates for safety inspector positions have extensive technical qualifications and experience, which is usually gained during careers in the aviation industry. For example, prospective manufacturing inspectors need experience in and knowledge of industrial technologies. Similarly, operations inspectors need pilot licenses to fly specific makes and models of aircraft; maintenance inspectors need to have certifications to repair the aircraft's airframe and power plant; cabin safety inspectors need extensive experience in aircraft cabin safety procedures; and avionics inspectors need extensive experience in servicing an aircraft's avionics system, which includes radar and other electrical systems. To supplement the skills inspectors bring with them from their previous careers in the aviation industry, FAA provides inspectors with extensive training in federal aviation regulations; inspection and investigative techniques; and technical skills, such as flight training required for operations inspectors. The services within FAA's Office of Aviation Safety that are responsible for conducting inspections of aircraft operators and aircraft repair stations[Footnote 9] (Flight Standards) and manufacturers (Aircraft Certification) have each established training units that develop curricula and specific courses for inspectors. Most of the regulatory, inspection, and investigative courses are taught by FAA instructors at the FAA Training Academy in Oklahoma City, Oklahoma. Much of the technical training (training that enhances skills concerning the production, maintenance, and operation of aircraft, aircraft parts, airworthiness, and systems) is contracted out to vendors, such as flight schools. (App. I provides information on the amount of inspector training provided by FAA and vendors.) FAA is also making increased use of nonclassroom training delivery methods, such as computer-based instruction, Web-based training, interactive video training, and correspondence courses. Inspectors also receive extensive on-the-job training, particularly when they are first hired. FAA has spent an average of $43 million per year on inspector training activities from fiscal years 2002 through 2004 and plans to spend $41 million in fiscal year 2005. The Flight Standards and Aircraft Certification training divisions have training priorities, which are set by the Associate Administrator for Aviation Safety. Determining training needs is based on the inspectors' job. Acquiring training is a shared responsibility between inspectors and their supervisors. (See fig. 2.) Each year inspectors and their supervisors meet to decide which training inspectors will request in the coming year. The inspectors are expected to choose training that will fulfill their mandatory training requirements in areas such as basic aircraft accident investigation, air carrier airworthiness, aviation safety inspector job functions, and data analysis and related skills needed to perform system safety inspections. Inspectors can also request training that they believe will further their professional development. FAA's directorate, regional, and headquarters offices then compile and fund the training requests. Headquarters training officials coordinate with the FAA academy and other training vendors to deliver the training. Inspectors provide feedback to headquarters and academy officials through surveys and course evaluations. Figure 2: FAA Safety Inspector Training Roles and Responsibilities: [See PDF for image] [End of figure] FAA's transition to the ATOS system safety concept represents a major change in the way the agency operates as it shifts the oversight emphasis from the traditional methods of inspection to identifying and assessing risks to safety. Under the traditional or compliance approach, inspectors rely upon random inspection activities, such as observing aircraft parked at departure gates. When applying the system safety approach inspectors develop comprehensive surveillance plans for each air carrier. Developing the plans requires using existing safety data, risk indicators and the inspector's knowledge of the operations to determine the priority and frequency of inspection activities. The resulting comprehensive surveillance plan includes a series of inspection tasks to determine whether an airline has systems in place to ensure safety and a second series of inspections to verify that the airline is actually using those systems. FAA has taken steps to introduce concepts used in ATOS into its traditional oversight process for the air carriers not in the ATOS program. In November 1999, FAA instructed its inspectors to begin adjusting planned inspections for new air carriers,on the basis of evaluation of areas of potential safety risks.[Footnote 10] In 2002, inspectors were instructed to perform safety risk evaluations of all other non-ATOS carriers using ATOS risk assessment principles as part of their inspections. However, the inspections of the non-ATOS carriers are still based on a determination of whether air carriers are complying with regulations rather than whether air carriers' systems are operating effectively.[Footnote 11] According to FAA, in the transition to the system safety concept, safety inspectors are learning new skills, such as data analysis, risk assessment, computer operations, auditing, systems thinking, and interpersonal skills. Inspectors will continue to need technical expertise in avionics, cabin safety, operations, maintenance, and aircraft production and design, and may need training in composites, basic accident investigation, and nondestructive inspections courses. FAA has concluded that it will take a significant training effort to develop and maintain both the system safety approach as well as the technical competencies. Strategic Planning Activities Generally Reflect Effective Practices and Focus on Reducing a Large Gap in System Safety Knowledge: FAA's strategic planning acknowledges the central importance of aviation safety inspectors and defines their role as mission critical. In its planning activities for training, FAA has, for the most part, followed effective management practices by developing strategic approaches to training that have established broad training priorities for inspectors, among other things. (See table 2.) Table 2: Extent That FAA Followed Effective Management Practices in Planning for Training: Effective management practice: Ensures training goals and related performance measures and targets are consistent with overall mission and goals; Extent followed: Fully. Effective management practice: Ensures human capital professionals work in partnership with agency leadership in addressing agency priorities, including training, in strategic and annual performance planning processes; Extent followed: Fully. Effective management practice: Determines skills and competencies its workforce needs to achieve current and emerging agency goals and identifies gaps--including those training strategies can help address; Extent followed: Mostly. Effective management practice: Identifies appropriate level of investment for training and prioritizes funding so that the most important training needs are addressed first; Extent followed: Partially. Effective management practice: Ensures agency strategic and tactical changes are promptly incorporated into training efforts; Extent followed: Fully. Source: GAO. [End of table] Establishing training goals and performance measures that further overall agency goals. One of the goals of the inspector training program is to provide the training required to support inspectors in the FAA transition to a system safety approach for meeting its goal of increased safety. Both Flight Standards and Aircraft Certification have developed training initiatives to support this training goal. Flight Standards also has eight training initiatives focused on improving aviation safety in general, each of which has related performance measurements or targets along with strategies, time lines, and resource estimates.[Footnote 12] In addition, to support the FAA safety goal, the Office of Aviation Safety, which contains both the Flight Standards and Aircraft Certification services, measures what the services have done to prepare their workforce to operate in a system safety environment. This measurement includes information on how the services have developed and delivered training, redesigned existing courses, and validated inspector competencies. Human capital professionals partnering with agency leadership. In June 2003 Flight Standards established a human capital council which brings senior managers together with training officials to oversee all human capital efforts to, among other things, establish priorities that will both maintain existing inspector technical competencies as well as new and emerging system safety competencies. In Aircraft Certification, the manager responsible for training programs is involved in the service's annual planning process and also participates in weekly meetings of senior level managers. The training manager keeps the training development staff informed of new or changing priorities that could affect the training program. Determining gaps in workforce skills and competencies. Both Flight Standards and Aircraft Certification conducted human capital analysis for their aviation safety inspectors that revealed significant gaps in their needed competencies and skills as FAA continues its implementation of a more risk assessment-based system approach to safety oversight. In Flight Standards the analysis involved a team of FAA senior managers and FAA subject-matter experts.[Footnote 13] The team reviewed the existing competency requirements for inspectors and then determined which competencies should be modified or added over the next 5 years. The list of competencies compiled was then reviewed by another group of subject-matter experts, primarily program managers, who estimated the relative importance of the competencies (existing and new) in the next 5 years as well as the gaps between the current workforce's actual and needed level for each competency. The largest critical competency gaps for the inspectors in Flight Standards included (1) risk assessment, (2) data analysis, (3) systems thinking, and (4) designee and industry oversight.[Footnote 14] Technical proficiency training was the only competency that Flight Standards did not identify as having a critical competency gap, and the list of competencies for Flight Standards field inspectors issued in February 2005 does not include technical proficiency. FAA Office of Aviation Safety officials said that inspectors do not need a substantial amount of technical training courses because inspectors are hired with a high degree of technical knowledge of aircraft and aircraft systems, and they can sufficiently keep abreast of many of the changes in aviation technology through FAA and industry training courses and on-the-job training. Flight Standards officials said that the list of competencies contains items that cut across all inspection specialties and that it will be the role of individual curriculum oversight teams to identify the technical skills and competencies for each inspector specialty. Flight Standards officials said they will establish these teams as part of an effort to develop specific curriculums for each inspector specialty (see the following section on developing training activities). Aircraft Certification subject-matter experts who manage inspectors identified four similar critical competency gaps for the implementation of system safety for its manufacturing inspectors. The gaps included (1) business and management, (2) data analysis/risk assessment, (3) system thinking skills, and (4) designee oversight. Aircraft Certification officials noted that these are skills that its inspectors need to perform their primary inspection function, which is ensuring that manufacturers meet design specifications for aircraft parts and components. Inspectors do this by inspecting the processes and quality assurance systems involved in aircraft and parts manufacturing. Prioritizing funding for training activities. Currently in Flight Standards, requests for course development projects come from the operational policy divisions to the training division. The training division then works closely with these individual policy divisions (such as the Air Transportation Division or the Aircraft Maintenance Division) to develop or revise the courses they request. However, the existing process does not explicitly consider which course development projects are most critical. During fiscal year 2005, Flight Standards plans to develop an approach that will consider the organizational factors necessary to prioritize requests for new courses and revision of current courses, including exploring ways to engage senior management. According to FAA officials, a curriculum oversight steering committee will provide strategic direction and prioritization for the service's training needs. Aircraft Certification already employs a process that prioritizes training activities on the basis of three factors: impact on aviation safety, inspector job functions, and the needs of the customer. Training division officials meet each year to establish training priorities and to determine the resources needed to meet these priorities. According to these officials, they are guided by FAA strategic plans and direction they receive from operational program managers. Promptly incorporating strategic and tactical changes into training and development. Flight Standards has recognized the need to quickly deploy its training in order to reduce the time lag between the identification of training needs and its delivery as inspector performance requirements change. Currently Flight Standards has quarterly and semiannual training program reviews between the training division, the Flight Standards organizations which sponsor the training, and the FAA academy to discuss the sponsoring organizations' training needs. It is also the responsibility of staff who oversee individual courses, known as course mentors, to ensure courses reflect changes in FAA policies and procedures or new developments in aviation technologies. In Aircraft Certification an executive-level mentor is selected from its management team for each course and is responsible for managing the development of new courses and the updating of existing courses to respond to changes in FAA policies and priorities. We did not attempt to assess the extent to which FAA incorporates strategic and tactical changes into its inspector training curriculum. FAA Follows Effective Management Practices in Developing Individual Courses but Recognizes the Need to Develop a Unified Curriculum: FAA has for the most part followed effective management practices for developing individual safety inspector courses. (See table 3.) These practices, such as establishing guidelines that call for the formation of course development teams for each new course and that require each team to follow a series of progressive course development steps, are aimed at enhancing course quality and ensuring that the content of the course meets the intended course goals and performance objectives. However, FAA has not systematically identified technical training needs because it develops courses on a course-by-course basis rather than as part of an overall curriculum framework. Table 3: Extent That FAA Followed Effective Management Practices in Developing Courses: Effective management practice: New courses developed to meet emerging demands and improve performance; Extent followed: Fully. Effective management practice: Course development teams enable stakeholders to provide input; Extent followed: Fully. Effective management practice: Guidelines provide progressive course development steps with ongoing evaluation at each step; Extent followed: Fully. Effective management practice: Merits of different course delivery methods are considered; Extent followed: Fully. Effective management practice: Criteria used for decisions regarding outside training providers; Extent followed: Fully. Effective management practice: Analysis of training needs and course development linked to overall curriculum approach[A]; Extent followed: Partially. Source: GAO. [A] This management practice is not specifically identified in our assessment guide. However, a management approach that assesses training needs holistically rather than on a course-by-course basis can provide for a more systematic assessment of whether and how training will help meet organizational needs. [End of table] FAA Follows Many Effective Management Practices for Developing Technical Courses: FAA follows many effective management practices for developing technical training courses for Flight Standards and Aircraft Certification safety inspectors. Ensuring new courses meet emerging demands and improve performance. At the very beginning of any new course development effort, Flight Standards and Aircraft Certification validate the need for new aviation safety inspector training by discussing (1) the facts that indicate the need for training, (2) the desired outcome of the training in terms of performance, and (3) the target audience of inspectors who will receive the training. Before any substantial course development activities occur, Flight Standards and Aircraft Certification training guidelines require that a task analysis be conducted. The purpose of the task analysis is to identify essential tasks, knowledge, and skills needed for effective safety inspector job performance. FAA then uses this task analysis as the basis for determining the scope, content, and sequencing of training topics for each new course. Flight Standards and Aircraft Certification have also created different ways for field and headquarters personnel to request the development of new aviation safety inspector training courses when a new training need has emerged. Field personnel who see a need for safety inspectors to perform a new task or acquire new knowledge can submit training development requests. In addition, officials in Flight Standards and Aircraft Certification can propose new training courses for inspectors when regulatory changes occur or when new FAA initiatives, such as the system safety approach for aviation safety inspectors, create a need for additional inspector knowledge and skills. Those proposing new courses must describe how the proposed course will contribute to FAA's mission, explain the inspector knowledge and skills that will be acquired by taking the course, define the target audience for the proposed course, and describe the impact on the inspector workforce if the course is not developed. Enabling qualified personnel to participate as stakeholders. When Flight Standards and Aircraft Certification begin to develop a new course, a training development team is formed and different policy, technical, and training personnel participate in team activities throughout new course development. As a result, each member of the course development team has different skills and unique perspectives that he or she can contribute to course development. Each course development team has a course mentor whose role is to work with other team members through all course development stages to ensure that the content of the course meets the intended course goals and performance objectives. In addition, Flight Standards and Aircraft Certification also encourage course development teams to have subject-matter experts. According to FAA, these subject-matter experts, who can be FAA employees or even outside consultants, can improve the quality and accuracy of the new course because they have specific knowledge and experience in one or more course topics. For example, Flight Standards is now developing a new course for the advanced avionics "glass cockpit" displays that are increasingly being used by air carrier and general aviation operators. (See sidebar.) The expertise and knowledge of the course development team for the new advanced avionics glass cockpit display course were significantly enhanced by subject-matter experts who were assigned to the team and had experience approving aircraft equipped with these advanced avionics displays. Besides the course mentor and subject-matter experts, other key team members on course development teams include instructional systems designers who provide expertise in training design and course developers who write the actual lesson plans for the new course. Experts outside of FAA can also provide input on course development in many technical subjects. For example, FAA established a partnership with universities and affiliated industry associations and businesses throughout the country to form Centers of Excellence, which conduct aviation research in a number of areas including advanced materials, aircraft emissions, and airworthiness. The General Aviation Center of Excellence, formed in 2001, has conducted research on aircraft seat- restraint systems, increasing aircraft landing safety, and aircraft de- icing. In addition to their technical expertise, many universities and private sector companies in the aviation industry have substantial experience conducting aviation training and education programs. For example, Embry-Riddle Aeronautical University, which is the lead university for the General Aviation Center of Excellence, has been in the aviation training industry since 1926; and in addition to pilot and maintenance training, it offers more than 30 degree programs, including programs in engineering, aviation management, and aviation safety science. FAA already contracts with Embry-Riddle for some inspector flight training and recently expanded the number of training locations with another General Aviation Center of Excellence program. Airlines also have substantial experience offering pilot, crew, and maintenance training. FAA receives input from training providers like Embry-Riddle on course development as part of contracted training courses. FAA officials who work with Centers of Excellence said that there could be more opportunities for the agency to utilize the technical and aviation training expertise of the Centers of Excellence in developing its inspector training program. Using a structured approach for course development. Flight Standards and Aircraft Certification both use a structured course development approach that calls for progressive course development steps and ongoing evaluation of the training at each step. This approach provides course development teams with a description of activities that should occur at each step, which helps to ensure that lesson plans, course materials, and course delivery methods enable the student to meet course objectives and increase job performance. (See fig. 3.) Figure 3: FAA's Structured Approach for Course Development: [See PDF for image] [End of figure] Flight Standards and Aircraft Certification have also built quality controls into their course development guidelines by requiring evaluation at each of these course development steps. Generally, after each stage in the course development process described in figure 3, the course development team reviews the work that occurred up to that point. For example, once the course developer has created lesson plans and any course materials, all members of the course development team review and make suggestions for revising them. Both Flight Standards and Aircraft Certification evaluate and test newly developed training in the final stages of course development. For example, Flight Standards conducts an operational tryout to see how effective the course is, with actual course instructors teaching the new course lesson plans and team members acting as observers. New FAA courses are tested again in the final prototype stage when instructors, with observers, teach the course in front of students from the course's planned target audience. These students provide feedback on each lesson in the course in such areas as clarity of objectives, appropriateness of the level of instruction, and the usefulness of training materials. These quality checks are aimed at ensuring that lesson plans flow smoothly and support the course objectives. Considering different approaches for presenting courses. Flight Standards and Aircraft Certification generally make decisions on the delivery approach or method to use for aviation safety inspector courses in the initial stages of course development. For example, the course development team will consider factors such as the complexity of the topic, how soon the course is needed, and how many students will need the training. In the case of training a large number of inspectors on a relatively simple topic or a quick refresher course, a short self- paced computer-based or Web-based training course might be selected. Because the course development process can take months to complete, self-paced training can also be used when the knowledge or information needs to be conveyed quickly to a large number of students. However, when a course requires interaction and hands-on learning and it covers a lengthy or complex topic, the course development team could decide that a classroom format followed by practical exercises is the most suitable delivery method. (See fig. 4.) For example, in developing the glass cockpit course discussed above, the course development team considered several factors, including the complexity and the rapid growth of the technology and the fact that relatively few students have had a chance to become familiar with glass cockpit systems. The course development team then decided to use a combination of classroom and practical exercises as the primary delivery methods. Under this course format, students participate in classroom lecture and discussion sessions for the introductory lesson on glass cockpit technology. The students then have practical exercises on flight simulators with glass cockpit displays to integrate and reinforce the knowledge gained in the classroom. Figure 4: FAA Inspectors Receiving Training in a Classroom Setting: [See PDF for image] Note: When covering a technical or complex subject FAA will often use a classroom format that allows for group interaction and practical exercises. [End of figure] Using criteria for decisions on outside training providers. Flight Standards and Aircraft Certification have developed and apply criteria for deciding whether to use outside training providers for their new aviation safety inspector courses. For example, one criterion is whether FAA or an outside training provider has more technical expertise. Generally, FAA will use its own instructors to teach many of the introductory courses that inspectors receive when they first join FAA. This is because many of these courses provide the new safety inspector with a familiarization of inspector responsibilities and job functions and a description of aviation regulations, and FAA is usually the most appropriate training provider to cover these topics. However, in a given aviation technology area, some private sector companies that concentrate in a technology will have more expertise than FAA. For example, because an outside training provider has more specialized technical knowledge in composite materials,[Footnote 15] Flight Standards contracts with this provider to deliver composites and composites repair training. FAA Plans to More Systematically Identify Technical Training Needs in Developing Its Inspector Training Curriculum: As discussed above, FAA's course development activities follow many effective management practices for developing individual courses, but it has not yet systematically identified its inspectors' overall training needs to ensure that the curriculum addresses the unique training needs of each type of inspector. However, FAA is developing a specific training curriculum for each type of inspector. Flight Standards recognizes that it manages courses as individual components and that it needs to develop courses and address training needs for each of its inspector specialties as part of an overall curriculum. In addition, our survey indicates that only 27 percent of inspectors said that the current set of FAA recommended training courses for each inspector type captures the training needed to do their jobs to a great or very great extent.[Footnote 16] Flight Standards recognizes that for curriculum transformation to work effectively, a strategy for curriculum management, as opposed to course management, needs to be clearly articulated. In response, Flight Standards is developing a new performance-based training initiative with the goal of systematically assessing the complete array of training to ensure it meets the performance requirements of the many specialties, disciplines, and positions in Flight Standard's ranks. In an effort to implement a more curriculum-based approach that addresses different inspector training needs, the curriculum transformation plan recommends creating curriculum oversight teams for each type of inspector made up of representative inspectors from the field and from headquarters. Rather than the current approach, in which course development teams focus on individual courses, these curriculum oversight teams would be responsible for the overall curriculum for each type of inspector, including defining training requirements and ensuring that curriculum and course content are current and consistent with Flight Standards policy and practices in the field. The Flight Standards steering committee is responsible for chartering these curriculum oversight teams and approving the curriculum they develop for each inspector type. Flight Standards estimates that it will complete implementation of its curriculum transformation plan in 2008. If effectively implemented, we believe that these approaches would allow Flight Standards to develop a more systematic method for identifying training needs and provide a curriculum that is more relevant to different types of inspectors and their needs. While Flight Standards is in the first stages of implementing its new curriculum-based approach to training, Aircraft Certification has recently taken steps to revise existing courses and develop new courses within an overall curriculum approach. For example, it formed a curriculum study team and completed a proposed curriculum for its manufacturing aviation safety inspectors and has revised inspector courses and other aspects of training according to its new curriculum plan. Aircraft Certification has only one type of aviation safety inspector; Flight Standards has avionics, maintenance, and operations inspectors for both general aviation and air carriers as well as other inspectors, such as cabin safety inspectors. Because Aircraft Certification has only one type of safety inspector, a permanent curriculum study team may not be absolutely necessary. FAA Provides Extensive Support for Delivering Training; However, Many Inspectors Believe Improvements Could Help Them Do Their Jobs More Effectively: FAA recognizes that effective delivery of quality inspector training is crucial to the success of the agency's mission to obtain industry compliance with safety standards and promote the continuing safety of air travel. FAA has generally followed effective management practices for training deployment to help ensure effective delivery of training, but improvements could be made. (See table 4.) Experts from the aviation and academic communities whom we consulted generally agreed that, for the most part, the courses FAA offers meet current and emerging technical needs. However, many inspectors question whether the training they receive is sufficient to provide them with the technical knowledge needed to perform their jobs. Table 4: Extent That FAA Follows Effective Management Practices in Delivering Technical Training: Effective management practice: Clearly delineates accountability for achieving agency training goals; Extent followed: Fully. Effective management practice: Uses a suitable and timely process for selecting inspectors for technical training given inspectors' current duties and existing skills; Extent followed: Partially. Effective management practice: Fosters an environment that is conducive to learning; Extent followed: Fully. Effective management practice: Takes steps to encourage employee buy-in to goals and priorities of technical training; Extent followed: Partially. Source: GAO. [End of table] FAA Generally Follows Several Effective Management Practices for Delivering Technical Training: FAA generally follows several of the effective management practices that are important for delivering technical training. Clearly delineating accountability for ensuring access to technical training. According to FAA officials, FAA program and training officials have developed a list of mandatory and recommended courses for each inspector position. These training lists contain some technical courses but focus mainly on courses involving the fundamentals of the inspection process (such as courses covering inspection of automation systems, compliance and enforcement procedures, and system safety concepts) and job tasks for each safety inspector specialty. FAA inspection program managers note that the recommended course lists are not more prescriptive for technical training because the need for technical training depends on the specific types of aircraft and equipment with which inspectors work. Thus, decisions on technical training needs are mainly the responsibility of the individual inspectors and their immediate supervisors, in accordance with FAA guidance that provides decision- tree criteria for approving training requests. After inspectors and their supervisors agree on inspectors' technical training requests, regional, headquarters, and academy training executives determine which courses will be taught. From there, FAA training divisions work with the FAA academy to implement the training by developing course schedules and inspector quota allocations. Using a suitable process for selecting inspectors for technical training. FAA's automated training request process provides inspectors with the opportunity to plan for, request, and be selected for the technical training necessary for their positions. The Flight Standards and Aircraft Certification lists of technical and other courses essential for each inspector position, as well as other courses that are available to further inspectors' professional development, are available for review and planning by the inspectors and their supervisors.[Footnote 17] The training system contains information on all the training courses previously completed by the inspectors and outlines their progress toward meeting training requirements. With the supervisor's guidance and approval, each year inspectors request training courses reflecting training needs related to the inspector's position, office inspection activity, and succession planning. However, both FAA and its inspectors recognize the need for more timely selection of inspectors for technical training, another aspect of technical training. This topic is discussed later in this report. Fostering an environment conducive to learning. FAA has provided many of the elements necessary to promote inspectors' learning of technical material, including allowing them time away from work to receive classroom or computer-based training. FAA also has invested in technologies such as computer-based and interactive-video training that help meet the demand for technical and other training. Similarly, FAA has moved some training closer to the inspector duty locations to facilitate and encourage training attendance and has begun experimenting with bringing training to the duty locations when appropriate. FAA's on-the-job training program also gives inspectors hands-on experience with the aircraft and components for which they are responsible. In addition, FAA streamlined the process for acquiring training opportunities that arise on short notice, such as when inspectors are assigned a new aircraft type to inspect. Finally, FAA's training management system allows inspectors to schedule available technical training courses tailored to their individual needs. Acting to obtain inspector buy-in for training goals and priorities. While believing that its inspectors have sufficient technical knowledge to perform inspections, FAA has recognized the need to facilitate communication between inspectors and management in order to gain inspector buy-in for a training program emphasizing system safety over technical courses. Currently, FAA primarily depends on its local office managers (the inspectors' supervisors) to communicate training goals and priorities to the inspectors, mostly during the annual training planning process. According to FAA training officials, this information is also disseminated to inspectors in strategic training plans and other guidance on training. FAA officials further note that inspectors have opportunities to communicate their views on training in course evaluations and employee surveys. Nevertheless, FAA recognizes the need to increase communication between inspectors and management with respect to the training program. Flight Standards recognizes that without inspector buy-in the safety inspectors will not be able to effectively execute system safety oversight and thus this buy-in is recognized as critical to Flight Standards success. FAA is concerned that inspectors have not fully bought into the system safety approach to inspections. In an attempt to gain support for and understanding of the system safety approach and the ways in which inspectors will be affected by the change, Flight Standards plans to host focus groups with management and inspectors, conduct individual interviews with all Flight Standards employees, and create an outreach and communication team to foster better understanding between FAA management and the inspectors. Similarly, Aircraft Certification has planned a number of steps to increase communication between management and the inspector workforce, including facilitated focus groups, individual interviews, and more effective employee feedback mechanisms. These actions may well be needed because, as discussed in the following section, inspectors are generally dissatisfied with the technical training that they receive. Inspectors Are Generally Dissatisfied with the Technical Training That They Receive: Although FAA has followed or is taking steps to follow many of the effective management practices in planning, developing, and delivering technical training, inspectors expressed widespread dissatisfaction with this training. Inspector dissatisfaction covered three areas: (1) having insufficient technical knowledge to do their jobs, (2) not being able to take training they say they needed, and (3) not receiving training in time to do their jobs. One possible explanation for this seeming contradiction is that, although FAA generally employs sound approaches for putting its technical training in place, its actual delivery falls short--the latter being the view of the bulk of its inspector workforce. We were not able to assess this possible explanation because, as discussed at the beginning of this report, we had no practical way to assess the amount of training necessary for inspector proficiency or the timeliness of the training provided. Another possible explanation is that the technical training that FAA provides meets the current and future needs of the agency to a large degree and its inspectors have unrealistic expectations about technical training. This is the view of FAA, and its reasons are discussed later in this section. Having Sufficient Technical Knowledge and Training: On the basis of our survey, we estimate that only about half of FAA inspectors believe, to a great or very great extent, that their technical knowledge is sufficient to enable them to do their jobs properly.[Footnote 18] (See fig. 5.) This belief varies somewhat among inspector specialties. Some inspectors--such as those who specialize in cabin safety and aircraft certification--told us that, to a great or very great extent (78 percent and 68 percent, respectively), they have enough technical knowledge to do their jobs.[Footnote 19] On the other hand, only a third of air carrier avionics inspectors told us that they currently have sufficient technical knowledge to do their jobs.[Footnote 20] Figure 5: Inspectors Responding that to a Great or Very Great Extent They Currently Have Enough Technical Knowledge to Do Their Jobs: [See PDF for image] Note: See table 13 in appendix II for additional results. [End of figure] One reason for the disparity of views concerning technical knowledge among inspectors of different specialties could be their perceived need for specialized knowledge. For example, cabin safety inspectors noted that much of their knowledge of the cabin environment comes from previous experience with airlines and through on the job experience. Similarly, according to FAA, Aircraft Certification inspectors bring with them a high degree of technical knowledge, gained in previous careers in the aviation industry; and typically these inspectors need less technical training than other types of inspectors. Our analysis of training received confirms that aircraft certification and cabin safety inspectors receive less technical training than other inspector specialties. Additionally, as shown above, they are the most satisfied of all inspector specialties that they have the technical knowledge needed to do their jobs. Alternatively, avionics inspectors--who were the least satisfied that they have received enough technical training to do their jobs--indicated that they believe they require specialized knowledge of the avionics systems they inspect. Inspector training data shows that these inspectors receive the most technical training of all inspector specialties. (This topic is discussed in more detail later in the report. See table 6.) However, from our survey, they believe that they need more. Our survey also indicates that most inspectors believe that the technical training they have recently received has not greatly contributed to their ability to perform inspections.[Footnote 21] Specifically, we estimate that about 35 percent of the inspectors believe that the technical training that they received in the last 2 years helped them do their current jobs to a great or very great extent.[Footnote 22] The results ranged from a high of 39 percent for air carrier operations inspectors to a low of 23 percent for general aviation avionics inspectors.[Footnote 23] The higher percentage for operations inspectors could be attributed to the fact that they are required to take flight training on an annual basis, whereas other inspector specialties such as avionics, maintenance, and cabin safety do not have similar requirements for annual training. In comments included with their surveys, inspectors expressed opinions on whether they have sufficient training to do their jobs. Of the 240 inspectors who took the time to write narrative responses about the sufficiency of training, 31 offered positive comments, 105 were strongly negative, and another 119 had weaker negative comments. In addition, 37 inspectors indicated they found themselves inspecting aircraft or components they had not been trained on.[Footnote 24] Our survey also indicates that inspectors believe that most of the technical knowledge they possess was gained in their previous careers in the aviation industry. For inspectors who said that to a great or very great extent they have sufficient technical knowledge to do their jobs, we estimate that 80 percent also noted that the knowledge and skills they brought to FAA from their previous careers contributed, to a great or very great extent, to this technical knowledge.[Footnote 25] We estimated that lower percentages of inspectors from this group rated technical training from FAA instructors (25 percent) and aviation industry sources (41 percent) as contributing, to a great or very great extent, to the technical knowledge needed to perform their jobs.[Footnote 26] Our analysis of survey responses indicates that the amount of time since inspectors left their careers in the aviation industry was not a factor in inspectors' views about their job-related technical knowledge. Newer inspectors were no more likely than longer- tenured inspectors to say that to a great or very great extent they have enough technical knowledge of the aircraft, systems, or operations they inspect to do their jobs. FAA officials indicated to us that inspectors will always believe they need more training. In addition, FAA officials further stated that inspectors need to have only enough technical knowledge of aircraft, systems, and components to be effective inspectors: they need to know enough to ask the right questions, recognize potential problems, and be able to understand issues that arise. Full proficiency with the aircraft and components is not necessary. However, FAA officials indicated that inspectors believe that full or near full proficiency is necessary. An FAA official attributed inspectors' views about the perceived insufficiency of technical training to many of them not fully accepting the agency's transformation to a system safety approach to inspections with its emphasis on risk analysis over technical knowledge. The traditional inspection system relied to a great extent on an individual inspector's technical expertise to identify safety problems with operations or aircraft systems. Although the system safety approach requires inspectors to have an understanding of aircraft and aircraft systems, it is more important for them to have the skills to analyze data to identify vulnerabilities in aircraft operators' and manufacturers' systems for ensuring safety. FAA officials believe that as inspectors gain experience with system safety, they will better understand the more limited role technical knowledge plays in this inspection approach. Most of the inspector training experts we consulted on the inspector technical training curriculum generally agreed with FAA's position. Seven of the 10 experts we contacted told us that the technical courses that FAA offers sufficiently covered existing and emerging technical areas.[Footnote 27] In particular, they noted that the flight training for operations inspectors was adequate for performing flight checks. However, two experts were concerned that there was not enough training in advanced aviation technologies for maintenance and avionics inspectors. These experts thought that maintenance and avionics inspectors should have periodic refresher training that would allow them to become more familiar with changes in the aircraft and systems they deal with during inspections. The experts did not comment on whether individual inspectors receive all the technical training necessary for their positions since this would have required an extensive, detailed review of training records. Most representatives from airlines we contacted were at least moderately satisfied that FAA inspectors have sufficient technical knowledge and training.[Footnote 28] On the basis of their experience with FAA inspectors, 19 of the 23 airline representatives we consulted said that to a great extent (7 responses) or moderate extent (12 responses) FAA inspectors had the technical knowledge to fulfill inspection responsibilities. Regarding training, 16 of the 23 thought that to a very great extent (1 response) or moderate extent (15 responses) FAA inspectors have the technical training to fulfill these responsibilities. Being Able to Take Requested Training: Also reflecting the disparity in views between inspectors and FAA management concerning technical training, many inspectors indicate that they are not greatly encouraged to take technical training and that to a large extent they do not get all the technical training they request. However, inspectors' views are not supported either by FAA training request data or from progress made in taking training courses deemed essential by FAA.[Footnote 29] On the basis of our survey, we estimate that less than half (43 percent) of inspectors think their supervisors encourage them to request the technical training needed to do their current jobs.[Footnote 30] We also estimate that about 28 percent of FAA's inspectors believe, to a great extent or very great extent, that they receive the technical training that they request.[Footnote 31] (See fig. 6.) If we include responses citing receiving requested training to a moderate extent, we then estimate that about 49 percent of inspectors overall believe that they receive the training they request at least to a moderate extent.[Footnote 32] Figure 6: Extent to Which Requested Technical Training Is Approved: [See PDF for image] Note: See table 14 in appendix II for additional results. [End of figure] Our survey also indicates that inspectors with longer tenures at FAA have more difficulty getting technical training they request than inspectors who have recently been hired. According to the survey, of those inspectors with 10 or more years with FAA, we estimate that 55 percent said that requested technical training was approved to some or no extent as compared to an estimated 35 percent of inspectors who had been with FAA 3 years or less.[Footnote 33] This may be due in part to newer inspectors' technical training opportunities tending to be essential courses, requests likely to be approved. In contrast, inspectors with more experience request courses outside of FAA requirements, requests more likely to be denied. As discussed earlier, FAA records did not allow us to assess the merits of the inspectors' views. FAA's data on the extent to which requested courses were taken do not support the inspectors' contention that they do not receive requested training. According to Flight Standards data, for fiscal years 2003 through 2005, on average FAA approved about 90 percent of requested technical training and has approved a similar percentage for upcoming fiscal year 2006. (See fig. 6.) FAA officials note that these data do not include late course cancellations that occur after the training schedule for the year is set. Though we were unable to obtain similar data from Aircraft Certification data on course cancellations or denials, these courses occasionally get cancelled or changed. In fiscal year 2005, one course was cancelled by Aircraft Certification. Officials told us that if a training course essential for an inspector to perform their job is cancelled, the inspector will be placed in the next available course. Aircraft Certification inspectors represent about 5 percent of the inspector workforce (excluding supervisors, managers, and others in the aviation safety inspector job series who do not perform front-line inspections). According to FAA, the agency tries to accommodate inspectors' requests for technical training to the extent possible. However, an inspector's request for technical training may be denied because (1) the inspector's need for the course was not adequately justified based on the inspector's current position, (2) the inspector had already completed a similar course, or (3) insufficient funding was available. Officials also said that, infrequently, a technical course requested by an inspector may be cancelled due to low enrollment or because its content is outdated. According to the officials, when a requested course is cancelled, inspectors can request it the next time it is offered, and in the meantime they can choose a replacement course from the list of courses for their position. Inspectors' views on why they did not get training that they requested corresponded somewhat with the reasons that FAA cited. We estimate that about 54 percent of the inspectors believe that lack of funds hindered or greatly hindered their ability to get requested technical training.[Footnote 34] (See fig. 7.) Inspectors cited other reasons somewhat less frequently: about 36 percent cited availability of courses, 28 percent cited impact on their workload, and 27 percent cited management's determination about the need for them to attend the course as hindering or greatly hindering their ability to receive the training they requested.[Footnote 35] Because it was impractical to investigate the reasons why thousands of training requests were granted or denied, we were not able to reconcile inspectors' views with FAA data. Figure 7: Inspectors' Views on Factors Hindering Their Ability to Take Requested Technical Training: [See PDF for image] Note: See tables 15 through 18 in appendix II for additional results. [End of figure] In addition to receiving most of the technical training they request, our analysis of inspectors' training records indicates that most are making good progress in taking the technical training FAA considers essential for their jobs. Our analysis of FAA training data[Footnote 36] indicates that over half of the inspectors have completed at least 75 percent of their essential technical training courses for their positions.[Footnote 37] (See table 5.) In addition, more than three- quarters have finished at least half of these essential technical courses. However, only 20 percent of air carrier avionics inspectors have completed 75 percent of their technical courses. Avionics inspectors have the most technical training requirements, due to the complexity of the aircraft components they inspect. Table 5: Percent of Inspectors Completing Essential Technical Courses: Type of inspector: Air carrier avionics; Percent of inspectors completing at least 75 percent of technical courses: 20; Percent of inspectors completing at least 50 percent of technical courses: 69. Type of inspector: Air carrier maintenance; Percent of inspectors completing at least 75 percent of technical courses: 46; Percent of inspectors completing at least 50 percent of technical courses: 82. Type of inspector: Air carrier operations; Percent of inspectors completing at least 75 percent of technical courses: 36; Percent of inspectors completing at least 50 percent of technical courses: 74. Type of inspector: Cabin safety; Percent of inspectors completing at least 75 percent of technical courses: 59; Percent of inspectors completing at least 50 percent of technical courses: 85. Type of inspector: General aviation avionics; Percent of inspectors completing at least 75 percent of technical courses: 52; Percent of inspectors completing at least 50 percent of technical courses: 78. Type of inspector: General aviation maintenance; Percent of inspectors completing at least 75 percent of technical courses: 53; Percent of inspectors completing at least 50 percent of technical courses: 84. Type of inspector: General aviation operations; Percent of inspectors completing at least 75 percent of technical courses: 69; Percent of inspectors completing at least 50 percent of technical courses: 88. Type of inspector: Aircraft certification[A]; Percent of inspectors completing at least 75 percent of technical courses: N/A; Percent of inspectors completing at least 50 percent of technical courses: N/A. Type of inspector: All inspectors; Percent of inspectors completing at least 75 percent of technical courses: 46; Percent of inspectors completing at least 50 percent of technical courses: 80. Source: GAO analysis of FAA data. Note: N/A = nonapplicable. [A] Aircraft certification inspectors have no essential technical courses according to the definition of technical training used in this report. Aircraft Certification considers all training related to the inspection process as technical training. [End of table] There are several reasons that may explain why inspectors have not completed most or all of their essential training--technical and other. First, a significant portion of the inspector workforce is relatively new to the agency and would thus not be expected to have completed the essential training. In fact, FAA data show that 28 percent of the inspectors have been employed by FAA for less than 5 years. Second, inspectors change specialties, which can affect their training requirements. Third, FAA allows inspectors to substitute prior experience for some essential technical training courses, and such substitutions are not always reflected in inspector training records. Finally, because the lists of essential training courses have been developed only within the past few years, some inspectors may not have had time to complete new essential courses. FAA officials emphasized that, especially in Flight Standards, training is carried out over the course of an inspector's career rather than occurring primarily at the beginning of the career. According to Aircraft Certification officials, although their service's inspectors receive training over the course of their careers, they receive the majority of their training within the first year on the job. This early training emphasizes the skills needed to perform inspections. Overall, according to our analysis, inspectors have taken an average of 3.4 technical training courses from fiscal years 2002 through 2004, or about one per year. (See table 6.) Avionics and maintenance inspectors have taken more technical training on average. Generally, these avionics and maintenance inspectors require more technical training than other inspector specialties because they often inspect several different models of aircraft. Table 6: Average Number of Technical and Nontechnical Training Courses Taken, Fiscal Years 2002 through 2004: Type of inspector: Air carrier avionics; Technical courses: 4.8; Nontechnical courses: 9.6. Type of inspector: Air carrier maintenance; Technical courses: 3.9; Nontechnical courses: 9.6. Type of inspector: Air carrier operations; Technical courses: 2.6; Nontechnical courses: 9.7. Type of inspector: Cabin safety; Technical courses: 1.2; Nontechnical courses: 8.5. Type of inspector: General aviation avionics; Technical courses: 4.0; Nontechnical courses: 8.8. Type of inspector: General aviation maintenance; Technical courses: 4.0; Nontechnical courses: 9.9. Type of inspector: General aviation operations; Technical courses: 3.1; Nontechnical courses: 9.6. Type of inspector: Aircraft certification; Technical courses: 1.2; Nontechnical courses: 5.4. Type of inspector: All inspectors; Technical courses: 3.4; Nontechnical courses: 9.4. Source: GAO analysis of FAA data. [End of table] Receiving Training in a Timely Manner: With the rapid development of aircraft and aircraft components, especially aircraft avionics, a training delivery mechanism that is responsive to these changes is critical. For the most part, FAA inspectors are dissatisfied with receiving the technical training they need in time to do their jobs. We estimate that only 20 percent of inspectors believe to a great or very great extent that they have received technical training in time to do their jobs.[Footnote 38] (See fig. 8.) No more than one-third of any type of inspector thought that technical training was timely to a great or very great extent, and none of the general aviation avionics inspectors who responded to our survey thought that this was so.[Footnote 39] Avionics are the most rapidly changing technological components of aircraft, which could account for this result. As discussed at the beginning of this report, FAA's records did not allow us to assess the extent to which inspectors received training before they conducted inspection activities related to that training. Figure 8: Inspectors' Views on the Extent to Which Technical Training Is Delivered in a Timely Manner: [See PDF for image] Note: See table 19 in appendix II for additional details. [End of figure] Similarly, we estimate that only about 23 percent of all inspectors indicated that they always or frequently received technical training on the equipment they were to inspect prior to scheduled inspection activities.[Footnote 40] (See fig. 9.) No more than 35 percent of inspectors in any specialty responded this way. In comments supplied with their surveys, many inspectors expressed the view that FAA is slow to react to changes in industry technology and slow to develop courses in response to the changes. Figure 9: Inspectors' Views on the Extent to Which They Received Technical Training Prior to Scheduled Oversight Activities: [See PDF for image] Note: Approximately 4 percent of inspectors responded that they had no basis to judge or did not know. See table 20 in appendix II for additional details. [End of figure] FAA has recognized the need to provide training on a timelier basis and has taken some actions that have yet to be fully implemented. One of the goals of Flight Standards is to establish a way to ensure that training is current and well designed, can be tailored to the needs of the individual employees, and is administered in a fast and flexible way in response to changing needs. Flight Standards plans to improve training delivery by taking advantage of new delivery mechanisms, increasing utilization of vendors where appropriate, and streamlining training programming and scheduling to reduce the lag time between the identification of training needs and the delivery of training. In addition, Flight Standards recently instituted a process to continuously monitor courses and to update their content when changes in FAA policy or aviation technology warrant doing so. Ensuring that course content is up to date and that courses are available when needed is an important aspect of delivering timely training. In fiscal year 2005,[Footnote 41] Flight Standards developed 5 new courses, revised 16 existing courses, and completed 13 course evaluations.[Footnote 42] Similarly, Aircraft Certification officials indicated that they have been evolving toward a more integrated approach to training delivery, mixing classroom training with Web-based technologies, on-the-job training and adding additional job aids, in part, for providing more timely training. These officials also noted that Aircraft Certification has a long history of providing just-in-time training when new work processes or job-related information needs to be disseminated to inspectors quickly. Although FAA Uses Several Approaches to Evaluate Technical Training Provided, Assessing Impact on Performance Remains to Be Done: For the most part, FAA has followed--or has begun to implement-- effective management practices in evaluating its efforts to provide technical training to inspectors and ensuring that this training leads to improved performance and results. (See table 7.) For example, it continuously assesses technical training through participant end-of- course evaluations and surveys of inspectors and supervisors that focus on the application of skills and knowledge to the job. While FAA's evaluation efforts provide information about these areas, these assessments have not measured the impact of training on FAA's mission goals, such as reducing accidents. Isolating improvements in mission performance that are a result of training programs is difficult for any agency. Table 7: Extent That FAA Followed Effective Management Practices in Evaluating Its Training Program: Effective management practice: Systematically plans for and evaluates the effectiveness of training and development efforts; Extent followed: Mostly. Effective management practice: Uses the appropriate analytical approaches to assess its training and development programs; Extent followed: Mostly. Effective management practice: Uses appropriate performance data (including qualitative and quantitative measures) to assess the results achieved through training and development efforts; Extent followed: Partially. Effective management practice: Incorporates evaluation feedback into the planning, design, and implementation of its training and development efforts; Extent followed: Fully. Effective management practice: Incorporates different perspectives (including those of line managers and staff, customers, and experts in areas such as financial, information, and human capital management) in assessing the impact of training on performance; Extent followed: Mostly. Effective management practice: Assesses the benefits achieved through training and development programs; Extent followed: Partially. Source: GAO. [End of table] FAA Mostly Follows Several Effective Practices for Evaluating Technical Training: FAA has taken several actions to evaluate the effectiveness of its technical training efforts. Collectively, the actions generally cover the effective management practices cited in table 7 by continuously and systematically evaluating the technical courses FAA provides for inspectors. In performing these evaluations, FAA has focused primarily on obtaining inspectors' and, to some extent, their supervisors' views on individual courses. FAA requires that participant evaluations be distributed after each training course for inspectors. The evaluations ask participants to rate the extent to which the course and course material (e.g., workbooks, slides, labs, and tests) met objectives as well as the extent to which the instructor provided assistance. According to FAA, participants return the evaluations 95 percent of the time. FAA also sends surveys to inspectors and their supervisors 90 to 180 days after course completion to obtain their perspectives on whether the course was needed and the extent to which the inspector is applying new skills and knowledge to the job. FAA reports that since the inception of post-course surveys, the return rate from inspectors and supervisors has ranged from 49 to 50 percent. The post-course survey results from the six most highly attended technical courses in the last 2 years reflected generally positive responses. These findings suggest that survey respondents generally think that the individual technical courses they received helped them in their jobs. Results from both the participant course evaluations and post-course surveys are automated and are available to training officials. In addition, according to Flight Standards training officials, they assess all complaints concerning a course and discuss the issues identified with the Flight Standards office that sponsors the course. According to FAA, it is the responsibility of the mentor for each training course to use the information from the participant evaluations and post-course surveys as well as other tools to determine if courses are meeting their objectives and enhancing inspectors' ability to do their jobs. In February 2005, Flight Standards established a policy that its course mentors evaluate each course for which they are responsible at least every 3 years using a standardized approach. According to the policy, course mentors should review the results of participant evaluations and post-course surveys as well as personally sit in the course to determine if a course is still current and is meeting objectives. Flight Standards has a performance plan initiative to track the completion of planned course evaluations. Flight Standards began training mentors on these and other mentor responsibilities and procedures in April 2005, and some course mentors have already begun thorough evaluations of their courses. Prior to this date, Flight Standards officials said that participant evaluations and post-course surveys were used by its Quality Assurance Branch in its annual course reviews and were routinely reviewed by FAA academy course managers and their supervisors to update or improve courses. However, because Flight Standards had not assigned specific individuals to be responsible for a particular course, some requests for updates were not tracked. According to Flight Standards officials, with the implementation of the course mentor program, each course will now have a point of contact for all course improvements and updates. Aircraft Certification is implementing a new approach for evaluating courses that officials believe will provide course mentors and other training officials more comprehensive information on technical courses sponsored by each office. The approach is based on the work of Dr. Robert O. Brinkerhoff, in particular his Success Case Method and High Impact Learning approach.[Footnote 43] This approach helps to increase and demonstrate organizational results from learning. With the Success Case Method, post-course surveys are used to gauge the extent of reported application of learning and are then validated through personal interviews with selected course participants. Using such interviews, Aircraft Certification seeks to determine whether and how much training was actually transferred to the job and if there was an impact on the organization as a result of the training. The High Impact Learning methodology allows for up-front evaluation prior to development or revision of courses to ensure that the objectives of the proposed training will lead to organizational impact through "impact mapping" of course objectives to organizational goals. Results of Success Case Method learning evaluations will be available to the course mentor, course managers at FAA's Training Academy, and program officials at FAA headquarters. Aircraft Certification began implementing the evaluation tool in spring 2005 and has thus far prototyped it with one course and plans to have it applied to all its technical courses within 2 years. FAA has also surveyed employees for their views on training in general, and one of these surveys will lead to revisions to the overall inspector curriculum, according to FAA. Every 2 years, FAA surveys all of its employees about many aspects of their employment, including the training they receive. This employee attitude survey asked employees about the extent to which they received the training they needed to effectively perform their jobs and whether or not they have been able to apply that training. However, the survey does not ask employees to differentiate between the types of training they receive, such as that relating to inspection processes or technical skills. In addition, although FAA isolates responses according to employee's work location- -such as headquarters and Flight Standards and Air Certification field offices--it does not ask respondents their position, so inspectors' responses cannot be identified. As a result, although the survey can be useful for FAA's workforce as a whole, it is not as useful for isolating safety inspectors' attitudes about their technical training. In order to obtain information on inspector training in particular, in August 2004, Flight Standards conducted a separate survey of 51 field inspectors and 8 field and 3 headquarters inspection program managers that revealed inspector dissatisfaction about several aspects of training. Because of its limited nature, Flight Standards recognizes that the survey does not necessarily represent the views of the entire inspector workforce. Of the 51 field inspectors who agreed or strongly agreed with certain statements--the only large group surveyed--21 percent indicated they received the training they needed, 10 percent said training is current and technically up to date, and less than 20 percent indicated training supports current or future job requirements in Flight Standards.[Footnote 44] Flight Standards officials expressed concern about inspectors' negative views toward the training they receive. Officials said that as part of their plans for a curriculum for inspectors, they will identify and implement measures necessary to monitor course satisfaction and content currency, and the data they gather from monitoring the course will provide the basis for continuous course improvement. According to Aircraft Certification officials, their office has not undertaken any similar surveys to field inspectors. FAA officials said that they also encourage inspectors to submit suggestions for revising existing courses or adding new courses to provide training in the technical skills not covered in the current curriculum. According to Flight Standards training program officials, inspector suggestions for new courses are placed in a pool of potential new courses, which are reviewed by program staff on the basis of need and the availability of funds. Suggestions for course revisions are now reviewed by course mentors as part of the course evaluation process.[Footnote 45] Although FAA has a process for inspectors to make recommendations regarding technical courses, our survey indicated that many inspectors are either not aware of or do not take advantage of this process. According to our survey, we estimate that 55 percent of the inspectors believe they have had an opportunity to recommend new courses for their position to some or no extent.[Footnote 46] In addition, 49 percent thought they had an opportunity to recommend new content in existing courses for their position to some or no extent.[Footnote 47] FAA Lacks Comprehensive Data on How Technical Training Contributes to Improved Performance and Results: The analytical approach FAA employed for evaluating technical training programs emphasized individual course evaluations and employee surveys that collect useful, but limited, information on the effectiveness of technical training courses. According to experts on training in government, agencies should adopt a balanced, multilevel approach to evaluating their training and development efforts. One commonly accepted model is the Kirkpatrick model, which consists of five levels of assessment.[Footnote 48] The first level measures the training participants' reaction to, and satisfaction with, the training program or planned actions to use new or enhanced competencies. The second level measures the extent to which learning has occurred because of the training effort. The third level measures the application of this learning to the work environment through changes in behavior that trainees exhibit on the job because of the training or development program. The fourth level measures the impact of the training program on the agency's program or organizational results. Finally, the fifth level--often referred to as return on investment--compares the benefits (quantified in dollars) with the costs of the training and development program. As discussed earlier, the course evaluations and surveys FAA uses to evaluate its technical and other training programs for inspectors cover, to some extent, the first three levels of assessment. Aircraft Certification has taken the first steps in evaluating the impact of training on organizational results (the fourth level) by linking course objectives to organizational goals, and Flight Standards is in the initial stages of implementing a process to assess the return on investment from the courses in its training program. Experts acknowledge that isolating performance improvements resulting from training programs and the cost-effectiveness of these programs is difficult for any organization. Federal agencies, such as FAA, have to consider the feasibility and cost-effectiveness of conducting these in- depth evaluations, along with budgetary and staffing circumstances that may limit the agencies' ability to complete such evaluations. The challenge of performing evaluations of the impact and cost- effectiveness of its training efforts is great for FAA. Along with undertaking these evaluations, FAA must also determine how its ongoing shift to a system safety approach to inspections is affecting its organizational goals of reducing accidents and increasing the overall safety of flying. Industry Provides Much of FAA's Technical Training; Additional Safeguards Needed to Prevent Real or Appearances of Conflicts of Interest: Over the past 3 years, the aviation industry has provided about 40 percent of the technical training for FAA's safety inspectors. To a limited degree, inspectors have received training from the aviation industry in exchange for in-kind services or at no cost to FAA. Although FAA has taken steps to address concerns over possible real or apparent conflicts of interest resulting from receiving this training, it has not consistently applied these policies. FAA Contracts with the Aviation Industry for Much of the Technical Training Provided to Inspectors: FAA provides technical training to its inspectors either through the FAA Academy (with courses taught in Oklahoma City or at FAA regional locations) or through contracts with the outside training providers, including some in the aviation industry that FAA regulates under FAA's gift authority. Technical training provided by the aviation industry includes: * pilot training; * aircraft maintenance training; * training covering inspection technologies and procedures; and: * training on aircraft systems, structures, and components. FAA has increasingly relied on industry to provide technical training to its inspectors over the past 3 fiscal years. In fiscal year 2004 (latest data available), industry delivered nearly half of FAA's technical training. (See fig. 10.) Industry-provided training occurs most frequently for air carrier and general aviation operations inspectors because it is often more economical to have flight training provided by an outside vendor than for FAA to maintain or lease its own aircraft for this purpose. Figure 10: Percent of Technical Training Provided by Industry as Reported by FAA, Fiscal Years 2002 through 2004: [See PDF for image] Note: Numbers reported as a percent of total FAA-and industry-provided training for each type of inspector. See table 21 in appendix II for additional details. Because aircraft certification inspectors receive limited technical training, the percent of this training provided by industry can vary widely from year to year. [End of figure] Under Limited Circumstances, FAA Receives Training in Exchange for In- Kind Services: In addition to paying industry to provide technical training for inspectors, FAA employs two arrangements by which inspectors obtain training from the aviation industry in exchange for in-kind services. In return for receiving training from the aviation industry, FAA delegates certain regulatory authority to qualified employees of the entity being overseen (called quid pro quo arrangements).[Footnote 49] Both programs apply to Flight Standards operations inspectors called aircrew program managers and training center program managers. FAA's Aircraft Certification service does not have any equivalent arrangements by which inspectors receive training in exchange for in- kind services. Under the aircrew designated examiner program, there is an arrangement between FAA and major passenger-carrying airlines, cargo-only carriers, and regional airlines by which FAA delegates, under the supervision of FAA inspectors, certain pilot certification authority and responsibility to pilots of the airline.[Footnote 50] In exchange, FAA inspectors receive training in the airline-specific programs and procedures in airline-specific aircraft or simulators at no cost to the agency. Airlines benefit from the increased flexibility of being able to certify their own pilots and not having to arrange and schedule certification by an FAA inspector. FAA also benefits from this flexibility because delegating the certification activities increases its capacity for and efficiency of its oversight and management activities. In addition, because the training received is airline- specific, it further enhances the inspectors' knowledge of the specific aspects of the airlines' operations that they are responsible for overseeing. The aircrew designated examiner program originated in 1982 as an agreement between a single airline and FAA, stemming from FAA's inability to meet industry's increasing demand for certification specialists. Each agreement between FAA and the airline is governed by a memorandum of understanding that outlines the reasons for establishing the specific aircrew designated examiner agreement, lists the aircraft types involved, and contains an overview of how the program requirements will be met by both parties. This arrangement was approved by the FAA ethics officer and was reviewed by the Department of Transportation Inspector General. FAA does not keep central records of training received through these arrangements, and it was not practical for us to gather the data from over 100 FAA field locations. Therefore, we asked FAA's nine regional office officials to contact their respective flight standards district offices and certificate management offices to compile these data. Some data was provided to us with incomplete information, sometimes without names or specific dates. Because of the many remote locations that gathered this information for us, it was not practical for us to independently verify the completeness or accuracy of these data. Overall, FAA regional office-supplied data indicate that FAA has memoranda of understanding with 61 airlines and 42 training centers under the two programs, encompassing about 300 fleets in total. (See table 8.) Table 8: Number of Memoranda of Understanding and Fleets Enrolled as Part of the Aircrew Designated Examiner Program and Agreements with Training Centers: Aircrew designated examiner program; Number of memoranda of understanding: 61; Number of fleets: 141. Training centers; Number of memoranda of understanding: 42; Number of fleets: 162. Total; Number of memoranda of understanding: 103; Number of fleets: 303. Source: GAO analysis of FAA supplied data. Note: This information may not be complete and was not independently verified. (See text.) [End of table] For the aircrew designated examiner program, FAA regional officials indicated that the agency has memoranda of understanding with 61 of the 134 airlines, and these agreements cover 141 aircraft fleets. More than 175 training activities took place per year, on average, from fiscal years 2002 through 2004, representing nearly 30 percent of all technical training received by air carrier operations inspectors. (See table 9.) Table 9: Numbers of Inspectors Trained under Aircrew Designated Examiner Program and Agreements with Training Centers, Fiscal Years 2002 through 2004: Aircrew designated examiners: Number of inspectors trained; 2002: 111; 2003: 114; 2004: 155; Average per year: 127. Aircrew designated examiners: Number of training activities; 2002: 155; 2003: 170; 2004: 200; Average per year: 175. Training centers: Number of inspectors trained; 2002: 43; 2003: 45; 2004: 59; Average per year: 49. Training centers: Number of training activities; 2002: 51; 2003: 52; 2004: 73; Average per year: 59. Source: GAO analysis of FAA supplied data. Note: This information may not be complete and was not independently verified. (See text.) [End of table] Similar to the aircrew designated examiner program, FAA's Flight Standards office also employs memoranda of understanding with private, FAA-certified training centers that provide training, testing, and pilot certification services to commercial and private pilots throughout the United States. Under these agreements, certain training center employees may be certified by FAA to serve as designees provided they meet FAA requirements. On behalf of FAA, these training center designees certify commercial and private pilots as qualified to operate an aircraft. FAA assigns one or more inspectors to each training center. The inspector is responsible for FAA regulatory management and oversight of the training center through periodic inspections of training center equipment, training courses, course materials, and instructors. As part of the agreement granting designee authority to the employee of the training center, the FAA inspector receives aircraft-specific training from the training center at no cost to the agency. This training benefits FAA by increasing inspector knowledge and familiarization with the actual equipment being inspected, thereby providing more effective oversight of the training center. In addition, FAA does not have to utilize inspectors to certify the individual pilots. The training center benefits from having its own employees authorized to certify the pilots attending training at the center, rather than having to schedule and wait for FAA inspectors to accomplish the certifications. According to FAA regional officials, FAA has agreements with 42 of the approximately 50 training centers across the United States that include 162 aircraft simulator fleets. From the data FAA regional offices supplied, we determined that an average of 59 instances of training per year occurred under these arrangements from fiscal years 2002 through 2004. (See tables 8 and 9.) In total, the technical training provided by industry sources in exchange for in-kind services through these two sets of arrangements accounts for approximately 17 percent of all industry-provided training and approximately 7 percent of all technical training provided to inspectors. The memoranda of understanding described above were formalized, in part, to eliminate actual or appearances of a conflict of interest. The U.S. Government Standards of Ethical Conduct precludes federal employees from accepting gifts, including training, from those whom they regulate. An exception includes anything that is paid for or secured by the government under contract. FAA considers the granting of check airmen authority to designated examiners to be payment in-kind for the training received by FAA inspectors. The purpose of the memoranda of understanding for the two arrangements discussed above is to outline the nature of the payment in-kind that FAA will provide to eliminate the appearance that FAA is receiving a service for free. The memoranda of understanding address the conflict-of-interest issue by explicitly outlining the duties and responsibilities of both FAA and the operator employees who are party to the agreement, and they also outline the specific nature of the in-kind exchange. For the aircrew designated examiner program, a sample memorandum of understanding was written into FAA guidance in 1989. In 1997, a Department of Transportation Office of the Inspector General report expressed concerns that the aircrew designated examiner program might entail a conflict of interest by precluding FAA from enforcing its safety regulations.[Footnote 51] As a result, FAA altered the memoranda of understanding by eliminating any language that could be construed as limiting FAA's oversight authority and by specifically adding language to the contrary. Unlike the memoranda of understanding for the aircrew designated examiner program, the memoranda of understanding between FAA and training centers do not contain a provision stating that FAA will take enforcement action against any individual who violates any regulation. When we brought this to their attention, FAA officials indicated that the absence of this enforcement language in the training center memoranda of understanding was most likely the result of a simple oversight. As a result of our inquiry, FAA officials told us that they are revising their guidance to incorporate this enforcement language for future arrangements under a memorandum of understanding. Although this action will address any concerns about future arrangements, it does not make enforcement authority explicit under existing arrangements. Some Inspectors Receive Free Training without Getting Approval from FAA Legal Counsel: Some safety inspectors have received training opportunities from aircraft manufacturers or operators that they regulate at no cost to FAA and without providing an in-kind service in exchange. FAA requires that any such free training opportunities be reviewed by FAA legal counsel, at the regional or headquarters level, to determine propriety of accepting the training. However, some Flight Standards inspectors have received free training for which FAA gave them training credit in the absence of prior approval from legal counsel. According to Aircraft Certification officials, its inspectors do not receive training credit for free training offered by aircraft manufacturers, although both Aircraft Certification and Flight Standards inspectors often audit classes, at no charge to FAA, without receiving credit. As mentioned above, U.S. Government Standards of Ethical Conduct generally preclude federal employees from accepting gifts, including training, from those they regulate. FAA generally does not accept offers of gifts unless there is some recognized need, or if acceptance will result in cost savings or other benefits in carrying out its work. FAA is allowed, however, to receive free training from the aviation industry in limited circumstances as a gift under the FAA Administrator's gift acceptance authority. Under this authority, the FAA Administrator can accept any gift of services in carrying out aviation duties and powers. FAA's Chief Counsel concluded in 1988 that FAA may accept free training if the session is necessary for the employee to perform his or her responsibilities, with respect to the provider's projects, and the information cannot be obtained from another source. Before inspectors can receive free training, they must obtain approval from FAA legal counsel, either at the regional or national level. Because FAA does not keep central records of free training received, and gathering this data was not practical, we asked FAA's nine regional offices to request and compile this information for us. In response, two of the nine regions indicated that they accept free training on a limited basis. One region cited 57 instances, over the past 4 years, of training accepted and credited to inspectors' personnel records that was provided free of charge. We were able to independently verify only 12 of these instances because some of the data lacked specific dates, some lacked inspector names, and some of the training was presented as completed but with no specific information at all. For the second region, although it could not identify specific records of such training, an official indicated that perhaps five instances of free training were accepted over the past 3 years. FAA's policy regarding acceptance of free training is not well known or uniformly applied by its regional offices. Though some regions indicated that the acceptance of free training of any type is not allowed under any circumstances, other regions were unsure how the policy for acceptance of this type of training is applied. Regarding the two regions that told us they had accepted free training, an official from the first region indicated that sometimes the regional legal counsel's office was asked to comment on the propriety of the training and sometimes not. In cases where legal counsel determined the training was improper, according to the official, the training was not accepted. An official from the second region indicated that free training opportunities were taken advantage of when normal FAA channels for obtaining the same training were often slow or difficult; therefore, accepting this training became necessary if inspectors were to receive it at all. Similar to the official from the first region where free training is routinely accepted, this official indicated that legal counsel was sometimes contacted for an opinion on the propriety of accepting a specific instance of training, and sometimes not. Because these opportunities generally arise at the local office level, whether such an offer is reviewed by legal counsel is dependent on the office manager, the manager's understanding of the FAA policy, and a judgment about whether a specific training opportunity raises any concern that should be reviewed by legal counsel. Both the government standards of ethical conduct and FAA policy address the propriety of accepting gifts and free training, but FAA has not clearly communicated this policy and the processes for accepting free training to its regional offices. None of the regions we contacted were able to cite any specific, relevant policy guidance governing this issue. Several of the FAA officials we interviewed cited "verbal policy" from FAA headquarters and a general, long-standing understanding that acceptance of such training is not allowed. Other regional officials indicated that although acceptance of free training is generally to be avoided due to conflict-of-interest considerations, they would treat each occurrence separately and likely consult with the regional legal counsel for an opinion on the propriety of accepting free training. In fact, one region supplied us with an opinion from the regional legal counsel, stating that as a general rule, FAA has long held that the agency must pay for its own training. The document goes on to say that it is permissible to accept such an opportunity to audit the class but warns that an inspector is not to consider it formal training. Many regional and headquarters officials we spoke with indicated that it is common practice for FAA inspectors to audit training in this manner, for informational purposes and not for formal FAA training credit. On the basis of our survey, about 37 percent of inspectors indicated that, in the past 2 years, they have attended or inspected a technical training course offered by an airline or manufacturer for which they did not receive credit.[Footnote 52] FAA headquarters officials agreed that the FAA order governing the acceptance of gifts and the government's standards of ethical conduct address the broad issue of gift acceptance. However, our work indicates that these policies may not be clearly and uniformly understood by the FAA regional offices. Conclusions: In providing training to its inspectors, FAA follows many of the effective management practices we have outlined in our guide for assessing training and development efforts in the federal government. In doing so, FAA has put in place thoughtful, structured processes for linking training to strategic goals, identifying and developing courses to improve individual and agency performance, actively encouraging and supporting technical training, ensuring that inspectors have opportunities to receive this technical training, and obtaining inspectors' and their supervisors' views on the extent to which technical training affects job performance. FAA also recognizes the need for improvements, including (1) systematically assessing inspectors' needs for technical and other training, (2) better timing of technical training so that inspectors receive it when it is needed to perform their jobs, and (3) better linking the training provided to achieving agency goals of improving aviation safety. FAA has begun to act in these areas, and we believe that, if effectively implemented, the actions should improve the delivery of training and ultimately help lead to fewer aviation accidents, fatalities, and injuries. Therefore, it is important for FAA to follow through with its efforts. FAA's plans for inspector training are premised on the assumption that inspectors currently have enough technical proficiency overall and that future training efforts should be geared toward closing gaps in proficiencies that the agency has determined inspectors require for system safety inspections, such as risk assessment and data analysis. However, FAA has not convinced inspectors of the merits of its approach nor has it systematically identified inspectors' training needs for conducting system safety inspections. Inspectors instead believe that they are not receiving all the training they need to stay current with rapidly changing aviation technologies. Many inspectors spoke out strongly on this issue--it is clearly a hot-button topic for them. Therefore, it is essential that as FAA continues to implement a system safety inspection process, it works closely with inspectors to demonstrate the benefits of the system safety approach, how inspectors' technical and other training needs will be met, and how aviation safety will benefit from a system safety approach. Finally, FAA has recognized that the manufacturers and operators of aircraft and aircraft systems can be the best source of much of the technical training for its inspectors. While FAA pays for most of the training its inspectors receive from aviation sources, some of this training is provided at no cost or in exchange for in-kind services. However, because FAA keeps only scattered records on the extent to which such training occurs, we cannot tell how widespread it is or whether FAA legal counsel reviewed each training activity for propriety. FAA has not communicated its policy on the acceptance of training without charge; and, as a result, some FAA regions have accepted training that has not been approved and could pose conflict-of- interest issues--or the appearance of such a conflict--for the agency. Recommendations for Executive Action: We are making five recommendations, three involving technical training and two involving industry provided training. Regarding technical training, we recommend that the Secretary of Transportation direct the FAA Administrator to complete the following two actions that are either planned or are in early stages of development or implementation: * To ensure that inspector technical training needs are identified and met in a timely manner, the Administrator should systematically assess inspectors' technical training needs, increase inspector involvement in the decision-making process for assessing the need for courses, including the need for more training for maintenance and avionics inspectors to familiarize them with recent changes in aviation technology, and ensure the technical curriculum meets those needs. The Administrator should also take the actions needed, including developing guidelines for inspectors, supervisors, and training managers, to ensure that technical training is requested and delivered closer to the time it is needed to help inspectors perform their jobs. * With a view toward maximizing the contributions of training to furthering FAA's safety mission, FAA's training organizations should determine the feasibility of developing measures of the impact of inspector training, including technical training, on achieving organizational goals. Third, to gain better acceptance from the inspector workforce for changes being made and planned for the inspector training curriculum, we recommend that the Secretary of Transportation direct the FAA Administrator to increase the focus of its training efforts on how system safety/risk management will improve inspections and aviation safety. Fourth, we recommend that the Secretary of Transportation direct the FAA Administrator to ensure that all existing and future memoranda of understanding pertaining to training received in exchange for in-kind services contain language stating that the agreement does not preclude FAA from fulfilling its oversight and enforcement role. Finally, to preclude situations where the provision of free training by the aviation industry may create a conflict of interest or result in the appearance of such a conflict, we recommend that the Secretary of Transportation direct the FAA Administrator to review its policies on the acceptance of free training accepted from the aviation industry to ensure they are understood by inspectors, supervisors, managers, and regional counsel; implement a process for monitoring field office compliance with these policies; and follow up on any noncompliance. Agency Comments and Our Evaluation: We provided a draft of this report to the Department of Transportation and received comments from FAA officials, including its Deputy Associate Administrator for Aviation Safety. FAA generally agreed with the report's findings and agreed to consider our recommendations. The FAA representatives appreciated the report's positive recognition of its efforts to provide safety inspectors with the technical training they need to effectively accomplish their mission. FAA officials suggested that we modify how we grouped our presentation of findings from our survey of inspectors. Specifically, they maintained that our analysis of results should have included "moderate extent" along with "very great" and "great extent" as a positive response because the inspectors would have viewed a "moderate extent" response as a positive response. Thus, in FAA's view, combining the "moderate extent" responses with "great extent" and "very great extent" responses would more accurately reflect the respondents' intent. The extent scale that we used in our survey represents a unidirectional scale. As such, it is possible to interpret any point along that scale, other than "no extent," as positive, depending upon how a question is worded. Generally, we presented information in the report with both "very great extent" and "great extent" combined to represent the clearly most positive responses. The combination of "very great extent" and "great extent" responses was intended to give FAA a clearer understanding of inspectors' perceptions and guidance as to where the application of its efforts is likely to have the greatest effect. Although this approach served our purposes best, there are naturally multiple ways in which one might combine response categories. As such, we have provided detailed results showing responses for each question by each response category in appendix II and the e-supplement to this report. The officials also noted that we defined technical training for the purpose of this report differently from what FAA considers to be technical training for inspectors. While these officials appreciated the recognition of the differences in the two definitions in our report, they said that the different definitions account for some disparity between what FAA considers the percentage of training achieved and that shown in the draft report. For example, the department considers the use of computer automation tools as a critical element of an inspector's ability to provide effective and efficient safety oversight. Because the Vision 100-Century of Aviation Reauthorization Act required that we focus on training in the latest aviation technologies--which we termed technical training--we did not include courses such as the use of computer tools in our assessment. Nevertheless, our draft and final reports acknowledge the importance of other training provided to inspectors, particularly training in skills relating to system safety and risk assessment. The department also provided several clarifying comments and technical corrections, which we have incorporated in this report as appropriate. We are sending copies of this report to congressional committees and subcommittees with responsibilities for transportation safety issues; the Secretary of Transportation; the Administrator, FAA; and the Director, Office of Management and Budget. We will also make copies available to others upon request. This report will be available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov]. If you have any questions about this report, please contact me at (202) 512-2834 or [Hyperlink, dillinghamg@gao.gov]. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. Staff who made key contributions to this report are listed in appendix IV. Signed by: Gerald L. Dillingham, Ph.D.: Director, Physical Infrastructure Issues: Appendixes: [End of section] Appendix I: Inspector-Reported Travel for Technical Training: The Vision 100-Century of Aviation Reauthorization Act required that we report on the amount of travel required of Federal Aviation Administration (FAA) inspectors in receiving training. To attempt to accomplish this requirement, we asked FAA to provide us with information on the number of times each inspector was in travel status for training, the location of the training, and the duration of the trips. FAA was not able to readily provide this information, citing limitations of its databases that track inspector travel. FAA told us that it is able to access and review individual inspector travel records, but its information systems are not set up to compile and analyze travel data for inspectors' travel for training, as a whole. In part, this information is not readily available because the data are stored in multiple databases, and the information is recorded differently, depending on how the training is arranged and budgeted. Further, FAA officials told us that (1) these data would be extremely time consuming to collect and compile and (2) a manual search for location of training would be necessary in some cases. On the basis of our inquiries, we concluded that it was not unreasonable for FAA to lack an easily accessible, comprehensive set of travel data. Thus to obtain information on inspectors' travel for training we used our survey of aviation safety inspectors (conducted in late 2004). We asked inspectors to tell us how many weeks they were on travel status for technical training in the past 12 months. On the basis of our survey, we estimate that inspectors spend an average of about 3.1 weeks per year on travel status for technical training.[Footnote 53] (See fig. 11.) We found that an estimated 54 percent of inspectors were on travel status for 1 to 3 weeks, and 27 percent spent 4 weeks or more on travel for technical training.[Footnote 54] About 19 percent of inspectors spent no time on travel status for technical training in the past year.[Footnote 55] Figure 11: Number of Weeks Inspectors Reported Spending on Travel for Technical Training within the Past 12 Months: [See PDF for image] [End of figure] On average, Flight Standards inspectors spent more time on travel for technical training than did Aircraft Certification inspectors, according to our analysis of survey responses. Flight Standards inspectors spent an average of approximately 3.2 weeks over the past year, and Aircraft Certification inspectors were on travel for training for approximately 2 weeks.[Footnote 56] The Vision 100-Century of Aviation Reauthorization Act contained a "Sense of the House" that stated that, if possible, FAA inspectors should be allowed to take training at the location most convenient for the inspector. As part of our survey, we asked the inspectors the extent to which there are opportunities for FAA to offer or contract for technical training closer to the inspectors' work location. According to our survey, we estimate that approximately 13 percent of inspectors indicated, to a great or very great extent, that such opportunities existed.[Footnote 57] (See fig. 12.) However, more than one-third of inspectors indicated they did not know if such training opportunities existed.[Footnote 58] We did not attempt to verify inspectors' views on opportunities for nearby technical training. Figure 12: Inspectors' Views on the Extent to Which Technical Training Opportunities Exist Closer to Their Work Location: [See PDF for image] Note: See table 22 in appendix II for additional details. [End of figure] [End of section] Appendix II: Additional Details on Training Data and Selected Inspector Survey Responses: Tables 10 through 22 provide additional inspector training data as well as additional detail on inspectors' views on FAA technical training, as discussed earlier in this report. The survey results, exclusive of inspector specialty breakouts, can be found at [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-704SP]. The stratified random sample of FAA inspectors was designed to have an overall margin of error of plus or minus 4 percentage points at a 95 percent level of confidence. Due to nonresponse, the actual overall margin of error is plus or minus 4.6 percentage points. The individual types of FAA inspectors represent strata in the sample. The precision for results within each stratum is less than the overall precision for population level estimates. Estimates for each individual type of safety inspector (stratum level) have margins of error greater than 4.6 percentage points. Estimates are more accurate for strata that have a larger number of responding inspectors than for those with fewer inspectors in them. For tables in this appendix that provide results of our survey of safety inspectors, we present both the estimated percentage of those responding in a certain way to each question and the confidence interval associated with that estimate. For example, in table 13, we report on the percentage of general aviation inspectors who responded to a great extent that they have enough technical knowledge to do their job as 39 (31-46). This means that we estimate that 39 percent of all general aviation inspectors believe this to a great extent. Had we surveyed the population of all general aviation inspectors, we are 95 percent confident that the percentage point responding "to a great extent" for this survey question would lie between 31 and 46 percentage points. The confidence interval reflects the sampling error that corresponds to the estimate of 39 percent. The tables associated with our survey in this appendix provide the number of respondents within each row. In some cases, the numbers are small because FAA has relatively few of these types of inspectors. See appendix III for more information on how we conducted our survey. Table 10: Percent of Essential Courses That Are Technical in Nature: Type of inspector: Air carrier avionics; Number of essential technical courses: 14; Number of essential courses overall: 28; Percent of essential courses that are technical: 50. Type of inspector: Air carrier maintenance; Number of essential technical courses: 8; Number of essential courses overall: 22; Percent of essential courses that are technical: 36. Type of inspector: Air carrier operations; Number of essential technical courses: 1; Number of essential courses overall: 10; Percent of essential courses that are technical: 10. Type of inspector: Cabin safety; Number of essential technical courses: 3; Number of essential courses overall: 12; Percent of essential courses that are technical: 25. Type of inspector: General aviation avionics; Number of essential technical courses: 12; Number of essential courses overall: 26; Percent of essential courses that are technical: 46. Type of inspector: General aviation maintenance; Number of essential technical courses: 8; Number of essential courses overall: 23; Percent of essential courses that are technical: 35. Type of inspector: General aviation operations; Number of essential technical courses: 2; Number of essential courses overall: 13; Percent of essential courses that are technical: 15. Type of inspector: Aircraft certification; Number of essential technical courses: 0; Number of essential courses overall: 7; Percent of essential courses that are technical: 0. Type of inspector: All inspectors; Number of essential technical courses: 48; Number of essential courses overall: 141; Percent of essential courses that are technical: 34. Source: GAO analysis of FAA data. [End of table] Table 11: Percent of Inspectors Completing Essential Courses: Type of inspector: Air carrier avionics; Percent of inspectors completing at least 75 percent of essential courses: 13; Percent of inspectors completing at least 50 percent of essential courses: 81. Type of inspector: Air carrier maintenance; Percent of inspectors completing at least 75 percent of essential courses: 25; Percent of inspectors completing at least 50 percent of essential courses: 83. Type of inspector: Air carrier operations; Percent of inspectors completing at least 75 percent of essential courses: 30; Percent of inspectors completing at least 50 percent of essential courses: 79. Type of inspector: Cabin safety; Percent of inspectors completing at least 75 percent of essential courses: 73; Percent of inspectors completing at least 50 percent of essential courses: 88. Type of inspector: General aviation avionics; Percent of inspectors completing at least 75 percent of essential courses: 30; Percent of inspectors completing at least 50 percent of essential courses: 81. Type of inspector: General aviation maintenance; Percent of inspectors completing at least 75 percent of essential courses: 36; Percent of inspectors completing at least 50 percent of essential courses: 86. Type of inspector: General aviation operations; Percent of inspectors completing at least 75 percent of essential courses: 47; Percent of inspectors completing at least 50 percent of essential courses: 88. Type of inspector: Aircraft certification; Percent of inspectors completing at least 75 percent of essential courses: 9; Percent of inspectors completing at least 50 percent of essential courses: 94. Type of inspector: All inspectors; Percent of inspectors completing at least 75 percent of essential courses: 30; Percent of inspectors completing at least 50 percent of essential courses: 84. Source: GAO analysis of FAA data. Note: See text for a discussion of why inspectors may not have completed essential courses. [End of table] Table 12: Average Number of Technical Training Courses Taken Outside of Requirements, Fiscal Years 2002 through 2004: Type of inspector: Air carrier avionics; Average number of technical courses: 1.5. Type of inspector: Air carrier maintenance; Average number of technical courses: 1.7. Type of inspector: Air carrier operations; Average number of technical courses: 2.3. Type of inspector: Cabin safety; Average number of technical courses: 0.2. Type of inspector: General aviation avionics; Average number of technical courses: 1.4. Type of inspector: General aviation maintenance; Average number of technical courses: 1.8. Type of inspector: General aviation operations; Average number of technical courses: 2.6. Type of inspector: Aircraft certification; Average number of technical courses: 1.2. Type of inspector: All inspectors; Average number of technical courses: 1.7. Source: GAO analysis of FAA data. [End of table] Table 13: Inspectors' Views on Extent to Which They Currently Have Enough Technical Knowledge to Do Their Jobs: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Very great: 13% (Confidence interval: 9-18); Great: 40% (Confidence interval: 34-46); Moderate: 33% (Confidence interval: 27-39); Some: 12% (Confidence interval: 9-17); None: 1% (Confidence interval: 0-4). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Very great: 7% (Confidence interval: 2-17); Great: 28% (Confidence interval: 17-42); Moderate: 37% (Confidence interval: 24-52); Some: 28% (Confidence interval: 17-42); None: 0% (Confidence interval: 0-6). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Very great: 8% (Confidence interval: 3-15); Great: 44% (Confidence interval: 35-54); Moderate: 35% (Confidence interval: 26-45); Some: 9% (Confidence interval: 4-17); None: 3% (Confidence interval: 1-9). Type of inspector: Air carrier operations; Unweighted sample size: 79; Very great: 20% (Confidence interval: 13-30); Great: 41% (Confidence interval: 31-51); Moderate: 30% (Confidence interval: 21-40); Some: 9% (Confidence interval: 4-16); None: 0% (Confidence interval: 0-4). Type of inspector: Cabin safety; Unweighted sample size: 18; Very great: 28% (Confidence interval: 11-50); Great: 51% (Confidence interval: 30-71); Moderate: 11% (Confidence interval: 2-31); Some: 11% (Confidence interval: 2-31); None: 0% (Confidence interval: 0-15). General aviation: General aviation; Unweighted sample size: 132; Very great: 10% (Confidence interval: 6-16); Great: 39% (Confidence interval: 31-46); Moderate: 36% (Confidence interval: 29-44); Some: 15% (Confidence interval: 10-22); None: 0% (Confidence interval: 0-2). General aviation: General aviation avionics; Unweighted sample size: 22; Very great: 0% (Confidence interval: 0-13); Great: 45% (Confidence interval: 25-67); Moderate: 36% (Confidence interval: 18-58); Some: 18% (Confidence interval: 6-39); None: 0% (Confidence interval: 0-13). General aviation: General aviation maintenance; Unweighted sample size: 56; Very great: 13% (Confidence interval: 6-23); Great: 41% (Confidence interval: 29-53); Moderate: 36% (Confidence interval: 24-49); Some: 11% (Confidence interval: 4-21); None: 0% (Confidence interval: 0-5). General aviation: General aviation operations; Unweighted sample size: 54; Very great: 11% (Confidence interval: 4-22); Great: 33% (Confidence interval: 22-47); Moderate: 37% (Confidence interval: 25-49); Some: 19% (Confidence interval: 10-31); None: 0% (Confidence interval: 0-5). Aircraft certification; Unweighted sample size: 25; Very great: 20% (Confidence interval: 7-39); Great: 48% (Confidence interval: 29-67); Moderate: 28% (Confidence interval: 13-48); Some: 4% (Confidence interval: 0-19); None: 0% (Confidence interval: 0-11). All inspectors; Unweighted sample size: 388; Very great: 12% (Confidence interval: 9-16); Great: 40% (Confidence interval: 35-45); Moderate: 34% (Confidence interval: 29-38); Some: 13% (Confidence interval: 10-16); None: 1% (Confidence interval: 0-2). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "To what extent do you currently have enough technical knowledge about the aircraft, systems, or operations you inspect to do your present job?" For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 5 for visual illustration. [End of table] Table 14: Inspectors' Views on Extent to Which Requested Technical Training Is Approved: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Very great: 8% (Confidence interval: 5-12); Great: 23% (Confidence interval: 18-28); Moderate: 24% (Confidence interval: 19-30); Some: 28% (Confidence interval: 22-33); None: 13% (Confidence interval: 9-18); Don't know: 4% (Confidence interval: 2-7). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Very great: 4% (Confidence interval: 1-14); Great: 22% (Confidence interval: 12-35); Moderate: 24% (Confidence interval: 13-38); Some: 37% (Confidence interval: 24-51); None: 9% (Confidence interval: 3-19); Don't know: 4% (Confidence interval: 1-14). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Very great: 3% (Confidence interval: 1-9); Great: 24% (Confidence interval: 16-33); Moderate: 27% (Confidence interval: 19-37); Some: 28% (Confidence interval: 20-38); None: 11% (Confidence interval: 6-19); Don't know: 6% (Confidence interval: 2-12). Type of inspector: Air carrier operations; Unweighted sample size: 80; Very great: 14% (Confidence interval: 7-23); Great: 21% (Confidence interval: 13-31); Moderate: 21% (Confidence interval: 13-31); Some: 24% (Confidence interval: 15-34); None: 17% (Confidence interval: 10-27); Don't know: 2% (Confidence interval: 0-8). Type of inspector: Cabin safety; Unweighted sample size: 17; Very great: 12% (Confidence interval: 2-33); Great: 34% (Confidence interval: 17-54); Moderate: 24% (Confidence interval: 8-47); Some: 18% (Confidence interval: 6-40); None: 6% (Confidence interval: 0-25); Don't know: 5% (Confidence interval: 0-22). General aviation: General aviation; Unweighted sample size: 132; Very great: 5% (Confidence interval: 2-9); Great: 19% (Confidence interval: 13-26); Moderate: 17% (Confidence interval: 12-24); Some: 30% (Confidence interval: 22-37); None: 24% (Confidence interval: 17-31); Don't know: 6% (Confidence interval: 3-11). General aviation: General aviation avionics; Unweighted sample size: 22; Very great: 5% (Confidence interval: 0-22); Great: 5% (Confidence interval: 0-22); Moderate: 18% (Confidence interval: 6-39); Some: 36% (Confidence interval: 18-58); None: 27% (Confidence interval: 11-49); Don't know: 9% (Confidence interval: 1-28). General aviation: General aviation maintenance; Unweighted sample size: 56; Very great: 7% (Confidence interval: 2-17); Great: 18% (Confidence interval: 9-30); Moderate: 20% (Confidence interval: 11-32); Some: 29% (Confidence interval: 18-41); None: 20% (Confidence interval: 11-32); Don't know: 7% (Confidence interval: 2-17). General aviation: General aviation operations; Unweighted sample size: 54; Very great: 2% (Confidence interval: 0-9); Great: 26% (Confidence interval: 15-39); Moderate: 15% (Confidence interval: 7-26); Some: 28% (Confidence interval: 17-41); None: 26% (Confidence interval: 15-39); Don't know: 4% (Confidence interval: 1-12). Aircraft certification; Unweighted sample size: 25; Very great: 4% (Confidence interval: 0-19); Great: 16% (Confidence interval: 5-35); Moderate: 28% (Confidence interval: 13-48); Some: 16% (Confidence interval: 5-35); None: 28% (Confidence interval: 13-48); Don't know: 8% (Confidence interval: 1-24). All inspectors; Unweighted sample size: 388; Very great: 6% (Confidence interval: 4-9); Great: 21% (Confidence interval: 17-25); Moderate: 22% (Confidence interval: 18-26); Some: 28% (Confidence interval: 24-32); None: 18% (Confidence interval: 14-21); Don't know: 5% (Confidence interval: 3-8). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "To what extent have the technical training courses you requested been approved?" For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 6 for visual illustration. [End of table] Table 15: Inspectors' Views on Whether Availability of Courses Helped or Hindered Their Ability to Take Requested Technical Training: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Greatly helped: 4% (Confidence interval: 2-7); Helped: 24% (Confidence interval: 18-29); Neither helped nor hindered: 27% (Confidence interval: 22-33); Hindered: 30% (Confidence interval: 24-36); Greatly hindered: 8% (Confidence interval: 5-12); No basis to judge: 5% (Confidence interval: 3-8); Don't know: 2% (Confidence interval: 1-4). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Greatly helped: 4% (Confidence interval: 1-14); Helped: 22% (Confidence interval: 11-35); Neither helped nor hindered: 28% (Confidence interval: 17-42); Hindered: 35% (Confidence interval: 22-49); Greatly hindered: 9% (Confidence interval: 3-20); No basis to judge: 0% (Confidence interval: 0-6); Don't know: 2% (Confidence interval: 0-10). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Greatly helped: 3% (Confidence interval: 1-9); Helped: 27% (Confidence interval: 19-37); Neither helped nor hindered: 20% (Confidence interval: 13-30); Hindered: 34% (Confidence interval: 25-43); Greatly hindered: 9% (Confidence interval: 4-17); No basis to judge: 2% (Confidence interval: 0-7); Don't know: 3% (Confidence interval: 1-9). Type of inspector: Air carrier operations; Unweighted sample size: 79; Greatly helped: 4% (Confidence interval: 1-10); Helped: 22% (Confidence interval: 13-32); Neither helped nor hindered: 34% (Confidence interval: 24-44); Hindered: 25% (Confidence interval: 17-36); Greatly hindered: 6% (Confidence interval: 2-14); No basis to judge: 9% (Confidence interval: 4-17); Don't know: 0% (Confidence interval: 0-4). Type of inspector: Cabin safety; Unweighted sample size: 18; Greatly helped: 6% (Confidence interval: 0-24); Helped: 16% (Confidence interval: 5-36); Neither helped nor hindered: 28% (Confidence interval: 11-50); Hindered: 12% (Confidence interval: 2-31); Greatly hindered: 17% (Confidence interval: 5-38); No basis to judge: 22% (Confidence interval: 8-43); Don't know: 0% (Confidence interval: 0-15). General aviation: General aviation; Unweighted sample size: 131; Greatly helped: 3% (Confidence interval: 1-7); Helped: 18% (Confidence interval: 12-26); Neither helped nor hindered: 34% (Confidence interval: 27-42); Hindered: 25% (Confidence interval: 18-32); Greatly hindered: 9% (Confidence interval: 5-15); No basis to judge: 8% (Confidence interval: 4-14); Don't know: 2% (Confidence interval: 0-5). General aviation: General aviation avionics; Unweighted sample size: 21; Greatly helped: 0% (Confidence interval: 0-13); Helped: 10% (Confidence interval: 1-29); Neither helped nor hindered: 33% (Confidence interval: 15-56); Hindered: 29% (Confidence interval: 12-51); Greatly hindered: 14% (Confidence interval: 3-35); No basis to judge: 14% (Confidence interval: 3-35); Don't know: 0% (Confidence interval: 0-13). General aviation: General aviation maintenance; Unweighted sample size: 56; Greatly helped: 4% (Confidence interval: 1-12); Helped: 23% (Confidence interval: 13-36); Neither helped nor hindered: 38% (Confidence interval: 26-49); Hindered: 21% (Confidence interval: 12-34); Greatly hindered: 7% (Confidence interval: 2-17); No basis to judge: 5% (Confidence interval: 1-14); Don't know: 2% (Confidence interval: 0-9). General aviation: General aviation operations; Unweighted sample size: 54; Greatly helped: 4% (Confidence interval: 1-12); Helped: 17% (Confidence interval: 8-29); Neither helped nor hindered: 31% (Confidence interval: 20-45); Hindered: 28% (Confidence interval: 17-41); Greatly hindered: 9% (Confidence interval: 3-20); No basis to judge: 9% (Confidence interval: 3-20); Don't know: 2% (Confidence interval: 0-9). Aircraft certification; Unweighted sample size: 25; Greatly helped: 8% (Confidence interval: 1-24); Helped: 20% (Confidence interval: 7-39); Neither helped nor hindered: 36% (Confidence interval: 19-56); Hindered: 4% (Confidence interval: 0-19); Greatly hindered: 20% (Confidence interval: 7-39); No basis to judge: 12% (Confidence interval: 3-30); Don't know: 0% (Confidence interval: 0-11). All inspectors; Unweighted sample size: 387; Greatly helped: 4% (Confidence interval: 2-6); Helped: 22% (Confidence interval: 18-25); Neither helped nor hindered: 30% (Confidence interval: 26-35); Hindered: 27% (Confidence interval: 23-31); Greatly hindered: 9% (Confidence interval: 7-12); No basis to judge: 7% (Confidence interval: 5-9); Don't know: 2% (Confidence interval: 1-3). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "Have the following factors helped or hindered your ability to take the technical training you requested to do your current job? Factor: Availability of courses." For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 7 for visual illustration. [End of table] Table 16: Inspectors' Views on Whether Availability of Funds Helped or Hindered Their Ability to Take Requested Technical Training: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 229; Greatly helped: 4% (Confidence interval: 2-8); Helped: 10% (Confidence interval: 6-14); Neither helped nor hindered: 21% (Confidence interval: 16-26); Hindered: 28% (Confidence interval: 22-33); Greatly hindered: 24% (Confidence interval: 18-29); No basis to judge: 7% (Confidence interval: 4-11); Don't know: 7% (Confidence interval: 4-10). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Greatly helped: 9% (Confidence interval: 3-20); Helped: 6% (Confidence interval: 2-16); Neither helped nor hindered: 15% (Confidence interval: 7-28); Hindered: 39% (Confidence interval: 26-54); Greatly hindered: 17% (Confidence interval: 8-30); No basis to judge: 0% (Confidence interval: 0-6); Don't know: 13% (Confidence interval: 5-25). Type of inspector: Air carrier maintenance; Unweighted sample size: 87; Greatly helped: 3% (Confidence interval: 1-9); Helped: 16% (Confidence interval: 9-25); Neither helped nor hindered: 22% (Confidence interval: 14-31); Hindered: 24% (Confidence interval: 16-34); Greatly hindered: 22% (Confidence interval: 14-31); No basis to judge: 6% (Confidence interval: 2-12); Don't know: 7% (Confidence interval: 3-14). Type of inspector: Air carrier operations; Unweighted sample size: 78; Greatly helped: 4% (Confidence interval: 1-11); Helped: 5% (Confidence interval: 2-12); Neither helped nor hindered: 23% (Confidence interval: 15-33); Hindered: 26% (Confidence interval: 17-36); Greatly hindered: 28% (Confidence interval: 19-39); No basis to judge: 10% (Confidence interval: 5-19); Don't know: 4% (Confidence interval: 1-10). Type of inspector: Cabin safety; Unweighted sample size: 18; Greatly helped: 0% (Confidence interval: 0-15); Helped: 6% (Confidence interval: 0-24); Neither helped nor hindered: 17% (Confidence interval: 5-38); Hindered: 33% (Confidence interval: 15-56); Greatly hindered: 28% (Confidence interval: 12-51); No basis to judge: 16% (Confidence interval: 5-36); Don't know: 0% (Confidence interval: 0-15). General aviation: General aviation; Unweighted sample size: 131; Greatly helped: 3% (Confidence interval: 1-7); Helped: 8% (Confidence interval: 4-13); Neither helped nor hindered: 16% (Confidence interval: 11-23); Hindered: 30% (Confidence interval: 22-37); Greatly hindered: 29% (Confidence interval: 22-36); No basis to judge: 10% (Confidence interval: 6-16); Don't know: 5% (Confidence interval: 2-9). General aviation: General aviation avionics; Unweighted sample size: 22; Greatly helped: 0% (Confidence interval: 0-13); Helped: 9% (Confidence interval: 1-28); Neither helped nor hindered: 18% (Confidence interval: 6-39); Hindered: 18% (Confidence interval: 6-39); Greatly hindered: 36% (Confidence interval: 18-58); No basis to judge: 14% (Confidence interval: 3-34); Don't know: 5% (Confidence interval: 0-22). General aviation: General aviation maintenance; Unweighted sample size: 56; Greatly helped: 2% (Confidence interval: 0-9); Helped: 11% (Confidence interval: 4-21); Neither helped nor hindered: 13% (Confidence interval: 6-23); Hindered: 32% (Confidence interval: 21-45); Greatly hindered: 30% (Confidence interval: 19-43); No basis to judge: 5% (Confidence interval: 1-14); Don't know: 7% (Confidence interval: 2-17). General aviation: General aviation operations; Unweighted sample size: 53; Greatly helped: 6% (Confidence interval: 1-15); Helped: 4% (Confidence interval: 1-12); Neither helped nor hindered: 19% (Confidence interval: 10-31); Hindered: 32% (Confidence interval: 21-46); Greatly hindered: 25% (Confidence interval: 14-37); No basis to judge: 13% (Confidence interval: 6-25); Don't know: 2% (Confidence interval: 0-9). Aircraft certification; Unweighted sample size: 25; Greatly helped: 12% (Confidence interval: 3-30); Helped: 4% (Confidence interval: 0-19); Neither helped nor hindered: 8% (Confidence interval: 1-24); Hindered: 28% (Confidence interval: 13-48); Greatly hindered: 28% (Confidence interval: 13-48); No basis to judge: 12% (Confidence interval: 3-30); Don't know: 8% (Confidence interval: 1-24). All inspectors; Unweighted sample size: 385; Greatly helped: 4% (Confidence interval: 3-7); Helped: 9% (Confidence interval: 6-12); Neither helped nor hindered: 18% (Confidence interval: 15-22); Hindered: 29% (Confidence interval: 24-33); Greatly hindered: 26% (Confidence interval: 22-30); No basis to judge: 8% (Confidence interval: 6-11); Don't know: 6% (Confidence interval: 4-9). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "Have the following factors helped or hindered your ability to take the technical training you requested to do your current job? Factor: Availability of funds." For more detail about the estimates and the corresponding confidence intervals (numbers in appendix), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 7 for visual illustration. [End of table] Table 17: Inspectors' Views on Whether Management's Determination of Need Helped or Hindered Their Ability to Take Requested Technical Training: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 230; Greatly helped: 7% (Confidence interval: 4-11); Helped: 29% (Confidence interval: 23-34); Neither helped nor hindered: 33% (Confidence interval: 27-39); Hindered: 15% (Confidence interval: 11-20); Greatly hindered: 7% (Confidence interval: 4-11); No basis to judge: 6% (Confidence interval: 4-10); Don't know: 3% (Confidence interval: 1-6). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Greatly helped: 4% (Confidence interval: 1-14); Helped: 37% (Confidence interval: 25-51); Neither helped nor hindered: 30% (Confidence interval: 19-44); Hindered: 15% (Confidence interval: 7-28); Greatly hindered: 6% (Confidence interval: 2-16); No basis to judge: 2% (Confidence interval: 0-11); Don't know: 4% (Confidence interval: 1-14). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Greatly helped: 7% (Confidence interval: 3-14); Helped: 28% (Confidence interval: 20-38); Neither helped nor hindered: 38% (Confidence interval: 28-47); Hindered: 15% (Confidence interval: 8-23); Greatly hindered: 7% (Confidence interval: 3-14); No basis to judge: 3% (Confidence interval: 1-9); Don't know: 2% (Confidence interval: 0-7). Type of inspector: Air carrier operations; Unweighted sample size: 79; Greatly helped: 8% (Confidence interval: 3-15); Helped: 25% (Confidence interval: 17-36); Neither helped nor hindered: 31% (Confidence interval: 21-40); Hindered: 15% (Confidence interval: 8-24); Greatly hindered: 8% (Confidence interval: 3-15); No basis to judge: 10% (Confidence interval: 5-19); Don't know: 4% (Confidence interval: 1-10). Type of inspector: Cabin safety; Unweighted sample size: 17; Greatly helped: 6% (Confidence interval: 0-25); Helped: 28% (Confidence interval: 14-46); Neither helped nor hindered: 18% (Confidence interval: 6-40); Hindered: 25% (Confidence interval: 10-46); Greatly hindered: 6% (Confidence interval: 0-25); No basis to judge: 17% (Confidence interval: 5-38); Don't know: 0% (Confidence interval: 0-16). General aviation: General aviation; Unweighted sample size: 132; Greatly helped: 6% (Confidence interval: 3-11); Helped: 23% (Confidence interval: 17-31); Neither helped nor hindered: 28% (Confidence interval: 21-35); Hindered: 19% (Confidence interval: 13-26); Greatly hindered: 13% (Confidence interval: 8-19); No basis to judge: 8% (Confidence interval: 4-14); Don't know: 2% (Confidence interval: 1-6). General aviation: General aviation avionics; Unweighted sample size: 22; Greatly helped: 0% (Confidence interval: 0-13); Helped: 27% (Confidence interval: 11-49); Neither helped nor hindered: 14% (Confidence interval: 3-34); Hindered: 23% (Confidence interval: 8-44); Greatly hindered: 18% (Confidence interval: 6-39); No basis to judge: 14% (Confidence interval: 3-34); Don't know: 5% (Confidence interval: 0-22). General aviation: General aviation maintenance; Unweighted sample size: 56; Greatly helped: 9% (Confidence interval: 3-19); Helped: 23% (Confidence interval: 13-36); Neither helped nor hindered: 34% (Confidence interval: 22-47); Hindered: 16% (Confidence interval: 8-28); Greatly hindered: 13% (Confidence interval: 6-23); No basis to judge: 4% (Confidence interval: 1-12); Don't know: 2% (Confidence interval: 0-9). General aviation: General aviation operations; Unweighted sample size: 54; Greatly helped: 6% (Confidence interval: 1-15); Helped: 22% (Confidence interval: 13-35); Neither helped nor hindered: 28% (Confidence interval: 17-41); Hindered: 20% (Confidence interval: 11-33); Greatly hindered: 11% (Confidence interval: 4-22); No basis to judge: 11% (Confidence interval: 4-22); Don't know: 2% (Confidence interval: 0-9). Aircraft certification; Unweighted sample size: 25; Greatly helped: 4% (Confidence interval: 0-19); Helped: 20% (Confidence interval: 7-39); Neither helped nor hindered: 20% (Confidence interval: 7-39); Hindered: 28% (Confidence interval: 13-48); Greatly hindered: 12% (Confidence interval: 3-30); No basis to judge: 16% (Confidence interval: 5-35); Don't know: 0% (Confidence interval: 0-11). All inspectors; Unweighted sample size: 387; Greatly helped: 6% (Confidence interval: 4-9); Helped: 26% (Confidence interval: 22-30); Neither helped nor hindered: 30% (Confidence interval: 26-35); Hindered: 17% (Confidence interval: 14-21); Greatly hindered: 9% (Confidence interval: 7-13); No basis to judge: 8% (Confidence interval: 5-10); Don't know: 3% (Confidence interval: 1-5). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "Have the following factors helped or hindered your ability to take the technical training you requested to do your current job? Factor: Management's determination of your need for the course." For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 7 for visual illustration. [End of table] Table 18: Inspectors' Views on Whether Inspection Workload Helped or Hindered Their Ability to Take Requested Technical Training: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Greatly helped: 1% (Confidence interval: 0-4); Helped: 5% (Confidence interval: 3-9); Neither helped nor hindered: 58% (Confidence interval: 52-64); Hindered: 21% (Confidence interval: 16-25); Greatly hindered: 7% (Confidence interval: 4-11); No basis to judge: 6% (Confidence interval: 4-10); Don't know: 2% (Confidence interval: 1-4). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Greatly helped: 0% (Confidence interval: 0-6); Helped: 9% (Confidence interval: 3-20); Neither helped nor hindered: 63% (Confidence interval: 48-76); Hindered: 15% (Confidence interval: 7-28); Greatly hindered: 9% (Confidence interval: 3-20); No basis to judge: 2% (Confidence interval: 0-10); Don't know: 2% (Confidence interval: 0-10). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Greatly helped: 2% (Confidence interval: 0-7); Helped: 7% (Confidence interval: 3-13); Neither helped nor hindered: 65% (Confidence interval: 55-74); Hindered: 15% (Confidence interval: 9-23); Greatly hindered: 5% (Confidence interval: 1-11); No basis to judge: 6% (Confidence interval: 2-12); Don't know: 1% (Confidence interval: 0-6). Type of inspector: Air carrier operations; Unweighted sample size: 79; Greatly helped: 1% (Confidence interval: 0-6); Helped: 3% (Confidence interval: 0-9); Neither helped nor hindered: 48% (Confidence interval: 38-59); Hindered: 29% (Confidence interval: 20-40); Greatly hindered: 9% (Confidence interval: 4-17); No basis to judge: 7% (Confidence interval: 3-15); Don't know: 2% (Confidence interval: 0-8). Type of inspector: Cabin safety; Unweighted sample size: 18; Greatly helped: 0% (Confidence interval: 0-15); Helped: 0% (Confidence interval: 0-15); Neither helped nor hindered: 62% (Confidence interval: 39-81); Hindered: 17% (Confidence interval: 5-38); Greatly hindered: 0% (Confidence interval: 0-15); No basis to judge: 22% (Confidence interval: 8-43); Don't know: 0% (Confidence interval: 0-15). General aviation: General aviation; Unweighted sample size: 132; Greatly helped: 2% (Confidence interval: 0-5); Helped: 6% (Confidence interval: 3-11); Neither helped nor hindered: 55% (Confidence interval: 47-63); Hindered: 18% (Confidence interval: 13-25); Greatly hindered: 11% (Confidence interval: 6-17); No basis to judge: 8% (Confidence interval: 4-13); Don't know: 1% (Confidence interval: 0-4). General aviation: General aviation avionics; Unweighted sample size: 22; Greatly helped: 0% (Confidence interval: 0-13); Helped: 5% (Confidence interval: 0-22); Neither helped nor hindered: 59% (Confidence interval: 37-78); Hindered: 5% (Confidence interval: 0-22); Greatly hindered: 23% (Confidence interval: 8-44); No basis to judge: 9% (Confidence interval: 1-28); Don't know: 0% (Confidence interval: 0-13). General aviation: General aviation maintenance; Unweighted sample size: 56; Greatly helped: 0% (Confidence interval: 0-5); Helped: 13% (Confidence interval: 6-23); Neither helped nor hindered: 57% (Confidence interval: 45-69); Hindered: 14% (Confidence interval: 7-25); Greatly hindered: 7% (Confidence interval: 2-17); No basis to judge: 7% (Confidence interval: 2-17); Don't know: 2% (Confidence interval: 0-9). General aviation: General aviation operations; Unweighted sample size: 54; Greatly helped: 4% (Confidence interval: 1-12); Helped: 0% (Confidence interval: 0-5); Neither helped nor hindered: 52% (Confidence interval: 39-64); Hindered: 28% (Confidence interval: 17-41); Greatly hindered: 9% (Confidence interval: 3-20); No basis to judge: 7% (Confidence interval: 2-17); Don't know: 0% (Confidence interval: 0-5). Aircraft certification; Unweighted sample size: 25; Greatly helped: 4% (Confidence interval: 0-19); Helped: 0% (Confidence interval: 0-11); Neither helped nor hindered: 56% (Confidence interval: 36-74); Hindered: 8% (Confidence interval: 1-24); Greatly hindered: 12% (Confidence interval: 3-30); No basis to judge: 20% (Confidence interval: 7-39); Don't know: 0% (Confidence interval: 0-11). All inspectors; Unweighted sample size: 388; Greatly helped: 2% (Confidence interval: 1-3); Helped: 5% (Confidence interval: 3-8); Neither helped nor hindered: 57% (Confidence interval: 52-62); Hindered: 19% (Confidence interval: 15-23); Greatly hindered: 8% (Confidence interval: 6-12); No basis to judge: 7% (Confidence interval: 5-10); Don't know: 1% (Confidence interval: 0-3). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "Have the following factors helped or hindered your ability to take the technical training you requested to do your current job? Factor: The impact on your workload of the time commitment required for the training." For more detail about the estimates and the corresponding confidence intervals (numbers in appendix), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 7 for visual representation. [End of table] Table 19: Inspectors' Views on the Degree to Which Technical Training Is Delivered in a Timely Manner: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 229; Very great: 1% (Confidence interval: 0-4); Great: 20% (Confidence interval: 15-25); Moderate: 27% (Confidence interval: 21-32); Some: 38% (Confidence interval: 32-44); None: 14% (Confidence interval: 10-18). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Very great: 0% (Confidence interval: 0-6); Great: 15% (Confidence interval: 7-28); Moderate: 15% (Confidence interval: 7-28); Some: 56% (Confidence interval: 43-70); None: 13% (Confidence interval: 5-24). Type of inspector: Air carrier maintenance; Unweighted sample size: 87; Very great: 0% (Confidence interval: 0-3); Great: 17% (Confidence interval: 10-26); Moderate: 37% (Confidence interval: 27-46); Some: 28% (Confidence interval: 19-38); None: 18% (Confidence interval: 11-28). Type of inspector: Air carrier operations; Unweighted sample size: 78; Very great: 4% (Confidence interval: 1-10); Great: 27% (Confidence interval: 18-38); Moderate: 20% (Confidence interval: 13-30); Some: 40% (Confidence interval: 30-50); None: 9% (Confidence interval: 4-17). Type of inspector: Cabin safety; Unweighted sample size: 18; Very great: 0% (Confidence interval: 0-15); Great: 6% (Confidence interval: 0-24); Moderate: 44% (Confidence interval: 24-66); Some: 34% (Confidence interval: 16-56); None: 16% (Confidence interval: 5-36). General aviation: General aviation; Unweighted sample size: 132; Very great: 2% (Confidence interval: 1-6); Great: 13% (Confidence interval: 8-19); Moderate: 22% (Confidence interval: 16-29); Some: 51% (Confidence interval: 43-59); None: 12% (Confidence interval: 7-18). General aviation: General aviation avionics; Unweighted sample size: 22; Very great: 0% (Confidence interval: 0-13); Great: 0% (Confidence interval: 0-13); Moderate: 27% (Confidence interval: 11-49); Some: 55% (Confidence interval: 33-75); None: 18% (Confidence interval: 6-39). General aviation: General aviation maintenance; Unweighted sample size: 56; Very great: 2% (Confidence interval: 0-9); Great: 16% (Confidence interval: 8-28); Moderate: 21% (Confidence interval: 12-34); Some: 46% (Confidence interval: 34-59); None: 14% (Confidence interval: 7-25). General aviation: General aviation operations; Unweighted sample size: 54; Very great: 4% (Confidence interval: 1-12); Great: 15% (Confidence interval: 7-26); Moderate: 20% (Confidence interval: 11-33); Some: 54% (Confidence interval: 41-66); None: 7% (Confidence interval: 2-17). Aircraft certification; Unweighted sample size: 25; Very great: 8% (Confidence interval: 1-24); Great: 24% (Confidence interval: 10-44); Moderate: 28% (Confidence interval: 13-48); Some: 32% (Confidence interval: 16-52); None: 8% (Confidence interval: 1-24). All inspectors; Unweighted sample size: 386; Very great: 2% (Confidence interval: 1-4); Great: 18% (Confidence interval: 14-21); Moderate: 25% (Confidence interval: 21-29); Some: 42% (Confidence interval: 37-47); None: 13% (Confidence interval: 10-16). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "During your FAA career, to what extent have you received technical training in a timely manner-meaning receiving training in time to do your current job?" For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. The number of inspectors responding "do not know" was 2 percent or less. These results are not presented. Some of the row percentages will not add up to 100 percent due to rounding. See figure 8 for visual representation. [End of table] Table 20: Inspectors' Views on the Extent That They Receive Technical Training Prior to Scheduled Oversight Activities: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Always: 4% (Confidence interval: 2-8); Frequently: 18% (Confidence interval: 14-23); Occasionally: 31% (Confidence interval: 26-37); Rarely: 29% (Confidence interval: 23-34); Never: 14% (Confidence interval: 10-19); No basis: 3% (Confidence interval: 1-6). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Always: 0% (Confidence interval: 0-6); Frequently: 7% (Confidence interval: 2-17); Occasionally: 26% (Confidence interval: 15-40); Rarely: 61% (Confidence interval: 46-74); Never: 7% (Confidence interval: 2-17); No basis: 0% (Confidence interval: 0-6). Type of inspector: Air carrier maintenance; Unweighted sample size: 88; Always: 6% (Confidence interval: 2-12); Frequently: 11% (Confidence interval: 6-19); Occasionally: 35% (Confidence interval: 26-45); Rarely: 22% (Confidence interval: 14-31); Never: 24% (Confidence interval: 16-33); No basis: 2% (Confidence interval: 0-7). Type of inspector: Air carrier operations; Unweighted sample size: 79; Always: 5% (Confidence interval: 2-12); Frequently: 29% (Confidence interval: 20-40); Occasionally: 29% (Confidence interval: 20-40); Rarely: 23% (Confidence interval: 14-33); Never: 9% (Confidence interval: 4-17); No basis: 5% (Confidence interval: 2-12). Type of inspector: Cabin safety; Unweighted sample size: 18; Always: 5% (Confidence interval: 0-21); Frequently: 28% (Confidence interval: 12-51); Occasionally: 40% (Confidence interval: 21-61); Rarely: 11% (Confidence interval: 2-31); Never: 11% (Confidence interval: 2-31); No basis: 5% (Confidence interval: 0-21). General aviation: General aviation; Unweighted sample size: 132; Always: 2% (Confidence interval: 1-6); Frequently: 19% (Confidence interval: 13-26); Occasionally: 27% (Confidence interval: 20-35); Rarely: 33% (Confidence interval: 26-41); Never: 11% (Confidence interval: 7-17); No basis: 7% (Confidence interval: 3-12). General aviation: General aviation avionics; Unweighted sample size: 22; Always: 0% (Confidence interval: 0-13); Frequently: 5% (Confidence interval: 0-22); Occasionally: 18% (Confidence interval: 6-39); Rarely: 36% (Confidence interval: 18-58); Never: 23% (Confidence interval: 8-44); No basis: 18% (Confidence interval: 6-39). General aviation: General aviation maintenance; Unweighted sample size: 56; Always: 4% (Confidence interval: 1-12); Frequently: 20% (Confidence interval: 11-32); Occasionally: 27% (Confidence interval: 16-39); Rarely: 32% (Confidence interval: 21-45); Never: 14% (Confidence interval: 7-25); No basis: 4% (Confidence interval: 1-12). General aviation: General aviation operations; Unweighted sample size: 54; Always: 2% (Confidence interval: 0-9); Frequently: 24% (Confidence interval: 14-37); Occasionally: 31% (Confidence interval: 20-45); Rarely: 33% (Confidence interval: 22-47); Never: 4% (Confidence interval: 1-12); No basis: 6% (Confidence interval: 1-15). Aircraft certification; Unweighted sample size: 25; Always: 4% (Confidence interval: 0-19); Frequently: 28% (Confidence interval: 13-48); Occasionally: 32% (Confidence interval: 16-52); Rarely: 20% (Confidence interval: 7-39); Never: 16% (Confidence interval: 5-35); No basis: 0% (Confidence interval: 0-11). All inspectors; Unweighted sample size: 388; Always: 4% (Confidence interval: 2-6); Frequently: 19% (Confidence interval: 15-23); Occasionally: 30% (Confidence interval: 26-34); Rarely: 30% (Confidence interval: 26-34); Never: 13% (Confidence interval: 10-17); No basis: 4% (Confidence interval: 3-7). Source: GAO survey of FAA inspectors. Note: The data in this table represent the responses from inspectors to the following question, "In thinking about the timing of when you received technical training during your FAA career, how often did the following situations apply? Situation: Technical and/or equipment training was received prior to scheduled oversight/surveillance activities." For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 9 for visual representation. [End of table] Table 21: Percent of Technical Training Provided by Industry as Reported by FAA, Fiscal Years 2002 through 2004: Type of inspector: Air carrier: Air carrier avionics; 2002: Number: 181; 2002: Percent[A]: 30; 2003: Number: 193; 2003: Percent[A]: 42; 2004: Number: 218; 2004: Percent[A]: 53. Type of inspector: Air carrier: Air carrier maintenance; 2002: Number: 475; 2002: Percent[A]: 39; 2003: Number: 313; 2003: Percent[A]: 38; 2004: Number: 323; 2004: Percent[A]: 53. Type of inspector: Air carrier: Air carrier operations; 2002: Number: 282; 2002: Percent[A]: 54; 2003: Number: 341; 2003: Percent[A]: 57; 2004: Number: 332; 2004: Percent[A]: 55. Type of inspector: Air carrier: Cabin safety; 2002: Number: 9; 2002: Percent[A]: 39; 2003: Number: 7; 2003: Percent[A]: 23; 2004: Number: 6; 2004: Percent[A]: 33. Type of inspector: Air carrier: Subtotal; 2002: Number: 947; 2002: Percent[A]: 40; 2003: Number: 854; 2003: Percent[A]: 45; 2004: Number: 879; 2004: Percent[A]: 54. Type of inspector: General aviation: General aviation avionics; 2002: Number: 77; 2002: Percent[A]: 24; 2003: Number: 65; 2003: Percent[A]: 33; 2004: Number: 59; 2004: Percent[A]: 33. Type of inspector: General aviation: General aviation maintenance; 2002: Number: 240; 2002: Percent[A]: 30; 2003: Number: 160; 2003: Percent[A]: 28; 2004: Number: 148; 2004: Percent[A]: 33. Type of inspector: General aviation: General aviation operations; 2002: Number: 167; 2002: Percent[A]: 41; 2003: Number: 248; 2003: Percent[A]: 56; 2004: Number: 222; 2004: Percent[A]: 40. Type of inspector: General aviation: Subtotal; 2002: Number: 484; 2002: Percent[A]: 32; 2003: Number: 473; 2003: Percent[A]: 39; 2004: Number: 429; 2004: Percent[A]: 37. Type of inspector: Aircraft certification; 2002: Number: 28; 2002: Percent[A]: 34; 2003: Number: 27; 2003: Percent[A]: 41; 2004: Number: 3; 2004: Percent[A]: 8. Total; 2002: Number: 1,459; 2002: Percent[A]: 37; 2003: Number: 1,354; 2003: Percent[A]: 42; 2004: Number: 1,311; 2004: Percent[A]: 46. Source: GAO analysis of FAA data. Note: See figure 10 for visual representation: [A] As a percent of total FAA-and industry-provided training for each type of inspector. [End of table] Table 22: Inspectors' Views on the Extent to Which Technical Training Opportunities Exist Closer to Their Work Location: Percent (confidence interval). Type of inspector: Air carrier; Unweighted sample size: 231; Very great: 4% (Confidence interval: 2-7); Great: 11% (Confidence interval: 7-15); Moderate: 5% (Confidence interval: 3-9); Some: 13% (Confidence interval: 9-18); None: 30% (Confidence interval: 24-35); Don't know: 38% (Confidence interval: 32-44). Type of inspector: Air carrier avionics; Unweighted sample size: 46; Very great: 2% (Confidence interval: 0-10); Great: 9% (Confidence interval: 3-20); Moderate: 2% (Confidence interval: 0-11); Some: 24% (Confidence interval: 13-37); None: 26% (Confidence interval: 15-40); Don't know: 37% (Confidence interval: 24-52). Type of inspector: Air carrier maintenance; Unweighted sample size: 87; Very great: 3% (Confidence interval: 1-9); Great: 7% (Confidence interval: 3-14); Moderate: 8% (Confidence interval: 4-15); Some: 10% (Confidence interval: 5-18); None: 26% (Confidence interval: 18-36); Don't know: 45% (Confidence interval: 35-54). Type of inspector: Air carrier operations; Unweighted sample size: 80; Very great: 5% (Confidence interval: 2-12); Great: 16% (Confidence interval: 9-25); Moderate: 4% (Confidence interval: 1-10); Some: 11% (Confidence interval: 6-20); None: 34% (Confidence interval: 24-44); Don't know: 30% (Confidence interval: 20-39). Type of inspector: Cabin safety; Unweighted sample size: 18; Very great: 0% (Confidence interval: 0-15); Great: 6% (Confidence interval: 0-24); Moderate: 0% (Confidence interval: 0-15); Some: 6% (Confidence interval: 0-24); None: 37% (Confidence interval: 22-54); Don't know: 51% (Confidence interval: 32-70). General aviation: General aviation; Unweighted sample size: 132; Very great: 2% (Confidence interval: 1-6); Great: 9% (Confidence interval: 5-15); Moderate: 10% (Confidence interval: 6-16); Some: 18% (Confidence interval: 12-24); None: 26% (Confidence interval: 19-33); Don't know: 35% (Confidence interval: 28-43). General aviation: General aviation avionics; Unweighted sample size: 22; Very great: 5% (Confidence interval: 0-22); Great: 0% (Confidence interval: 0-13); Moderate: 14% (Confidence interval: 3-34); Some: 5% (Confidence interval: 0-22); None: 23% (Confidence interval: 8-44); Don't know: 55% (Confidence interval: 33-75). General aviation: General aviation maintenance; Unweighted sample size: 56; Very great: 2% (Confidence interval: 0-9); Great: 7% (Confidence interval: 2-17); Moderate: 11% (Confidence interval: 4-21); Some: 11% (Confidence interval: 4-21); None: 27% (Confidence interval: 16-39); Don't know: 43% (Confidence interval: 31-55). General aviation: General aviation operations; Unweighted sample size: 54; Very great: 2% (Confidence interval: 0-9); Great: 15% (Confidence interval: 7-26); Moderate: 7% (Confidence interval: 2-17); Some: 30% (Confidence interval: 19-43); None: 26% (Confidence interval: 15-39); Don't know: 20% (Confidence interval: 11-33). Aircraft certification; Unweighted sample size: 25; Very great: 4% (Confidence interval: 0-19); Great: 0% (Confidence interval: 0-11); Moderate: 12% (Confidence interval: 3-30); Some: 20% (Confidence interval: 7-39); None: 28% (Confidence interval: 13-48); Don't know: 36% (Confidence interval: 19-56). All inspectors; Unweighted sample size: 388; Very great: 3% (Confidence interval: 2-5); Great: 10% (Confidence interval: 7-13); Moderate: 7% (Confidence interval: 5-10); Some: 15% (Confidence interval: 12-18); None: 28% (Confidence interval: 24-32); Don't know: 37% (Confidence interval: 32-41). Source: GAO survey of FAA inspectors. Note: The data in this table represents the responses from inspectors to the following question, "To what extent are there opportunities for FAA to offer or contract for technical training, such as recurrency training, closer to your work location that is currently held at a central location far from your work location?" For more detail about the estimates and the corresponding confidence intervals (numbers in parentheses), please see the text at the beginning of this appendix. Some of the row percentages will not add up to 100 percent due to rounding. See figure 12 for visual representation. [End of table] [End of section] Appendix III: Scope and Methodology: To assess the extent to which FAA followed effective management practices in planning for, developing, and delivering up-to-date technical training, and ensuring that the technical training for inspectors contributes to improved performance and results, we identified key elements for assessing effective training and development efforts in the federal government using our recent guide on this subject.[Footnote 59] We identified the elements of this guidance that were most relevant to the training activities at FAA for aviation safety inspectors and then determined the extent to which FAA followed these practices. In determining the extent to which FAA followed a practice, we used the following scale: "fully" indicated that in our judgment all or virtually all aspects of the practice were followed; "mostly" indicated that more than half were followed; "partially" indicated that less than half were followed; and "not followed" indicated that few or no aspects of the practice were followed. For each element, we obtained information from FAA on its plans and activities and compared this information with the published criteria. We discussed this information with FAA training and program officials to gain their perspectives. In addition to gaining an understanding of these plans, and activities generally, we applied the elements in our training guidance to two emerging technologies (glass cockpits and composite materials) and determined how training needs in these areas were incorporated into training courses for FAA inspectors. We supplemented these activities in several ways to gain additional perspectives of inspector technical training needs and FAA's efforts to meet these needs. First, we collected materials from and interviewed FAA managers, supervisors, and inspectors at 7 of approximately 130 locations across the United States where FAA inspections take place. These efforts illuminated a mix of FAA inspector responsibilities for air carrier and general aviation operations and maintenance, new aircraft certifications, and oversight of manufacturing facilities. Second, we discussed technical training needs and FAA's actions with senior management of the Professional Airways System Specialists, the collective bargaining unit for air safety inspectors. Third, we sought the advice of two sets of experts, one to provide advice on the overall design of our study and a second to help us assess FAA's technical training curriculum and the extent to which FAA ensures that safety inspectors receive needed technical training. (See table 23.) We selected these experts on the basis of their knowledge of FAA safety inspectors and aviation technologies. We also sought the views of 23 member airlines of the Air Transport Association and the Regional Airline Association on the technical training of inspectors. Fourth, we visited the FAA Training Academy in Oklahoma City, Oklahoma, and Embry- Riddle Aeronautical University in Daytona Beach, Florida, to learn more about how courses are delivered to inspectors. Fifth, we reviewed National Transportation Safety Board recommendations concerning FAA safety inspector technical training. Lastly, we reviewed our studies and those of the Department of Transportation's Inspector General concerning inspector training and human capital issues. (See the Related Products section of this report for a list of our products.) Table 23: Experts Consulted for Our Work: Design experts: Mr. Gary Kiteley, Executive Director, Council on Aviation Accreditation; Curriculum experts: Mr. Brian Finnegan, President, Professional Aviation Maintenance Association. Design experts: Mr. Kent Lovelace, Chairman and Professor, Department of Aviation, University of North Dakota; Curriculum experts: Mr. David Lotterer, Vice President of Technical Services, Regional Airline Association. Design experts: Dr. Thomas Q. Carney, Professor and Department Head, Department of Aviation Technology, Purdue University; Curriculum experts: Mr. Basil Barimo, Vice President, Operations and Safety and Mr. Mont J. Smith, Director, Safety; Air Transport Association of America. Design experts: Mr. Anthony J. Broderick, Independent Aviation Safety Consultant; Curriculum experts: Mr. David Wright, Director of Training, Aircraft Owners and Pilots Association, Air Safety Foundation. Curriculum experts: Mr. Theodore Beneigh, Professor, Aeronautical Science; Mr. Charles Westbrooks, Assistant Professor, Aeronautical Science; Mr. Fred Mirgle, Director, Aviation and Avionics Training; Mr. Neill Fulbright, Associate Program Coordinator, Avionics Line Maintenance; Embry-Riddle Aeronautical University. Curriculum experts: Mr. Walter Desrosier, Vice President, Engineering and Maintenance; Mr. Jens Hennig, Manager, Operations; Mr. Gregory Bowles, Manager, Engineering and Maintenance; General Aviation Manufacturers Association. Curriculum experts: Dr. Michael Romanowski, Vice President, Civil Aviation and Mr. Ronald R. Baker, Jr., Manager, Civil Aviation Programs; Aerospace Industries Association. Curriculum experts: Ms. Sarah MacLeod, Executive Director and Mr. Paul Hawthorne, Vice President, Operations; Aeronautical Repair Station Association. Source: GAO. [End of table] To determine the type and amount of technical and other training that FAA inspectors receive, we obtained course descriptions from FAA and data from the Flight Standards training management system database and spreadsheets from Aircraft Certification for fiscal years 2002 through 2004. FAA officials indicated their belief that a study of inspector technical training should encompass training records over the whole of the inspectors' careers. However, because the Vision 100-Century of Aviation Reauthorization Act asked us to study up-to-date training on the latest technologies, we analyzed only the most recent 3 fiscal years of data. The data that we obtained included (1) essential and recommended courses by type of inspector; (2) training completed by each inspector; and (3) inspector specialty, location, and date of employment. We then calculated the amount of inspector training completed by course category and by type of inspector. In addition, we used the training records and course requirements to determine the extent to which inspectors have completed essential FAA courses. To assess the reliability of the training data, we (1) interviewed knowledgeable agency officials about the data, (2) performed electronic testing of relevant data fields for obvious errors in accuracy and completeness, and (3) collected and reviewed documentation from data system managers about the data and the systems that produced them. We determined that the data were sufficiently reliable for the purposes of this report. To gather information about inspectors' perspectives on the technical training available to them, we conducted a Web-based survey of a representative sample of FAA safety inspectors. The survey asked a combination of questions that allowed for open-ended and close-ended responses. We drew a stratified random probability sample of 496 inspectors from the population of 2,989 aviation safety inspectors across the United States.[Footnote 60] We stratified the population into 12 groups on the basis of the type of work the inspector performed. Each sample element was subsequently weighted in the analysis to account statistically for all the members of the population. Because we followed a probability procedure based on random selection, our sample is only one of a large number of samples that we might have drawn. Since each sample could have provided different estimates, we express our confidence in the precision of our particular sample's results as a 95 percent confidence interval (e.g., plus or minus 4.6 percentage points). This is the interval that would contain the actual population value for 95 percent of the samples we could have drawn. As a result, we are 95 percent confident that each of the confidence intervals in this report will include the true values in the study population. The percentage estimates for all survey respondents have a margin of error of plus or minus 4.6 percentage points. However, the percentage estimates for the subgroups are larger with a range of margin of error of plus or minus between 9.7 and 20.0 percentage points. Survey estimates presented as comparisons between groups are statistically significant when the 95 percent confidence intervals do not overlap. The surveys were conducted using self-administered electronic questionnaires accessible on the Internet through a secure Web browser. We sent e-mail notifications to 496 inspectors, beginning on December 4, 2004. We then sent each potential respondent a unique password and username to ensure that only members of the target population could participate in the survey. The initial version of the questionnaire that was posted on December 4, 2004, did not include three questions. A revised version was posted on December 14, 2004, before most respondents had answered the questionnaire. Because approximately one- quarter of the respondents did not answer these three new questions (questions 9, 20, and 25d), these results are not included in the report. To encourage respondents to complete the questionnaire, we sent a subsequent e-mail message to further prompt each nonrespondent approximately 2 weeks after the initial e-mail message. We sent nonrespondents two more notices and closed the survey on February 4, 2005. Of the 496 inspectors whom we surveyed, we received 392 useable responses (79 percent). In addition to these sampling errors, the practical difficulties in conducting surveys of this type may introduce other types of errors, commonly referred to as nonsampling errors. For example, questions may be misinterpreted, or the respondents' answers may differ from those of the inspectors who did not respond. We took steps to reduce these errors. Finally, we pretested the content and format of the questionnaire with safety inspectors at local FAA offices in Baltimore, Los Angeles, and Seattle. During the pretests we asked the inspectors questions to determine whether (1) the survey questions were clear, (2) the terms used were precise, (3) the questionnaire placed an undue burden on the respondents, and (4) the questions were unbiased. We made changes to the content and format of the final questionnaire on the basis of the pretest results. To determine the amount of training FAA receives from the aviation industry, we analyzed the training records of all FAA safety inspectors. We obtained FAA's course numbering and categorization system and used it to determine whether individual courses were provided by FAA or by the aviation industry. We computed the total number of technical courses attended by FAA inspectors from fiscal years 2002 through 2004 and identified those provided by the aviation industry. We discussed our results with FAA training officials. See the discussion above for our actions to assess the completeness and reliability of these data. To determine the amount of training safety inspectors received from industry either (1) in return for in-kind services or (2) for free, we reviewed training records and interviewed FAA headquarters officials and regional officials about FAA policies for accepting these types of training. We also asked about procedures used when such training is requested, including steps taken to ensure that any real or apparent conflict-of-interest issues are addressed. FAA does not keep separate records of these two types of training, and these data cannot easily be identified from the central training data files. Therefore, we instead interviewed officials at FAA's nine regional officials and requested these training data from them. Subsequently, we used the safety inspector training records to validate some of these data. We relied on FAA's nine regional office officials to contact over 100 Flight Standards and Air Certification offices to collect these data for fiscal years 2002 through 2004. Some regions indicated that their offices did not keep full records. Other regions provided us with incomplete data records, sometimes without names or specific dates. Because of the large number of offices from which the data were gathered, it was not practical for us to independently verify the completeness or accuracy of these data. As a result, we cannot be sure that the information FAA supplied includes all industry-provided training received for the 3 fiscal years. [End of section] Appendix IV: GAO Contact and Staff Acknowledgments: GAO Contact: Gerald L. Dillingham, Ph.D. (202) 512-2834: Acknowledgments: In addition to the contact named above, James Ratzenberger, Assistant Director; Carl Barden; Nancy Boardman; Brad Dubbs; Alice Feldesman; Jim Geibel; Kim Gianopoulos; David Hooper; Michael Krafve; Ed Laughlin; Donna Leiss; Jean McSween; Minette Richardson; and Sandra Sokol made key contributions to this report. [End of section] Related GAO Products: FAA Safety Inspector Training: National Aerospace System: Reauthorizing FAA Provides Opportunities and Options to Address Challenges. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-473T] Washington, D.C.: February 12, 2003. Major Management Challenges and Program Risks: Department of Transportation. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-108] Washington, D.C.: January 1, 2003. Aviation Safety: FAA's New Inspection System Offers Promise, but Problems Need to Be Addressed. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-99-183] Washington, D.C.: June 28, 1999. Aviation Safety: Weaknesses in Inspection and Enforcement Limit FAA in Identifying and Responding to Risks. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-6] Washington, D.C.: February 27, 1998. Aviation Safety: FAA Oversight of Repair Stations Needs Improvement. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-21] Washington, D.C.: October 24, 1997. Aviation Safety: Targeting and Training of FAA's Safety Inspector Workforce. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-96-26] Washington, D.C.: April 30, 1996. FAA Technical Training. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-94-296R] Washington, D.C.: September 26, 1994. FAA Budget: Important Challenges Affecting Aviation Safety, Capacity, and Efficiency. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-93-33] Washington, D.C.: April 26, 1993. FAA Budget: Key Issues Need to Be Addressed. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-92-51] Washington, D.C.: April 6, 1992. Aviation Safety: Commuter Airline Safety Would Be Enhanced with Better FAA Oversight. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-92-40] Washington, D.C.: March 17, 1992. Aviation Safety: FAA Needs to More Aggressively Manage Its Inspection Program. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-92-25] Washington, D.C.: February 6, 1992. Aviation Safety: Problems Persist in FAA's Inspection Program. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-92-14] Washington, D.C.: November 20, 1991. Serious Shortcomings in FAA's Training Program Must Be Remedied. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-90-91] Washington, D.C.: June 21, 1990. Serious Shortcomings in FAA's Training Program Must Be Remedied. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-90-86] Washington, D.C.: June 6, 1990. Staffing, Training, and Funding Issues for FAA's Major Work Forces. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-90-42] Washington, D.C.: March 14, 1990. Aviation Training: FAA Aviation Safety Inspectors Are Not Receiving Needed Training. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-89-168] Washington, D.C.: September 14, 1989. FAA Training: Continued Improvements Needed in FAA's Controller Field Training Program. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-89-83] Washington, D.C.: March 29, 1989. FAA Staffing: Recruitment, Hiring, and Initial Training of Safety- Related Personnel. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-88-189] Washington, D.C.: September 2, 1988. Aviation Safety: Measuring How Safely Individual Airlines Operate. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-88-61] Washington, D.C.: March 18, 1988. Aviation Safety: Needed Improvements in FAA's Airline Inspection Program Are Underway. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-87-62] Washington, D.C.: May 19, 1987. FAA Work Force Issues. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-RCED-87-25] Washington, D.C.: May 7, 1987. Human Capital: Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-546G] Washington, D.C.: March 1, 2004. Human Capital: Selected Agencies Experiences and Lessons Learned in Designing Training and Developing Programs. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-291] Washington, D.C.: January 30, 2004. Human Capital Management: FAA's Reform Effort Requires a More Strategic Approach. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-156] Washington, D.C.: February 3, 2003. A Model of Strategic Human Capital Management. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-373SP] Washington, D.C.: March 15, 2002. Human Capital: A Self-Assessment Checklist for Agency Leaders. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/OCG-00-14G] Washington, D.C.: September 1, 2000. Human Capital: Design, Implementation, and Evaluation of Training at Selected Agencies. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-00-131] Washington, D.C.: May 18, 2000. Human Capital: Key Principles From Nine Private Sector Organizations. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-28] Washington, D.C.: January 31, 2000. Related FAA Training: Federal Aviation Administration: Challenges for Transforming into a High-Performing Organization. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-770T] Washington, D.C.: May 18, 2004. Aviation Safety: Data Problems Threaten FAA Strides on Safety Analysis System. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-95-27] Washington, D.C.: February 8, 1995. Aircraft Certification: New FAA Approach Needed to Meet Challenges of Advanced Technology. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-93-155] Washington, D.C.: September 16, 1993. Aviation Safety: Progress on FAA Safety Indicators Program Slow and Challenges Remain. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/IMTEC-92-57] Washington, D.C.: August 31, 1992. Aviation Safety: FAA's Safety Inspection Management System Lacks Adequate Oversight. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-90-36] Washington, D.C.: November 13, 1989. [End of section] (540089): FOOTNOTES [1] In addition, FAA delegates about 90 percent of its safety inspection activities to about 13,600 private persons and organizations, known as designees. The designees augment FAA's inspection workforce by allowing inspectors to concentrate on what FAA considers to be the most critical safety areas. For example, while designees conduct routine functions, such as approvals of aircraft technologies that the agency and designees have had previously experience with, FAA inspectors focus on new and complex aircraft designs or design changes. For an assessment of the designee programs, see GAO, Aviation Safety: FAA Needs to Strengthen the Management of Its Designee Programs, GAO-05-40 (Washington, D.C.: Oct. 8, 2004). [2] System safety is a multidisciplinary, integrated, and comprehensive regulatory approach using engineering and management principles, criteria, and techniques to identify and mitigate high-risk areas. When FAA uses it in the oversight of airlines, the system safety approach covers every aspect of an airline's operations, from the design of the hardware to the culture and attitudes of the airline's personnel. The approach calls for a systematic review of an airline's policies and procedures to ensure that they incorporate such basic safety principles as clear lines of responsibility and written documentation. According to FAA, the approach allows it to concentrate and target inspector resources where there is the greatest safety risk. The success of a system safety approach to regulation depends on comprehensive safety data, sophisticated analysis tools, and a workforce well trained in risk assessment, auditing, systems thinking, and communications. [3] In addition to training involving aviation technologies (such as use of new materials in aircraft and aircraft electronic systems), FAA includes in its definition of technical training, topics such as inspector job skills, risk analysis, data analysis and training in software packages, such as spreadsheets. Our use of the term "technical" is limited to aviation technologies. [4] We considered all training that FAA classifies as either mandatory, position essential, or continuing development as essential training. Mandatory is training that is required for all newly hired inspectors and previous experience may not be substituted for this training. Position essential is training or a skill that is required based on an inspector's current position (e.g., training required for maintenance inspectors). To determine which courses were technical, we reviewed the description for each course taken from 2002 through 2004 and determined whether it was primarily technical in nature, within our use of the term. [5] GAO, Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government, GAO-04-546G (Washington, D.C.: Mar. 1, 2004). [6] FAA inspectors specialize in conducting inspections of various aspects of the aviation system, such as aircraft and parts manufacturing, aircraft operations, aircraft airworthiness, and cabin safety. See the background section of this report for more information on inspector specialization. [7] Because of the statistical survey techniques we employed in surveying FAA's inspectors, we are 95 percent confident that the results we present are within 4.6 percentage points of the results that we would have obtained if we had surveyed all 3,000 front-line inspectors. That is, we are 95 percent confident that had we surveyed all inspectors, between 48 and 57 percent of them would have told us that, to a great or very great extent, they have the technical knowledge to do their jobs. All percentage estimates from the survey have a margin of error of plus or minus 4.6 percentage points or less, unless otherwise noted. Throughout the survey we used a 5-point scale (very great, great, moderate, some, and no extent). For the most part, we report on the degree to which inspectors expressed their views to a very great or great extent because we believe that "a moderate extent" does not represent a strong positive or negative view and does not represent a level of performance to which a high-performing organization should aspire. [8] The goal of ATOS is to identify safety trends in order to spot and correct problems at their root cause before an accident occurs. This program allows FAA inspectors to now look at an airline as a whole, to see how the many elements of its operations--from aircraft to pilots to maintenance facilities to flight dispatch to cabin safety--interact to meet federal standards. The program will ultimately encompass all of the approximately 120 American airlines that operate in the United States, at any given time. [9] A repair station is an FAA-certified maintenance facility that is authorized to perform maintenance or alterations on U.S.-registered aircraft. [10] Air carriers are considered new entrants (or new air carriers) for their first 5 years of operation. [11] We will issue a report on FAA's oversight of non-ATOS carriers later this year. [12] In addition, Flight Standards has recently moved forward with a curriculum transformation strategy that will fundamentally change how its training program is managed. Acknowledging that its training activities tend to be fragmented, sometimes working at cross purposes, and sometimes leaving major gaps, Flight Standards' transformation strategy calls for a transition from the current course management structure to one that is curriculum based. The new structure will integrate individual training courses into a logical curriculum for each type of inspector that incorporates the specialty needs of inspectors. Flight Standards believes this new approach will cover inspectors' technical training needs, including their training on current technologies. Under this transformation strategy, training curriculum oversight teams will be formed for each inspector specialty to ensure that those inspectors receive the appropriate technical training. [13] We did not attempt to assess how Flight Standards conducted its analysis because it could not locate documentation associated with it. [14] Flight Standards officials said that the office is validating the inspector competencies. [15] Composite materials are materials that when combined are stronger than the individual materials by themselves. The benefits of using composite materials in aviation include light weight, durability, and corrosion resistance. [16] The 95 percent confidence interval for this estimate is from 23 to 32 percent. [17] About 34 percent of all essential courses are technical and range from 0 percent for Air Certification inspectors to 50 percent for air carrier avionics inspectors. See table 10 in appendix II for additional results. [18] The 95 percent confidence interval for this estimate is from 48 to 57 percent. [19] The 95 percent confidence interval for the cabin safety inspector responses is from 62 to 94 percent. The 95 percent confidence interval for the aircraft certification inspector responses is from 51 to 85 percent. [20] The 95 percent confidence interval for this estimate is from 22 to 48 percent. The survey also asked inspectors about their knowledge of the automated systems used in their jobs, such as ATOS or the Performance Tracking and Reporting Subsystem, because these are important tools for the system safety approach to inspections. We estimate that about 46 percent of inspectors believe, to a great or very great extent, that they have enough knowledge of automated systems to do their jobs. The 95 percent confidence interval for this estimate is from 42 to 51 percent. [21] FAA notes that its course evaluations support that 78 percent of its employees report that training has improved their job performance. However, its survey results are not comparable to ours. First, FAA's results represent responses to evaluations for all courses, both technical and nontechnical. In addition, it represents inspectors who responded that FAA training greatly improved, improved, and somewhat improved their job performance. In analyzing the results of our survey, we did not include the third category, as it does not represent a strong endorsement for the results of FAA training. [22] The 95 percent confidence interval for this estimate is from 30 to 39 percent. FAA officials stressed that training occurs over the span of a career and cautioned that asking inspectors' views about 2 years experience would present a distorted view. According to our analysis, FAA inspectors have been with the agency an average of 9.3 years, according to our survey inspectors have been in their current position an average of 5.3 years. We recognize FAA's concern. However, it is not reasonable to expect inspectors to recall their views on training received over a large time span, as doing so could lead to unreliable results. In addition, since this report focuses on FAA's current actions to ensure up-to-date technical training, we believe it is more useful to measure inspectors' views about the training that they are receiving or have recently received. [23] The 95 percent confidence intervals for these estimates are from 29 to 49 percent and from 6 to 39 percent, respectively. [24] Our survey provided the opportunity for inspectors to relate anything they wanted us to know about technical training. Some inspectors submitted both positive and negative comments. [25] The 95 percent confidence interval for this estimate is from 76 to 84 percent. [26] The 95 percent confidence intervals for these estimates are 21 to 29 percent and 36 to 45 percent, respectively. [27] See appendix III for a list of these experts. Experts commented in the area of their expertise. [28] We contacted representatives from the airlines that belong to the Air Transport Association and the Regional Airline Association to obtain their perspectives on FAA inspector technical training. [29] See footnote 4 for how we defined essential training. [30] The 95 percent confidence interval for this estimate is from 38 to 48 percent. [31] The 95 percent confidence interval for this estimate is from 23 to 32 percent. [32] The 95 percent confidence interval for this estimate is from 45 to 54 percent. [33] The 95 percent confidence intervals for these estimates are 47 to 64 percent and 27 to 43 percent, respectively. [34] The 95 percent confidence interval for this estimate is from 50 to 59 percent. [35] The 95 percent confidence intervals for these estimates are 32 to 41 percent, 23 to 32 percent, and 23 to 31 percent, respectively. [36] We analyzed the training records for FAA's approximately 3,000 front-line inspectors only. The analysis did not include supervisors, managers, and others in the aviation safety inspector job series who do not perform front-line inspections. [37] About 81 percent of the inspectors have completed at least half of their essential courses, both technical and nontechnical. (See table 11 in app. II.) In addition to the essential courses, most inspectors were also able to take other technical training courses that they, their supervisors, and FAA management have determined are related to their jobs. Inspectors averaged 1.7 technical courses outside of their list of essential courses over the past 3 fiscal years. (See table 12 in app. II.) [38] The 95 percent confidence interval for this estimate is from 16 to 24 percent. [39] The maximum value for the upper end of the 95 percent confidence interval for all other inspectors is 49 percent. The 95 percent confidence interval for general aviation avionics inspectors is between 0 and 13 percent. [40] The 95 percent confidence interval for this estimate is from 19 to 26 percent. Inspector dissatisfaction with the timeliness of training delivery is not limited to technical training. Inspectors also expressed concern about the timeliness of automation training. According to our survey, about 29 percent of inspectors indicated, to a great or very great extent, that this type of training was received in time to do their current job. The 95 percent confidence interval for this estimate is from 24 to 33 percent. [41] As of June 2005. [42] Flight Standards estimates that in addition it will develop 7 new courses, revise 9 existing courses, and complete 9 course evaluations by the end of fiscal year 2005. [43] Robert O. Brinkerhoff and Anne M. Apking, High Impact Learning: Strategies for Leveraging Business Results from Training, (Cambridge: Perseus, 2001) and Robert O. Brinkerhoff, The Success Case Method, (San Francisco: Berrett-Koehler Publishers, Inc., 2003). [44] We do not present the results of the 8 field and 3 headquarters managers because of the small numbers, and some did not answer all questions. [45] According to data provided by Flight Standards, there have been over 60 requests for revised, updated, or new courses to be developed since 2000, with more than 33 of them submitted since 2003 (plus a possible 11 more that were undated requests). Of all that have been submitted, 23 were technical. [46] The 95 percent confidence interval for this estimate is from 50 to 60 percent. [47] The 95 percent confidence interval for this estimate is from 44 to 54 percent. [48] Donald L. Kirkpatrick, Evaluating Training Programs: The Four Levels (San Francisco: Berrett-Koehler Publishers, Inc., 1994). Kirkpatrick conceived a commonly recognized four-level model for evaluating training and development efforts. The fourth level is sometimes split into two levels, with the fifth level representing a comparison of costs and benefits quantified in dollars. The fifth level, return on investment, is attributed to Jack Phillips and is taught in education and training seminars linking the two methodologies. [49] In addition, FAA receives a very limited amount of training for free from aircraft manufacturers during the development and deployment of a new or reconfigured aircraft type. FAA inspectors and representatives from the aircraft manufacturer certify each other initially by flying in the new aircraft. During these initial flights, the FAA inspector works closely with a manufacturer's test pilot to learn to operate the aircraft while also identifying requirements for special training or operation of the aircraft that will be necessary to certify future pilots. In addition, the criteria for pilot certification on the unique characteristics of the aircraft are identified during this process. As this is accomplished, the FAA inspector also learns to fly the aircraft at no cost to FAA. These instances account for less than 10 training sessions per year for FAA inspectors, according to FAA. [50] These qualified pilots become designated check airmen who are then permitted to conduct flight checks or instruction in an airplane for the purpose of certifying other air carrier pilots to ensure they are properly trained and able to fly the aircraft. [51] U.S. Department of Transportation, Office of Inspector General, Free Industry Flight Training of Inspectors: Federal Aviation Administration, AV-1998-042 (Washington, D.C.: Dec. 9, 1997). [52] The 95 percent confidence interval for this estimate is from 33 to 42 percent. [53] The 95 percent confidence interval for this estimate is from 2.7 to 3.5 weeks. [54] The 95 percent confidence intervals for these estimates are from 50 to 59 percent and 22 to 31 percent, respectively. [55] The 95 percent confidence interval for this estimate is from 15 to 23 percent. [56] The 95 percent confidence intervals for these estimates are from 2.8 to 3.6 weeks and 1.0 to 3.0 weeks, respectively. [57] The 95 percent confidence interval for this estimate is from 9 to 16 percent. [58] The 95 percent confidence interval for this estimate is from 32 to 41 percent. [59] GAO-04-546G. [60] Our population included only those inspectors who actively participate in inspection activities as part of their regular job duties. It did not include managers, supervisors, or inspectors detailed to headquarters or regional offices. FAA employs a total of approximately 3,700 safety inspectors. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548: