17
AD-A277 549 II IBUll I1I1|ill llimlI Ilt DOT/FAAIAM-94/4 Validation of the Federal Aviation Administration Air Office of Aviation Medicine Washington, D.C. 20591 Traffic Control Specialist Pre-Training Screen Dana Broach Civil Aeromedical Institute Federal Aviation Administration D T IC Oklahoma City, Oklahoma 73125 S ELECTE R 2 91994 Jan Brecht-Clark 8D Federal Aviation Administration S F Office of Aviation Medicine Washington, DC 20591 February 1994 I0- Pb releseased povd Final Report This document is available to the public through the National Technical Information Service, Springfield, Virginia 22161. 94-09542 0 11 INIllllI 1111l l il 0U.S. Department of Transportation Federal Aviation Administration 28 100

Validation of the Federal Aviation Administration Air ... of the Federal Aviation Administration Air ... suggested that predictor tests ... proposed test battery was developed by ASI

  • Upload
    ngodieu

  • View
    214

  • Download
    0

Embed Size (px)

Citation preview

AD-A277 549II IBUll I1I1|ill llimlI IltDOT/FAAIAM-94/4 Validation of the Federal

Aviation Administration AirOffice of Aviation MedicineWashington, D.C. 20591 Traffic Control Specialist

Pre-Training Screen

Dana BroachCivil Aeromedical InstituteFederal Aviation Administration

D T IC Oklahoma City, Oklahoma 73125

S ELECTER 2 91994 Jan Brecht-Clark8D Federal Aviation AdministrationS F Office of Aviation Medicine

Washington, DC 20591

February 1994

I0- Pb releseased povd Final Report

This document is available to the publicthrough the National Technical Information

Service, Springfield, Virginia 22161.

94-09542 011 INIllllI 1111l l il 0U.S. Department

of TransportationFederal AviationAdministration

28 100

BestAvailable

Copy

NOTICE

This document is disseminated under the sponsorship ofthe U.S. Department of Trar-portation in the interest of

information exchange. The United States Governmentassumes no liability for the contents or use thereof.

Technical Documentation Page1. Report No. 2. Government Accesion No. 3. Recipient's Catalog No.

DOT/FAA/AM-94/44. Title and Subtitle 5. Report Date

Validation of the Federal Aviation Administration Air February 1994Traffic Control Specialist Pre-Training Screen 6. Performing Organization Code

7. Author(s) 8. Performing Organization Report No.

Dana Broach, Ph.D., and Jan Brecht-Clark, Ph.D.

9. Performing Organization Name and Address 10. Work Unit No. (TRAIS)

FAA Civil Aeromedical InstituteP.O. Box 25082Oklahoma City, OK 73125 11. Contract or Grant No.

12. Sponsoring Agency Name and Address 13. Type of Report and Period Covered

Office of Aviation MedicineFederal Aviation Administration800 Independence Avenue, S. W.Washington, DC 20591

15. Supplementary Notes

This research was conducted under task AAM-C-93-HRR-1 29 and AAM-C-HRR-1 6316. Abstract

Two formal validation studies of the Air Traffic Control Specialist Pre-Training Screen(ATCS/PTS), a 5-day computer-administered test battery, are described. The ATCS/PTSwas designed to replace the 9-week U.S. Federal Aviation Administration (FAA) AcademyATCS Nonradar Screen program that served as the second major test in the ATCS selectionsystem. Review of ATCS job analyses suggested that predictor tests should assesscognitive constructs such as spatial reasoning and short-term memory, and require dynamic,concurrent performance. A proposed test battery was developed, consisting of 2 computer-administered information processing tests and a simplified radar-based air traffic controlwork sample. In study 1, predictive, criterion-related validation (N = 438) found that theproposed test battery explained additional variability in scores earned in the 9-week FAAAcademy program, after taking into account student aptitude. In study 2, criterion-relatedvalidation (N = 297) demonstrated that the proposed test battery was as valid as the 9-week FAA Academy ATCS Nonradar Screen for predicting progress in field training.Preliminary data from a third study conducted after validation of the ATCSIPTS seem tosuggest that the abilities assessed by the new computerized tests reflect the abilitiesrequired on the job. However, implementation of the ATCS/PTS for actual employmentdecisionmaking in June 1992 was based on results obtained in the second concurrent,criterion-related validation study. The U.S. controller selection system since June 1992 hasconsisted of the 4-hour written ATCS aptitude test battery followed by, for those applicantsearning a qualifying score and dependent upon agency manpower requirements, second-level screening on the ATCS/PTS. Additional research requirements as part of an aviationhuman factors research program are described.

17. Key Words 18. Distribution StatementAir Traffic Control Specialist, Selection, Document is available to the publicValidation, Tests, Ability, Job Analysis, through the National TechnicalComputer-administered Test Information Service, Springfield, VA

1 2216119. Security Classif. (of this report) 20. Security Classif. (of this page) 21. No. of Pages 22. Price

UNCLASSIFIED UNCLASSIFIED 1 16 IForm DOT F 1700.7 (8-72) Reproduction of completed page authorized

ACKNOWLEDGMENTS

Research reported in this article was performed under FAA research tasks AAM-92-C-HRR- 128 andAAM-93-C-HRR-163. The views expressed in this article are those of the authors and do not necessarily

reflect the views or policy of the Federal Aviation Administration.The authors would like to thank Mr. Barry Lambert Harris, the former Deputy Administrator of the

FAA, for sponsoring this project. The authors also acknowledge the contributions of Dr. Jay Aul of theFAA and Dr. Brian O'Leary of the U.S. Office of Personnel Management for the civil service andtechnical oversight and guidance they provided throughout the project. Finally, the authors wish tothank Dick Rodine and Charlie Collier for their work on contracts, logistics, coordination, andmanagement throughout the course of this project.

Accesion For _•

NTIS C. , &!

OT • -- : -

By...... °...............By

D,.. . . . . .ist................Dist tc i

Uiii

VALIDATION OF THE FAA AIR TRAFFIC CONTROL SPECIALIST

PRE-TRAINING SCREEN

The United States Federal Aviation Administration at the end of the program. That risk may have discour-(FAA) is charged with managing the U.S. airspace. Air aged potentially qualified women and minority per-traffic controllers are at the heart of a web of radars, sons from pursuing an air traffic career (ASI, 1991).computers, and communication facilities that ensure The FAA undertook a major review of its ATCSthe safety and efficiency of an increasingly complex air selection and training programs in 1990 to addresstransportation system. Appropriate selection of per- these agency and applicant costs and other concerns.sonnel into a training program for these critical posi- Three major ATCS selection policy goals were identi-tions is an important human factors problem. This fied for the project: (1) reduce the costs of ATCSpaper describes research conducted by the FAA to selection; (2) maintain the validity of the ATCS selec-validate a cost-effective air traffic control specialist tion system; and (3) support agency cultural diversity(ATCS) selection procedure. The project resulted in goals. The first step toward achieving these goals was tothe implementation of a new selection test for ATCS develop and validate a test battery to replace the 9 weekapplicants in June 1992 that was radically different ATCS Screen.from any previous ATCS selection program under-taken in the U.S. Proposed test battery

Development of the new test battery began in lateProject background 1990 by reviewing available information about the

The ATCS selection process between fiscal years cognitive requirements of the ATCS job. As described1986 and 1992 consisted of two major tests: (a) a 4 in one recent cognitive task analysis, controllers attendhour written aptitude examination administered by to multiple information sources, assess and integratethe United States Office of Personnel Management the data, develop and prioritize plans of action, and(OPM); and (b) a 9-week initial training program implement those plans under time pressure while main-administered by the FAAAcademy. Between 1984 and taining situational awareness (Human Technology,1992, over 200,000 applicants took the written OPM Inc., 1991). To assess the cognitive and sensory at-aptitude examination across the country at a cost of tributes required to perform these job functions, aabout $20 per examinee (J. Aul, personal communica- proposed test battery was developed by ASI. Thetion). Between October 1985 and January 1992, just battery was developed within the conceptual framework12,869 of those 200,000+ applicants were selected to provided by Multiple Resources Theory (Rodriquez,attend the FAA Academy ATCS Nonradar Screen Narayan, & O'Donnell, 1986; Shingledecker, 1984;("ATCS Screen"). The direct cost of this second-stage Wickens, 1984). Two computer-administered infor-in the selection process was about $10-12,000 per mation processing tests were designed to dynamicallystudent (Gwen Sawyer, June 1990). Of those students assess cognitive attributes such as spatial reasoning,entering the ATCS Screen, 7,091 successfully gradu- short-term memory, movement detection, pattern rec-ated and entered into on-the-job training. This two- ognition, and attention allocation (ASI, 1991). Instep selection process cost the FAA annually between addition, a low-fidelity radar simulation of air traffic$20 and 25 million to obtain approximately 1,400 control vectoring and separation tasks was also devel-trainee or "developmental" controllers. oped as a computer-administered work sample. The

The written aptitude tests - ATCS Screen selection information processing tests and the work sampleprocess also imposed significant costs on applicants, required performance of concurrent, multiple tasks byApplicants selected to attend the ATCS Screen had to candidates to reflect the job demands placed on con-leave their current jobs and families for 9 weeks with a 55 trollers.- 60% chance of remaining in the controller occupation

Validation of the FAA A TCS Pre- Training Screen

The 2 computerized information processing tests now the bottom call sign before responding, for aswere (a) the Static Vector/Continuous Memory test soon as an answer was made, a new set of call signs(SV/CM) and (b) the Time Wall/Pattern Recognition appeared. The attention director at the bottom centertest (TW/PR). In the Static Vector (SV) component of of the SV/CM screen informed the subject which taskthe first test, a pair of simulated aircraft were presented (SV or CM) was to be performed for each trial. A fixedon the left half of the computer screen (Figure 1). A number of trials for each component (SV and CM)quasi-data block for each target gave speed ("S250" were administered in a 5 minute SV/CM session. Thewas 250 knots), altitude ("A250" meant 25,000 feet). speeds, altitudes, and spatial relationships betweenThe subject's task was to determine as rapidly and aircraft in the SV and the call signs in the CM variedaccurately as possible if the simulated aircraft were in from trial to trial within the session. Performanceconflict based on their altitude, speed, and spatial feedback was provided at the end of each sessX,ýn onrelationship. The Continuous Memory (CM) compo- each component (SV and CM).nent on the right side of the screen presented 2 aircraft The TW/PR test also consisted of a set of pairedcall signs, one above and the other below a line. The tasks (Figure 2). In the Time Wall (TW) component,subject's task was to remember the bottom call sign a square target appeared first, moving from left to right("Target call sign" in Figure 1), for in the next CM at a steady speed toward the "wall" on the far right oftrial, the subject had to indicate if the call sign above the screen. After an initial time interval, the movingthe line ("Probe call sign" in Figure 1) was the same as target and wall disappeared and were replaced by pairshad been presented below the line in the previous CM of patterns. The Pattern Recognition (PR) task was totrial. However, the subject had to encode what was decide if the patterns were identical while keeping in

Data block Aircraft target Probe call sign

SV -123 SVUA312

Target call sign

Attention director

FIGURE 1. STA TIC VECTOR (SV)/CONTINUOUS MEMORY (CM) SCREEN. SV test is shown onthe left-hand side of the screen, CM test on the right. When the attention director was to the left,the subject's task was to decide if the aircraft targets would collide or not, based on the altitude("A230") and speed ("S300") information in the data blocks and spatial relationships of the targets.When the attention director was to the right, the subject's task was to first, memorize the targetcall sign below the line, and second, indicate if the probe call sign above was the same, ordifferent, as the target call sign that had been presented below the line in the previous CM trial.

2

Validation of the FAA A TCS Pre- Training Screen

mind the continuing movement of the TW target The Air Traffic Scenario Test (ATST; Figure 3), thetoward the wall. The TW task was to stop the now computer-administered work sample component ofinvisible t rget as close as possible to, without actually the proposed test battery, was developed by 4 subjecthitting or 3assing through, the wall. Subjects were matter experts with more than 30 years of air trafficpresenled w+:h a fixed number of TW/PR trials within control experience (ASI, 1991). The task required thea nominal 5-minute test session; the actual length of subject to control aircraft within a simplified syntheticthe session was a function of subject response time. For airspace, directing them to their destinations accord-example, consistently stopping the moving target in ing to a small set of rules. There were 6 destinations: 4the TW short of the wall by a large margin reduced outbound gates, A, B, C, and D; and 2 airports, E andtotal session time proportionately. Measures from both F. The direction of travel, speed, and altitude of thethe SV, CM, and PR components included the mean aircraft, represented by small arrows next to the quasi-percent correct and mean reaction time for correct data blocks, were controlled by mouse. Three alphanu-responses across trials within the 5-minute sessions for meric characters comprised the quasi-data blocks: first,each test pair; the TW measure was the absolute aircraft speed (Ulow, Medium, fast); second, altitudedistance (in milliseconds) between the wall and target (1 = Lowest, 4 = Highest); and third, destination. Thewhen stopped by the subject. Performance feedback orientation of the aircraft arrow indicated its currenton these measures was provided to the subjects at the direction of flight. An open circle in an upper cornerend of each 5 minute session, of the data block indicated an aircraft waiting to be

Moving target

Tw-

Wall

Patterns tobe matched

TW -- Stoppedtarget

FIGURE 2. TIME WALL (TW)/PATTERN RECOGNITION (PR) SCREENS. First, the target appeared,moving from left to right at a steady speed toward the "wall' (Top screen). After an initial timeinterval, the target and wall were masked by a pair of patterns (Middle screen). The subject's taskwas to decide if the patterns were the same or different. A new pair of patterns appeared after eachresponse was made. However, the subject had to keep in mind the continuing movement of theTW target toward the wall, as the TW task was to stop the target (Bottom screen) as close aspossible to, without actually hitting or passing through, the wall.

3

Valadasy of fke FAA A TCS Pre- Trix ig Senms

activated by ("handed off to*) the subject. The large en route delay time was summed with the time eacharrow in the lower right hand corner of the screen aircraft spent waiting to be activated as a measure of

indicated the landing direction at tirports E and F, overall controller efficiency. Performance feedback onwhile the bottom horizontal bar icon represented the these measures was provided to subjects at the end ofminimum lateral separation distance. Aircraft landed each of 20 practice scenarios.

at airports E and F at the lowest altitude and slow speedin the required direction; aircraft exited gates A, B, C, Study 1:and D at the fastest speed and highest altitude. A Predictive, Criterion-related validation

difference in altitude between any two aircraft was Two validation studies of this proposed test batteryconsidered adequate separation; aircraft at the same were conducted by the FAA in 1991. The purpose ofaltitude had to be separated by at least 5 nautical miles the first study was to assess the predictive, criterion-

as represented by the separation icon. In addition, all related validity of the proposed test battery, and toaircraft had to be separated from the airspace boundary determine the incremental validity of the proposedby at least 5 nautical miles. Error counts were obtained computerized tests over the existing written test. Theand summed to create an overall error score. In addi- sample in the first predictive, criterion-related valida-tion, the system automatically computed the differ- tion study consisted of the 423 newly hired air trafficence between the actual time to reach destination for control students who entered the ATCS Screen in

each aircraft and the time required for the optimum March and April 1991 in accordance with existingflight path as determined by the system software. This FAA procedures and policies. The sample was pre-

Boundary

At

Airport - EF3D DirectionAEcontrol icon

434-- Altitude

Gate----* D S4A/1* B 2 control icon1

F 4-.Speed

12F-0-F M control iconA -SAircraft M2F i - Landing

'/ T C direction

X7 H-and-offData block indicator Separation

distance icon

FIGURE 3. AIR TRAFFIC SCENARIO TEST (ATST) SCREEN. The boundary encloses a simplifiedairspace, with 4 outbound gates, A, B, C, and D and 2 airports, E and F. The aircraft and directionof flight are represented by the arrows adjacent to a data block. The alphanumeric data blockindicates aircraft speed (S, M, or F) and altitude (1 = lowest, 4 = highest). Aircraft waiting to behanded off are tagged with a small open circle in the upper right hand corner of the data block.Aircraft are controlled with a mouse. First, the candidate clicks on an aircraft, and then clicks on

the appropriate element of either the direction control, altitude control, or speed control icons tochange that flight parameter. Subjects are reminded of the required landing direction at airportsand minimum horizontal separation distance by the landing direction and separation distanceicons respectively.

4

Validation of the FAA A TCS Pre- Training Screex

dominantly male (77.1%) and non-minority (83.0%); SCREENscore of 71.8 (SD = 11.8) for this validationmost (88.7%) had entered federal service by competi- sample of 423 students was not significantly differenttive examination rather than by non-competitive spe- from that of the population students that had enteredcial appointment. There were significantly more women the ATCS Screen between October 1985 and January(22.9%) in this validation sample compared to the 1991.population of students that had entered the ATCSScreen between October 1985 and January 1991 Procedure(18.9%; Z= 2.05,p <_.05). Similarly, minorities were The proposed test battery was administered in 2also over-represented in this validation sample (17.0%) waves to subjects the week prior to beginning thein comparison to the population of ATCS Screen ATCS Screen. The subjects were tested in March andstudents (10.2%; Z = 2.38, p :S .05). The majority April 1991 at the FAA Civil Aeromedical Institute(65%) had no prior aviation-related experience which (CAMI) in Oklahoma City. Instructions on the testwas representative of the population that reached the battery were given on Monday morning. A total of 202nd stage of the ATCS selection system. Aptitude SV/CM and 20 TW/PR practice sessions were admin-scores for the ATCS occupation, represented in this istered to subjects across 3.5 days (Monday afternoonstudy by the variable RATING, are based on the civil through Thursday). The SV/CM and TW/PR tests didservice test scores earned by an applicant on the written not change in difficulty across sessions. Subjects alsoaptitude test plus any statutory veteran's preference were given 20 practice scenarios for the ATST, build-points. The general development, psychometric char- ing in complexity and difficulty from about 12 aircraftacteristics, and validity of the written aptitude test in 30 minutes to over 40 aircraft in less than 30battery has been extensively described (Sells, Dailey, & minutes in the final practice sessions. PerformancePickrel, 1984). RA TINGwas used to rank-order com- feedback was provided to subjects after each practicepetitive applicants within statutory guidelines such session. On Friday, subjects received a final 4 SV/CM,that hiring was done on the basis of merit (Aul, 1991). 4 TW/PR sessions, and 6 ATST scenarios. Measures

were averaged across these final graded sessions withinCriterion for predictive validation test, yielding 8 proposed test scores: (1) SV average

The criterion for this predictive study was the final percent correct; (2) SV average correct response reac-composite score earned in the ATCS Screen. The tion time; (3) CM average percent correct; (4) CMATCS Screen was originally established in response to average correct response reaction time; (5) TW averagerecommendations by the U.S. Congress House Coin- absolute error; (6) PR average correct response reac-mittee on Government Operations (U.S. Congress, tion time; (7) average ATST error score; and (8)1976) to reduce field training attrition rates. The summed delay and waiting times in the ATST sce-ATCS Screen was based upon a miniaturized training- nario. Aptitude ratings and ATCS Screen scores weretesting-evaluation personnel selection model (Siegel, extracted for the 423 subjects from the CAMI research1978, 1983). Thirteen performance assessments, in- data bases after all subjects had completed the ATCScluding classroom tests, laboratory simulations of Screen. These data were matched with proposed testnonradar air traffic control, and a final written exami- scores for analysis; proposed test scores were not usednation, were made during the course of the ATCS in any way to make employment decisions about theScreen (Della Rocco, Manning, & Wing, 1990). The subjects.final summed composite score (SCREEN) of theseATCS Screen performance measures was weighted Results20% for classroom tests, 60% for laboratory scores, On one hand, performance on the SV/CM andand 20% for the final examination, with a minimum TW/PR tests appeared to reach differential stabilityscore of 70 out of 100 required to pass. In this sample, (Bittner, 1979) at about the 1 5th session. The average56.0% passed the ATCS Screen, 27.7% failed, and performance within test component across the final ses-16.3% withdrew prior to completion. The mean sions represented a reasonable measure of asymptotic

5

Vadetie of the FAA A TCS Pre- Trmiasug Smme

individual differences on those tests. On the other tion study was composed of 297 trainee ("develop-hand, learning curve analyses were not possible with mental') and FPL controllers. While this sample wasthe ATST because scenario difficulty increased across predominantly male (64.6%) and non-minoritysessions. However, average performance across the (61.6%), women and minorities were over sampledfinal 6 scenarios was still computed as the index of relative to their representation in the ATCS workforce.individual differences on that test component. Mul- The majority of the sample was drawn from en routetiple regression analysis was used to assess how well the centers (58.2%), reflecting the historical employmentproposed test battery predicted student performance patterns in the workforce; 49.2% had attained FPLin the ATCS Screen after taking into account student certification. The final composite SCREEN score foraptitude. First, RATINGwas entered into the regres- each participant was extracted from the CAMI ATCSsion equation predicting SCREEN. There was a statis- Selection data base and used as the current predictor intically significant linear relationship between RATING this study. The SV/CM, TW/PR, and ATST averageand SCREEN of R .23, p .001, where R was the test scores described in the first study were the alterna-multiple correlation between predictor (RATING) and tive predictors in this validity study. The ATCS Pre-criterion (SCREEN) and p < .001 indicated that an R Training Screen (ATCS/PTS), as the proposed batteryof this magnitude would be expected by chance alone had come to be known, was administered to subjectsin less than 1 in a thousand times. The relationships of during late summer 1991 using the same test adminis-proposed test battery average final scores to SCREEN tration protocols as in the first study.were analyzed in the second step of the multipleregression analysis using a forward stepwise procedure Criterion for concurrent validationto determine the optimal combination of predictor This study was constrained to use available trainingvariables. In a forward stepwise multiple regression performance indices as validation criteria; no otheranalysis, the proposed test score accounting for the criteria were developed or collected. These indicesmost variability left in the criterion SCREEN entered included the number of days spent in particular phasesthe regression equation; then, one at a time, proposed of field training and hours of formal, documented on-test scores which accounted for the most of the remain- the-job training (OJT) provided under the supervisioning unexplained variability in SCREENwere added to of a designated OJT Instructor within those phases, asthe equation, until the amount of variability explained reported by field ATC facilities in accordance withby a new score became insignificant. The optimal national policy (FAA, 1985). Subjective ratings oflinear combination of proposed test scores accounted developmental performance in that phase of trainingfor an additional 20% (RI = .20, p < .001) of the (1 = Bottom 10% compared to all other controllersvariability in SCREENover the proportion ofvariabil- observed in training, 6 = Top 10% compared to all otherity already explained by student aptitude scores (RA T- controllers observed in training) were also available forING). There were no statistical differences in the each participant in this second validation study. Dataprediction equation by sex and minority status (ASI, for the ground, local, and radar control phases of1991), suggesting that the proposed test battery might instruction were extracted from the CAMI ATCSnot adversely impact protected classes of applicants. Training Tracking data base for subjects drawn from

FAA terminal facilities. The ground control phaseStudy 2: qualified a developmental to control the movement of

Concurrent, criterion-related validation departing and arriving aircraft on the airport surface,Encouraged by the results of the initial predictive including ramps and taxiways. Local control devel-

study, the FAA conducted a concurrent, cr;terion- oped the skills to control arriving and departing air-related validation study to assess the validity of the craft on the active runways and in the immediate visualproposed test battery as a replacement for the ATCS airspace of the terminal. Radar control taught tech-Nonradar Screen (Weltin, Broach, Goldbach, & niques and procedures for the control of aircraft arriv-O'Donnell, 1991). The sample for this second valida- ing in and departing from the terminal's extended

6

Validetiox of the FAA A TCS Pre- Traiximig Screex

aitspace for facilities equipped with radar. Data on the current SCREENpredictor (see Ghisclli, Campbell, &initial radar associate and initial radar qualification Zedeck, 1981) and submitted for regression analysis.phases of training were collected for en route subjects. The corrected multiple correlation between the ATCS/The en route radar associate phase qualified the devel- PTS average final scores and TRNGPERFwas R = .25opmental controller to initiate and accept radar hand- (uncorrected R = .21, p .05) compared to R = .19offs and point-outs, perform flight data entries, (uncorrected R .11, p.05) for the current SCREENmaintain flight progress strips, and communicate with predictor. While modest, the validity coefficient of.25aircraft and other facilities by interphone and radio as for the ATCS/PTS indicated that a prediction aboutdirected by the radar controller on a position. In con- probable performance in field training for an indi-trast, the goal of the radar qualification phase of vidual could be made from knowledge of his or herinstruction was to qualify the developmental as the scores on the computerized test battery. Moreover, theradar controller on two positions or sectors within the validity of the proposed 5-day test battery was at leastassigned area of specialization. The radar controller equal to that of the existing 9-week training-as-screen.has overall responsibility for the safe, orderly, and Subsequent analyses again suggested that the validitiesexpeditious movement of air traffic within the as- of the ATCS/PTS and ATCS Screen did not vary as asigned sector of airspace. Performance assessments function of sex or minority group status (Weltin, et4afrom additional radar training conducted by the FAA 1992).Academy were also extracted from the research databases where available for subjects. FAA Academy radar Study 3:training provided instruction in critical radar tech- Comparison of ATCS/PTS to jobniques and procedures in the safety of a simulated attribute requirementsairspace. A third analysis (Broach & Aul, 1993) of the ATCS/

An overall standardized composite score for each of PTS was undertaken after it was validated in order to297 participants in this validation study was created independently compare test constructs with job cogni-from these time-to-complete, performance assessment tive attribute requirements. During the data collectionmeasures, and FAA Academy radar training. This phase of the second study, FAA psychologists andtraining performance (TRNGPERF composite crite- technicians interviewed 52 of the incumbent FPLrion represented the rate and quality of progress in controllers from all types and levels of air traffic con-training for an individual relative to peers assigned to trol facilities. Example facility types included Air Routethe same type and level of facility that had completed Trafr 'nntrol Centers (ARTCC), also known as Enthe same curriculum. The mean TRNGPERF score Route centers, Terminal Radar Approach Controlwas 0.44 (SD - .30), with a range of 0 to 1. A criterion (TRACON) terminals with high traffic densities, Levelscore of 0 indicated consistently poorer (longer than 3 radar terminals (L3R) with intermediate traffic den-average times to complete and lower assessments of sities, and Level 1 and 2 Nonradar (e.g., VFR Non-quality). A score of I reflected consistently higher approach) towers (L12NR) with lower traffic counts.performance than peers (shorter than average times The job analysts then completed a Position Analysisand higher assessments); an intermediate score of.50 Questionnaire (PAQ; McCormick, Mecham, &indicated consistently average performance relative to Jeanerett, 1977) for each interview. Estimated require-peers assigned to the same type and level of facility. ments for worker attributes of an ability or aptitude

nature were computed from the 52 sets of job ratingsResults by PAQ Services, Incorporated, based on their data

Correlations were computed between the current base on over 2,000 jobs in the U.S. economy. Prelimi-predictor SCREEN, alternative ATCS/PTS predic- nary data from this third analysis in the form oftors, and the criterion. The correlation matrix was estimated percentiles for cognitive and general intelli-corrected for explicit and incidental restriction in gence attributes are illustrated in Figure 4 for tworange due to prior selection of the sample on the selected air traffic control facility types and levels. The

7

Vakdatin of Aw FAA A TCS P*- Traixmn Scree

Long.termon ... .....

Short-twm marory -----

llm~ues ng -~---- - ---

Spai visualizaon - ---0-- -- - L12NRSptial orentation 7. . f ....

Numiericalcomputation -- -i--- -- ---- --

Arttmeftc reasoning --- --- ~------*- ---- *---/l~~hmetic re~~o.... ........... I.....

FIGURE 4. ATCS JOB ABILITIES PROFILE. Ability attributes are listed along the vertical axis. The

0 - 100 scores along the horizontal axis indicate the estimated proportion of jobs in the PAQServices, Inc. data base for which an attribute received the same or lower relevance scores thanA TCS jobs. Ability requirements for the A TCS job in Level 1 & 2 nonradar/nonapproach (VFR)towers are contrasted with the profile for en route Air Route Traffic Control Centers.

percentile estimates the proportion of the jobs in the U.S. economy to which spatial and attention alloca-PAQdata base for which an attribute received the same tion abilities are more relevant than to the controlleror lower relevance scores as the job being analyzed occupation. Finally, with the exception of spatial abili-(Mecham & McCormick, 1969; Mecham, McCormick, ties as illustrated in Figure 4, the cognitive abilities& Jeanneret, 1977; McCormick, Jeannerett, & requirements for controllers appeared to be reasonablyMecham, 1972). homogenous across facility types and levels. The re-

These analyses by Broach and Aul (1993) suggested quirement for spatial visualization appeared to bethat perceptual speed, closure, simple reaction time, more relevant to terminal facilities than to en routeand short-term memory were more relevant to the facilities.controller job than to many other jobs in the U.S. Overall, tests that represented perceptual speed,economy. Numerical computation, arithmetic reason- closure, reaction time, memory, arithmetic reasoning,ing, convergent and divergent thinking also appeared and some degree of spatial ability would be expected toto be more relevant to performance in the ATCS predict performance in both en route and terminaloccupation. But contrary to expectation, time sharing, environments. In order to evaluate the correspondenceselective attention, spatial visualization, and spatial between test and job requirements, PAQratings of theorientation were not more relevant to air traffic conrol proposed test battery were completed by a single,than to other U.S. occupations. In other words, there highly experienced PAQ consultant from Jeanerett

appears to be a substantial proportion of jobs in the and Associates. The resulting cognitive attribute re-

8

Validation of the FAA A TCS Pre- Training Screen

Simple reactiontm Urns- -

Long~tefm memoriaShomt-term memory -

Time 0aring • . .

Spatial ottentatlon --- _ _

Numeitclcomputation ------- _,_. ____

Arit-mtc masoning TestS........ ,- -' B-SV/CM (0)

Divergent thlnd .ng - 1W/-- (0)Intelligence --- --

Ideationu fluency - - -,_ATST(M)

Nrgina~lty-- -- - -- _ _

Pmlm es"' sroe n M e rleD v~ant0 10 20 30 40 s0 60 70 80 90 100

FIGURE 5. ATCS/PTS ABILITIES PROFILE. Ability attributes are listed along the vertical axis. The0 - 100 scores along the horizontal axis indicate the estimated proportion of jobs in the PAQServices, Inc. data base for which an attribute received the same or lower relevance scores. Abilityrequirements for the SV/CM, TW/PR and A TST are illustrated.

quirements for the test battery are illustrated in Figure DISCUSSION5 for this preliminary study of the correspondancebetween test and job. While no formal statistical analy- Two formal validation studies on a total of 720ses have been conducted as yet, there appeared to be subjects demonstrated that the ATCS/PTS was a vi-some degree of similarity between the test and job able replacement for the ATCS Screen as the 2ndprofiles in kind, if not degree. For example, the re- hurdle in the FAA's ATCS selection system. The firstquirement for perceptual speed and simple reaction study demonstrated that the computer-administeredtime were similar between the ATCS job and the TW/ test battery explained some of the variability in scoresPR and ATST tests. While the attribute percentile earned in the ATCS Screen, even after taking intoscores for the ATCS/PTS were generally lower, the account student aptitude. The second study foundshape of the profile across basic mental abilities such as that ATCS/PTS was about as valid as the ATCS Screenmemory and attention and higher-order skills such as in predicting relative performance in ATCS field tech-numerical computation and divergent thinking was nical training. The new test battery was objectivelybroadly similar to that of the job. Overall, these early administered and scored, and the validity of the newdata suggested at least some degree of correspondence test battery did not appear to vary as a function of sexbetween proposed test battery and job attribute re- and minority status. Finally, the ATCS/PTS achievedquirements; further analyses, using multiple raters to the major policy goal of reducing the cost of selection atevaluate the test battery, will provide a basis for a more the 2nd hurdle in the ATCS selection process from aboutdefinitive assessment. $10,000 to about $2,000 per candidate. Therefore, the

9

V, id4atou of the FAA A TCS Pre- Training Screen

FAAAcademyATCS Nonradar Screen was terminated performance, can be inferred from test scores (Guion,in March 1992 and the ATCS/PTS became opera- 1992). For example, given the nature of the criteriontional as the FAA's 2nd stage selection test in June in the concurrent validation study, the only fully1992 on the basis of the results of the second. concur- justified inference that currently can be drawn fromrent validation study. The ATCS selection sysi-em now ATCS/PTS scores is how rapidly a person might beconsists of the 4-hour written ATCS aptitude test expected to complete field ATCS training relative tobattery followed by, for those applicants earning a other developmentals (slower or faster than average,qualifying score, second-level screening on the ATCS/ overall). Inferences about probable technical job per-PTS. The final ATCS/PTS protocol provides 20 SV/ formance, such as efficiency in separating aircraft andCM, 20 TW/PR, and 20 ATST practice sessions over orderliness of the flow of aircraft, will require develop-2.5 days (Monday afternoon through Wednesday), ment of different criterion measures. Similarly, infer-followed by the final 4 SV/CM, 4 TW/PR, and 6 ences about attrition from the ATCS occupation fromATST "for grade" testing sessions on Thursday. Can- ATCS/PTS scores will have to await results of longitu-didates are informed of the outcome of screening on dinal evaluations of the Study 1 students and Study 2Friday. Those that successfully complete the ATCS/ developmentals as they progress through the fieldPTS are then eligible for hiring by the FAA and training program. Other important ATCS selectionsubsequent enrollment in the FAA Academy ATCS issues include differential assignment to facility typestraining programs. In this new system, all selection is and levels based on rest score profiles and assessment ofaccomplished prior to the actual hiring and subse- selection system utility. Finally, as the controller occu-quent training of entry-level controllers. pation changes, the ATCS selection process must also

The ATCS/PTS represents a major policy and re- change. The emerging Advanced Automation Systemsearch initiative for the FAA. As noted by Ackerman may (or may not) have profound implications for(1991), ATCS selection research represents a praxis of ATCS selection (see Manning & Broach, 1992 for an

public policy, psychological theory, and psychometric early exploratory study). Systematic and continuouspractice. Continuing research is required to assess the selection-oriented research is strongly recommendedlongitudinal fairness of the new battery in order to as an integral part of ATC systems design specificallysatisfy legal and human resource policy requirements. and the national aviation human factors research planAn additional research requirement is to develop and generally.validate an expanded test battery. Only cognitive abili-ties are assessed by the current version of the ATCS/PTS. But non-cognitive factors such as biographicaldata have been shown to be useful predictors of near-term criteria such as the ATCS Screen (Collins, Nye,& Manning, 1992) and criteria such as performance inradar-based training I to 2 years after entry into theoccupation (Broach, 1992). Personality has similarlyshown promise in several studies as a predictor of near-term performance (Schroeder, Broach, & Young, 1992;Nye & Collins, 1991). Development of a expandedtest battery might enable the agency to implement asingle-hurdle selection system, further reducing thefinancial costs of ATCS selection. A third importantresearch requirement is the development of appropri-ate measures of ATCS job performance. What is vali-dated in personnel selection research is the hypothesisthat job performance, or important aspects of job

10

Validation of the FAA A TCS Pre- Training Screen

REFERENCES Ghiselli, E. E., Campbell, J. P., & Zedeck, S. (1981).Measurement theory for the behavioral sciences. San

Ackerman, P. D. (1991). Discussion of selection of air Francisco, CA: Freeman.

traffic controllers: Complexity, requirements, and Guion, R. M. (1992). Personnel assessment, selection,the public interest. In H. Wing & C. A. Manning and placement. In M. D. Dunnette & L M. Hough(Eds.), Selection ofAir Traffic Controllers: Complex- (Eds.), Handbook of industrial and Organizationality, Requirements and Public Interest. (DOT/FAA/ Psycholok: Volume 2 (2d Ed.) (pp. 3 2 7 - 3 9 7 ). PaloAM-91/9). Washington, DC: Federal Aviation Ad- Alto, CA: Consulting Psychologists Press.ministration Office of Aviation Medicine. Human Technology, In,. "991). Cognitie task analys

Aerospace Sciences, Inc. (1991). Air Traffic Control Spe- of en route air traffic controL Model extension andcialit Pre- Training Screen Preliminary Validation: validation. Volume I: Report. (Final report deliveredFinal report. (Final report delivered under contract under Office of Personnel Management contractDTFA01-90-Y-0 1034). Washington, DC: Federal OPM-87-9041). Washington, DC: Federal Avia-Aviation Administration Office of the Deputy Ad- tion Administration Office of Training and Higherministrator. Education.

Aul, J. C. (1991). Employing air traffic controllers. In H. Lautenschlager, G. J., & Mendoza, J. L (1986). A step-Wing & C. A. Manning (Eds.), Selection of Air down hierarchical multiple regression analysis forTraffic Controllers: Complexity, Requirements and examining hypotheses about test bias in prediction.Public Interest. (DOT/FAA/AM-91/9). Washing- Applied Psychological Measurement, 10, 133- 139.ton, DC: Federal Aviation Administration Officeof Aviation Medicine. Manning, C.rA., & Broach, D. (1992). Identfjringab lit7requirements for operators of future automated air

Bittner, A. C. Jr. (1979). Statistical Tests for Diferential traffic control systems. (DOT/FAA/AM-92/26).Stability. Proceedings of the Human Factors Society Washington, DC: Federal Aviation Administra-23rdAnnualMeeting(pp. 541 - 544). Santa Monica, tion Office of Aviation Medicine.CA. Human Factors Society. Mecham, R. C. & McCormick, E. J. (1969). The Rated

Broach, D. (1992). Non-cognitive Predictors of Perfor- Attribute Requirements ofJob Elements in the Posi-mance in Radar-basedAir Traffic Control Training. tion Analysis Questionnaire. (Report No. 1 underPaper presented at the 4th Annual Convention of ONR Contract NONR-1100(28)). West Laayette,the American Psychological Society, San Diego, IN: Purdue University.CA. McCormick, E. J., Jeanneret, P. R., & Mecham, R. C.

Broach, D., &Aul,J. C. (1993). Analysis ofthe air traffic (1972), A study of job characteristics and job di-control specialit (ATCS) occupation Wuing the Posi- mensions as based on the Position Analysis Ques-tion Analysis Questionnaire (PAQ). Unpublished tionnaire (PAQ). Journal ofApplied Psycholog, 56,manuscript. 347 - 368.

Collins, W. E., Nye, L G., & Manning, C. A. (1992). McCormick, E. J., Mecham, R. C., & Jeanneret, P. R.Poststrike air traffic control trainees: Biodemo- (1977). Technical Manualfor the Position Analysisgraphic predictors ofsuccess in selection and screen- Questionnaire (System II). Logan, UT: PAQ Ser-ing. International Journal ofAviation Psycholo, 2, vices, Inc.213 -223. Nye, L. G., & Collins, W. E. (1991). Some personality

Della Rocco, P. S., Manning, C. A., & Wing, H. (1990). characteristics ofair traffic control specialist trainees:Selection of controllers for automated systems: Appli- Interactions of personality and aptitude test scorescations from current research. (DOT/FAAIAM-901 with FAA Academy success and career expectations.13). Washington, DC: Federal Aviation Admin- (DOTIFAAJAM-91/8). Washington, DC: Federalistration Office of Aviation Medicine. Aviation Administration Office of Aviation Medi-

Federal Aviation Administration. (1985). NationalAir cine.

Traffic Training Tracking System. (FAA order Rodriquez, E., Narayan, K., & O'Donnell, R. D. (1986).3120.22A). Washington, DC: Federal Aviation An Approach to Testingfor NeurobehavioralEffcts ofAdministration Office of Air Traffic Program Man- Organic Solvents. Los Angeles, CA: Atlantic Richfieldagement. Co.

11

Validstion of tke FAA A TCS Pre- TrainiWg Sreem

Schroeder, D. J., Broach, D., & Young, W. C. (1992). Siegel, A. I. (1983). The miniature job training andContributions of Personality Measures to Predicting evaluation approach: Additional findings. Person-Success of Trainees in the Air Traffic Control Special- nel Psychology, 36; 41 - 56.istNonradar Screen Program. Paper presented at the1992 Annual Meeting of the Aerospace Medical U.S. Congress. (January 20, 1976). House Committee on1992As aion, Miamig of thGovernment Operations Recommendations on AirAssociation, Miami, FL. Traffic Control Training. Washington, DC: Au-

Sells, S. B., Dailey, J. T., & Pickrel, E. W. (Eds.) (1984). thor.Selection of air traffic controllers. (DOT/FAA/AM- Weltin, M., Broach, D., Goldbach, K., & O'Donnell, R.84/2). Washington, DC. Federal Aviation Admin- (1992). Conennent criterion-related validation ofistration Office of Aviation Medicine, the air traffic control specialist Pre- Training Screen.

Shingledecker, C. A. (1984). A task battery for applied (Final report delivered under contract DTFAO1-human performance assessment research. 89-Y-01019). Washington, DC: Federal Aviation(AFAMRL-TR-84-071). Wright-Patterson AFB, Administration Office of the Deputy Administra-OH: Armstrong Aerospace Medical Research Labo- tor.ratory. Wickens, C. D. (1984). Engineering psychology. Colum-

Siegel, A. 1. (1978). Miniature job training and evalua- bus, OH: Bobbs-Merrill Inc.tion as a selection/classification device. HumanFactors. 20, 189 - 200.

12 *U.S. GOVERNMENT PRINTING OMCe: 1"4 - S~3InUW