25
Contractor selection innovation: examination of two decades’ published research Gary Holt Department of Civil and Building Engineering, Loughborough University, Loughborough, UK Abstract Purpose – The “problem” of selecting a contractor has attracted significant academic research endeavour over the last two decades. The principal aim here is to examine that research via published academic outputs for the period circa 1990-2009. Design/methodology/approach – A sample of published contractor selection (CSn) research is critically appraised. Aspects highlighted include: stated aims and research justification; methodological approaches employed; research tools used; and products of CSn research. Findings – Main research foci are observed as: modelling the CSn process; studying selection criteria; and “interrogation” of existing CSn systems. Foci justifiers are linked mainly to the “importance” and “difficulties” of CSn decision making. Deterministic modelling of CSn is the favoured methodological approach, followed by documentary synthesis then questionnaire surveys. Preferred research tools are found to be system interrogation, rank order analysis and Likert scale/importance indices, with hypothesis testing and “other” methods used less so. Almost two-thirds of research products are CSn models, with derived or proffered processes, and knowledge relating to CSn criteria, between them representing approximately the remaining third of output. Research limitations/implications – It is suggested that many of the CSn models exhibit as much complexity as the original “problem” they sought to resolve, while the reliability and longevity of suggested “cocktails” of CSn criteria (in practice), might be questioned. A call for future research products to more closely consider end-user impact and potential for “take-up” by industry is made. An empirical follow-on study to assess (inter alia) practitioner use and “value” of CSn research is proposed. Practical implications – The paper signals a possible need for greater industrial engagement in the research domain. Originality/value – The findings are novel to this paper. Keywords Procurement, Contractor workers, Tendering, Subcontractoring, Clients Paper type Literature review Introduction The construction industry is recognised for many distinct characteristics and in a procurement context, its separation of design from production and resulting transient fragmentation, arguably reigns supreme (Langford and Male, 2001, Chapter 2; Walker, 2002, Chapter 1; Loosemore et al., 2003, p. 2; Cooke and Williams, 2004, p. 2). Resultantly, selecting the most appropriate contractor for the purpose of their undertaking the production function[1] has for decades, been the subject of research, debate and published guidance. The current issue and full text archive of this journal is available at www.emeraldinsight.com/1471-4175.htm The combined constructive comments of the anonymous reviewers are acknowledged, for helping develop and improve the paper. CI 10,3 304 Received 16 April 2009 Accepted 15 October 2009 Construction Innovation Vol. 10 No. 3, 2010 pp. 304-328 q Emerald Group Publishing Limited 1471-4175 DOI 10.1108/14714171011060097

Contractor Selection

  • Upload
    cepehr

  • View
    125

  • Download
    2

Embed Size (px)

Citation preview

Page 1: Contractor Selection

Contractor selection innovation:examination of two decades’

published researchGary Holt

Department of Civil and Building Engineering, Loughborough University,Loughborough, UK

Abstract

Purpose – The “problem” of selecting a contractor has attracted significant academic researchendeavour over the last two decades. The principal aim here is to examine that research via publishedacademic outputs for the period circa 1990-2009.

Design/methodology/approach – A sample of published contractor selection (CSn) research iscritically appraised. Aspects highlighted include: stated aims and research justification;methodological approaches employed; research tools used; and products of CSn research.

Findings – Main research foci are observed as: modelling the CSn process; studying selection criteria;and “interrogation” of existing CSn systems. Foci justifiers are linked mainly to the “importance” and“difficulties” of CSn decision making. Deterministic modelling of CSn is the favoured methodologicalapproach, followed by documentary synthesis then questionnaire surveys. Preferred research tools arefound to be system interrogation, rank order analysis and Likert scale/importance indices, withhypothesis testing and “other” methods used less so. Almost two-thirds of research products are CSnmodels, with derived or proffered processes, and knowledge relating to CSn criteria, between themrepresenting approximately the remaining third of output.

Research limitations/implications – It is suggested that many of the CSn models exhibit as muchcomplexity as the original “problem” they sought to resolve, while the reliability and longevity ofsuggested “cocktails” of CSn criteria (in practice), might be questioned. A call for future researchproducts to more closely consider end-user impact and potential for “take-up” by industry is made.An empirical follow-on study to assess (inter alia) practitioner use and “value” of CSn research isproposed.

Practical implications – The paper signals a possible need for greater industrial engagement in theresearch domain.

Originality/value – The findings are novel to this paper.

Keywords Procurement, Contractor workers, Tendering, Subcontractoring, Clients

Paper type Literature review

IntroductionThe construction industry is recognised for many distinct characteristics and in aprocurement context, its separation of design from production and resulting transientfragmentation, arguably reigns supreme (Langford and Male, 2001, Chapter 2; Walker,2002, Chapter 1; Loosemore et al., 2003, p. 2; Cooke and Williams, 2004, p. 2). Resultantly,selecting the most appropriate contractor for the purpose of their undertakingthe production function[1] has for decades, been the subject of research, debate andpublished guidance.

The current issue and full text archive of this journal is available at

www.emeraldinsight.com/1471-4175.htm

The combined constructive comments of the anonymous reviewers are acknowledged, forhelping develop and improve the paper.

CI10,3

304

Received 16 April 2009Accepted 15 October 2009

Construction InnovationVol. 10 No. 3, 2010pp. 304-328q Emerald Group Publishing Limited1471-4175DOI 10.1108/14714171011060097

Page 2: Contractor Selection

Contractor selection (CSn) normally comprises a “theoretically infinite” group ofcontractors (Holt, 1996) whose attributes (El-Sawalhi et al., 2008) are – often on the basisof past performance (Kadefors, et al., 2007; Holt et al., 1994a) – assessed against clientobjectives or “criteria” (Russell and Skibniewski, 1988; Wong and Holt, 2003).

Effectively, this is the “judging” of contractor(s’) capabilities (Ng and Skitmore, 1999;Marzouk, 2008) but that judgment – often performed as a “snapshot in time” – must beconsidered against the potential for these capabilities to significantly change over shortperiods. In practice, CSn procedures vary significantly (Palaneeswaran andKumaraswamy, 2001; Egemen and Mohamed, 2005) and are typified by: non-linearity(Lam et al., 2005); uncertainty (Watt et al., 2009); subjectivity (Holt et al., 1993; Watt et al.,2009); and volatility/competitiveness (Fong and Choi, 2000). These aspects will beexpanded upon as this study unfolds.

Given the importance of CSn to project performance (Holt, 1998), numerousconstruction industry reports have focused on it (Cooke and Williams, 2004, p. 1.2). Forexample, The Simon Committee (1944) strongly advocated selective rather than opentendering (Masterman, 1996, p. 8), while Emmerson (1962) highlighted the issue ofconstruction design being “removed” from production as referred to above. Shortly afterEmmerson, Banwell (1964) reiterated aspects of The Simon Committee (et al.),advocating what were then “non-conventional” procurement methods such asnegotiation, and a follow-on report from the National Economic Development Office,also supported greater use of serial tendering and negotiated contracts (NEDO, 1967). Inthe title of its report, The Tavistock Institute (Tavistock, 1966), reflected that thebuilding industry was characterised by “[. . .] interdependence and uncertainty”concluding somewhat ironically (given said calls for increased negotiation), that theindustry “thrived” on the latter (Woodward, 1997).

The oft-cited Latham (1993, 1994) reports were quite influential in procurementterms, raising issues of mistrust between stakeholders (Cooke and Williams, 2004, p. 4),whilst among other aspects, promoting greater use of selection based on quality inaddition to price (i.e. value). Egan (1998) subsequently encouraged longer termprocurement relationships in favour of selective methods, as a mechanism to achievingquality and efficiency improvement; while in a second report (Egan, 2003), arguablyrooted the term “integrated supply chains” into procurement dialogue.

Aside from government, the Construction Industry Institute offered explicit CSnguidance (CII, 1988) along with direction on choosing contractors when striving toimprove construction safety (CII, 2003), while the issue of long-term (contractor/client)relationships surfaced yet again, when the European Construction Institute highlighted“value enhancement” via strategic alliances (ECI, 1999). Given the adoption ofinformation technology within construction, it was inevitable that problems (and theiravoidance) of CSn and e-tendering have also been commented upon (CRC, 2006).

Since the middle of the last century, a shift from promoting broadest CSn competitiontowards integrated supply chain mechanisms that encourage mutual benefit,organisational learning, partnerships, profitability and quality – while embracinginformation technology – is witnessed (Li et al., 2000; Love et al., 2002b; Walker andJohannes, 2003; Akintoye and Main, 2007). The emphasis within the literature hasmoved from “lowest cost” to “value” (CIRIA, 1998, 2009) (while the author’s anecdotalevidence suggests that, among smaller contractors at least, “cost” often remains theultimate selection criterion). The “credit crunch” has certainly placed strain on the

Contractorselection

innovation

305

Page 3: Contractor Selection

“literature’s call” for value, with price competition presently fierce[2] (RICS, 2008, p. 3),resulting from unprecedented negative forces acting on international constructionactivity (The Times, 2008; CECA, 2008; CIOB, 2009; ASCE, 2009).

The issue of CSn – and academic research in the field for that matter – has been anevolving one over the last two decades; moulded by reaction to changes in theprocurement environment and in the case of academic CSn research, to advancements inthe science of methodology. Regarding the latter, developing techniques such as neuralnetworks, fuzzy decision making, and web-based technologies have witnessedapplication among a range of construction management problems (Graham et al.,2006; Vries and de and Steins, 2008; Han et al., 2008, respectively). Somewhat expectedlytherefore, examples of their use in CSn research are equally easy to find, for example, seeAlbino and Garavelli (1998), Singh and Tiong (2005) and Palaneeswaran andKumaraswamy (2005), respectively.

Aim and objectives of this studyThis backcloth of “evolvement” brings matters conveniently to the focus of this study,the aim of which was to critically appraise academic research within the field ofconstruction CSn, published over a time window of circa 20 years prior to 2009. In sodoing, such review and concomitant synthesis sought accord with contention of Fellowsand Liu (2003, p. 64) that a literature review must provide, “[. . .] a summary of thestate-of-the-art”.

Objectives related to this aim included particular consideration of:

(1) the foci of that research;

(2) stated research drivers;

(3) favoured methodological approaches;

(4) research tools employed; and

(5) the products of research effort.

That is, the study aspired to capture a temporal transit through the observed window ofliterature, with particular clusters of synthesis, emanating from focus on five areas of thesubject defined by objectives (1)-(5) above. Based on observation of outcomes resultingfrom satisfying these objectives, the contribution of published academic research isconsidered, and future research direction intimated.

The literature review: rationale and methodThe literature review is discussed in terms of its:

. rationale; and

. method.

RationaleLevin (2007, p. 75) asserted that to write a good literature review requires one to ask:“What am I reviewing the literature for?” In the present instance, some answers to thisquestion mirror “uses” of a review included in O’Leary’s (2004, p. 78) list, these being to“inform”, “argue rationale”, and “develop questions”. Notwithstanding that, someregard a review as simply an “essay” (Davies, 2007, p. 39), for this study the “answer” toLevin connects with two particular definitions:

CI10,3

306

Page 4: Contractor Selection

(1) that a review can be, “an expert’s general review of current literature on aparticular topic” (Walliman, 2006, p. 182); and

(2) that it should overview, “[. . .] who the key writers are, [. . .] prevailing theoriesand hypotheses [. . .] what questions are being asked, [. . .] what methods andmethodologies are appropriate [. . .] and show relationships [. . .] so that keythemes emerge” (Emerald, 2009).

Literature reviews have frequently been used – to good effect – in constructionmanagement research (CMR). For example, Sidwell and Budiawan (2001) undertook areview to highlight problems with competitive tendering in relation to contractor-ledinnovation; Wong et al. (2005) reviewed the literature to investigate intelligent buildingresearch and discover three research foci as a basis for summarising future researchdirection; Donohoe (2005) reviewed literature and case law to consider a (then) recentappeal case and its implications for building surveyors; and constraints on informationcommunication technology were identified through the literature by Peansupap andWalker (2006). More recently, common measures for improving constructability havebeen studied via review (Wong et al., 2007). This is but an indicative sample.

MethodThe academic literature, upon which this study focussed, was identified via onlinesearch among selected construction management journals and other research databases(see later). Additional kinds of literature such as trade journals, comments, magazinearticles and theses[3] do exist and are not excluded herein; but academic journal paperswere afforded emphasis, because they represent the most important wealth of literatureavailable (Fellows and Liu, 2003, p. 62). Further, journal papers are (should be)academically “robust” in that most are subject to blind review; a process stated as key toscientific progress (Smith, 1999, p. 9).

Search keywords were limited to “CSn” intending only to identify research directlyrelating to CSn – or to consider this conversely – to avoid contractor research in abroader sense such as that relating to: performance measurement per se (Luu, et al.,2008); specific contractor assessment such as safety performance (Hinze and Gambatese,2003); or contractor metrics such as financial stability (Chan, et al., 2005).

While it is difficult to definitively justify which journals should, or should not beincluded in this kind of (targeted) literature search, those chosen (Table I) wereconsidered to represent contemporary eminence in the field of constructionmanagement, and hence, able to yield a representative sample (note journal acronymsin Table I which are used hereafter). By way of supporting this contention, several ASCEpublications, BRI, CME, CI, ECAM, and some in the “others” category of the table are allincluded in the Australian Business Deans Council list of “top” journals in the field(ABDC, 2009); while a list of journals “most respected” by UK academics – based on ananalysis by Hughes (2009) relating to papers submitted for the 2007 UK ResearchAssessment Exercise (RAE, 2008) – ranked virtually all journals targeted in the presentstudy highly. Notwithstanding this emphasis on journal papers, some conference papersin the field were also identified via the Royal Institution of Chartered Surveyorsconference series (RICS, 2009); the Association of Researchers in ConstructionManagement database (ARCOM, 2009); and the American Society of Civil Engineersdatabase (ASCE, 2008).

Contractorselection

innovation

307

Page 5: Contractor Selection

The decision regarding a paper’s relevance to this study was a subjective onebased entirely on the author’s experiential judgement. This decisional characteristic ofliterature reviews has previously been recognised. For example, Birmingham (2000)suggested that the problem of what to include or otherwise is an, “[. . .] evaluativedecision” based on the researcher’s own judgement of “adequacy”; while Fink (2005, p. 5)confirmed its being a subjective task that might use screening criteria for evaluating apaper’s coverage of a subject, along with its scientific quality. Similarly, the categorisationof papers (regarding, for example, subject focus and methodological standpoint), was aresult of subjective evaluation and harmonized with Cooper’s (1998) suggestion that “[. . .]literature reviews can focus on research, research methods, theories, applications [. . .] toidentify the central issues in a field” while also noting that, “identifying related work[assists in helping] to rationalise the work and identify a focus” (Hart, 1998).

Table I summarises the search sources and respective numbers of worksidentified/included from each. The sample does not purport to be exhaustive or adefinitive listing in the field, taking into account the above issue of “inclusion” and thefact that new research is always working through to publication. However, given thescope, depth, and targeted nature of the review, the sample was considered appropriateto meaningful analysis.

Analysis of the literatureThe analysis is focussed under each of the headings: “CSn research focus andjustification” (to observe favoured aspects of study and their motives for attention);“Methodological approaches to CSn research” (to consider methodological predilection);

Source/databaseNo. papers returned

from searchaNo. papers relevant

to studyb

ASCE Databasehttp://pubs.asce.org/research/ 86 24Building and Environmentwww.sciencedirect.com/science/journal/03601323ISSN: 0360-1323 (print) 133 13Building Research and Informationwww.tandf.co.uk/journals/rbri ISSN: 0961-3218 (print) 35 6Construction Management and Economicswww.tandf.co.uk/journals/titles/01446193.htmlISSN: 0144-6193 (print) 121 14Construction Innovationhttp://info.emeraldinsight.com/products/journals/journals.htm?id¼CI ISSN: 1471-4175 (print) 8 2Engineering, Construction and Architectural Managementhttp://info.emeraldinsight.com/products/journals/journals.htm?id¼ecam ISSN: 0969-9988 (print) 164 10International Journal of Project Managementwww.sciencedirect.com/science/journal/02637863ISSN: 0263-7863 (print) 312 12Other, non-specified journals 12Total 93

Notes: aUsing search terms “Contractor þ Selection”; bsubjectively chosen – refer to narrativeTable I.Results of searches

CI10,3

308

Page 6: Contractor Selection

“Research tools used to solve CSn problems” (to identify solutions chosen); and“Products of research” (to highlight the results of this combined effort). Accordingly, adiscussion ensues; salient observations are highlighted; and suggestions for futureresearch direction in the field are proffered.

CSn research focus and justificationThe majority of CSn research has focussed on the selection “problem” as a process, withjustification centred on decisional aspects relating to (inter alia) its issues of difficulty,uncertainty, and subjectivity announced in the introduction. Coupled with this, theimportance of the selection decision is often mentioned. Differing Csn models have beenproffered as solutions thereto.

On justification, Hatush and Skitmore (1998) highlighted the important link betweenCSn and project success in terms of achieving schedule, cost and quality; such linkadditionally highlighted by many others (Russell and Jaselskis, 1992). Kumaraswamy(1996) stated that, “The right choice of contractor is crucial to success”; further examplesconfirming that CSn, “[. . .] is important to ensure the success of projects” (Bubshait andAl-Gobali, 1996); “[. . .] is a decisive event for project success” (Alarcon and Mourgues,2002); “[. . .] the project’s success level is largely dependent on it” (Li, et al., 2005); and(selecting a capable bidder) is crucial to, “ensure the success of construction projects”(Li, et al., 2007). Fong and Choi (2000) meanwhile, identified that CSn is one of the mostimportant tasks facing a client who wishes to achieve successful project outcomes. AsKumaraswamy et al. (2000) summed it up, “[. . .] smarter selection does in fact matter”.

The issue of importance is closely linked to the decisional difficulties of CSn andhence, addressing these difficulties, has been the ambition of many researchers. Hatushand Skitmore (1997) confirmed, “One of the most difficult decisions taken by the client[. . .] is selecting the contractor”; while Albino and Garavelli (1998) referred to thecomplexity of the decision in terms of, “analysis and judgement”. El-Sawalhi et al. (2007)identified with the need to, “provide uncertain, incomplete, or imprecise assessments dueto lack of information” in harmony with Watt et al. (2009), who stated that the decision is,“[. . .] plagued with uncertainties”. Similar rhetoric includes, “[. . .] the task ischallenging” (Fong and Choi, 2000); and its description as, “[. . .] a complicatedtwo-group non-linear classification problem” by Lam et al. (2005). Elsewhere, Pongpengand Liston (2003) recognised that that whether in public or private sector arenas,multiple decision makers are involved.

In contrast, others have justified (and focused) their work differently. Khosrowshahi(1999) for instance, in developing a CSn model for the public sector, suggested thatprequalification should be an, “[. . .] important issue for contractors who seek to obtainwork” and that they might therefore direct [more] attention to their qualifying attributes.The need to aggregate both quantitative and linguistic selection data was thejustification of a model suggested by Sonmez et al. (2001); while Phillips et al. (2007)described the need to achieve best value in delivering savings identified in the Gershon(2004) review, as the basis for their study. Juan et al. (2009) addressed the, “[. . .] fewefforts [previously] focused on CSn in refurbishment projects”; while Holt and Edwards(2005) justified their treatise, by highlighting problems of selecting from apreponderance of “cowboy” firms in the domestic building sector.

Some research has substantiated the targeting of a country-specific CSn setting orproblem. Examples include Lai et al. (2004) who studied Chinese bid evaluation;

Contractorselection

innovation

309

Page 7: Contractor Selection

Ferguson et al. (1995) focussing on Dutch tendering practice; Kumaraswamy (1996) whopresented a Hong Kong perspective; and observation of European contract award byLambropoulos (2007) and Topcu (2004). Other, setting-specific justification includes thework of Yiu et al. (2002) that addressed the “dearth of research” aimed specifically atbuilding and maintenance CSn; Aziz (2008) whose study related to highway pavementproject contractors; a study aimed at apartment block maintenance contractors byZavadskas and Vilutiene (2006); and Juan et al. (2009) whose focus was housingrefurbishment contractors. Finally, regarding research focus, it was perhaps inevitablethat the broader aspect of research into main CSn would eventually be applied tosubcontractors; and recent outputs confirm this. For example, see Kumaraswamyand Mathews (2000), Ng and Tang (2008), Mbachu (2008), Ng et al. (2008) andArslan et al. (2008).

By categorising each paper making up the sample (n ¼ 93) into one principalresearch focus, the results shown in Table II were achieved. Two caveats areacknowledged regarding this categorisation:

(1) it was a subjective, “evaluative” process (as described in the introduction); and

(2) many papers spanned more than one research focus.

Nonetheless, the exercise does provide thematic insight (Emerald, 2009) that iselucidated within the “Discussion” later.

Methodological approaches to CSn researchAnalysis of methodological approaches was based on the chosen categories of:

. statistical/deterministic modelling;

. literature/documentary analysis;

. surveys;

. other (non-deterministic) forms of statistical enquiry; and

. interviewing.

The sixth classification designated “Other” (Figure 1), represented the “note” offered bySkitmore and Mills (1999), which did not fall into a methodological category.

Statistical/deterministic modelling generally sets out to mathematically model aprocess, such that the derived model can replicate a previous event as a means ofpredicting a future one (Fellows and Liu, 2003, p. 76); and being different from stochasticmodelling, which draws on probability (Fellows and Liu, 2003). Although, as elsewherehighlighted, “[. . .] models whose solutions include chaotic dynamics are a link betweendeterministic and stochastic [types]” (Metcalfe, 1996, p. 208). In the present context,typical deterministic examples tend to comprise an algorithm that given specific inputs,will yield contractor classification output and/or optimum choice.

Statistical/deterministic modelling of CSn was the most favoured methodologicalapproach among the sample (almost 40 per cent of papers); being represented by a rangeof selection models. For instance, Tam and Harris (1996) used a discriminant function toclassify contractors into performance groups; Wong et al. (2003) used a similar functionand logistic regression to achieve contractor classification; Lam et al. (2005) employedprincipal component analysis during prequalification decision making to, “alleviatemulticollinearity and largely reduce the dimensionality of the prequalification data set”;

CI10,3

310

Page 8: Contractor Selection

Res

earc

hfo

cus

Des

crip

tion

Ex

amp

lep

aper

sN

um

ber

%a

Sel

ecti

onm

odel

Mod

elem

bra

cin

gth

eco

ntr

acto

rev

alu

atio

nan

dse

lect

ion

pro

cess

toy

ield

opti

mu

mre

sult

Pon

gp

eng

and

Lis

ton

(200

3)an

dA

bd

elra

hm

anetal.

(200

8)39

42

Ass

essm

ent

crit

eria

Inv

esti

gat

ion

ofas

sess

men

tcr

iter

ia(e

.g.

for

pre

qu

alifi

cati

on,p

roje

ctsp

ecifi

cev

alu

atio

n,o

rte

nd

erev

alu

atio

n)

Ru

ssel

lan

dS

kib

nie

wsk

i(19

88),

Waa

raan

dB

roch

ner

(200

6)an

dH

oltet

al.

(199

4b)

1617

Su

rvey

ofp

ract

ice

Su

rvey

offo

rex

amp

le,

pre

sen

tse

lect

ion

orev

alu

atio

nm

eth

ods

Fer

gu

sonet

al.

(199

5)an

dP

alan

eesw

aran

and

Ku

mar

asw

amy

(200

1)13

14

Dec

isio

nsu

pp

ort

Mod

elto

emb

race

sele

ctio

nor

ap

art

ther

eof,

exp

lici

tly

stat

edas

dec

isio

nsu

pp

ort

or‘a

id’

Lam

etal.

(200

0)an

dS

hen

etal.

(200

3)8

9

Con

trac

tor

per

form

ance

Con

trac

tor

eval

uat

ion

dir

ectl

yre

late

dto

pas

tor

pre

dic

ted

con

trac

tor

per

form

ance

Nget

al.

(200

5)an

dM

bac

hu

(200

8)7

8

Op

inio

nsu

rvey

Su

rvey

ofop

inio

n(f

orex

amp

le,

ofcl

ien

ts,

con

trac

tors

orco

nsu

ltan

ts)

Jen

nin

gs

and

Hol

t(1

998)

and

Sin

gh

and

Tio

ng

(200

6)4

4

Cas

est

ud

yIn

dep

than

aly

sis

ofa

pas

tca

seor

anex

isti

ng

/p

rop

osed

syst

emH

olt

(199

7)an

dL

aiet

al.

(200

4)3

3

Oth

erF

oran

yp

aper

sn

otfi

ttin

gin

toth

eab

ove

foci

Sk

itm

ore

and

Mil

ls(1

999)

33

Tot

als

9310

0

Note:

aP

erce

nta

ge

ofsa

mp

le,

rou

nd

ed

Table II.Analysis of

“research focus”

Contractorselection

innovation

311

Page 9: Contractor Selection

while Bendana et al. (2008) presented a fuzzy-logic-based selection method.Deterministic techniques that embrace CSn variables represented on differentmeasurement scales are appropriate to the multivariate character of CSn input data.

Literature/documentary synthesis, is taken here to include examination of all kinds ofexisting (written) knowledge or data, as a means of conceptualising, developing theory,or presenting propositions (Davies, 2007, p. 3). Almost one-third of the sample employedthis methodology, such as within the review of research on transaction costs related toCSn method (Lingard et al., 1998); the comparative overview of documented knowledgeof practices in design/build (sic) CSn (Palaneeswaran and Kumaraswamy, 2000b); thebenchmarking of public sector selection practices (Palaneeswaran and Kumaraswamy,2000a); and an examination of the management and tender evaluation literature byWatt et al. (2009). Many texts expand the virtues of this “interrogative” researchmethodology, see for example, Silverman (2001) and May (2001, Chapter 8).

Within the broader field of CMR, questionnaires are favoured, having even been usedto study the journals within which the results of these kinds of surveys are oftenpublished (Chau, 1997). But, such persistent usage has been questioned of late.For instance, McCaffer (2008) asked, “How do you move from analyses of questionnairereturns to making a real change in industry’s practices?” Surveys were used in 22 per centof the papers making up the sample, such as to rate the importance of a list of35 prequalification criteria (Ng et al., 1999); to study the perceived importance of factorsaffecting choice of contractors in the UK (Holt et al., 1994c); and to compare use ofproject-specific vis-a-vis “value-specific” tender selection criteria (Wong et al., 2000).

Figure 1.Summary ofmethodologicalapproaches employed

36

27

20

6

3 10

5

10

15

20

25

30

35

40

Stat./detmnst.model

Lit./doc.synthesis

Postalsurvey

Statisticalinquiry

Interviews Other(note)

No. Percent

CI10,3

312

Page 10: Contractor Selection

Other, non-deterministic forms of statistical enquiry such as correlation analysis(Colman and Pulford, 2006); hypothesis testing (Colman and Pulford, 2006, sections 7-8);and ordinal or rank analysis (Meddis, 1984) were used in six studies (Chinyio et al., 1998;Holt, 1998). Interviewing and/or case studies were employed among three studies,including Hatush and Skitmore (1997) who developed a questionnaire for structuredinterviews to investigate project success factors; and Palaneeswaran andKumaraswamy (2001) who combined “correspondence” with interviews to seekinternational expert opinion on CSn.

Research tools used to solve CSn problemsSpecific research tools used among the sample were examined (in respect of16 classifications), to look in more detail at the mechanisms by which research problemswere resolved and the “popularity” of these. These classifications and respective usagewere ordered by rank based on the number of times each was observed among thesample (see column 1, Table III).

The number of “usage observations” (column 3, total ¼ 113) is greater than thesample (n ¼ 93) because some researchers used more than one classification of tool.To offer some examples, Assaf et al. (1996) used rank order analysis with regressionanalysis; while Chau et al. (2003) combined Likert scale/importance ranking withmulti-attribute utility theory. Often, the combining of several methods in this wayunderpins a triangulated CMR methodology (Love, et al., 2002a; Edwards and Holt, 2010).

Here, system interrogation was defined as any method of inquiry that observedqualitative data and/or observation of existing systems, and was the favoured researchtool (40 observations). Examples include Holt et al.’s (1995b) review of UK selectionpractice; a cross-sectional review of CSn methodologies embedded in the literature(Holt, 1998); and the study of, “[. . .] current literature and actual experience refinedthrough a review of public-sector owners’ selection procedures” by Al-Reshaid andKartam (2005).

Forms of rank order analysis were second-favourite tool. Zavadskas and Vilutiene(2006) presented values of qualitative performance criteria obtained from a survey, andin part, rank analysed these using a quadrant analysis matrix. Marzouk (2008) applied asuperiority and inferiority ranking model to CSn, while several other researchesdemonstrated rank analysis of one form or another (Hatush and Skitmore, 1997;Wong et al., 2000; Chau et al., 2003). The latter two studies (as did several others)employed Likert/importance ranks as a means of abstracting subjective personalopinion (Likert, 1932; Oppenheim, 2000). This was the third favoured research tool and(given that such analysis is often used to analyse survey returns), mirrors the earlieridentified popularity of questionnaires.

Other tools used include: muilti-attribute analysis (Assaf and Jannadi, 1994) andmulti-attribute utility theory (Holt et al., 1994d, 1995a) (especially among “earlier”studies in the field), along with several interpretations of fuzzy theory (Singh andTiong, 2005). Artificial neural networks (Lam et al., 2000), mutivariate discriminantanalysis (Wong and Holt, 2003), analytical hierarchy process (Abdelrahman et al., 2008)and cluster analysis (Chinyio et al., 1998) were all observed four times each. Slightlylesser use (three observations) was noted for regression analysis (Tam and Harris, 1996),variance analysis (Hatush and Skitmore, 1997), correlation analysis (Zavadskas andVilutiene, 2006) and evidential or case-based reasoning (Juan, 2009).

Contractorselection

innovation

313

Page 11: Contractor Selection

Ran

kb

yu

seR

esea

rch

tool

No.

ofob

serv

atio

nsa

Per

cen

tag

eof

obse

rvat

ion

sbE

xam

ple

pu

bli

cati

ons

1S

yst

emin

terr

ogat

ion

4035

.4H

olt

(199

8)an

dA

l-R

esh

aid

and

Kar

tam

(200

5)2

Oth

erra

nk

anal

ysi

s11

9.7

Ass

afet

al.

(199

6)an

dW

onget

al.

(200

1)3

Lik

ert/

imp

orta

nce

ran

ks

108.

8W

onget

al.

(200

0)an

dC

hau

etal.

(200

3)¼

4M

AA

65.

3A

ssaf

and

Jan

nad

i(1

994)

and

Hol

tet

al.

(199

4d)

¼4

MA

UT

65.

3H

oltet

al.

(199

5a)

and

Hol

tet

al.

(199

4e)

6F

uzz

yth

eory

54.

4B

end

anaet

al.

(200

8)an

dJu

anet

al.

(200

9)¼

7A

NN

43.

5K

hos

row

shah

i(1

999)

and

Lam

etal.

(200

0)¼

7M

DA

43.

5T

aman

dH

arri

s(1

996)

and

Nget

al.

(199

9)¼

7A

HP

43.

5M

ust

afa

and

Ry

an(1

990)

and

Mah

diet

al.

(200

2)¼

7C

lust

eran

aly

sis

43.

5H

olt

(199

7)an

dC

hin

yio

etal.

(199

8)¼

11R

egre

ssio

nan

aly

sis

32.

7W

ong

(200

4)an

dA

ziz

(200

8)¼

11V

aria

nce

anal

ysi

s3

2.7

Hat

ush

and

Sk

itm

ore

(199

7)an

dW

onget

al.

(200

3)¼

13C

orel

atio

nan

aly

sis

21.

8Z

avad

skas

and

Vil

uti

ene

(200

6)¼

13E

vid

enti

al/C

Bre

ason

ing

21.

8Ju

an(2

009)

¼13

Hy

pot

hes

iste

stin

g2

1.8

Ng

and

Tan

g(2

008)

16O

ther

1(£

7in

stan

ces)

c6.

2D

arv

ishet

al.

(200

8)an

dIp

etal.

(200

4)T

otal

s11

310

0

Notes:

aN

um

ber

ofob

serv

atio

ns

dif

fers

from

sam

ple

size

du

eto

som

ep

aper

su

tili

sin

gm

ore

than

one

rese

arch

tool

;bm

inor

erro

rin

tota

ld

ue

toro

un

din

g;

c oth

er,o

ne

obse

rved

inst

ance

each

ofan

aly

tica

ln

etw

ork

pro

cess

,pri

nci

pal

com

pon

ent

anal

ysi

s,fa

ctor

anal

ysi

s,co

mb

ined

mat

rix

det

erm

inat

ion

,gen

etic

neu

ral

net

wor

k,

gra

ph

theo

ryan

dm

atri

xm

eth

ods,

bra

nch

and

bou

nd

alg

orit

hm

;k

ey:

MA

A,

mu

lti-

attr

ibu

tean

aly

sis;

MA

UT

,m

ult

i-at

trib

ute

uti

lity

theo

ry;

AN

N,

arti

fici

aln

eura

ln

etw

ork

;M

DA

,m

ult

ivar

iate

dis

crim

inan

tan

aly

sis;

AH

P,

anal

yti

cal

hie

rarc

hy

pro

cess

;C

B,

case

-bas

ed(r

easo

nin

g)

Table III.Analysis of research tools

CI10,3

314

Page 12: Contractor Selection

Seven papers used “novel” tools in this setting, viz.: analytic network process (Cheng andLi, 2004), principal component analysis (Lam et al., 2005), factor analysis (Ng and Tang,2008), combined radix determination (Wang et al., 2007), genetic-neural network(El-Sawalhi et al., 2007), graph theory and matrix methods (Darvish et al., 2008), andbranch and bound algorithm (Ip et al., 2004).

Products of researchFigure 2 shows that the research products of the sample may be broadly grouped intofour categories, the largest of which (63 per cent) comprised CSn models of one form oranother. Of products presented, 17 per cent derived, or proffered, processes (derivedrepresenting or conceptualising an existing process; proffered suggesting animprovement or variation of existing process).

Suggested groupings of selection criteria for given CSn scenarios and/or developedimportance metrics attached to these, evolved as a result of 15 per cent of the studiesundertaken; with the remaining four per cent of studies comprising the “miscellaneous”grouping whose products did not fit with these three former classifications of output.

By considering authors’ own descriptions of the 59 CSn models contained within thesample (Figure 2), emphases of intended use were determined:

. 26 models were described as intended for contractor (or sub-contractor, or tender)“selection” ( Jaselskis and Russell, 1992; Fong and Choi, 2000; Kashiwagi andByfield, 2002; Cheng and Li, 2004; Arslan et al., 2008);

. eight were related to assessing or predicting contractor (or sub-contractor)“performance” (Holt et al., 1994a; Tam and Harris, 1996; Yasamis et al., 2002;El–Sawalhi et al., 2008);

. seven were related to “prequalification” (Russell and Skibniewski, 1990;Russell et al., 1990; Assaf and Jannadi, 1994; Lam et al., 2005; Palaneeswaran andKumaraswamy, 2005; El-Sawalhi et al., 2007);

Figure 2.Broad classifications of

research products

59

1614

40

10

20

30

40

50

60

Model Process Criteria Misc.

No. Percent

Contractorselection

innovation

315

Page 13: Contractor Selection

. six addressed tender or bid “evaluation” (Mustafa and Ryan, 1990; Wong et al.,2001; Pongpeng and Liston, 2003; Watt et al., 2009); and

. four were designed for contractor/client “classification” (Holt, 1996; Chinyio et al.,1998; Wong and Holt, 2003; Wong et al., 2003).

The remainder had specific intended application such as for contractor “rating” (Albinoand Garavelli, 1998), contractor “ranking” (Darvish et al., 2008), or “assessingcompetitiveness” (Shen et al., 2003, 2006). See also Lo et al. (1998) in this respect.

Process products have contributed to a range of issues such as for betterunderstanding of: practices in the public sector (Palaneeswaran and Kumaraswamy,2000a); geographically specific CSn tradition (Kumaraswamy, 1996); procurementspecific synthesis (Gransberg and Molenaar, 2003); and contrast between CSnstakeholders (Egemen and Mohamed, 2005). Equal diversity is evident in criteriaproducts, including guidance on prequalification criteria (Hatush and Skitmore, 1997);with respect to cost-benefit (Ng and Skitmore, 2001); and reporting practitionerpreference (Singh and Tiong, 2006).

DiscussionFigure 3 shows graphically the above findings of the review and may becross-referenced with the following discussion.

CSn importanceThe significance of selecting the most appropriate contractor for a given selectionsetting is of exceptional importance, largely, because of the positive “association”between employing a good contractor and achieving project aspirations. (Or itsantipode the “association” between employing a poor contractor and likelihood of

Figure 3.Model of CSn research

Observed research foci:

1. Selection models2. Assessment criteria3. Survey of practice4. Decision support5. Contractor performance6. Opinion survey7. Case study8. Other

Observed research justification:1. Importance of selection decision2. Difficulties of selection decision3. Other, selection setting specific challenges

Observed methodologicaldesign:

1. Deterministic models2. Literature and documentary synthesis3. Postal surveys4. Statistical enquiry5. Interviews

Observed research tools:1. System interrogation2. Rank analysis3. Likert / importance data4. MAA5. MAUT6. Fuzzy theory7. ANN8. MDA9. AHP

Observed researchproducts

Misc.

Criteria

Process

Models C.2010

Future research products?

Closer dialogue with industry andpractitioners?

More use of face-to-face research?

Greater end-user impact and take-up?

Web based technologies – transitional interface betweenacademic products and industry?

Interrelationships

C.1990 Increasing complexity

Existing: mapping, formalisation, understanding

Proffered: improvements, variations on existing

Relevance to selection setting or scenario

Importance:ranks, indices to selection setting

Knowledge advancement in specific CSn contexts

Inte

rrel

atio

nshi

ps

10. Cluster analysis11. others…

CI10,3

316

Page 14: Contractor Selection

unsatisfactory project outturn). Such interrelationship, between an event that (normally)occurs at the outset of a project and one representing its conclusion, demonstrates CSn’srelevance in the overall context of construction “product” procurement – and helpsexplain why “importance” is a significant research justifier in this field.

Uniqueness of the CSn environmentThe uniqueness of each construction product is matched by the (resulting) individualityof its CSn decisional parameters (such as CSn criteria to be applied). Accordingly,appropriate CSn criteria and their degrees of importance will fluctuate; both as afunction of “objective” factors related to the project (such as time, quality, and valuetargets), along with more “subjective” counterparts, such as decision makers’predilection, disposition and perceived utility of (those) criteria.

To this, uniqueness may be added the dynamic state within which contractorsoperate and within which selection decisions are made – impacted for example – byinconstant macroeconomic, political, and commercial ambience. These characteristics,intuitively, explain the focus of much research observed in this study being the CSnprocess itself, and equally, why “importance” of that process is an oft-cited principalresearch justifier.

Research foci and justifiersThe uniqueness of construction “product procurement” however, demonstrates whysome research has focussed on (and been justified by), distinctive CSn settings – setapart for example, by type of project, form of contractual arrangement, geographicalcharacteristics, or some other peculiarity. This also helps illuminate the decision byseveral researchers, to study selection criteria for given selection scenarios. To seek forexample, their applicability and concomitant importance levels which equally – even forthe same criteria – will differ between these varying selection environments. A caveathere however, relates to the general applicability of any CSn research outputs, acrossdiffering selection settings and the “dynamic state” referred to above.

Methodological considerationsThe methodological approaches observed lean toward deterministic modelling,reflecting the leading desire to model the selection process per se. That is, todetermine optimal output (contractor identification) given numerous multivariatedecisional input parameters. The predominance of literature and other documentaryanalysis underlining methodological design (the basis of almost one-third of the sample),corresponds with the third most popular research focus, of surveying existing practice.The former “desire” can often, in part be informed by results of the latter approaches.

Postal surveys were the research tool of choice among one-quarter of the sample,whereas interviews only represented 4 per cent. It seems that in line with earlier calls for,“[. . .] closer dialogue with practitioners and more demonstrations of application [. . .]”within CMR generally (McCaffer, 2008); face-to-face research “tools” appearunder-utilised in this branch of the field also. This is somewhat surprising in thatgiven a desire to produce optimum selection models – that by definition are intended forpractitioner use – one might have expected to witness more discourse with “end-users”if these particular products of research are to be (of maximum utility to and ultimately)adopted by them.

Contractorselection

innovation

317

Page 15: Contractor Selection

Favoured research tool was system interrogation of different kinds (35 per cent ofresearch tool observation on the sample), relating to the above recurring research focusof surveying existing practice. The use of rank analysis (10 per cent of observations)relates mainly to those studies that investigated and subsequently positioned in variousways, selection criteria. Predominant remaining research tools were statistical oralgorithmic in character, being applied principally to said modelling aspirations andother forms of interrogating data, accrued by way of the 25 per cent of studies employingquestionnaire and interview surveys.

General observationsOne general observation is that earlier studies tended toward more “simpler” modellingtechniques; predominantly of an additive multi-attribute nature, and sometimesembracing utility theory. Later models have employed “newer” research tools such asartificial neural networks and fuzzy theory. A few recent studies meanwhile, havedemonstrated use of tools such as “Combined Matrix Determination” and “GraphTheory and Matrix Models”. Referring to the earlier statement concerning “end-user”take-up, then given complexity of some models, one might speculate as to how“acceptable” these are to practitioners unless presented in a “user-friendly” – forinstance, “ready-to-use” computer package or web-based form (Ng et al., 2003).

That is, algorithmic solutions alone require some form of transitional interface bywhich non-academic users can employ them in practice. Of note on this issue, White(2005) suggested that with respect to tendering procedures in the public sector, anenvironment exists that, “[. . .] mandates procedures which make the implementation ofsuperior methods unlikely” – so one might reasonably ask, “How many of these “new”models are used in public sector practice?” The issue of industry take-up, or “impactvalue” of CSn research products, is perhaps best answered definitively by survey of(dialogue with) practitioners, and points to future research direction as a follow-on to thepresent study.

Other research products of the sample include formulation of existing, and profferingof proposed CSn systems (18 per cent of observations), along with knowledge of selectioncriteria (15 per cent). If the dialogue with practitioners mentioned above is performed, itwould be valuable to also assess end-user adoption of these suggested “cocktails” ofcriteria, to identify degree of linkage between what academia suggests should impactcontractor assessment decisions, and to what extent practitioners are actually aware of,or employ such knowledge in practice.

ConclusionMuch academic research has targeted CSn over the last two decades. The focus of thishas, predominantly, been the decisional processes involved in CSn and decisional criteriaapplied thereto, with principal research “drivers” often cited as being to optimally modelthat complex decisional process, linked to the importance of the decisional task.Nonetheless, where equally complex statistical or algorithmic CSn “solutions” evolvefrom that research (as often observed among the sample), then paradoxically, thesesolutions may simply be replacing one “demanding” state with another.

Similarly, one may question the reliability or relevance of cocktails of “optimal”selection criteria, given the transient and dynamic environments in which they areapplied. That is, while a set of criteria might be optimal in one setting; will this be the

CI10,3

318

Page 16: Contractor Selection

case in a different setting, or a similar setting at a different point in time? With the benefitof hindsight (i.e. since CSn research began in earnest) there is justification in askingwhether such “recommended” criteria groupings can make the transition; from theiracademic evolvement, to sustained and reliable application in practice. Only dialoguewith end-users could determine the answer.

Statistical and/or deterministic modelling was the most prominent methodologicalapproach observed; with earlier models tending towards (arguably simpler)multi-attribute analysis; while later models have employed (arguably more complex)deterministic techniques, such as artificial neural networks, and fuzzy logic. The abovecomments regarding model complexity and relevance, similarly, apply here too.

Specific research tools employed within CSn methodological research frameworksare equally varied. System interrogation – primarily as a function of qualitative dataanalysis – has frequently been used. This, predominantly subjective, interpretive toolmay be welcomed by those within the broader field of CMR that espouse a shift from(more mechanistic) survey-based interrogative approaches; while simultaneouslyencouraging discourse between researchers and practitioners. Accrual of Likert scaledata via questionnaire surveys and analysis of (often resulting) interval data using rankanalysis methods, were also popular, but might be viewed as more convenient to theresearcher than relevant to the research problem, in light of the above comments.

Given said complexity of some evolving CSn models and an arguably under-utilisationof dialogue-based research methods in this subject domain, a call is made for future CSnresearch to more enthusiastically employ the latter techniques, as a means of:

(1) Assessing the take-up of CSn research products by practitioners. For instance,what products actually filter through to practice? And which aspects of researchoutput can actually demonstrate real end-user impact? Accordingly, the answershere would better inform CSn research design if industry “impact” is the intentionof research.

(2) Identifying exemplars and/or barriers associated with (1). Such practicalknowledge could then be fed back to the CSn research community, informingmethodological designs with a view to maximising research value.

(3) Helping to steer future CSn research direction – in the context of maximising itsend-user impact and potential for industrial take-up.

In sum, CSn research should henceforth be targeted at satisfying, and perhaps its“success” be measured more by, the latter criterion (3). These propositions represent anatural extension of this review of the subject, and will form the basis of a follow-on,empirical study in the field.

Notes

1. Here, the term “production” is used loosely; procurement variants embracing contractordesign for example, are not overlooked.

2. Asserted partly on anecdotal evidence from the author’s own network of constructionprofessionals.

3. Many research theses have addressed the CSn issue but are not considered here, becausemuch of that research ultimately filters down into published papers. A searchable electronicindex of theses is however, available (ITT, 2009).

Contractorselection

innovation

319

Page 17: Contractor Selection

References

ABDC (2009), Australian Business Deans Council – Journal Ratings List, Australian BusinessDeans Council, Sydney, available at: www.abdc.edu.au/download.php?id¼120836,177,1(accessed August 2009).

Abdelrahman, M., Zayed, T. and Elyamany, A. (2008), “Best-value model based on projectspecific characteristics”, Journal of Construction Engineering and Management, Vol. 134No. 3, pp. 179-88.

Akintoye, A. and Main, J. (2007), “Collaborative relationships in construction: the UK contractors’perception”, Engineering, Construction and Architectural Management, Vol. 14 No. 6,pp. 597-617.

Alarcon, L.F. and Mourgues, C. (2002), “Performance modelling for contractor selection”, Journalof Management in Engineering, Vol. 18 No. 2, pp. 52-60.

Albino, V. and Garavelli, C. (1998), “A neural network application to subcontractor rating inconstruction firms”, The International Journal of Project Management, Vol. 16 No. 1,pp. 9-14.

Al-Reshaid, K. and Kartam, N. (2005), “Design-build pre-qualification and tendering approach forpublic projects”, The International Journal of Project Management, Vol. 23, pp. 309-20.

ARCOM (2009), The Association of Researchers in Construction Management Database,The Association of Researchers in Construction Management, Nottingham, available at:www.arcom.ac.uk/abstracts.html (accessed via February 2009).

Arslan, G., Kivrak, S., Birgonul, M.T. and Dikmen, I. (2008), “Improving sub-contractor selectionprocess in construction projects: web-based sub-contractor evaluation system (WEBSES)”,Automation in Construction, Vol. 17, pp. 480-8.

ASCE (2008), ASCE Publications. Research Databases, American Society of Civil Engineers,Reston, VA, available at: http://pubs.asce.org/research/ (accessed November 2008).

ASCE (2009), Civil Engineers Present Guidelines for Economic Stimulus InfrastructureInvestment, The American Society of Civil Engineers, Reston, VA, available at: www.asce.org/pressroom/news/display_press.cfm?uid¼4510 (accessed February 2009).

Assaf, S. and Jannadi, O. (1994), “A multi-criterion decision-making model for contractorprequalification selection”, Building Research and Information, Vol. 22 No. 6, pp. 332-5.

Assaf, S.A., Al-Hammad, A. and Ubaid, A. (1996), “Factors affecting construction contractors’performance”, Building Research and Information, Vol. 24 No. 3, pp. 159-63.

Aziz, A.M.A. (2008), “Minimum performance bounds for evaluating contractors performanceduring construction of highway pavement projects”, Construction Management andEconomics, Vol. 26, pp. 507-29.

Banwell, H. (1964), “The placing and management of contracts for building and civil engineeringwork”, Report Chaired by Sir H. Banwell, Ministry of Public Buildings and Works, HMSO,London.

Bendana, R., del Cano, A. and de la Cruz, M.P. (2008), “Contractor selection: fuzzy controlapproach”, Canadian Journal of Civil Engineering, Vol. 35 No. 5, pp. 473-86.

Birmingham, P. (2000), “Reviewing the literature”, in Wilkinson, D. (Ed.), The Researcher’sToolkit: The Complete Guide to Practitioner Research, Routledge, London.

Bubshait, A.A. and Al-Gobali, K.H. (1996), “Contractor prequalification in Saudi Arabia”, Journalof Management in Engineering, Vol. 12 No. 2, pp. 50-4.

CECA (2008), “CECA surveys show contractors hit hard times”, CECA Communicates. EconomySpecial, Civil Engineering Contractors Association, London, available at: www.ceca.co.uk/NewsDetail.aspx?NewsID¼188&NewsTypeID¼1 (accessed February 2009).

CI10,3

320

Page 18: Contractor Selection

Chan, J.K.W., Tam, C.M. and Cheung, R. (2005), “Monitoring financial health of contractors at theaftermath of the Asian economic turmoil: a case study in Hong Kong”, ConstructionManagement and Economics, Vol. 23 No. 5, pp. 451-8.

Chau, C.K., Sing, W.L. and Leung, T.M. (2003), “An analysis on the HVAC maintenancecontractors selection process”, Building and Environment, Vol. 38, pp. 583-91.

Chau, K.W. (1997), “The ranking of construction management journals”, ConstructionManagement and Economics, Vol. 15 No. 4, pp. 387-98.

Cheng, E.W.L. and Li, H. (2004), “Contractor selection using the analytic network process”,Construction Management and Economics, Vol. 22, pp. 1021-32.

Chinyio, E.A., Olomolaiye, P.O., Kometa, S.T. and Harris, F.C. (1998), “A needs-basedmethodology for classifying construction clients and selecting contractors”, ConstructionManagement and Economics, Vol. 16, pp. 91-8.

CII (1988), “Project materials management primer”, Best Practice Guidance, The ConstructionIndustry Institute, Austin, TX, Document Ref. RS7-2, available at: www.construction-institute.org/scriptcontent/more/7_2_more.cfm (accessed August 2009).

CII (2003), “The owner’s role in construction safety”, The Construction Industry Institute,Document Ref. RR190-11, available at: www.construction-institute.org/scriptcontent/more/rr190_11_more.cfm (accessed August 2009).

CIOB (2009), “The impact of the global financial crisis on the construction industry”, Policy Brief,2009, The Chartered Institute of Building, Ascot, available at: www.ciob.org.uk/about/ciobpolicies (accessed February 2009).

CIRIA (1998), “Selecting contractors by value”, CIRIA, London, Document Ref. SP150.

CIRIA (2009), “Selecting contractors by value update (CON167)”, available at: www.ciria.org/service/current_projects/AM/ContentManagerNet/ContentDisplay.aspx?Section¼current_projects&ContentID¼11900 (accessed April 2009).

Colman, A. and Pulford, B. (2006), A Crash Course in SPSS for Windows, 3rd ed.,Blackwell, Oxford.

Cooke, B. and Williams, P. (2004), Construction Planning, Programming and Control, 2nd ed.,Blackwell, Oxford.

Cooper, H.M. (1998), Synthesising Research: A Guide for Literature Reviews, Sage, New York, NY.

CRC (2006), E-tendering – Security and Legal Issues, The Co-operative Research Centre forConstruction Innovation, Brisbane.

Darvish, M., Yasaei, M. and Saeedi, A. (2008), “Application of the graph theory and matrixmethods to contractor ranking”, The International Journal of Project Management, Vol. 27No. 6, pp. 610-19.

Davies, M.B. (2007), Doing a Successful Research Project, Palgrave Macmillan, Basingstoke.

de Vries, B. and Steins, R.J. (2008), “Assessing working conditions using fuzzy logic”,Automation in Construction, Vol. 17 No. 5, pp. 584-91.

Donohoe, S. (2005), “Building surveyors and adjudication: the implications of Hurst Stores v. M LEurope Property Ltd (2004)”, Structural Survey, Vol. 23 No. 1, pp. 55-62.

ECI (1999), Summaries of ECI Value Enhancement Practices, The European ConstructionInstitute, Loughborough.

Edwards, D.J. and Holt, G.D. (2010), “The case for ‘3D Triangulation’ when applied toconstruction management research”, Construction Innovation: Information, Process,Management, Vol. 10 No. 1, pp. 25-41.

Contractorselection

innovation

321

Page 19: Contractor Selection

Egan, J. (1998), Rethinking Construction, Report of the Construction Task Force to the DeputyPrime Minister, John Prescott, on the scope of improving the quality and efficiency of UKconstruction, Department of Environment Transport and the Regions, London, Chaired bySir J. Egan.

Egan, J. (2003), “Accelerating change”, A report by the Strategic Forum for Construction,Rethinking Construction, London.

Egemen, M. and Mohamed, A. (2005), “Different approaches of clients and consultants tocontractors’ qualification and selection”, Journal of Civil Engineering and Management,Vol. 11 No. 4, pp. 267-76.

El-Sawalhi, N., Eaton, D. and Rustom, R. (2007), “Contractor prequalification model:state-of-the-art”, The International Journal of Project Management, Vol. 25, pp. 465-74.

El-Sawalhi, N., Eaton, D. and Rustom, R. (2008), “Forecasting contractor performance using aneural network and genetic algorithm in a pre-qualification model”, ConstructionInnovation, Vol. 8 No. 4, pp. 280-98.

Emerald (2009), How to Write a Literature Review, Emerald, available at: http://info.emeraldinsight.com/authors/guides/literature.htm (accessed August 2009).

Emmerson, H. (1962), “Survey of problems before the construction industries”, Report Chaired bySir H. Emmerson, Ministry of Works, HMSO, London.

Fellows, R. and Liu, A. (2003), Research Methods for Construction, 2nd ed., Blackwell Science,Oxford.

Ferguson, N.S., Langford, D.A. and Chan, W.M. (1995), “Empirical study of tendering practice ofDutch municipalities for the procurement of civil-engineering contracts”, The InternationalJournal of Project Management, Vol. 13 No. 3, pp. 157-61.

Fink, A.G. (2005), Conducting Research Literature Reviews: From the Internet to Paper, Sage,New York, NY.

Fong, P.S. and Choi, S.K. (2000), “Final contractor selection using the analytical hierarchyprocess”, Construction Management and Economics, Vol. 18, pp. 547-57.

Gershon, S.P. (2004), Releasing Resources to the Front Line. Independent Review of Public SectorEfficiency, HMSO, London, available at: www.hm-treasury.gov.uk/spending_sr04_efficiency.htm (accessed February 2009).

Graham, L.D., Forbes, D.R. and Smith, S.D. (2006), “Modelling the ready mixed concrete deliverysystem with neural networks”, Automation in Construction, Vol. 15 No. 5, pp. 656-63.

Gransberg, D.G. and Molenaar, D. (2003), “A synthesis of design-builder selection methods forpublic infrastructure projects”, Journal of Construction Procurement, Vol. 9 No. 2, pp. 40-51.

Han, S.H., Kim, D.Y., Kim, H. and Jang, W. (2008), “A web-based integrated system forinternational project risk management”, Automation in Construction, Vol. 17 No. 3,pp. 342-56.

Hart, C. (1998), Doing a Literature Review: Releasing the Social Science Research Imagination,Sage, New York, NY.

Hatush, Z. and Skitmore, R.M. (1997), “Criteria for contractor selection”, ConstructionManagement and Economics, Vol. 15, pp. 19-38.

Hatush, Z. and Skitmore, R.M. (1998), “Contractor selection using multicriteria utility theory:an additive model”, Building and Environment, Vol. 33 Nos 2/3, pp. 105-15.

Hinze, J. and Gambatese, J. (2003), “Factors that influence safety performance of specialitycontractors”, Journal of Construction Engineering and Management, Vol. 129 No. 2,pp. 159-64.

CI10,3

322

Page 20: Contractor Selection

Holt, G.D. (1996), “Applying cluster analysis to construction contractor classification”, Buildingand Environment, Vol. 31 No. 6, pp. 557-68.

Holt, G.D. (1997), “Classifying construction contractors: a case study using cluster analysis”,Building Research and Information, Vol. 25 No. 6, pp. 374-82.

Holt, G.D. (1998), “Which contractor selection methodology?”, The International Journal ofProject Management, Vol. 16 No. 3, pp. 153-64.

Holt, G.D. and Edwards, D.J. (2005), “Domestic builder selection in the UK housing repair andmaintenance sector: a critique”, Journal of Construction Research, Vol. 6 No. 1, pp. 123-37.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1993), “A conceptual alternative to currenttendering practice”, Building Research and Information, Vol. 21 No. 3, pp. 167-72.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1994a), “Applying multi-attribute analysis tocontractor selection decisions”, European Journal of Purchasing & Supply Management,Vol. 1 No. 3, pp. 139-48.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1994b), “Evaluating performance potential in theselection of construction contractors”, Engineering, Construction and ArchitecturalManagement, Vol. 1 No. 1, pp. 29-50.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1994c), “Evaluating prequalification criteria incontractor selection”, Building and Environment, Vol. 29 No. 4, pp. 437-48.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1994d), “Factors influencing UK constructionclients’ choice of contractor”, Building and Environment, Vol. 29 No. 2, pp. 241-8.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1994e), “Incorporating project-specific criteria andclient utility in the evaluation of construction tenderers”, Building Research andInformation, Vol. 22 No. 4, pp. 214-21.

Holt, G.D., Olomolaiye, P.O. and Harris, F.C. (1995a), “Application of an alternative contractorselection model”, Building Research and Information, Vol. 23 No. 5, pp. 255-64.

Holt, G.D., Olomoloaiye, P.O. and Harris, F.C. (1995b), “A review of contractor selection practicein the UK construction industry”, Building and Environment, Vol. 30 No. 4, pp. 553-61.

Hughes, W. (2009), “New wine in old bottles: how about the ranking of journals in constructionengineering and management”, Reply to debate on the Co-operative Network for BuildingResearchers (CNBR) web site, June 12, available at: http://tech.groups.yahoo.com/group/cnbr-l/ (accessed August 2009), (registration required for access).

Ip, W.H., Yung, K.L. and Wang, D. (2004), “A branch and bound algorithm for sub-contractorselection in agile manufacturing environment”, International Journal of ProductionEconomics, Vol. 87, pp. 195-205.

ITT (2009), Index to Theses, A comprehensive listing of theses with abstracts accepted forhigher degrees by universities in Great Britain and Ireland since 1716, available at: www.theses.com/ (accessed February 2009).

Jaselskis, E.J. and Russell, J.S. (1992), “Risk analysis approach to selection of contractorevaluation method”, Journal of Construction Engineering and Management, Vol. 118 No. 4,pp. 814-21.

Jennings, P. and Holt, G.D. (1998), “Prequalification and multi-criteria selection: a measure ofcontractors’ opinions”, Construction Management and Economics, Vol. 16, pp. 651-60.

Juan, Y. (2009), “A hybrid approach using data envelopment analysis and case-based reasoningfor housing refurbishment contractors selection and performance improvement”, ExpertSystems with Applications, Vol. 36 No. 3, pp. 5702-10.

Contractorselection

innovation

323

Page 21: Contractor Selection

Juan, Y., Perng, Y., Castro-Lacouture, C. and Lu, K. (2009), “Housing refurbishment contractorsselection based on a hybrid fuzzy-QFD approach”, Automation in Construction, Vol. 18No. 2, pp. 139-44.

Kadefors, A., Bjorlingson, E. and Karlsson, A. (2007), “Procuring service innovations: contractorselection for partnering projects”, The International Journal of Project Management,Vol. 25, pp. 375-85.

Kashiwagi, D. and Byfield, R.E. (2002), “Selecting the best contractor to get performance: on time,to budget, meeting quality expectations”, Journal of Facilities Management, Vol. 1 No. 2,pp. 103-16.

Khosrowshahi, F. (1999), “Neural network model for contractors’ prequalification for localauthority projects”, Engineering, Construction and Architectural Management, Vol. 6 No. 3,pp. 315-28.

Kumaraswamy, M.M. (1996), “Contractor evaluation and selection: a Hong Kong perspective”,Building and Environment, Vol. 31 No. 3, pp. 273-82.

Kumaraswamy, M.M. and Mathews, J.D. (2000), “Improved subcontractor selection employingpartnering principles”, Journal of Management in Engineering, Vol. 16 No. 3, pp. 47-57.

Kumaraswamy, M.M., Palaneeswaran, E. and Humphreys, P. (2000), “Selection matters – inconstruction supply chain optimisation”, International Journal of Physical Distribution& Logistics Management, Vol. 30 Nos 7/8, pp. 661-80.

Lai, K.K., Liu, S.L. and Wang, S.Y. (2004), “A method used for evaluating bids in the Chineseconstruction industry”, The International Journal of Project Management, Vol. 22,pp. 193-201.

Lam, K.C., Hu, T.S. and Ng, S.T. (2005), “Using the principal component analysis method as a toolin contractor pre-qualification”, Construction Management and Economics, Vol. 23,pp. 673-84.

Lam, K.C., Ng, S.T., Hu, T., Skitmore, R.M. and Cheung, S.O. (2000), “Decision support system forcontractor pre-qualification – artificial neural network model”, Engineering, Constructionand Architectural Management, Vol. 7 No. 3, pp. 251-66.

Lambropoulos, S. (2007), “The use of time and cost utility for construction contract award underEuropean legislation”, Building and Environment, Vol. 42, pp. 452-63.

Langford, D. and Male, S. (2001), Strategic Management in Construction, 2nd ed., BlackwellScience, Oxford.

Latham, M. (1993), “Trust and money”, Joint Government/Industry Review of Procurement andContractual Arrangements in the United Kingdom Construction Industry – Interim Report.Chaired by Sir M. Latham, HMSO, London.

Latham, M. (1994), “Constructing the team”, Final Report of the Joint Government/IndustryReview of Procurement and Contractual Arrangements in the United KingdomConstruction Industry. Chaired by Sir Michael Latham, HMSO, London.

Levin, P. (2007), Excellent Dissertations, McGraw-Hill International (UK), Maidenhead.

Li, H., Cheng, E.W.L. and Love, P.E.D. (2000), “Partnering research in construction”, Engineering,Construction and Architectural Management, Vol. 7 No. 1, pp. 76-92.

Li, Y., Nie, X. and Chen, S. (2007), “Fuzzy approach to prequalifying contractors”, Journal ofConstruction Engineering and Management, Vol. 133 No. 1, pp. 40-9.

Li, Y., Shouyu, C. and Xiangtian, N. (2005), “Fuzzy pattern recognition approach to constructioncontractor selection”, Fuzzy Optimization and Decision Making, Vol. 4, pp. 103-18.

Likert, R.A. (1932), “Techniques in the measurement of attitudes”, Archives of Psychology,Vol. 140, pp. 1-55.

CI10,3

324

Page 22: Contractor Selection

Lingard, H., Hughes, W. and Chinyio, E. (1998), “The impact of contractor selection method ontransaction costs: a review”, Journal of Construction Procurement, Vol. 4 No. 2, pp. 89-102.

Lo, W., Chao, C., Hadavi, A. and Krizek, J. (1998), “Contractor selection process for Taipei massrapid transit system”, Journal of Management in Engineering, Vol. 14 No. 3, pp. 57-65.

Loosemore, M., Dainty, A. and Lingard, H. (2003), Human Resource Management in ConstructionProjects. Strategic and Operational Approaches, Spon Press, London.

Love, P.E.D., Holt, G.D. and Li, H. (2002a), “Triangulation in construction management research”,Engineering, Construction and Architectural Management, Vol. 9 No. 4, pp. 294-303.

Love, P.E.D., Tse, R.Y., Holt, G.D. and Proverbs, D.G. (2002b), “Transaction costs, learning,and alliances in construction”, Journal of Construction Research, Vol. 3 No. 2, pp. 193-207.

Luu, V.T., Kim, S. and Huynh, T. (2008), “Improving project management performance of largecontractors using benchmarking approach”, International Journal of Project Management,Vol. 26 No. 7, pp. 758-69.

McCaffer, R. (2008), “Editorial”, Engineering, Construction and Architectural Management,Vol. 15 No. 2.

Mahdi, I.M., Riley, M.J., Fereig, S.M. and Alex, A.P. (2002), “A multi-criteria approach tocontractor selection”, Engineering, Construction and Architectural Management, Vol. 9No. 1, pp. 29-37.

Marzouk, M. (2008), “A superiority and inferiority ranking model for contractor selection”,Construction Innovation, Vol. 8 No. 4, pp. 250-69.

Masterman, J.W.E. (1996), An Introduction to Building Procurement Systems, E. & F. N. Spon,London.

May, T. (2001), Social Research Issues, Methods and Process, Open University Press,Buckingham.

Mbachu, J. (2008), “Conceptual framework for the assessment of subcontractors’ eligibility andperformance in the construction industry”, Construction Management and Economics,Vol. 26, pp. 471-84.

Meddis, R. (1984), Statistics Using Ranks: A Unified Approach, Basil Blackwell, Oxford.

Metcalfe, A. (1996), “Mathematical models and simulation”, in Greenfield, T. (Ed.), ResearchMethods, Guidance for Postgraduates, Arnold, London.

Mustafa, M.A. and Ryan, T. (1990), “Decision support for bid evaluation”, The InternationalJournal of Project Management, Vol. 8 No. 4, pp. 230-5.

NEDO (1967), “The Economic Development Committee for Building”, Action on the BanwellReport, National Economic Development Office, London.

Ng, S.T. and Skitmore, R.M. (1999), “Client and consultant perspectives of prequalificationcriteria”, Building and Environment, Vol. 34, pp. 607-21.

Ng, S.T. and Skitmore, R.M. (2001), “Contractor selection criteria: a cost-benefit analysis”, IEEETransactions on Engineering Management, Vol. 48 No. 1, pp. 96-106.

Ng, S.T. and Tang, Z. (2008), “Delineating the predominant criteria for subcontractor appraisaland their latent relationships”, Construction Management and Economics, Vol. 26,pp. 249-59.

Ng, S.T., Cheng, K.P. and Skitmore, R.M. (2005), “A framework for evaluating the safetyperformance of construction contractors”, Building and Environment, Vol. 40, pp. 1347-55.

Ng, S.T., Luu, C.D.T. and Chu, A.W.K. (2008), “Delineating criteria for subcontractorsregistration considering divergence in skill base and scales”, The International Journal ofProject Management, Vol. 26 No. 4, pp. 448-56.

Contractorselection

innovation

325

Page 23: Contractor Selection

Ng, S.T., Palaneeswaran, E. and Kumaraswamy, M.M. (2003), “Web-based centralized multiclientcooperative contractor registration system”, Journal of Computing in Civil Engineering,Vol. 17 No. 1, pp. 28-37.

Ng, S.T., Skitmore, R.M. and Smith, N.J. (1999), “Decision makers’ perceptions in the formulationof prequalification criteria”, Engineering, Construction and Architectural Management,Vol. 6 No. 2, pp. 155-65.

O’Leary, Z. (2004), The Essential Guide to Doing Research, Sage, Thousand Oaks, CA.

Oppenheim, A.N. (2000), Questionnaire Design, Interviewing and Attitude Measurement,Continuum International, London.

Palaneeswaran, E. and Kumaraswamy, M. (2001), “Recent advances and proposed improvementsin contractor prequalification methodologies”, Building and Environment, Vol. 36,pp. 73-87.

Palaneeswaran, E. and Kumaraswamy, M. (2005), “Web-based client advisory decision supportsystem for design-builder prequalification”, Journal of Computing in Civil Engineering,Vol. 19 No. 1, pp. 69-82.

Palaneeswaran, E. and Kumaraswamy, M.M. (2000a), “Benchmarking contractor selectionpractices in public-sector construction – a proposed model”, Engineering, Construction andArchitectural Management, Vol. 7 No. 3, pp. 285-99.

Palaneeswaran, E. and Kumaraswamy, M.M. (2000b), “Contractor selection for design/buildprojects”, Journal of Construction Engineering and Management, Vol. 126 No. 5, pp. 331-9.

Peansupap, V. and Walker, D.H.T. (2006), “Information communication technology (ICT)implementation constraints. A construction industry perspective”, Engineering,Construction and Architectural Management, Vol. 13 No. 4, pp. 364-79.

Phillips, S., Dainty, A. and Price, A. (2007), “The development of a tender analysis support toolfor use in social housing best value procurement”, in Boyd, D. (Ed.), Proceedings of the23rd Annual ARCOM Conference, 3-5 September, The Association of Researchers inConstruction Management, Belfast, pp. 797-808.

Pongpeng, J. and Liston, J. (2003), “TenSem: a multicriteria and multidecision-makers’ model intender evaluation”, Construction Management and Economics, Vol. 21, pp. 21-30.

RAE (2008), rae 2008 (sic), Results and Analysis of the UK’s Research Assessment Exercise2008, available at: www.rae.ac.uk/ (accessed August 2009).

RICS (2008), “Expected profit margins. Graph of expected pressure on construction profitmargins”, RICS Construction Market Survey United Kingdom. Third Quarter 2008,The Royal Institution of Chartered Surveyors, London, available at: www.rics.org/NR/rdonlyres/7FDB18D6-B0E4-4777-A777-5FA044873D71/0/RICSConstructionMarketSurveyQ32008.pdf (accessed February 2009).

RICS (2009), Royal Institution of Chartered Surveyors Research Publications Database,available at: www.rics.org/Newsroom/Researchandreports/Researcharchive/ (accessedFebruary 2009).

Russell, J.S. and Jaselskis, E.J. (1992), “Quantitative study of contractor evaluation programs andtheir impact”, Journal of Construction Engineering and Management, Vol. 118 No. 3,pp. 612-24.

Russell, J.S. and Skibniewski, M.J. (1988), “Decision criteria in contractor selection”, Journal ofManagement in Engineering, Vol. 4 No. 2, pp. 148-64.

Russell, J.S. and Skibniewski, M.J. (1990), “QUALIFIER-1: contractor prequalification model”,Journal of Computing in Civil Engineering, Vol. 4 No. 1, pp. 77-90.

CI10,3

326

Page 24: Contractor Selection

Russell, J.S., Skibniewski, M.J. and Cozier, D.R. (1990), “QUALIFIER-2: knowledge based systemfor contractor pre-qualification”, Journal of Construction Engineering and Management,Vol. 116 No. 1, pp. 155-69.

Shen, L.Y., Lu, W. and Yam, M.C.H. (2006), “Contractor key competitiveness indicators: a Chinastudy”, Journal of Construction Engineering and Management, Vol. 132 No. 4, pp. 416-24.

Shen, L.Y., Lu, W., Shen, Q. and Li, H. (2003), “A computer-aided decision support system forassessing a contractor’s competitiveness”, Automation in Construction, Vol. 12, pp. 577-87.

Sidwell, A.C. and Budiawan, D. (2001), “The significance of the tendering contract on theopportunities for clients to encourage contractor-led innovation”, Construction Innovation,Vol. 1 No. 2, pp. 107-16.

Silverman, D. (2001), Interpreting Qualitative Data: Methods for Analysing Talk, Text andInteraction, Sage, London.

(The) Simon Committee (1944), “The placing and management of building contracts”,The “Simon” Report, Central Council for Building and Works, Chairman Sir E. Simon,HMSO, London.

Singh, D. and Tiong, L.K. (2005), “A fuzzy decision framework for contractor selection”, Journalof Construction Engineering and Management, Vol. 131 No. 1, pp. 62-70.

Singh, D. and Tiong, L.K. (2006), “Contractor selection criteria: investigation of opinions ofSingapore construction practitioners”, Journal of Construction Engineering andManagement, Vol. 132 No. 9, pp. 998-1008.

Skitmore, M. and Mills, A. (1999), “Note. A needs based methodology for classifying constructionclients and selecting contractors: comment”, Construction Management and Economics,Vol. 17, pp. 5-7.

Smith, A.J. (1999), “The task of the referee”, IEEE Computer, Vol. 23 No. 4, pp. 65-73, available at:http://etrij.etri.re.kr/Cyber/Doc/Referee.pdf (accessed February 2009).

Sonmez, M., Yang, J.B. and Holt, G.D. (2001), “Addressing the contractor selection problem usingan evidential reasoning approach”, Engineering, Construction and ArchitecturalManagement, Vol. 8 No. 3, pp. 198-210.

Tam, C.M. and Harris, F.C. (1996), “Model for assessing building contractors’ performance”,Engineering, Construction and Architectural Management, Vol. 3 No. 3, pp. 187-203.

Tavistock (1966), Interdependence and Uncertainty – A Study of the Building Industry, TavistockInstitute of Human Relations, Tavistock Publications, London.

(The) Times (2008), “Galliford says construction downturn is spreading”, Times Online,available at: http://business.timesonline.co.uk/tol/business/industry_sectors/construction_and_property/article4735858.ece (accessed February 2009).

Topcu, Y.I. (2004), “A decision model proposal for construction contractor selection in Turkey”,Building and Environment, Vol. 39, pp. 469-81.

Waara, F. and Brochner, J. (2006), “Price and nonprice criteria for contractor selection”, Journal ofConstruction Engineering and Management, Vol. 132 No. 8, pp. 797-804.

Walker, A. (2002), Project Management in Construction, 4th ed., Blackwell, London.

Walker, D.H.T. and Johannes, D.S. (2003), “Preparing for organisational learning by HKinfrastructure project joint ventures organisations”, The Learning Organisation, Vol. 10No. 2, pp. 106-17.

Walliman, N. (2006), Social Research Methods, Sage, Thousand Oaks, CA.

Contractorselection

innovation

327

Page 25: Contractor Selection

Wang, Y., Yu, X. and Xue, X. (2007), “An application of the method of combined radixdetermination for selecting construction supply chain partners”, The International Journalof Project Management, Vol. 25, pp. 128-33.

Watt, D.J., Kayis, B. and Willey, K. (2009), “Identifying key factors in the evaluation of tenders forproducts and services”, The International Journal of Project Management, Vol. 27 No. 3,pp. 250-60.

White, A.S. (2005), “Public sector tendering – issues and analysis”, in Khosrowshahi, F. (Ed.),Proceedings of the 21st Annual ARCOM Conference, 7-9 September, The Association ofResearchers in Construction Management, University of London, London, pp. 1143-52.

Wong, C.H. (2004), “Contractor performance prediction model for the United Kingdomconstruction contractor: study of a logistic regression approach”, Journal of ConstructionEngineering and Management, Vol. 130 No. 5, pp. 691-8.

Wong, C.H. and Holt, G.D. (2003), “Developing a contractor classification model using amultivariate discriminate analysis approach”, RICS Research Papers, Vol. 4 No. 20.

Wong, C.H., Holt, G.D. and Cooper, P.A. (2000), “Lowest price or value? Investigation of UKconstruction clients’ tender selection process”, Construction Management and Economics,Vol. 18, pp. 767-74.

Wong, C.H., Holt, G.D. and Harris, P.T. (2001), “Multi-criteria selection or lowest price?Investigation of UK construction clients’ tender evaluation preferences”, Engineering,Construction and Architectural Management, Vol. 8 No. 4, pp. 257-71.

Wong, C.H., Nicholas, J. and Holt, G.D. (2003), “Using multivariate techniques for developingcontractor classification models”, Engineering, Construction and ArchitecturalManagement, Vol. 10 No. 2, pp. 99-116.

Wong, F.W.H., Lam, P.T.I., Chan, E.H.W. and Shen, L.Y. (2007), “A study of measures to improveconstruct ability”, International Journal of Quality & Reliability Management, Vol. 24 No. 6,pp. 586-601.

Wong, J.K.W., Li, H. and Wang, S.W. (2005), “Intelligent building research: a review”,Automation in Construction, Vol. 14 No. 1, pp. 143-59.

Woodward, J.F. (1997), Construction Project Management. Getting It Right First Time,Thomas Telford, London.

Yasamis, F., Arditi, D. and Mohammadi, J. (2002), “Assessing contractor quality performance”,Construction Management and Economics, Vol. 20, pp. 211-23.

Yiu, C.Y., Lo, S.M., Ng, S.T. and Ng, M.M.F. (2002), “Contractor selection for small building worksin Hong Kong”, Structural Survey, Vol. 20 No. 4, pp. 129-35.

Zavadskas, E.K. and Vilutiene, T. (2006), “A multiple criteria evaluation of multi-familyapartment block’s maintenance contractors: I-model for maintenance contractor evaluationand the determination of its selection criteria”, Building and Environment, Vol. 41,pp. 621-32.

Corresponding authorGary Holt can be contacted at: [email protected]

To purchase reprints of this article please e-mail: [email protected] visit our web site for further details: www.emeraldinsight.com/reprints

CI10,3

328