15
Workbrain Scalability Assessment Test Completion Report Reference Number: CD-805·0005 Stream: Vers ion: 2.0 Approvers . •!.-·. - Title Brett Matthews SDA Emwum<e>Director .z}. Paul Hickey Program 42 Program /1 __________ PI)JJ Shared SeNice Sof tJtion s Commercial-in-Confidence

Workbrain Scalability Assessment Test Completion Report · Workbrain Scalability Assessment Test Completion Report ... and leave accrual engine is extremely ... a Workbrain Scalability

Embed Size (px)

Citation preview

Workbrain Scalability Assessment

Test Completion Report

Reference Number: CD-805·0005

Stream:

Version: 2.0

Approvers

. •!.-·. -Title

Brett Matthews SDA Emwum<e>Director

.z}. Paul Hickey Program 42 Program

/1 L-------~----~D_irn_ct_or __________ ~~~~~~~--~+c7-~~ -:J~:rmc:.s &~tJ,J PI)JJ

Shared SeNice SoftJtions

Commercial-in-Confidence

Program LJ2

Approvals

PREPAREO BY:

APPROVED BY:

Amir Mahmoudi and John Sinclair Name Role

Name Role

Commercial-in-Confidence

CD-805-0005 Test Completion Report

V2.0

Date

Date

Page 2 of 13

Program 42

Document Information

Docum~nt Source

CD-805-0005 Test Completion Report

V2.0

The source of this document is named Program 42 Work:brain Scalability Assessment Tes.t

Completion Report il2.0.doc.

Author : Amir Mahmoudi and John Sinclair

Reference Number: CD-805-0005

Revision History ve~ion Revision d~te sun#n~ry of ch~·~ges Priqi~-r.y Author ~)J.rij~er ., .. 0.1 28/04/08 Initial version Amir Mahmoudi

0.2 30/04/08 Review and update John Sinclair

0.3 06/05/08 Updates based on lntf~mal review John Sinclair

1.0 15/05/08 Updated based on comments from the Amir Mahmoudi client and minor update to section 4,3.2

1.1 23/05/08 Updates the recommendations section Amir Mahmoudi

2.0 25/07/08 Final version PDO

Reviewer List I N,.ame ·

Distribution

~ ~ame . _.;-·.: I Title

Related D.ocuments Test• ~ele~~-~ Titi~· of OOcli~~rttlLocati~~

.. ···. v~rslon Typ~· .. p~te.

. .. - 5.:. '-~ . . ' ~.· .'!5-· ··- . .

Project 26/03/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime 0.1 Execution Contractor- IBM\23. Worl<brain Scalability: Assessmeni\Pian\Proiecf Execution Plan Plan\Workbrain Scalabilit~ Assessment PEP v0.1 .doc

Test Plan 23/04/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime 1.1 Contractor- IBM\23. Workbrain Scalability: Assessment\Deliverables\Test Plan\Workbrain Scalability Assessment Test Plan ver 1.1 final.doc

1 07/02/08 \\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime 0.2

Cotnmercial•in-Confidence Page 3 of 13

Program42

Test Release. Type I) ate ······

2a 06/03/08

2b 28/03/08

2c 07/04/08

3-.online 18/04/08 to 24/04/08

3-batch

CD-805-0005 Test Completion Report

V2.0

· · Title of Oqiu-Mentll ocation Versioil ... ,._ _ _ • • 1 • , , _.:

Contractor- IBM\23. Workbrain Scalability ASsessment\Pian\Test1\Results\P42 Scalability Testing Test1 Re11ort V0.2.doc ·

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor - IBM\23. Workbrain Scalability AssessmentiPian\Test 2a1Results\Workbrain POC rostering Test 2 summa!Y of find ings. doc

\INW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 2a1Results1Workbrain POC rosteri!HJ Test 2 summar:y of [!roblems encountered. doc

\INW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor- IBM\23.Workbrain Scalabilitll Assessment\Pian\Test 2a1Results\Workbrain Scalabililll - Test 2 first rostering test- Anal:tsis v0.1 .doc

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime Contractor- IBM\23. Workbrain Scalabililll Assessment\Pian\Test 2a1Results\Workbrain Scalabili!ll - Test 2a - Anal~is vb.2:doc

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS - Prime Contractor- IBM\23. Workbrain Scalabililll AssessmentiPian\Test 2b\Results\Workbrain POC rostering Test 2b Executive Summa[Y.doc

\\NW-CLUSTER-01 CTDATA SERVERICTDATA\DATA\Cn 17 - SSS - Prime Contractor ~ IBM\23. Workbrain Scalability Assessment\Pian\Test 2c\Results\Microsoft Word -Test2c 500VU Run1 3Server 03042008 Re(!ort.odf

\\NW-CU..iSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor - IBM\23. Workbtain Scalabilitll Assessment\Pian\Test 2c\Resuits\Microsoft Word -Test2c 1000VU Run1 3Server 03042008 Re(!ort.Qdf

\\NW-CLUSTER-01 CTDATA S!;RVERICTDATA\DATA\Cn17- SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 2c\Results\Microsoft Word -Test2c 1500VU Run2 3Server NS RR 04042008 ReQort.Qdf

1\NV\i-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor" IBM\23. Workbrain Scalabililll Assessment\Pian\Test 3\Results\Test3 500VU Run1 1Server 17042008 Re[!ort.!!df

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor- IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results\Test3 500VU Run1 .2Server 18042Q08 Re11ort.doc

\INW-CUJSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor- IBM\23. Woil<brain Scalabii[!Y Assessment\Piari\Test 3\Results\Test3 1000VU Run1 1Server 17042008 Rellorl.(!df

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17- SSS- Prime Contractor- IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results1Test3 1000VU Run1 2 Server 18042008 ReQort.doc

\\NW-CLUSTER-{)1 CTDATA SERVER\CTDATA\DATA\Cn 17- SSS - Prime Contractor - IBM\23. Workbrain Scalabilitll Assessment\Pian\Test 3\Resulls\Test3 1200VU Run1 1Server 18042008 Rellort.doc

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn1 7 - SSS - Prime Contractor - IBM\23: Workbrain Scalabilitll Assessment\Pian\Test 3\Resulls\Test3 1500VU Run1 1Server 18042008 Re(!orl.doc

\\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS - Prime Contractor - IBM\23. Workbrain Scalability AssessmentiPlan\Test 3\ResuHs\Test3 1500VU Run1 3Server 22042008 Re12ort.!1df

\\NW-CLUSTER-01 CTDATA SERVER\~TDATA\DATA\Cn17- SSS- Prime Contractor - IBM\23. Workbrain Scalabililll AssessmentiPian\Test 3\Results\Test3 2000VU Run1 3Server 22042008 Re12ori.Qdf

\\NW-Cl USTER-01 CTDATA SERVERICTDATA\DATA\Cn 17 - SSS - Prime Contractor - IBM\23. Workbrain Scalabililll AssessmentiPlan\Test 3\Results\Test3 2500VU Run1 3Server 22042008 Re12ort.!ldf

1\NW-CLUSTER-01 CTDATA SERVER\CTDATA\DATA\Cn17 - SSS- Prime Contractor- IBM\23. Woil<brain Scalabilitl! Assessment\Pian\Test 3\Results\Test3 3000VU .Run1 3Server 2204200S ReQort·.!ldf

\\NW-CLUSTER-{)1 CTDATA SERVER\CTDATA\DATA\Gn17 - SSS - Prime 0.2 Contractor - IBM\23. Workbrain Scalability Assessment\Pian\Test 3\Results\P42

Commercial-in-Confide.nce Page 4 of 13

Program 42 CD-805-0005

Test Completion Report V2.0

T.~~t Tifp~ .

Rel.e.as~ oaiij"·

Version

·. , · ;,: -..rl .•

Scalability Testing Test 3 Batch Report V0.2.doc

Definitions

Term Definition

CPU Central Processing Unit.

End-state This term refers to the end of the SSI, SSS design, build and implementation program; when all Queensland Government agencies anQ Shared Service Providers have .gone live with the new SSS b,usiness solutions.

HTTP Hypertext Transfer Protocol.

J2EE Java 2 Platform, Enterprise Edition.

LDAP Lightweight Directory Access ProtocoL

POC Proof-of-Concept

SDA Solution Design Authority

SSI Shared Service Initiative,

SSL Secure Socket Layer.

sss Shared Service Solutions.

TPH Transactions per hauL

TPS Transactions per second.

Vuser Virtual User, HP/Mercury LoadRunner user that emulates the. business transaction of a normal user.

WoG Whole-Of-Government

CoinmerciaJ-in-Confidence P9ge 5 of 13

PrQgram 42

Table of Contents

CD-805-0005 Test Completion Report

V2.0

1. Introducti.on .................................... , ............................................................................................ 7 .

1.1 Purpos.e ...... ........ ..... ......................................................... ..... ................. ..... ....... .. ................... 7

1.2 Audience· ... ..... .. .... .... ... .... ................... .. .. ...... ..................... .... ............ .. .... .. ....... .. ........ .......... . 7

2. o·escription .................................................................................................................................... 7

2.1 Ovetview .. .... .......... .. .... .. .. .......... .. .. .. .. .................. .. .. .. ...................... .................. .. ................. 7

3. Test Scope ................... , ................................................................................................................. 8

4. Test Results ............................... , .................................................................................................. 8

4.1 Test 1 .. ............... ......... ............................ .......... ....... ; .. .......... ; ....... ... ...................... ..... ..... ..... 8

4.2 Test 2 ............ .. ... ................. .. , ... .............. ... ..... ..... .... ... ... .. ....... .. ....... .... ... ....... .... .... ... ........ .... 9

4.3 Test 3 .. ..... .. .. .... ......... ..... ...... .. .. .. ... ... ... ... ........... .. ... ....... .. ...... .... .... ... ....... .... .. .... .. ... ... .. .... .. .. 11

4.4 Any variance to originai plan .. .... ..................................... .. .. .. .. ............ , .... .. .. ..... .......... .. 13

4.5 Observations ......... ................. ............. .. ...... .. ......... ...... ...................................... ... ........ : ... . 13·

5. Conclusions ................................................................................................................................ 13

5.1 Exit criteria .... ..... ... ..................... ..... ................. .. ................................ .. .. ....... ........ .. .. , .. .. , .. .'13

5.2 Recommendations ...... ..... ............. ...... ; ... ........ .. .. .. .. ............ ................. ... ...... ... .... ............. 13

·Commercial-in-Confidence Page 6 of 13

Program 42

1 ~ Introduction

1.1 Purpose The purpose of this document is to:

~ Provide a summary of the Work brain Scalability Assessment testing effort;

CD-805-0005 Test Completion Report

V2.0

~~ Enabl~ management to m,ake an informed decision regarding whether Workbrain is the right solution for award interpretation, or if it is not, to identify the associated risks;

• Ensure exit criteria are met, or if they are not, that risk is accepted.

This document does not cover the detailed testing results. Rather it provides a high level summary of the results. For further detail, please refer to the individual testing results.

1.2 Audience The target audience includes all those who have had a role to play both during test preparation and test execution as well as those who will be involved in making the go/no go decision. This will include:

e Corp Tech SDA and SSS Technology Solutions • IBM • Program I Project Managers

2. Description

2.1 Overview The original SSS solution design for non rostering Agencies relied solely .on the SAP application for award interpretation processing. Based on actual payroll processing results at the Department of Housing, and e)(trapolating these to include the remainder of non rostering agencies, it became evident that the SAP application would be unable to process the award interpretation function within an acceptable timefram_e to me~t SSS business requirements. In IBM's Invitation to Offer, it was proposed to repiC~ce the SAP centric approach to award interpretation with the Workbrain application to alleviate processing time constraints and to realise the additional following benefits: · • Economies of scale through developing only one set of ri.!les that will be utilised across all Agencies.

The Workbrain architecture allows pay rules to be de~igned so they are fully configurable across multiple awards and across multiple Agencies. The award configurations have a large number of similarities allowing a small set of rule components to be built that can then simply be reconfigured to fit any number of awards. Currently ~wards i?)nd their associated pay ru les will be developed in Workbrain as part of the Phase One roll out to Health. Therefore, huge efficiencies can be gained through re-using these existing rules and configurations for other Agencies, eliminating the costly and time consuming development phase iri SAP; .

e Implementing all awards in Workbrain provides a single system of rec:ord for their configuration, therefore providing significant efficiencies for maintenance, EBA updates, training, etc;

• The Workbrain rule and leave accrual engine is extremely efficient in providing considerable time savings, particularly when all awards will already have been interpreted when SAP payroll runs are made, vastly improving the time window required for this;

o The system ar<;::hiteclure of tne Workbrain rule engine is designed to allow simple and efficient integration with custom components. This allows for a very short development cycle where a.ll custoinisations are written in Java and fully extendable and re-usable; ·

" Workbrain provides a large number of standard pay ru les within the core system, considerably reducing the size of custom development required;

G Changes to award agreements can be made quickly and easily via the pay rule editor in Worl<brain, eliminating the need for further custom developmemt and associated costs;

• Configuration of pay rules in Workbrain is simple relative to SAP, with all configurations fully exportable via XML.

Commercial-in-Confidence Pa~e 7 of 13

Program42 CD-805-0005 Test Completion Report

V2.0

In principle, SSS accepted IBM's proposal to utilise the Workbrain application for both rostering and non rostering agency award interpretation. Before proceeding with this approach, SSS has requested that IBM perform a series of tests that must evidence Workbrain's ability to scale, providing comfort that b~siness requirements can be satisfied. This scalability test requirement is documented as a deiiverable in SOW 5 -Priority Core HR & Finance Development. On test completion, a Workbrain Scalability Assessment Test Completion Report (contracted deliverable, this document) will be presented to SSS, outlining the tests performed, results obtained and interpreted findings. A go/no-go d.ecision on the proposed solution will be made by SSS after review and consideration of this document.

3. Test Scope

The scope of this testing exercise is to assess the scalability capability of the Workbraih epplication. As such any applica.tion functional testing, performance testing, stress and volume testing, and any tests to confirm production sizing is out of scope of this activity. ··

There are three types of testing involved in this assessment: • Non-rostering agency processing (Test 1) will execute award interpretation in batch utilising an

imported employee ·file as input. This test will be executed on interim sized batch and production sized database servers;

• Rostering agency processing (Test 2) will te$t concurrent user scalability by submitting timesheets through the onlihe Workbrain application fr9nf end and subsequently perform the Award interpretation. This test will be conducted on a production sized database server and on interim hardware (approximately - 50% productiqn) for the application servers;

• Combined rostering and non-rostering (Test 3) scenarios on production sized hardware using a range of transactions per hour and a range of concurrent users. Results will be assessed against anticipated volumes. The rostering and non-rostering tests will not overlap in time, i.e. The rostering and non-rostering tests will be executed at different times.

Important Note: Initially Test 3 included a file tr<!nsfer portion via SAP XI. A deCision has been made to exclude the same from the scope of Test 3, A CR has been raised by Corp Tech in order to remove this from the scope of the Test 3. Below is a representation of the testing types:

Workbraln

RcstDringeinpiO'JI!is

1'1-- ~~·-~ 1::- A•""

/..-;:st 3 inc/ud~';,\ .... . .. : combination or Test 1 J

\ and 2tes1 cases / .. '.:; ...... _____ ......

tnterpretallon

I Stllodule Vl~rk , ___, H PayroH Export I Delails Import

'---· . . t .. . . L------l - +-.! _· _ _...,

Non-rt;~oi~iing t-d.: [F.'~ ~o:;teling and

iimp\_0'1~' It~; CSV CSV ~~ ~'j ":;~~~~~~!)

~~~ . c.__ __ ~I,·_J ______ ~x, _______ ,+f_' ~ Le{]end - t

i~ csv csv 'IT~ .. ~ . ~~~t t R;:;::,~~~.·; · [~ _ _ t::::.~~=...::--·:::....j ____ s_·A_P ____ l::..:·.}::..::~=--4-~

Ta t 2

4. Test Results

4.1 Test 1

The results of Test 1 demonstrate the ability of the Workbrain batch processing solution to scale horizontally and vertically. In addition, they indicate very fast processing speeds for two of the three

Commercial-in-Confidence Page 8 of 13

!Program 42 CD-805-0005 Te~t Completion Report

V2.0

tests- full recalculation and creating the export fife . Both show increased performance proportionate to changes in hori.?;ontal and. vertic~! configuration.

The results ofthe schedule work detail import test, were inconsistent and required further database tuning and test configuration before developing a final assessment on either processing speeds or th~ solution's ability to scale to meet d.ata import requirements. The test results to date. demonstrate the ability to scale vertically. They als.o indicate fast processing times which, when extrapolated, would very easily fall within the window for pay processing. However, an issue was surfaced with horizontal scaling (i:e., addition of job schedulers). The database timed out in an irregular manner when running 40 or more import process threads simultaneously, even though processing completed within required windows.

Schedule Work Detail Import File

Full Recalculation Process

Export File

4.2 Test 2

4.2.1 Test 2a

112 minutes

48 minutes

12 minutes

• 0

• • 0 .. 0

• 0

• • • • • • • •

Batch Servers: 1 Job schedulers: 4 Tasks per schedl,ller: 10 Total Tasks: 40 Periop: 7 Days (Q5May- 12May)

Batch servers: 2 Jcib schedulers: 2 Tasks pEir scheduler: 2 Threads per task: 4 Total threads: 16

Batch servers: 2 Job schedulers: 8 Tasks per scheduler: 1 Threads per iask: 1 Total threads: 8

8

Tests were performed to determine the maximum number of users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shows the results witl11, 2 and 3 servers:

1

2

3

. ·: 4.:~: . ·:~JL .· ~'~·~· Sucri~'k~

100

200

300

Commercial-in-Confidence

300

450

· ~:~ ... ~:·· • I · ~ ~!; ·:·

600

• " .b~~·.·

<.'dj~~· . ;..-·

Page 9 of 13

Program 42

~ C1J Ill

~ .... c:: ~ :I (.) c 0 0

3000 2700 2400 2100 1800 1500 1200 900 600 300

0

Test2a

2

Servers

3

CD-805-0005 Test Completion Report

V2.0

It was determined that Workbrain was installed using default settings for WebSphere and that a WebSpi:Jere Specialist should be engaged to tune WepSphere. It was also determined that the Load Runner VUsers were executing at the rate of 32 timesheets submitted per VUser per hour. It was agreed that a more realistic scenario was that each of the concurrent users was executing at the. rate of 12 timesheets per VUser per hour, and this definition of concurrent users was used for subsequent tests, and VUsers were calibrated to this rate.

4.2.2 Test 2b

Tests were performed to determine the maximum number of concurrent users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shoWs the results with 1, 2 and 3 servers:

·-· ")' • · _ '1 '0 1 •• tl 0 o o' 0 ',\ t~> ,• ~?~. !. :.Ll~ ~ ·- .... ~ ':r:. ;..{' G. '

·': Wci.r~Brai~rr :-· >- .: ~·-· · , ··:q.oncur~nt 0$~-rs, ;~,~~ ".',->: '):.:. 1 ·:~ s':. . ~~ ~ . ~ · ,.,..--~· . · . ~ ;· · .c;t. ~ ~~ ~~ ~. \ c._-n·. ~~ ~fit ~· . :A!·.~ ~;: ~~ers ,~:.>. -c:-._.}succ'e~~f . •: , Boj7fj,~tli~lf~.f. ( · ~- F~!~,;: , . :. ·

1 380 400

2 500 630 675

760 775 800

Test2b

2 3

Se~vers

Commercial-in-Confidence Page 10 of 13

Progr(tm 42 CD-805-0005 Test Completion Report

V2. 0

It was determined that the cause of the problem was the load balancing method used by the W eb Server Plug-in. During this test; the "Weighted Round Robin Equal Weighf' method was used. The method was changed to Random and subsequently Test 2c was executed.

4.2.3 Test2c

After completion and certification of test 2b, further tuning was done to the interim landscape. Capacity was do!Jbled by employing an alternative load balancing algorithm on the weq server plug-in. Where "Weighted Round Robin Equal Weighf' had been employed previously, the use of "Random" doubled the performance. Tests were performed to determine the maximum number of users that could be successfully run on 1 server, 2 servers and 3 servers. Below matrix shows the re.sults with 1. 2 and 3 servers:

Test2c

3000 2700 ·--·-":·'

'· 2400

112 2100 C1l. (/)

~ 1800 +> .. . ,· c 1500 : ·~:-

~ ::2 1200 u c 900 0 0 600

300 0

2 3

L Servers

--· -- - - - - - -

With this configuration, the interim environment showed scaling as additional servers were added, more concurrent users could be handled, and the configuration could successfully handle the 1500 concurrent users that it was sized for.

4.3 Test 3

4.3.1 Test 3 online

The online portion of Test 3 was e~ecuted against the Workbrain environment on the Production infrastructure. Note that SSL is fully configured from Client all the way up to the WebSphere Application Servers, due to securitY requirements in the Production infrastructure.

Tests were performed to qetermine the maximum number of users that could be successfully ·run on 1 server, 2 servers and 3 servers. Below matrix shows the results with 1, 2 and 3 servers:

Commercial~in-Confidence Pa~e 11 of 13

Program42

rl Q) en :J ... c: ~ :I u c:

I 0 0

i

L

;3000

2900

2000

1500

1000

500

0

----Test3

2

Servers

3000

3

CD-805-0005 Te~t Completion Report

V2.0

This test was executed using 3, 000 concurrent virtual w~ers on three Production Workbrain servers. 3,000 concurrent users represent a total user base of 90,000 and i.s considered to be the end state figure for rostered employees. This test was successful.

4.3.2 Test 3 batch

The results of Test 3 demonstrate the ability of the Workbraih batch processing solution to meet the exit criteria on production~sized hardware.

Full Recalculation Process

Export File . 28minvtes

-' ~ ,; .. .

!- ,1 \ . · . ·' .·

Commercial-in-Confidence · 1

Two sequential jobs of • Batch Servers: 1 0 Schedulers per Server: 4

" Ti!sks per Scheduler: 2 e Total Tasks: a

• Batch Servers:. 2 0 Schedulers per Server: 4 0 Tasks per Scheduler: 2 • Threads per Task:,a • Total Threads: 128 • • Batch servers: 2 • Schedulers per Server: 2 • Tasks per Scheduler: 1 .. Threads per Ta:>k 1 • Total Tasks: 4 • 4

. . . ~ ·· ; . .

Paqe 12 of 13

Progra~42

4.4 Any variancE:> to original plan

CD-805-0005 Test Completion Report

V2.0

The original plan for n~st 3 included a file transfer portion to SAP via XI. Corp Tech raised a Change Request in order to remove this from th'e scope of Test 3, hence Test 3 execution included the Online Award processing and Batch Awarq processing. The B~tch Award processing included an import to Workbrain, processing by Workbrain and export from Workbrain, without transferring files to and ffom Workbrain from SAP (via XI).

4.5 Observations

Ouring Test 2c, an additional approach was also undertaken. The web servers were eliminated from the solution and Citrix Netscaler performed all load balancing. All traffic was routed directly from Citrix Netscaler to the Workbrain application servers. The three load balancing algorithms were:

• Simple round robin; • Fastest response time; • Least connections,

The first algorithm provided performance comparable to (if not marginally better than) the performance when using the web servers. The remaining algorithms did not exhibit good performance.

During Test 3, it was found that the Oracle SQL execution plan for Payroll Export was not optimal, and the registry parameter was changed from FiRST_ROWS to ALL_ROWS.

During Te$t 3, it was found that the Oracle SQL execution plan was not optimal, snd that Oracle was choosing the Wrong index for the query. The workaround used to was to drop the IDX_EMPSKD_WORKDT index. A Workbrain Support Hub Incident was raised to add a config1.1rable registry parameter for this tuning (Incident number 1978603).

5. Conclusions

5.1 Exit criteria

All exit criteria, as defined in the Test Plan, have been reached.

Online

Batch

3000 concurrent users on 3 Servers -Application scales linearly -

Work Schedule Import less than 4 hours­Full Awarq ~ecalculalion less than 1 hour ­Payroll Export less than 30 minutes -

reached reached

reached reached reached

5.2 Recommendations

Recommendation 1: Based on tt1e results of the scalability assessment tests, it is recommended that a "Go" decision be made to adopt Workbrain as the whole-of-government solution for awards interpretatiqn. The tests did hot identify any unacceptable risks. The Workbrain scalability assessment has been sucoessfuliy completed.

Recommendation 2: It is re.commended that further architectural analysis and testing be undertaken to determine if the Web Server functiOnality can b~ handJed by the NetScaler device.

Recommendation 3: It is recommended that detailed Non-Functional Requirements be gathered aild architectural work products devetop~d for Program 42. ·

This is the last page of this ,document.

· CommerCial-in-Confidence PaQe 13 of 13

Deliverable Acceptance Sheet -A ·separate Acceptance Sheet is to be completed for each deliverable. Upon completion, please submit to Vendor Management

~ -· - -- -

D~liverable Descript_ion

l)eliver.i~te IQ ...

CD 805 0005

Deliverable N;~me Workbrain Scalability Assessment Test Report

~ow N1,un~~r SOW5

Qeliverabi.e L-ea(l Tom Lambert

Purpose Submission of Workbrain Scalability Assessment for SOW 5

Date of Deliverable .. .-,

26/05108 I Date d!:l~··t9. IJ3M J 30/06/08

Deljverable Cqmposition Word document .

Dep~pdent Deliverables N/A

Prere.guisite Deli·.J~~ables N/A

Linked. to Payment Milestone . __ [gj Yes 0 No . .-,, Mileiito~e Date: -3bY1 June ZOOS

Acceptance Details

breat~ii; -

· Oatei · ro· ''' ' .... ,., ···t .. ::. ... ,:_ ::

; . •; ... , . ~i· ••

~Acc'eP.ta~ce Lea& lillike Robin~on · - P.e~t~i -...... '

!_:;::_: .·., • .. ·n.--· .. . : .. ~ .-•---!_!( _ • . ·-:-:' !

ACCEPTED (refer to 'Comments' Section) NOT ACCEPTED (refer to 'Comments' Section)

181 D ;:Accell~ii~~ Lead .s·i~~ature: ' '· D~t~;·:~p/o.S/2008 ·_.

··:: :_·'·; ,, I

· !- •·' - e

. -~ . ; ~.~:.\ .. ... .. !--~ '

Acceptance Board

Acceptance Panel Area of Acceptance Date (list Board Members) (e.g. Technical, Quality, Audit, Resourcing etc)

Mike Robinson HR Lead 30/05/2008 - ---

-

-

-

~-

-

:: :

C:\Documents and Settings\rzgott\l ocal Settings\Temp\notesE259CD\CD_B05...:.0005_0eliverable Acceptance Sheet.doc Page 1 of2

,.,: ·-

.. .. . • .·. \ .

. ;;;:: . ,-.· . .:.:/., .

- -

--

- - --

C~mments

Comments /Issues I Defects (add notes or attach document. If not accepted, please provide list of defects and issues). - -

- --··

- -

- -

r-- · -

f-- - -- -

f-----.- . -

- - -

- - · ___:____ -

-

t---- -

r- -

r--

- -·

-

t---

C:\Documerits and Settings\rzgolt\Local Settings\Temp\notesE259CD\CD_B05_0005_Deliverable Aci:eptarice Sheet.doc Page 2 ·or2

-

-

·· -

- -

-

-

-

-

-

--

-

- -