Upload
letruc
View
214
Download
0
Embed Size (px)
Citation preview
A RESPONSE TO CHAPTER 6 OF THE DEPARTMENT OF
EDUCATION AND SKILLS DOCUMENT –
Better Literacy and Numeracy for Children and Young People:
A Draft National Plan to Improve Literacy and
Numeracy in Schools
Submitted by
The Educational Research Centre
St Patrick’s College
Dublin
February 28, 2011
iii
Contents
About the Educational Research Centre iv
Preface vi
1 Introduction 1
2 Enhancing Teaching and Learning through Classroom
Assessments
5
3 Linking Classroom Assessments to Curriculum Standards 7
4 Standardised Testing at Four Key Points during Compulsory
Schooling
9
5 Contextualising the Outcomes of Standardised Test Results 13
6 National and International Assessments of Educational
Achievement
18
7 Conclusion 21
References 24
iv
About the Educational Research Centre
The Educational Research Centre was established on the campus of St Patrick's College,
Dublin in January, 1966. The setting up of the Centre was a means of widening the scope of,
and making for greater continuity in, research efforts in Ireland. The Centre works at all
levels of the education system, from pre-school to third level.
Research is undertaken by the Centre in the following areas:
Evaluations of Initiatives and New Programmes
The Centre conducts evaluations of initiatives and new programmes in schools to improve the
learning and life-chances of students in socially-disadvantaged situations. Completed projects
include evaluations of the Rutland Street Project, Breaking the Cycle, Early Start, and the
Home-School-Community Liaison Scheme. Current projects include an evaluation of School
Support Programme under DEIS (Delivering Equality of Opportunity in Schools).
Evaluations of Services and Existing Provision in Schools
The Centre conducts surveys of services, conditions and practice in schools. These include
topics as diverse as teachers‟ perceptions of girls and boys in primary schools, learning
support provision in primary schools, and dropout in third-level institutions.
Critical Analysis of Issues in Education
The Centre conducts research on a number of critical issues in education, including factors
associated with early school leaving. The designs of a number of Department of Education
and Skills schemes for addressing educational disadvantage have been influenced by studies
of the extent and geographical distribution of disadvantage (e.g., Giving Children an Even
Break, Delivering Equality of Opportunity in Schools). These studies also provided a basis
for assessing levels of disadvantage in individual schools and were used to select schools for
participation in the schemes.
Analysis of Public Examination Results
The Centre has on several occasions carried out analyses of the results of public examinations
(Junior Certificate and Leaving Certificate). Studies have been carried out on: (i) the impact
of examination components on total scores in the Leaving Certificate Examination; (ii)
relationships between performance on Certificate examinations and performance on
international assessments of educational achievement; (iii) the comparability of grades in the
Leaving Certificate Examination; and (iv) the relationship between performance on the Junior
and Leaving Certificate examinations.
National Assessments of Educational Achievement
As part of the Centre‟s work in monitoring the outcomes of education, national surveys of
achievement at primary level have been carried out in association with the Department of
Education and Skills. The sample-based surveys have been conducted in the areas of English
(reading), Irish (reading and oral language), and mathematics. In 2009, pupils in Second and
v
Sixth classes in primary schools were assessed on English and mathematics, while in 2010,
pupils at the same class levels in Irish-medium schools were assessed in these subjects.
International Assessments of Educational Achievement
The Centre has been involved in international assessments of educational achievement since
the 1980s. Current work includes the Organisation for Economic Co-operation and
Development‟s (OECD) Programme for International Student Assessment (PISA) in which
15-year olds in 34 OECD member countries, and several partner countries are assessed every
three years on reading literacy, mathematical literacy and scientific literacy. Current work
also includes two studies organised by the International Association for the Evaluation of
Educational Achievement (IEA) – the Progress in International Reading Literacy Survey
(PIRLS) and the Trends in International Mathematics and Science Study (TIMSS) – both of
which involve pupils in Fourth class in primary schools.
Development of Assessment Instruments
Work is carried out on an ongoing basis on the development of tests and profiling systems
which teachers can use to assess pupils in primary and post-primary schools. These include
the Drumcondra English Profiles, the Drumcondra Primary Reading Test, the Drumcondra
Primary Mathematics Test, the Drumcondra Reasoning Test, and Trial Ghaeilge Dhroim
Conrach do Scoileanna Lán-Ghaeilge agus Scoileanna Gaeltachta. Most recently, the Centre
has published sets of screening and diagnostics tests for children in Senior Infants and First
classes in primary schools – the Drumcondra Tests of Early Reading Literacy and the
Drumcondra Tests of Early Numeracy. The Centre has recently begun work on the
development of standardised tests in English and mathematics for Second-level schools.
vi
Preface
The Educational Research Centre welcomes the publication by the DES of the Better Literacy
and Numeracy for Children and Young People: A Draft National Plan to Improve Literacy
and Numeracy in Schools.
The results of PISA 2009, which were published in December 2010, have received a lot of
attention and provoked some negative comment about education in Ireland. While it is
important to be cautious about PISA 2009 findings relating to declines in performance (see
Cosgrove et al., 2010; Perkins et al., 2010; LaRoche & Cartwright, 2010), it is clear from all
PISA cycles, and from other assessments including National Assessments of English reading
and mathematics conducted over many years, that a range of issues relating to performance in
reading and mathematics need to be addressed. These include the poor performance of
students in mathematics at both primary and post-primary levels, with mathematical problem
solving being an area of particular concern. They also include the relatively poor performance
of higher-achieving students in Ireland, who do not do as well as might be expected in either
reading literacy or mathematics. PISA 2009 also draws attention to the challenges facing
students for whom English or Irish is not the language of the home.
The Draft Plan is a welcome attempt to address these and other issues in the educational
system that impact in a negative way on students‟ performance and, by inference, their life
chances.
In framing our response to the Plan, we have focused on Chapter 6, which presents detailed
proposals for assessment. We hope that our comments will be considered by the DES as it
develops its strategy to improve student performance, and we look forward to co-operating
with the DES in its efforts to ensure that students in Ireland achieve their potential in all
aspects of the curriculum, including literacy and numeracy.
1
1. Introduction
In this section, we describe the different types of assessments referred to in Chapter 6 of the
Draft Plan. We define each assessment type, summarise its purpose, and link it to the
corresponding proposals in the Draft Plan (see Table 1-1). We then identify the section of this
response where each assessment will be considered in greater detail.
Table 1-1
Types of Assessments Referred to in the Draft Plan
Type of Assessment Description Proposals in Draft Plan
Classroom Assessments Based on
Curriculum Standards
Medium-stakes assessment for
schools in which teachers rate or
grade student achievement in L1,
L2 and mathematics with reference
to national standards, using
exemplars, scoring rubrics, or other
tasks.
Assessment to be implemented at
five stages during compulsory
schooling. Results to be shared at
school level, and with Boards of
Management.
Schools Like Ours Standardised
Testing
High-stakes assessment of English
reading and mathematics for
schools in which they administer
standardised tests and submit
results to a central unit, for
comparison with other schools
sharing similar intake
characteristics.
Tests to be administered at four
points in compulsory schooling.
Results to be used at school level
for monitoring and target setting,
presented to Board of
Management, and shared with
Inspectorate. Aggregated national
results to be published.
National Assessments Low-stakes assessments of English
reading, mathematics (and
sometimes Irish) for schools and
students. Feedback on each
school‟s performance provided to
school only; system level
performance published.
Outcomes to be used to monitor
achievement of national targets
specified in the Draft Plan.
Currently administered in Second
and Sixth classes every four years.
To be extended to Fourth class
(Primary) and Second year (Post-
primary).
International Assessments Low-stakes for schools; High-
stakes for the system. Performance
reported at national and
international levels.
Participation in international
assessments to continue. Ireland
currently participates in three such
assessments – PIRLS reading
literacy (4th
class), TIMSS maths
and science (4th
), and PISA (15-
year olds attending post-primary
schools).
Classroom Assessments Based on Curriculum Standards
Currently, teachers are not required to assess student performance with reference to national
standards. However, First Steps (which is used in many schools in the SSP under DEIS)
includes an assessment component that is based on teacher judgements about student success
on key learning outcomes presented along a developmental continuum.
2
Assessment of student performance according to national curriculum standards would
represent a significant challenge for schools and teachers, since, as noted above, there is no
current requirement for this type of assessment. In 2000, the Educational Research Centre
published a framework for standards-based assessment, the Drumcondra English Profiles
(Shiel & Murphy, 2000). The Profiles cover oral language, reading and writing, and can be
used to assess pupils in primary schools (Junior Infants to Sixth class) against key curriculum
outcomes. Although the Profiles were widely disseminated, subsequent surveys of teachers‟
assessment practices suggested modest uptake (e.g., Eivers et al., 2005, where teachers of
54% of pupils in First class reported that they never used curriculum profiles). However,
usage levels need to be considered in the context of concerns raised by the Irish National
Teachers Organisation soon after dissemination about the time required by teachers to
implement the Profiles. Profiles for Irish-medium schools were published by the Centre in
2004 (O‟Siaghail & Déiseach, 2004).
In Table 1-1, we have categorised classroom assessments based on national standards as
„medium stakes‟. This is because the assessments may have some consequences for schools,
if, as proposed in the Draft Plan, schools are required to report the outcomes to parents and
school management.
We return to teacher assessments based on curriculum standards in later sections of this
response. In Section 2, we consider the use of classroom-based assessments for formative
purposes, without specific reference to curriculum-based national standards. In Section 3, we
look at classroom-based assessments linked to standards. We believe that classroom
assessments with and without curriculum standards need to be developed and implemented in
classrooms.
Standardised Tests
Norm-referenced standardised tests are widely used in primary schools. According to Eivers
et al. (2010), standardised tests of reading are administered to 95% of pupils in Second class
and to 88% in Sixth, while standardised tests of mathematics are administered to 95% of
pupils in Second and 90% in Sixth. Part of the impetus for this comes from Circular
0138/2006 (DES, 2006), which requires schools to administer standardised tests at two points
during primary school: the end of First class or the beginning of Second class, and the end of
Fourth class or the beginning of Fifth class. However, the high levels of usage at other class
levels also reflect an acceptance of standardised tests by most schools, at least for the
purposes for which they are currently used.
The Draft Plan proposes that standardised tests would be administered at four points:
Second, Fourth and Sixth classes (primary) and Second year (post-primary). The requirement
to administer standardised tests in the Second year of post-primary schooling appears to
reflect increased concerns about literacy and numeracy standards at post-primary level. Such
concerns arises despite the fact that just 1.7% of candidates achieved a grade E or lower in
Junior Certificate English in 2010, and 5.6% achieved a similar level in Mathematics (State
3
Examinations Commission, 2010a)1. However, in 2009, 17% of 15-year olds in post-primary
schools achieved at or below Level 1 on PISA reading literacy, indicating poor performance
in that domain, and 21% had a similarly low performance on PISA mathematics (Perkins et
al., 2010).
We have labelled the proposed standardised testing programme in the Draft Plan as Schools
Like Ours Standardised Testing. This is because it is proposed to gather test data from
schools and to use the data to enable schools to benchmark their average performance against
other schools with similar characteristics. We have categorised Schools Like Ours
Standardised Testing as „high stakes‟ because aggregated test results, and information about
performance targets based on the results, are to be reported to school management, and test
results are to be used by members of the Inspectorate to identify schools that are performing
above or below expectations (Draft Plan, p. 46).
In a recent report prepared for the NCCA, Shiel, Kellaghan and Moran (2010) examined a
number of options for the introduction of standardised tests of literacy and numeracy in post-
primary schools. Two key options were put forward:
Standardised tests of achievement in literacy (English/Irish) and numeracy with Irish
norms developed for the three years of lower secondary schooling, and made
available to schools to be administered when they consider it appropriate, to support
monitoring of student progress (the „Netherlands model‟)
Standardised tests developed and mandated for use at one point in lower-secondary
schooling, such as the first term of first year („the French model‟, but without central
reporting), or at the end of Second/beginning of Third year, when results might be
used for guidance purposes [within schools] (p. 198).
Significantly, the options did not include census-style every-pupil testing where results would
be submitted to a central body for further analysis and comparison with other schools.
However, it was noted that the issue of whether the results of standardised tests should be
available to the DES (in summary form), as occurs at primary level in the context of Whole
School Evaluation, might arise.
In this response, we consider the implementation of standardised testing from two
perspectives. In Section 4, we look at issues arising from the extension of standardised testing
to Fourth class (primary) and Second year (post primary) without reference to the Schools
Like Ours initiative. In Section 5, we look more closely at implications of the possible
introduction of Schools Like Ours, including the proposal to compare performance across
schools with similar student intake, and to „report on the national achievement trends in
literacy and numeracy at each of the four stages using Schools Like Ours data‟ (p. 46).
1 The corresponding percentages at Leaving Certificate level in 2010 were 2.4% for English and 8.3% for
mathematics (State Examinations Commission, 2010b)
4
National and International Assessments
The Draft Plan proposes that national assessments of achievement in English reading and
mathematics, which are currently administered in Second and Sixth classes in primary
schools, be expanded to include Fourth class (primary) and Second year (post-primary). The
national assessments can be categorised as „low stakes‟ because there are no consequences
for schools, teachers or students. As discussed in Section 6, national assessments can provide
useful information on performance levels, as well as on the performance of students in key
subgroups. Recent national assessments have also seen the development of proficiency levels
in English and mathematics, and these may be relevant for other forms of assessment
proposed in the Draft Plan such as standards-based classroom assessments.
The Draft Plan refers to Ireland‟s continued participation in international studies of reading
literacy and mathematics. Currently, Ireland participates in three such studies – two at
primary level (PIRLS and TIMSS) and one at post-primary level (PISA). Section 6 addresses
how these assessments might be integrated with current national assessments to provide a
balanced picture of performance and instructional practices across schools.
The concluding section of our response (Section 7) considers the proposals for assessment of
literacy and numeracy in the Draft Plan in their totality and identifies some possible
implications for the educational system if all of them are implemented within the proposed
timescale. In the same section, we propose an alternative timeline that allows for some of the
assessment proposals to be examined in more detail before they are implemented across all
schools in the system.
5
2. Enhancing Teaching and Learning through Classroom Assessments
We welcome the emphasis on classroom-based assessment (also termed assessment for
learning or AfL) in the Draft Plan. We agree that “gathering evidence about how well
students are learning and using this information to improve learning opportunities” is an
essential element of literacy and numeracy development (Draft Plan, p. 39). We also agree
that „good practice in assessment means using a variety of assessment methods in order to
provide a full picture of a learner‟s achievement‟ (ibid, p. 39). Indeed, we would point to the
NCCA document, Assessment in the Primary School Curriculum – Guidelines for Schools
(NCCA, 2007) as an important source that outlines strategies for integrating assessment into
teaching and learning, including strategies for assessing aspects of literacy and numeracy.
It is important, however, to draw a clear distinction between the assessments that teachers
conduct on a day-to-day basis to support teaching and learning in the classroom and the
standards-based assessment of learning proposed in the Draft Plan. Classroom-based
assessments, which teachers administer on an ongoing basis, may include:
Sharing goals, expectations, targets and success criteria with students
Analysing student talk (discourse)
Evaluating students‟ mastery of objectives
Identifying students‟ errors
Assessing students‟ prior knowledge and knowledge structures
Assessing students‟ problem-solving abilities (linguistic, schematic, strategic and
algorithmic knowledge)
Questioning students to explore their knowledge, understanding and misconceptions
Evaluating critical thinking
Providing feedback – whether norm-referenced, criterion-referenced or self-
referenced
Developing students‟ self assessment capacity
Promoting peer-assessment.
The outcomes of classroom-based assessments can be used by teachers on a daily basis to
inform teaching and learning without accountability consequences for teachers or students.
There is strong support in the research on assessment for classroom-based assessments or
AfL (Black & Wiliam, 1998), and many benefits are associated with AfL. According to
Afflerback and Cho (2011), they include the following:
AfL can contribute to enhanced student learning, as teachers accurately identify
students‟ learning strengths and needs, and teach within students‟ zones of proximal
development.
AfL conducted in classrooms can take into account the situated nature of learning
(i.e., the specific contexts in which teaching and learning occur, including students‟
states of affective and emotional development).
AfL can accommodate broad definitions of literacy (and numeracy) that include
cognitive as well as non-cognitive outcomes of instruction (self-concept, self-esteem,
attributions of performance, locus of control, motivation and attitude).
6
AfL can provide teachers with information that enhances their role as „reflective
practitioners‟ which, in turn, can contribute to better-informed teaching.
AfL can improve students‟ ability to assess their own learning through the
enhancement of their megacognitive skills. Access to the scoring rubrics that teachers
use to rate performance assessments can inform students about the criteria against
which they will be assessed, increasing the likelihood that they will internalise those
criteria (i.e., self-assess).
AfL can modify the socio-cultural space of classroom discourse. Teacher discourse
patterns may change as teachers focus on higher-level thinking and problem solving,
since traditional discourse patterns may not support assessment of such learning.
AfL could enable teachers to focus assessment efforts more fruitfully on students with
the greatest needs (e.g., children with learning difficulties, children from
disadvantaged backgrounds, English language learners).
Specific strategies for classroom-based assessments have been outlined in detail in
Assessment in the Primary School Curriculum – Guidelines for Schools (NCCA, 2007).
These include self-assessment, portfolio assessment, questioning, teacher observation etc.
These assessments have obvious implications for teachers‟ continuous professional
development (CPD), and should be included in CPD on literacy and numeracy (as proposed
in Chapter 2 of the Draft Plan). Indeed, an initial focus on classroom-based assessment in
CPD might lay a foundation for future work involving standards-based assessment as the two
types of assessment share several features. The immediate introduction of standards-based
classroom assessments, without first establishing a strong underlying knowledge base about
classroom assessment strategies among teachers, may not be successful.
7
3. Linking Classroom Assessments to Curriculum Standards
The Draft Plan includes specific proposals for the development of standards-based
assessment. It is envisaged that this will consist of two steps – developing standards based on
the curriculum, and providing supports to teachers (e.g., exemplars of performance at
different levels of achievement, Internet-based item banks) so that they can rate their
students‟ achievements against the standards (page 44). It is further envisaged that this work
will occur on a formal basis in Senior Infants, Second, Fourth and Sixth classes at primary
level, and in Second year at post-primary level. There is an expectation that teachers can
„readily, validly and reliably interpret and communicate their students‟ progress in achieving
national standards‟ (page 44).
We welcome the proposal to establish curriculum-based standards for student learning in the
context of classroom assessments and see it as something that builds on our earlier work on
the development of a structured programme for the infant classes (Gleeson, Kelly & Archer,
1988a, 1988b, 1988c) and the Drumcondra English Profiles (Shiel & Murphy, 2000).
A number of issues arise when considering the development of standards:
It is envisaged that national standards for English at primary level will have been
completed by the end of 2012 (p. 44). However, on page 29 there is a target for
review of curriculum content of L1 (English or Irish) at primary level and the
specification of learning outcomes is to be completed in the 2012-13 school year.
Clarification is needed on whether learning outcomes and national standards are one
and the same. If it is envisaged that national standards will be based on learning
outcomes, there may be a sequencing problem.
It would seem important to revise the content of L1 curricula before developing
national standards. The current L1 and mathematics curricula at primary level have
now been in place for over a decade, implementation of these curricula has been
evaluated, and information is available on their strengths and weaknesses, through
evaluation reports (e.g., NCCA, 2005, 2008; DES, 2005), and reports on national
assessments (Eivers et al., 2005; Eivers et al., 2010). The DES Inspectorate has
looked in detail at the teaching of English at post-primary level (DES, no date).
Can standards be derived from proficiency levels already established in the context of
existing assessments of English reading and mathematics? Considerable work has
already been done on the development of proficiency levels for English reading and
mathematics for Second and Sixth classes in the context of national assessments
(Eivers et al., 2010), while proficiency levels are already available for mathematics in
Fourth class (Shiel et al., 2006), and for standardised tests of mathematics that are
available to schools (Educational Research Centre, 2008). It would seem useful to
explore possible alignment between national standards and the proficiency levels that
have already been developed. The approach adopted in England has been to link
teacher assessments and standardised tests to the same underlying set of level
statements (standards). Is it envisaged to do the same here? If two different sets of
8
standards are in use (curriculum-based standards and proficiency levels based on
current curriculum-based national assessments), there is potential for confusion.
If standards are established independently of students‟ current levels of achievement,
there is a risk that the standards will be too difficult to achieve (i.e., large numbers of
students will not reach them), or too easy (almost all students will reach the highest
standards for their class level).
Moderating Classroom Assessments Based on Standards
As noted earlier, the standards-based teacher assessments proposed in the Draft Plan are
summative in nature and are medium-stakes to the extent that the results are to be shared at
school level and reported to a school‟s Board of Management. In this context, it would seem
important that a system of moderation be put in place, at least within schools or clusters of
schools, so that high levels of consistency can be achieved in rating students‟ performance.
While support materials are to be provided – exemplars representing different levels of
achievement, and a bank of Internet-based assessment resources – there will still be a need to
ensure that teachers within and across schools interpret the standards in similar ways.
It can be argued that a system of moderating teachers‟ judgements through professional
collaboration benefits teaching and learning as well as assessment (Assessment Reform
Group, 2006). Such moderation might involve planning assessments, collecting relevant data,
and analysing the data in a collaborative setting. This, in turn, could enhance teachers‟
understanding of learning goals and criteria indicating progress towards them.
9
4. Standardised Testing at Four Key Points during Compulsory
Schooling
In this section, we consider the proposed extension of standardised testing in schools, without
reference to Schools Like Ours, which is looked at in the next section. Here, the focus is on
implications of the Draft Plan for the expanded use of standardised tests within schools.
Standardised tests can serve a variety of functions. These range from evaluating the effects of
new interventions to monitoring standards across the educational system. However, it is often
difficult and sometimes impossible for the same testing programme to serve multiple
functions effectively. Standardised tests that aim to provide teachers with information on the
performance of individual students in their classes may not be suitable for monitoring
standards across schools, or evaluating the educational system.
Use of Tests within Schools
Proposals in the Draft Plan call for the extension of standardised testing in L12 and
mathematics to four points during compulsory schooling – the end of the Second, Fourth and
Sixth classes at primary level, and the end of Second Year at post-primary level – using
standardised achievement tests of literacy and numeracy (p. 45). They also refer to a
requirement for schools to report outcomes for individual students to parents and for
principals to report aggregated data to Boards of Management on an annual basis (p. 45).
CPD is to be provided to schools on how to use the outcomes to support school self-
evaluation and target setting (p. 45).
In general, these proposals are uncontroversial. Primary schools are currently required to
administer and report to parents on the outcomes of standardised tests at the end of First class
or the beginning of Second class, and at the end of Fourth class or the beginning of Fifth class
(Circular 0136/2006), while target setting is a component of planning for schools in the
Schools Support Programme under DEIS. A shift in standardised testing to the end of
Second, Fourth and Sixth classes should be accomplished without difficulty since schools
already administer standardised tests at most class levels. According to Eivers et al. (2010),
standardised tests of reading are administered to 95% of pupils in Second class and to 88% in
Sixth, while standardised tests of mathematics are administered to 95% of pupils in Second
and 90% in Sixth.
There are issues about reporting test results to Boards of Management in the case of small
schools (mainly rural primary schools), where there may be just a handful of pupils
(sometimes as few as one) enrolled at a class level. In such cases, it would be problematic to
issue standardised test result such as the percentages of students at a class level performing at
each of several proficiency levels, since the identity of pupils may be compromised.
2 We assume the Draft Plan intends that standardised testing will involve L1, whether English or Irish. While
reference is made to L1 in the context of classroom assessments based on curriculum standards, the situation is
less clear in respect of standardised tests.
10
Testing at Post-primary Level
There are no standardised tests of reading or mathematics with current national norms
available at post-primary level. Hence, it will be necessary to develop such tests, which
should be curriculum-related3. The indicative date in the Draft Plan for implementation of
standardised testing in post-primary schools (2011-12) seems over-ambitious in this respect,
as high-quality standardised tests for Second year would take two to three years to develop.
Furthermore, since a new mathematics course is currently being implemented at Junior Cycle
level (Project Maths), it may take two to three years before the content and focus of the new
syllabus is fully understood (with the exception of pilot schools, 2011-12 will be the first year
in which Project Maths will be taught in Second year). Hence, the ERC would suggest
delaying implementation of standardised testing at Second Year until at least 2013-14.
According to the Draft Plan, standardised tests at post-primary level are to be connected to
the outcomes of the Junior Cycle syllabus. However, it is unclear if key outcomes of Junior
Certificate English, especially those dealing with higher levels of text interpretation and
evaluation, can be assessed using standardised tests where, for reasons of cost and efficiency,
many if not most items must be in multiple-choice format. Perhaps clarity about this issue
will be achieved, when, as proposed in the Draft Plan, the current Junior Cycle English
syllabus is revised and specific curriculum outcomes are identified. However, it is as yet
unclear how proposed syllabus revisions will align with the overall review of the Junior
Certificate currently being conducted by the NCCA.
Age vs. Grade Norms
The Draft Plan states that scores of individual children on standardised tests should be
reported to parents with reference to „the norm for children of their age‟ (p. 45). Since it is
specified that tests would be administered at particular class levels, this suggests that
performance would be reported with respect to age norms within class levels. This would
require separate norms for each age range within a class level and, in some cases, a
renorming of existing tests. There is little obvious benefit to such a change, and the ERC
recommends that reference to age-based norms be replaced by a reference to grade-based
norms.
Target Setting
The proposal that schools would engage in self-evaluation and set targets for improvement
based on standardised tests needs some refinement:
Many primary schools are small, and numbers in Second, Fourth and Sixth classes
fluctuate dramatically from year to year. Furthermore, even if numbers stay stable
(e.g., 8 or 9 students at a grade level), the characteristics of the students may fluctuate
from year to year (different socio-economic composition; different gender
composition; different proportion of ESL students and so on), limiting the inferences
3 Some preliminary work on the development of standardised tests began at the Educational Research Centre in
October 2010.
11
that schools, teachers and Boards of Management could make, using average scores or
the percentages of students at each proficiency level. In Ontario, Canada, school-
level standardised test scores are not reported outside of schools if there are fewer
than 15 students enrolled at the grade level tested. In Ireland, this would exclude
almost 60% of primary schools.4 In the case of small schools, it may be more useful to
establish targets in respect of individual students rather than class-level cohorts.
What targets are realistic? At this time, we don‟t know how much progress the
highest-performing schools can reasonably make from year to year. Thomas, Peng
and Gray (2007) examined a GCSE data-set over a period of ten cohorts (1993-2002)
and found that only one in 16 schools managed to improve continuously for more than
four years at some point over the decade in terms of value added. Also, they found
that schools with a lower value-added starting point in 1993 were more likely to make
significant improvement.
It would seem important, as a prerequisite to providing CPD to schools on target
setting, that some research be undertaken on what schools can reasonably expect to
achieve in terms of raising scores from year to year relative to their intake, the
resources that are available to them, and whether schools can be expected to make
similar levels of improvement over several years.
The targets for literacy and numeracy on page 12 of the Draft Plan (a 5% increase at
each of Levels 3 and 4 on the National Assessments of English and Mathematics at
Second and Sixth classes between now and 2020, and a reduction of 5% in those
scoring at or below Level 1)5 seem reasonable, through it should be noted that,
between 1980 and 2004, no change was observed in overall performance in English
reading at Fifth class level (Eivers et al., 2005). Similarly, no increase in overall
performance was recorded for mathematics in Fourth class between 1999 and 2004,
despite the introduction of the revised Primary School Mathematics Curriculum
during that time (Shiel et al., 2006). At this point it is unclear what the targets mean
for individual schools, especially those schools where student numbers are small, and
performance might be expected to fluctuate from year to year.
The experience gained in setting and maintaining targets in DEIS schools should prove
helpful in achieving the refinements of the proposals in the Draft Plan that we see as
necessary.
4 In the UK, results for schools with 10 or fewer pupils are not published in league tables of GSCE results. Ray
(2006) states that "there is no easy solution to the problem of interpreting value added for small schools" but
suggests that one possibility is to combine results across years. 5 It would be important to provide a justification in the Plan for these targets. For example, why was 5% selected
rather than 10%?
12
Provision of Diagnostic Information
We would encourage the provision of diagnostic information to teachers – i.e., information
on the strengths and weaknesses of individual pupils – in the context of reporting
standardised test results. In their work on the effects of standardised tests in low-stakes
contexts, Kellaghan et al. (1980) found that more progress was made by students of teachers
who received both norm-referenced and diagnostic test results in respect of individual
students, compared with students of teachers who received norm-referenced results only, or
no results at all. With increased access to computer scoring of standardised test performance,
it should be possible to provide more detailed diagnostic information, at school and student
levels, than has been possible in the past. While schools and teachers may find that the
descriptions of achievement embedded in the proficiency levels that accompany some
standardised tests are useful in terms of identifying which tasks or processes a student is
likely to be successful on, there may be a case for generating more detailed, item-level
information as well.
We also recognise the use of tests specifically developed to serve diagnostic purposes.
Indeed, the ERC has recently completed work on the development of diagnostic tests of early
literacy and numeracy for at-risk students in Senior Infants and First class (ERC, 2010,
2011).
Potential of Computer-based Assessment
We recommend that consideration be given to assessing reading and mathematics using a
computer-based system, particularly at Sixth class and Second year. This would greatly ease
the test administration burden on schools, and have the potential to facilitate adaptive testing,
where most of the items that students are asked to respond to are at a level at which they have
a reasonable chance of being successful. This can also reduce the time required for testing
and produce more accurate scores as students would not be expected to answer questions that
are too easy or too difficult for them (something that happens when current print-based tests
are administered).
In PISA 2009, Ireland was among 19 countries that participated in a computer-based
assessment of reading. The outcomes of this (to be released in 2011) should be instructive in
illuminating similarities and differences in student performance in paper-based and electronic
models, and inform decisions on how best to proceed with computer-based assessment of
reading in the Irish context. It might also be pointed out that, over time, PISA hopes to move
fully to computer-based assessment – something that has already been done in the case of
national assessments in Denmark (see Wandall, 2008).
13
5. Contextualising the Outcomes of Standardised Test Results
The Draft Plan contains a number of proposals relating to the aggregation of standardised test
results generated in schools. The proposals include collection by an external agency of
standardised test results for all pupils at the end of Second, Fourth and Sixth classes in
primary schools, and the end of Second year in post-primary schools (p. 45). It is proposed
that schools would then benchmark their performance against other schools with similar
background characteristics, using online tools. Schools would be required to report on
„comparative performance of the school‟s students against national trends as well as
standards in schools in similar contexts‟ (p. 45). The outcomes would also be used to „report
on national achievement trends at each of four stages‟ (p. 45).
There are a number of issues that should be taken into account in considering how this
proposal might operate in practice:
In the case of English reading and mathematics, schools currently use a number of
nationally-normed standardised tests. There are obvious difficulties in comparing the
performance of schools using different tests. For example, although two tests may be
nationally-normed, one may be easier or more difficult than the other at the outset or
may become easier or more difficult over time, giving an unfair advance to some
schools and inadvertently penalising others. In the same vein, high (or low) achievers
may be at an advantage on one test, compared with the other.
Although statistical linkages may be established between two or more tests by
administering them to the same set of students and placing performance on a common
underlying scale, such a process would inevitably result in an increase in the error
associated with individual and aggregated test scores, making changes in performance
more difficult to detect. It is worth noting here that one underlying test only is used in
each target grade in National Curriculum Assessment in England, NAPLAN (National
Assessment Program – Literacy and Numeracy) in Australia and No Child Left
Behind at state level in the US, with some content changing from year to year.
Schools and teachers may become over-familiar with current standardised tests of
English reading and mathematics in the context of high-stakes testing (with
aggregated results to be reported to a central agency and to school management).
When familiarity with a test increases, teachers may inadvertently teach to the test, at
the expense of teaching well. In such circumstances, increases in performance may
not reflect actual changes in underlying proficiency.
14
Continuous vs. Categorical Social Context Measures
We have a number of concerns about the proposal to cluster schools by category (for
example, DEIS Band 2 urban schools where English is not the first language of 25% of
students – Draft Plan, p. 41):
Assuming that a measure based on average socioeconomic status (such as DEIS
status) is used to categorise schools, those schools with the lowest SES levels will
find themselves at the bottom of a category, and will inevitably find it more difficult
to achieve average performance for the category, compared with schools that are
closer to the middle of the category, or those near the top. This could have a negative
effect on morale in lower-SES schools in a category, especially if they are viewed as
performing below expectations. Another potential difficulty is that, if too crude a
measure of school socioeconomic status is used, schools of differing socioeconomic
status could be incorrectly grouped together.
The Draft Plan does not consider the use of continuous variables to adjust school
average achievement scores (e.g., an updated version of the measure used to identify
schools in terms of levels of disadvantage in the lead-in to DEIS in 2005). If a
continuous measure of school socio-economic status were available, it and other
relevant variables could be used to adjust school average scores. The ERC believes
that adjusting average scores by modelling achievement using a range of relevant
school and student-level background variables is a more accurate and reliable method
than that proposed for Schools Like Ours.
The Draft Plan does not consider the possibility of comparing the performance of
schools over time, using a relevant pre-test measure (and hence implementing a value-
added approach). There is now a significant and growing body of research on
approaches to contextualising value added change in schools (e.g., OECD, 2008) that
might be applicable to the Irish situation if suitable and current measures of school
and student background were available for all schools and students. The more fine-
grained information that would derive from value-added growth modelling could be
more useful to schools than the less sensitive category-based information that would
emerge from Schools Like Ours. 6
However, the complexity of value-added modelling
should not be overlooked either (e.g., Braun, Chudowsky & Koenig, 2010; Ray,
2006).
We return to the matter of value-added models of growth in Section 7.
Frequency of Gathering Aggregated Test Data
The Draft Plan specifies that school-level test data (and possibly contextual data) will be
gathered on an annual basis. Gathering data on such a frequent basis may result in schools
6 For example, multi-level modelling, in the context of value-added analysis, allows for an examination of
whether schools are differentially effective for different subgroups of students.
15
allocating undue amounts to time to administering and scoring tests, and sourcing
background information on students. This is time that could be allocated to other aspects of
teaching and learning, including classroom assessments (whether or not they are linked to
curriculum standards). A longer interval between data collection points (e.g., three years)
would also reduce the resource demands on schools, and perhaps sustain the testing
programme over a greater number of years. It is relevant in this regard that the Draft Plan
suggests that schools be required to establish three-year plans to improve literacy and
numeracy.
The Situation of Small Schools
As noted earlier, the situation of small schools needs to be examined carefully. In some small
schools, numbers of students are likely to fluctuate substantially from year to year, while in
others, the social composition of a cohort may fluctuate. It would be unsafe to report average
performance, progress towards reaching targets, or comparative performance for such
schools. Moreover, the challenge presented by having a large number of small schools in the
system would arise in model-based approaches to tracking performance changes in the same
way that it would arise in Schools Like Ours.
A related issue concerns the treatment of students who enrol in a school during the year of
testing. Should they be included in school average scores, given that their prior schooling
occurred elsewhere, and, especially in the case of smaller schools, could they have a
disproportionate influence on average scores?
Exemptions from Testing and Provision of Support to Students with Disabilities
The Draft Plan is unclear on who will be tested, and who will be exempted from testing (i.e.,
excluded from aggregate scores submitted by schools). A clear policy on this issue needs to
be developed, as does a policy on dealing with students who are absent on the date of testing.
At post-primary level, between 80% and 85% of students selected to participate in the OECD
Programme for International Student Assessment (PISA) in Ireland present themselves on the
day of the test. The vast majority of non-presenting students are absent on the day of testing
and we know that on average their achievement on English, mathematics and (where
applicable) science on the Junior Certificate examination is lower than that of students who
present. At primary level, the situation is somewhat better, with participation in national
assessments generally in excess of 90% (Eivers et al., 2010, Table 2.9). Many testing
programmes similar to that proposed in the Draft Plan require a participation rate in excess of
95% so that results across schools are directly comparable. In some programmes (e.g., No
Child Left Behind in US states), there is a requirement to assess students with learning
disabilities and students for whom English is not the language spoken at home.
Related to the exemptions issue is whether second-level students with severe learning
difficulties are to be provided with the same types of supports they can access when they take
the Junior Certificate Examinations. Will schools have discretion in determining what levels
of support are to be given? If so, how can we know that there is consistency from school to
school, and over time?
16
Extent of Deviation of Schools from Expected Performance Levels
Proposals for Schools Like Ours are based on identifying schools that display average
achievement that is significantly above or below what would be predicted on the basis of
their socioeconomic composition. Currently, we cannot quantify this phenomenon. An
analysis completed in the context of the 2009 National Assessment of Mathematics and
Reading (Eivers et al., 2010) shows that, when error bands are established around a fit line
defining the relationship between socioeconomic status and reading, the majority of schools
fall within the error bands, with just a handful of schools outside. Further work could be
undertaken using existing datasets (Junior Certificate grades, DEIS data) to establish the
extent to which schools deviate from expected performance, based on socioeconomic status
or other intake factors. The following questions could inform the research:
What proportion of schools „underperform‟ or „over-perform‟ in reading literacy and
mathematics, after adjusting for the socioeconomic status of the school?
Do these schools consistently under- or over-perform in different areas (subjects) and
over time?
What are the characteristics of „under-performing‟ and „over-performing‟ schools7?
Use of Schools Like Ours Data to Report on National Trends
Finally, we would caution against using Schools Like Ours standardised test results to report
on achievement trends in literacy and numeracy, as suggested on page 46 of the Draft Plan.
First, the national assessments (described in Section 6) are intended to report on achievement
trends in these important areas. Current national assessments undertaken by the ERC share
the following characteristics:
Detailed assessment frameworks
Test items designed to assess a broad range of skills
A balance between multiple-choice and constructed response items
Clear rules on exclusion of students from testing
Monitored administration of the tests in schools
Use of secure tests that are not available to schools
Translation of test materials into Gaeilge
Controlled scoring of constructed response items to ensure reliability.
Use of complex statistical procedures to link test performance over time.
Extensive quality control including independent review of statistical analyses by an
international consultant
Support in interpreting outcomes from a national advisory group
Schools Like Ours standardised testing would not have any of these features. Therefore, it
would be unsafe to publish aggregated national outcomes, or to use those outcomes to draw
inferences about national performance standards.
7 An exercise of the type described here in relation to the population of schools in general is planned as the next
phase in the evaluation of the SSP under DEIS, focusing on schools in that programme.
17
Impact of Whole Cohort Testing
In thinking about the scale of the testing programme implied by Schools Like Ours, it is
instructive to note that there has been a growing realisation in England that full-cohort testing
(such as that proposed under Schools Like Ours) represents a significant burden to the
educational system (House of Commons Select Committee on Children, Schools and
Families, 2008). Indeed, full-cohort testing no longer applies to the end of Key Stage 3 (age
14) in England. Instead, schools now submit the results of moderated classroom-based
assessments to a central unit.
18
6. National and International Assessments of Educational Achievement
The Educational Research Centre has been involved in implementing national assessments of
educational achievement on behalf of the Department of Education and Science since the
early 1980s. In 2009, national assessments in English and mathematics were administered to
representative national samples of pupils in Second and Sixth classes, with results reported in
November 2010 (see Eivers et al., 2010). In 2010, the 2009 assessments were administered
to representative samples of students in Scoileanna Lán-Ghaeilge and in Gaeltacht schools. A
report on this study will be ready later in 2011.
In general, national assessments are used to monitor the performance of the educational
system as a whole. This is reflected in the aims of the 2009 national assessments:
To establish current reading and mathematics standards of Second and Sixth class
pupils;
To provide high-quality and reliable data to assist the DES in policy review and
formulation, and in decisions regarding resource allocation;
To examine school, teacher and home background and pupil characteristics, and
teaching methods which may be related to reading and mathematics achievement;
To provide a basis against which to compare outcomes of future national assessments
of English reading and mathematics at Second and Sixth classes (Eivers et al., 2010,
p. 1).
Since 2009 was the first year in which students were assessed in Second and Sixth classes, it
was not possible to interpret the outcomes with respect to trend data. However, a structure
has been established so that future national assessments can link back to performance in
2009.
According to the Draft Plan, it is proposed to continue with national assessments of reading
and mathematics in a representative sample of schools on a four-yearly basis (p. 46). It is also
proposed to extend the national assessments to Fourth class at primary level and Second year
at post-primary level (p. 12). If this occurs, there will be two parallel testing programmes
designed to gather national level data. In one, Schools Like Ours. aggregated data based on
standardised tests administered by teachers would be collected and submitted to a contractor
of the DES in respect of all pupils in the Second, Fourth and Sixth classes at primary level,
and Second year at post-primary level. In the other (National Assessments), broadly similar
data would be generated at the same class levels every four years, albeit in a sample of about
150 schools rather than across the population of students at these class levels (Table 6-1).
There is a danger that schools might not be able to accommodate the two testing programmes
on those years in which they overlap (every four years according to the Draft Plan). Since one
(Schools Like Ours) is a high-stakes testing programme, with results to be reported to Boards
of Management, and to members of the inspectorate engaged in school evaluation, and the
other (the National Assessments) is low stakes, schools may be under too much pressure to
19
engage fully with the latter. Again, the frequency of Schools Like Ours testing might be
reconsidered to ensure that schools are not over-burdened, and have the time and energy to
engage with other assessments as well.
Table 6-1: Comparison of Schools Like Ours Standardised Testing and
National Assessments
Schools Like Ours National Assessments
Purpose To hold schools accountable for
improving achievement in reading
and mathematics by comparing their
performance to other similar schools,
and setting targets in the context of a
three-year plan.
To monitor achievement in English
reading and mathematics over time
relative to targets specified on page 10 of
the Draft National Plan.
Domains assessed English reading and mathematics English reading and mathematics
Stakes High-stakes, with school-level
outcomes reported to Boards of
Management, Inspectorate
Low stakes, with system-level outcomes
reported
Frequency Annually Every four years
Grade levels involved 2nd, 4th, 6th Primary, 2nd Year post-
primary
2nd, 4th, 6th Primary, 2nd Year post-
primary
National reporting Report nationally-aggregated data on
performance standards
Report nationally-aggregated data on
performance standards.
Proposed CPD and
support
CPD opportunities and information
to teachers and Boards of
Management on the use of
aggregated data from standardised
tests to enable reflection on the
impact of their teaching and learning
None proposed
On page 39 of the Draft Plan, it is noted that „schools also have access to reports published by
the ERC on national assessments of reading and mathematics‟. Our view is that there would
be value in including detailed information on the types of texts and items used in national
assessments, and on the outcomes, as part of the CPD in literacy and numeracy for teachers.
This occurs in New Zealand, in the context of tests administered as part of the National
Monitoring Project. It ensures that teachers are familiar with the purpose and uses of the
national assessments and that they can use released materials to compare the performance of
their students against national performance in a low-stakes context (International Reading
Association, 2010).
International Assessments
In spring 2011, Ireland will participate in two international studies of educational
achievement – the Progress in International Literacy Study (PIRLS) and the Trends in
International Mathematics and Science Study (TIMSS). Rather than expanding the national
20
assessment programme to include Fourth class, there may be value in participating in PIRLS
and TIMSS on a more regular basis than has been the case in the past (these studies are
implemented by the International Association for the Evaluation of Educational Achievement
every four or five years, but the last time Ireland participated was in 1995). Regular
participation in PIRLS and TIMSS would provide valuable comparative information against
which to benchmark the performance of Fourth class students in Ireland, and to verify if any
increases in performance on tests of literacy and numeracy administered by schools are
transferred to international tests and vice versa8. It might also be pointed out that access to the
outcomes of international assessments such as PIRLS and TIMSS could inform curriculum
development and renewal as PIRLS and TIMSS compare the curricula of participating
countries in a systematic way (e.g., Kennedy, Mullis, Martin & Trong, 2007).
Reviews of the Irish outcomes of PISA 2009 revealed some issues relating to the procedures
used by the OECD and its contractors to scale PISA reading literacy across assessment cycles
(Cosgrove et al., 2010; LaRoche & Cartwright, 2010). It would seem important to see that
these technical issues are addressed in future PISA cycles, so that greater confidence can be
placed in trend results issued by the OECD. Commentary in the media and elsewhere on the
outcomes of PISA also supported the view that there may be a gap between what
assessments like PISA actually do, and what people think they can do.
There is some evidence that the outcomes of PISA have had an impact on curriculum
development at junior cycle in Ireland (see Conway & Sloane, 2005 in respect of
mathematics and DES, 2003 in respect of science). There is value in considering implications
of the PISA reading literacy framework in reviewing and revising the Junior Certificate
English syllabus, particularly the recent widening of the concept of reading literacy in PISA
to include both print-based and electronic reading (OECD, 2009).
8 An evaluation of the benefits of participating in PIRLS and TIMSS in Fourth class could be done after the
reporting stage of PIRLS and TIMSS 2011.
21
7. Conclusion
While we are supportive of many of the proposals for assessment in the Draft Plan, we have
concerns about the schedule for implementing new assessment procedures, and believe that
some of the proposals for assessment should be the subject of further study before
implementation across all schools.
In Table 7-1, we have outlined the time scale for the introduction of all the assessments
initiatives proposed in Chapter 6 of the Draft Plan. We are not sure that the system can
sustain such a large number of additional assessments at this time, and suggest a more
gradual approach to introducing these measures. We also have reservations about the
methodology proposed for Schools Like Ours, and believe that some analysis work, mainly
using existing data sources, should be done before implementing Schools Like Ours.
The Centre proposes a more gradual introduction of assessment initiatives, with time
available to study the impact of some measures in a subset of schools before extending them
to all class levels and all schools:
Expand the programme of national assessments of English reading and mathematics
for the 2013 cycle. Some development work will be required if the assessments are to
be extended to Fourth class (where a secure English reading test is not currently
available) and Second year in post-primary schools (though the Educational Research
Centre has already begun work on tests for Second years).
Although national assessments involving samples of schools and students are low
stakes, they have the potential to provide valuable information on the functioning of
the educational system and on how students interact with the reading and mathematics
tasks they are assigned. We would like to see more information (outcomes, examples
of tasks etc.) from the national assessments featuring in CPD for teachers.
The scope of national assessments at some class levels might be broadened to look at
non-test outcomes, such as the teaching of English and mathematics in classrooms.
For example, a video-study of a subset of participating schools could provide valuable
quantitative and qualitative information on teaching methods in classrooms that could
then be disseminated to all teachers.
As part of the national assessments, the introduction of computer-based testing for
students in Second-year post-primary should be considered, with an extension of this
to other class levels in later cycles.
A study should be conducted that examines in more detail some of the assumptions
underlying the Schools Like Ours initiative. Some of the data required for such a study
already exist (e.g., Junior Certificate examination results9 for the same schools for
consecutive years, data on the same DEIS schools for 2007 and 2010). Additional
9 It might be noted that Junior Certificate results have the potential to provide far more detailed information
on student performance than standardised test results. In the JC examinations, students are assessed over a far broader range of outcomes than on a typical standardised test.
22
data, if needed, could be obtained by administering the national assessments in
consecutive years in a subsample of primary schools.
Table 7.1 Indicative Timeline for Implementation of Assessment Proposals in Draft Plan*
Standards-based
Assessments
Administered by
Teachers
Schools Like Ours
Standardised
Testing
National
Assessments of
English reading and
mathematics
International
Assessments
2010-2011 PIRLS and TIMSS to
be implemented in
Fourth class in
primary schools
2011-2012 Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary.
PISA 2012 (Major
domain: Mathematics)
15-year olds Post
Primary
2012-2013 Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary.
Implementation of
National Assessments
in Second, Fourth and
Sixth classes
(Primary) and Second
Year (Post-primary)
(not yet confirmed)
2013-2014 Implementation of
English Standards-
based Assessments
(Primary) (Senior
Infants, 2nd, 4th, 6th
classes)
Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary. Collection of
aggregated data –
English (Primary)
2014-2015 Implementation of
English and
Mathematics
Standards-based
Assessments
(Primary) (Senior
Infants, 2nd, 4th, 6th
classes)
Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary. Collection of
aggregated data –
English & Maths
(Primary)
TIMSS in Fourth
class in primary
schools (Participation
not confirmed);
PISA 2015 Main
Domain: Science. 15-
year olds Post
Primary
2015-2016 Implementation of
English and
Mathematics
Standards-based
Assessments
(Primary) and English
(Post-primary)
Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary. Collection of
aggregated data –
English & Maths
(Primary); English
(Post-primary)
PIRLS in Fourth class
in primary schools
(Participation not
confirmed)
2016-2017 Implementation of
English and
Mathematics
Standards-based
Assessments (Primary
and Post-primary)
Implementation of
Testing in 2nd, 4th,
6th Classes Primary
and 2nd year Post-
primary. Collection of
aggregated data –
English & Maths
(Primary); English &
Maths (Post-primary)
Implementation of
National Assessments
in Second, Fourth and
Sixth classes
(Primary) and Second
Year (Post-primary)
(not yet confirmed)
*Excludes piloting, test development, field trialling etc.
23
Some of the issues that could be addressed are variation in performance in schools
(especially smaller schools) over a two or three year period; numbers of schools that
deviate significantly from the average performance of other similar schools; average
expected progress in different school types; and, where available, information on the
effects of standard setting on performance and the engagement of different
stakeholders (school management, teachers, students, parents and community) in
supporting improvement.
Using the same databases, value-added studies should be conducted to examine the
effects of adjusting average school achievement scores using a continuous measure of
school average socioeconomic status and other relevant contextual variables10
. Is
multi-level modelling involving both school and student variables better than single-
level modelling that involves adjusting school average scores using school-level
average socioeconomic status only? How much error is associated with year-on-year
measures of value added using different approaches? Are there any additional benefits
to be had by measuring value-added differences from one year to the next, compared
with a longer interval such as three years? Should school scores be averaged over a
number of years? How can the difficulties associated with small cohort sizes be dealt
with within a value-added framework (for example, can or should clusters of adjacent
small schools be collapsed into a single group)?
Examine how new measures of social context that can be updated regularly might be
developed. Summary school scores on the DEIS contextual variables were computed
in 2005 and are probably now out of date in respect of some schools. If schools are re-
evaluated for participation in DEIS or other such initiatives in the next few years, it
would be important to examine how a measure of school social context that can be
updated on a more regular basis could be constructed. A primary pupil database,
updated on an annual basis, would yield enhanced records of pupil characteristics that
could be anonymised and then used in analysing school-level performance. Other
methods, such as use of street addresses to provide an indicator of socio-economic
status, can be expected to emerge over time.11
Starting in 2011-12, design a programme for strengthening teachers‟ classroom-based
assessments. Begin implementation of the programme, perhaps as part of CPD in
literacy and numeracy, in 2012-13. While such a programme might initially focus on
general assessment skills (such as those outlined in Section 2 of this response, and
elaborated on in Assessment in the Primary School Curriculum – Guidelines for
Schools), it could be broadened over time to include outcomes of national assessments
as well as implementation and interpretation of classroom assessments based on
national standards.
10
The Centre has already done some relevant analyses. 11
The work of the group put in place by the Social Inclusion Unit of the DES is relevant here.
24
References
Afflerbach, P., & Cho, B. Y. (2011). The classroom assessment of reading. In M. L. Kamil, P. D.
Pearson, E.B. Moje & P. P. Afflerbach (Eds.), Handbook of reading research (Vol. 4), 487-
514. New York: Routledge.
Assessment Reform Group. (2006). The role of teachers in the assessment of learning. Cambridge:
Author.
Black P.J. & Wiliam. D. (1998), Assessment and classroom learning. Assessment in Education, 5(1),
7-74.
Braun, Chudowsky, & Koenig, (2010). Getting value out of value-added: Report of a workshop.
Washington, DC: National Academic Press.
Conway, P. F. and Sloane, F. C. (2006) International Trends in Post-Primary Mathematics
Education: Perspectives on Learning, Teaching and Assessment. Dublin: National Council for
Curriculum and Assessment.
Cosgrove, J., Shiel, G., Archer, P., & Perkins, R. (2010). Comparisons of performance in Ireland on
PISA 2000 and PISA 2009: A preliminary report to the Department of Educational and Skills.
Dublin: Educational Research Centre.
DES. (Department of Education and Science). (2005). An evaluation of curriculum implementation in
primary schools: English, mathematics and visual arts. Dublin: Government Publications.
DES. (2006). Circular 0138/2006. Supporting assessment in primary schools. Dublin: Author.
DES. (2003). Junior Certificate science syllabus. Ordinary level and higher level. Dublin: Stationery
Office.
DES. Inspectorate. (No date). Looking at English: Teaching and learning English at post-primary
level. Dublin: Author.
Educational Research Centre. (2008). Drumcondra primary mathematics test-revised. Progress in
mathematics: Using class-level performance bands and monitoring progress over time.
Dublin: Author. http://www.erc.ie/documents/dpmt_combined_documents.pdf
Educational Research Centre. (2010). Drumcondra tests of early literacy – screening and diagnostic.
Dublin: Author.
Educational Research Centre. (2011). Drumcondra tests of early numeracy – screening and
diagnostic. Dublin: Author.
Eivers, E., Shiel, G., & Shortt, F. (2004). Reading literacy in disadvantaged primary schools. Dublin:
Educational Research Centre.
Eivers, E., Shiel, G., Perkins, R., & Cosgrove, J. (2005). The 2004 National Assessment of English
Reading. Dublin: Educational Research Centre.
Eivers, E., Close, S., Shiel, G., Millar, D., Clerkin, A., Gilleece, L., & Kiniry, J. (2010). The 2009
National Assessments of mathematics and English reading. Dublin: Stationery Office.
Accessed at: http://www.erc.ie/documents/na2009_report.pdf
25
Gleeson, J., Kelly, J., & Archer, P. (1988a). A structured program for infant classrooms in
mathematics, reading and oral language. Volume 1 – Manual for junior infants. Dublin: The
Department of Education in association with the Educational Research Centre.
Gleeson, J., Kelly, J., & Archer, P. (1988b). A structured program for infant classrooms in
mathematics, reading and oral language. Volume 2 – Manual for senior infants. Dublin: The
Department of Education in association with the Educational Research Centre.
Gleeson, J., Kelly, J., & Archer, P. (1988c). A structured program for infant classrooms in
mathematics, reading and oral language. Volume 3 – Manual for first class. Dublin: The
Department of Education in association with the Educational Research Centre.
House of Commons. (2008). Third Report of the Select Committee on Children, Schools and Families.
UK Parliament, May.
International Reading Association. (2010). Standards for the assessment of reading and writing.
Newark, DE: Author.
http://www.reading.org/general/currentresearch/standards/assessmentstandards.aspx
Kellaghan, T., Madaus, G.F., & Airasian, P.W. (1980). Standardised testingin elementary schools:
Effects on schools, teachers, and students. Washington DC: National Institute of Education,
US Department of Health, Education & Welfare.
Kennedy, A.M., Mullis, I.V.S., Martin, M.O. & Trong, K.I. (2007). PIRLS 2006 Encyclopaedia: A
guide to reading education in 40 PIRLS 2006 countries. Chestnut Hill, MA: TIMSS and
PIRLS International Study Centre.
LaRoche, S., & Cartwright, F. (2010). Independent review of the 2009 PISA results for Ireland:
Report prepared for the Educational Research Centre at the request of the Department of
Education and Skills. Dublin: Department of Education and Skills.
National Council for Curriculum and Assessment (NCCA). (2005). Primary curriculum review.
Phase 1. Report. Dublin: Author.
National Council for Curriculum and Assessment. (2007). Assessment in the primary school
curriculum – Guidelines for schools. Dublin: Author.
National Council for Curriculum and Assessment (NCCA). (2008). Primary curriculum review.
Phase 2. Report with recommendations. Dublin: Author.
Ó Siaghail, G., & Deiseach, C. (2004). Próifílí measúnachta don Ghaeilge sna scoileanna Gaeltachta
agus scoileanna lán-Ghaeilge. Lámhleabhar. Baile Átha Cliath: Foras Taighde ar Oideachas.
OECD. (Organisation for Economic Cooperation and Development). (2008). Measuring
improvements in learning outcomes: Best practices to assess the value-added of schools.
Paris: Author.
OECD. (2009). PISA 2009 Assessment framework: Key competencies in reading, mathematics and
science. Paris: Author.
Perkins, R., Moran, G., Cosgrove, J., & Shiel, G. (2010). PISA 2009: The performance and progress
of Irish 15-year olds in Ireland. Dublin: Educational Research Centre.
Queensland Studies Authority. (2009). Student assessment regimes: Getting the balance right for
Australia.
http://www.qsa.qld.edu.au/downloads/publications/qsa_paper_assess_balance_aust.pdf
26
Ray, A. (2006). School Value Added Measures in England: A Paper for the OECD Project on the
Development of Value-Added Models in Education Systems. London: Department for
education and Skills.
Shiel, G., & Murphy, R. (2000). Drumcondra English profiles. A framework for assessing oral
language, reading and writing in primary schools. Dublin: Educational Research Centre.
Shiel, G., Surgenor, P., Close, S., & Millar, D. (2006). The 2004 National Assessment of Mathematics
Achievement. Dublin: Educational Research Centre.
Shiel, G., Kellaghan, T., & Moran, G. (2010). Standardised testing in lower secondary education
(Research report no. 12). Dublin: National Council for Curriculum and Assessment.
http://www.ncca.ie/en/Publications/Reports/Standardised_Testing_In_LowerSecondary_Educ
ation.pdf
State Examinations Commission. (2010a). State examination statistics – Provisional results for Junior
Certificate 2010. http://www.examinations.ie/statistics/statistics_2010/JC_619_2010.pdf
State Examinations Commission. (2010b). 2010 Leaving Certificate statistics.
http://www.examinations.ie/statistics/statistics_2010/gender_ard_2010_excluding_less_than_
10.pdf
http://www.examinations.ie/statistics/statistics_2010/gender_gnath_2010_excluding_less_tha
n_10.pdf
http://www.examinations.ie/statistics/statistics_2010/gender_bonn_2010_excluding_less_than
_10.pdf
Thomas, SM, Peng, W-J & Gray, J. (2007). Modelling patterns of improvement over time: value
added trends in English secondary school performance across ten cohorts. Oxford Review of
Education, 33 (3), 261-295.
Wandall, J. (2008). National tests in Denmark. CAT as a pedagogical tool. In F. Scheuermann & J.
Björnsson (Eds.), The transition to computer-based assessment: New approaches to skills
assessment and implications for large-scale assessment. ISPRA, Italy: European Commission
Joint Research Centre/Institute for the Protection and Security of the Citizen.