Advantages and Down sides of System Testing

Keywords: features of system evaluation, education system testing

Assessment and reporting will be the means by which learning can be monitored and stakeholders can be enlightened of achievements. Its educational aspect perceives results used to recognize strengths and weaknesses and improve future learning, and its own instrumental aspect includes grouping of students matching to accomplishment. Parents, teachers and students are thinking about its educational function, whereas external stakeholders such as governments are worried with the instrumental aspect.

Movement towards a worldwide and digital current economic climate has necessitated skilled and competent school leavers, crucial for Australia's public and economic wealth. Government authorities therefore require classes to demonstrate university student achievement at satisfactory levels to justify their economic support. This accountability also ensures the community is aware of provision of financing and services to classes. To provide these details, evaluation must be performed on a nationwide scale. As the information required differs to that required in the class, strategies for assessment differ in design, execution and reporting. Country wide assessment must be inexpensive, fast and externally mandated, and results must be clear and accessible.

Herein lie the problems with national tests. Authentic assessment is becoming popular in the school room, testing real-life encounters and functional knowledge over numerous evaluation tasks. In contrast, national tests determine students on one occasion and rely over a 'pen-and-paper' method of deliver, resulting in debate over validity.

Benefits of system-wide testing

Over the past 40 years, international and countrywide screening has increased greatly. While early implementation assisted selection of students for advanced schooling, more recent national assessment can be used to judge curriculum execution. As different curricula operate throughout Australia and internationally, benchmarking has been developed to facilitate comparisons between countries or students and identify talents and weaknesses.

In Australia, the Country wide Examination Program (NAP) contains twelve-monthly NAP literacy and numeracy (NAPLAN), and three yearly test assessments in research literacy, civics and citizenship, and information and communication technology literacy. Most controversy surrounds NAPLAN, hence it will be reviewed further.

NAPLAN proceeds under course of the Ministerial Council for Education, Early on Youth Development and Youth Affairs (MCEECDYA, recently MCEETYA) and it is federally funded. It was developed to check skills "essential for each child to progress through school and life". Each year, all students in Years 3, 5, 7 and 9 are assessed in reading, writing, terms conventions and numeracy. NAPLAN endeavours to provide data enabling Government to
  • analyse how well colleges are performing
  • identify institutions with particular needs
  • determine where resources are most had a need to lift attainment
  • identify best practice and innovation
  • conduct national and international comparisons of techniques and performance
  • develop a substantive facts base on what works.

NAPLAN claims to do this by collecting a breadth of information that can't be obtained from class assessment. Government benefits from research on such large data examples: benefits for organizations including men/females, Indigenous and low socio-economic status students offer an evidence-base to see coverage development and learning resource allocation.

Comparing specific students to others in their talk about, and countrywide benchmarks provides comprehensive information for educators to see future learning. Individual students may also be 'mapped' over time, to identify areas of improvement or those needing intervention. Furthermore, national testing assists students moving colleges for the reason that it allows immediate identification of the learning level by their new university.

Strict guidelines encompass reporting of results to ensure benefits are gained. THE FEDERAL GOVERNMENT have committed to ensuring that general population reporting
  • focuses on improving performance and learner outcomes
  • is both locally and nationally relevant
  • is timely, regular and comparable.

If NAPLANs implementation follows these rules, it will provide great profit to Australia. Yet, in these first stages of implementation, it's important to consider the stressed experience of other countries regarding national testing.

Lessons to be learnt

National analysis was presented in Great britain in 1992 to determine national goals for education. Students are evaluated at ages 7 and 11 in British and mathematics, with 14 also in technology. The 'no child kept behind' legislation was carried out in the USA in 2001 to reduce the disparity between high and low ends of student accomplishment, focussing on literacy and numeracy. Students are evaluated yearly between Calendar year 3 and 8, as soon as between Time 9 and 12. Email address details are analysed on the basis of socioeconomic and cultural background, and shared as school category dining tables by the press. Federal funding is associated with school performance. The common issues with both circumstances will be discussed below.

Being a topical ointment issue, nearly all literature on national tests is highly biased for the author's judgment. However if or when these results occur, they have got the capability to negatively effect on students. As such, they also have to be considered within the Australian framework.

Narrowing of the curriculum

With funding linked to success, professors are appreciated to ensure students achieve the best effect possible in assessed subjects, and can finish up 'coaching to the test'. Those educators who 'produce' successful students using this strategy are rewarded, deepening the situation. Within assessed topics, increased category time is spent teaching students to adopt lab tests and increasing give attention to tested areas, leading to reduced emphasis on skills such as creativeness and higher order thinking. Furthermore, time spent on subjects not analyzed is low in preference for the ones that are. This type of teaching has been labelled 'defensive pedagogy' and contributes to narrowing of the curriculum.

Excluding low-achieving students

Reports suggest that some low-performing students are excluded from enrolment or suspended during screening to improve college performance. In one example, students with low ratings were avoided from re-enrolling, but were officially labelled as having withdrawn. Compounding this result, successful institutions then have further capacity to choose students, leading to a widening gap between low and high performing institutions; in immediate opposition to the reasons for implementing nationwide assessment.

disregarding high-achieving students

High-achieving students can be adversely affected, as much results are reported only as ratio achieving benchmarks. Top priority is therefore directed at students just underneath benchmarks to ensure they reach them. It has been described as expanding 'cookie-cutter' students, all with similar skills. By doing this, students achieving above benchmarks aren't challenged, reducing drive and triggering disengagement.

Lowered self-esteem

In one analysis, for the three years after national assessment was implemented student self-esteem was significantly reduced compared to students the previous 2 yrs. Furthermore, attainment in countrywide checks correlated with self-esteem, recommending that both pressure of testing and the student's accomplishment can impact self-esteem.

Increased drop-out rates

When in comparison to schools of similar socio-economic qualifications but without national testing, a significant increase in Calendar year 8-10 students dropping out of institution was observed. This may be linked to pressure to suspend students or reduced self-esteem and motivation associated with high-stakes assessment.

Reporting of category tables

National testing results are often reported as 'league tables', delivering average ratings allowing direct comparison between universities. However, results tend to reflect socio-economic status alternatively than true accomplishment, and the depiction of schools as successes or failures contributes to even more inequity between socio-economic teams. Importantly, the dining tables give no information regarding the cause of low accomplishment or opportinity for improvement, and therefore do not fulfil their planned purpose.

Recent trends have observed publication of 'value-added' data, adjusted for socio-economic position, however the ways of calculation aren't explicit, hence their benefit is debatable.

disparity from class room assessment

Classroom assessment is becoming increasingly real, with students being assessed on real-world responsibilities, giving them the perfect chance of demonstrating knowledge and skills. The usage of national tests opposes this model, evaluating students on one solo occasion and going out of teachers uncertain as to appropriate pedagogy. Results obtained during school room inspections of traditional styles of analysis have been proven to differ to people from national testing, resulting in questioning over validity.

Ensuring consistency and validity in Australia

The issues detailed above need to be considered to ensure reliability and validity of countrywide testing in the Australian framework.


Reliability refers to consistency of evaluation, where results ought to be the same irrespective of when, where and the way the assessment was used and marked. The principal issue is marking uniformity throughout Australia. Information technology facilitates accurate marking of simple answers, and Newton implies computer-based rating algorithms for constructed responses also improve reliability. Moderation ensures all assessors use the same strategies, and marking by more than one person could also improve reliability. Moderation also assists in preserving threshold levels as time passes.


Validity refers to the assessment screening what it was designed to test.

Construct validity: diagnosis is relevant, significant and fair and provides correct information about scholar knowledge

Content validity: assessment is associated with a specific curriculum outcome

Consequential validity: analysis does not bring about a specific group of students consistently doing poorly

Concurrent validity: students obtain similar results for similar responsibilities.

Debate develops over the capacity of national examination to show real-world tasks in significant contexts, or deep thinking and problem handling. With diverse cultural and dialect backgrounds, Australian students bring to school a number of experiences and values and illustrate learning differently. The solitary occasion, 'pen and newspaper' style delivery of countrywide testing will not capture this variety and can result in anxiety.

This is evident specifically for students from Indigenous and low socioeconomic backgrounds. One tutor advised that the examination is overwhelming, and skills valued in their culture aren't seen as relevant. The concept of silent and specific examination is international due to their social value of cooperation, and the numeracy assessments are unfair because of their low English literacy (G. Guymer, personal communication, Apr 2011). Much time is spent instructing students how to complete forms, reducing teaching time already tied to low attendance.

The aspiration for equality in Australian education is visible. However, this facts suggests that alternatively than 'closing the distance' national trials may actually be increasing it.

Reporting of results

In the past, rather than submitting league dining tables Australia has 'value-added' to data by grouping classes with similar characteristics, to trail individual students, and identify colleges in need. However, this grouping has been challenged, as each school is essentially unique. To handle this, the My College website was released in 2010 2010 (http://www. myschool. edu. au), posting a school account including home elevators staffing, facilities and financial resources. The NAPLAN email address details are reported for every school against countrywide averages as well as against 60 universities with similar socio-economic characteristics throughout Australia.

Using results to improve learning

Despite the overwhelming negative reaction to national evaluation, it is improbable to disappear. Consequently, using results to improve learner learning is the best response. Some methods used effectively are defined below.

Diagnostic application

Although not made for the purpose, results can be used to identify talents and weaknesses for folks or groups of students. By analysing specific questions, common problems can be identified and inadequacies in pondering inferred. In doing this, national examination results can be utilized as a formative diagnosis to guide future teaching.

As NAPLAN is undertaken every 3 years, results for individual students may also be analysed over time to recognize improvement or decline.

Consistency of schooling

Together with the Country wide Curriculum, results from NAPLAN will ensure students have the same schooling across Australia. This will reduce difficulty associated with students changing colleges, as their achievement level will be immediately accessible.

Incorporation of content in the classroom

NAP assessment jobs will be based on Country wide Curriculum content once carried out. As students will be exposed to content during course, national testing shouldn't pose an extra burden for educators. Educators at Ramingining College ensure all worksheets include question formats just like those on NAPLAN assessments, and in key school checks are undertaken each week in British or mathematics under test conditions (G. Guymer & B. Thomson, personal communication, April 2011). The institution therefore does explicitly train students to consider the test.

Allocation of financing and resources

Arguably the most important outcome of countrywide assessment is to 'identify universities with particular needs' and 'determine where resources are most had a need to lift attainment'. Appropriate distribution of funding and resources will mean NAPLAN has shipped on these pledges. In turn, there must be a 'shutting of the distance' between low and high attaining schools, and a decrease in many of the issues reviewed.

Hopefully, execution of the National Curriculum will support the purposes of NAPLAN, alongside one another resulting in equality for young Australians.

Also We Can Offer!

Other services that we offer

If you don’t see the necessary subject, paper type, or topic in our list of available services and examples, don’t worry! We have a number of other academic disciplines to suit the needs of anyone who visits this website looking for help.

How to ...

We made your life easier with putting together a big number of articles and guidelines on how to plan and write different types of assignments (Essay, Research Paper, Dissertation etc)