KM-05,-Learner Guide
KM-05,-Learner Guide
KM-05,-Learner Guide
LEARNER GUIDE
1
Name
Contact Address
Telephone (H)
Telephone (W)
Facsimile
Cellular
2
Note to the learner
This Learner Guide provides a comprehensive overview of the module. It is designed
to improve the skills and knowledge of learners, and thus enabling them to
effectively and efficiently complete specific tasks.
Skills Rationale
Assessment skills in practice is the basis of many other qualifications and skills sets
and is a skill that leads to employability. This skills programme will equip people to
be able to supervise performance and provide feedback in a professional manner.
There is a single unit standard 115753 Conduct Outcomes Based Assessment, Level
5, 15 Credits registered on the NQF which will expire 2023-06-30 which is used to
train assessors currently.
Assessment Practitioner status can lead to self-employment
opportunities. Typical learners include post school learners and college graduates
on
NQF Level 4 or equivalent with work experience that enables them to observe
learners in practice and make judgements on competence and quality in skills
deliver. This skills programme is for people with an interest in using assessment
tools to promote learning.
Entry Requirements
NQF Level 4 with Communication
Quality Assurance
QCTO will facilitate the assessment and quality assurance
Physical Requirements
3
Learning and development resources aligned to the scope and content of the
module
Learner assessment resources (database, tools, records)
Human Resource Requirements:
Access to facilitators of learning with subject matter expertise as evident
from professional designations, or experiential competence or prior
qualifications
Access to assessment expertise in terms of subject matter related prior
learning, experience and credibility
Legal and Safety Requirements:
Access to facilities that comply with occupational health and safety standards
where contact learning or assessment sessions are delivered
The only way to establish whether you are competent and have accomplished the
learning outcomes is through continuous assessments. This assessment process
involves interpreting evidence about your ability to perform certain tasks. You will
be required to perform certain procedures and tasks during the training
programmer and will be assessed on them to certify your competence.
4
This module includes assessments in the form of self-evaluations/activities and
exercises. The exercises, activities and self-assessments will be done in pairs,
groups or on your own. These exercises/activities or self-assessments (Learner
workbook) must be handed to the facilitator. It will be added to your portfolio of
evidence, which will be proof signed by your facilitator that you have successfully
performed these tasks.
Listen carefully to the instructions of the facilitator and do the given activities in the
time given to you.
Assessment in Practice
5
Although assessments are currently used for many purposes in the educational
system, a premise of this report is that their effectiveness and utility must
ultimately be judged by the extent to which they promote student learning. The aim
of assessment should be “to educate and improve student performance, not merely
to audit it. To this end, people should gain important and useful information from
every assessment situation. In education, as in other professions, good decision
making depends on access to relevant, accurate, and timely information.
Furthermore, the information gained should be put to good use by informing
decisions about curriculum and instruction and ultimately improving student
learning.
Assessment methods
Assessment methods are the strategies, techniques, tools and instruments for
collecting information to determine the extent to which students demonstrate
desired learning outcomes. Several methods should be used to assess student
learning outcomes.
Relying on only one method to provide information about the program will only
reflect a part of students’ achievement. Additionally, SLO may be difficult to assess
using only one method. For each SLO, a combination of direct and indirect
assessment methods should be used. For example, responses from student surveys
may be informative, however, when combined with students’ test results they will
be more meaningful, valid, and reliable.
6
Direct methods of assessment ask students to demonstrate their learning while
indirect methods ask students to reflect on their learning. Tests, essays,
presentations, etc. are generally direct methods of assessment, and indirect
methods include surveys and interviews.
Even though course grades are a source of information about student achievement,
they are generally insufficient in measuring the student learning outcomes of the
program.
Grades may not identify whether the SLO have been achieved, may include factors
not related to SLO such as class participation, and faculty members may differ in
their grading policies and practices. Considering these limitations, however, grades
MAY be able to be used for program assessment IF they relate to the program’s SLO
and if grading methods are consistent across program faculty and courses. There is
a book dedicated to the use of grades as an assessment measure
Each program will select the assessment methods that will provide the most useful
and relevant information for the purposes that the program and its faculty have
identified.
When selecting which assessment methods to use, consider what questions need to
be answered, the availability of resources, and the usefulness of the results.
Programs may find it valuable to identify what information currently exists in the
program that can be utilized as well as what assessment methods have been used
for past assessments.
7
Utilize capstone courses or “second-year” projects/assignments to directly
assess student learning outcomes
Use established accreditation criteria/standards when developing the
assessment plan
Concepts in Assessment
We are continually faced with the challenge of assessing the progress of our
students as well as our own effectiveness as teachers. Educational Measurement
The first step towards elevating a field of study into a science is to take
measurements of the quantities and qualities of interest in the field. Types of
Measurement Objective measurements- are measurements that do not depend on
the person or individual taking the measurements. Subjective measurements- often
differ from one assessor to the next even if the same quantity or quality is being
measured. Underlying principle in educational measurement is summarized by the
following formula: Measurement of quantity or quality of interest = true value plus
random error.
Good tests should be used to guide instruction, helping teachers understand where
students are excelling and where they need more work. But improving student
8
learning and fairly assessing teacher quality should be based on a more broad and
whole assessment. Standardized tests are only one piece of the assessment tools
utilized in the classroom. Instead of spending millions of dollars on standardized
tests, we should be investing in schools to give every child the opportunity to
succeed. That means smaller class sizes and more funding for schools.
Teachers are spending too much time teaching to the test when they could be
giving students more individual attention to help students think critically and solve
problems. That’s the kind of learning that will prepare them for college and career.
Diagnostic assessment
Summative assessment
Formative assessment
9
usually takes place during day to day learning experiences and involves
ongoing, informal observations throughout the term, course, semester or unit
of study
is used to monitor students’ ongoing progress and to provide immediate and
meaningful feedback
assists teachers in modifying or extending their programmes or adapting
their learning and teaching methods
is very applicable and helpful during early group work processes.
Assessment plan
10
There are a variety of reasons why it is a good idea to create an assessment plan in
your program.
Having a well-written assessment plan reduces the amount of time you will
ultimately have to devote to conducting your assessment project, since it can clarify
tasks and timelines.
Allows for the consideration of all parts of an assessment project and critical
analysis of how the different pieces fit together and can best be implemented
achieving the desired goals of assessment.
An assessment plan provides a dynamic map for your assessment project, allowing
for better flexibility as you implement your project and more reasoned responses to
issues that may arise.
11
How does assessment fit into other aspects of the program and
programmatic priorities for the future
Do you have the appropriate resources (ex. faculty time, financial resources,
space, etc.) lined up to implement your assessment project?
Pre-assessment briefing
This is the second video in a series of six which aims to help candidate assessors
understand what is expected during a vocational assessment. Follow each
installment and create your own perfect assessment.
It is important that you always carry out the pre assessment briefing when
conducting your own vocational assessments. Here you can check that nothing has
been missed out of the planning stage. This could just be minor issues that are
cleared up during conversation. However it could also be about vital medical
information, or individual learning requirements, which could put the learner at risk.
Feedback
The assessor will provides feedback to the learner with regard to the assessment
decision. If the learner is deemed to be “competent” then there is no further action
required other than his/her feedback on the assessment process and confirmation
of receipt of the feedback from the assessor. Should the learner be deemed to be
“not yet competent” then the assessor should explain where (which SO and AC in
particular) and what remedial action is required. The learner would also be advised
by when the remediation is required. The learner would also be advised of his or her
12
right to appeal if they are unhappy with the decision. Receipt of this feedback
should also be confirmed by the learner.
Evidence
The assessor will then conduct the assessment and collect the required evidence as
presented by the learners. On receipt of the assessments in whatever form (e.g.
written test or portfolio of evidence) the assessor will conduct the assessment
making use of the marking memos and related recording sheets. All of the
requirements are usually summarized in a matrix that covers formative, summative
and critical cross field assessment requirements and all of the relevant areas should
be included.
Once the assessment has been marked the assessor will make and record a
judgement about the evidence presented. The evidence will be judged against the
criteria as set out in the unit standard – all SO’s and AC’s must be considered and
the principles of assessment must be adhered to. On completion the judgment of
the assessor will be one of “competent” or “not yet competent”.
Competence
The assessment criteria derived from the learning objectives are used to identify
learning outcome. Assessment of competence depicts how well set learning
objectives have been achieved.
Developing the students study process takes into account the learning and the
knowledge gained through academical perspective. The key factors in study process
are self-evaluation and feedback received from teacher or supervisor as well
sometimes from fellow students.
13
Practical competence is manifested as a systematic, smooth command of
tasks.
Academic competence indicates a command of theoretical knowledge, an
understanding of the connections between facts and processes and the
ability to grasp entities.
Social competence is related to interactive competence; the ability to work as
a member and a leader of a group.
Evaluative competence describes learning, evaluation and development skills
related to one’s own work and the ability to develop and influence joint
activity and activity that takes place in the surroundings.
Assessment methods are selected so that they truly measure achievement and the
level of competence.
The feedback collected from the students will develop the quality of the degree
programme. You have a change to give feedback about the implementation of the
study units as well as overall yearly feedback.
Assessment concepts within the QCTO model (Internal and final integrated
summative external assessment)
Wherever accredited learning takes place there is a need for assessment so that the
quality and transfer of learning can be evaluated. Such assessment has long been
associated with trades worldwide, and there are usually specific rules governing
how assessment is handled. Often, in fact, there is national policy, which governs
the process, with highly specific requirements for how things should be run.
14
minimize disruption of operations; and a formalized process by which roles are
linked to standard operating procedures, against which measurement of the ability
to perform competently can be made.
Assessment centres in South Africa, therefore, need to tick these boxes while taking
national policy into account. In the case of trades, the situation is relatively simple,
with the setting of trade tests being sufficient. However, in South Africa there needs
to be an extension, or adaptation, of this process t includes occupations as well.
As we set out on the search for an assessment centre model, in terms of the new
Quality Council for Trades and Occupations (QCTO) concept of Occupational
In most environments these days, there are new opportunities for using technology
to improve and streamline processes. It is no different with the assessment of
learning. In keeping with such advances, the new assessment model should
embrace new methodologies, but also stay true to national policy, and to
Assessment Specifications.
Each learning area is assessed separately. Each assessment is 2.5 hours long and
consists of 60 multiple-choice questions. Assessments are machine scored.
Assessments are conducted at iEB assessment centres.
15
Exemplar assessment papers for both subject areas are available for download on
the website. Many candidates are not familiar with multiple choice testing, thus it is
advisable that Skills Development providers familiarize their learners with the
assessment method.
Learners will be enrolled for external summative assessment with the providers. Will
issue a statement of results to FLC skills development providers. The will upload
learner results to the qcto. Learners need to be found competent in both learning
areas in order to achieve the part qualification. Competence is achieved if a learner
gets 50% or above in each learning area.
16
Principles of good assessment practice (Fairness, Validity, Reliability,
Sufficiency, Practicability, Currency of evidence)
Assessment works best when the programs it seeks to improve have clear, explicitly
stated purposes. Assessment is a goal-oriented process. It entails comparing
educational performance with educational purposes and expectations -- those
derived from the institution's mission, from faculty intentions in program and course
design, and from knowledge of students' own goals. Where program purposes lack
specificity or agreement, assessment as a process pushes a campus toward clarity
about where to aim and what standards to apply; assessment also prompts
attention to where and how program goals will be taught and learned. Clear,
shared, implementable goals are the cornerstone for assessment that is focused
and useful.
17
Assessment requires attention to outcomes but also and equally to the experiences
that lead to those outcomes. Information about outcomes is of high importance;
where students "end up" matters greatly. But to improve outcomes, we need to
know about student experience along the way -- about the curricula, teaching, and
kind of student effort that lead to particular outcomes. Assessment can help us
understand which students learn best under what conditions; with such knowledge
comes the capacity to improve the whole of their learning.
18
problems, including tensions between the summative and formative purposes of
assessment and doubts surrounding the validity and reliability of teacher-
constructed assessment tasks. This article examines ways in which these problems
have been manifested and addressed, using two recent examples from school and
adult immigrant education. The first example concerns a recent controversy
surrounding the use of national literacy benchmarks for primary school learners.
Analysis of the issues suggests that some learner groups may be disadvantaged by
the practice of reporting aggregate outcomes in terms of minimum standards, but
that government policy is unlikely to change as long as the accountability function
of assessment remains paramount in the public eye.. It is argued that problems of
consistency and comparability that have been identified by research can be
addressed through the development of fully-piloted task banks and the provision of
appropriate forms of professional development. Greater attention needs to be given
to the role of the teacher if outcomes-based assessments are to provide high quality
information.
Learning targets, competencies, and outcomes–these are all ways of describing how
teachers define what it is that students need to know. The basic premise of
outcomes-based assessment is that teachers need to be able to define explicitly, in
language their students can understand and adopt the key content for which they
and their students will be responsible.
19
For students, outcomes-based assessment means
With the ever evolving and diversifying business challenges, the approach to the
management of human resources has also undergone a paradigm shift. The
competitive advantage achieved through technology, new products and information
is short lived and vastly evaporating. The only distinguishing feature from the
competition which remains, are the skills and contribution from the employees. The
organizational leadership therefore plays an important part as they directly
influence the performance and the people of the organization. A well-defined and
uniform competency framework is the first step towards an organized approach to
the human resource management of the organization
There is a constant need to increase efficiency and create and deliver value in each
transaction. Therefore it is imperative that a more scientific approach like
competencies be used to define and understand the knowledge, skills and attitude
required to perform a job effectively. Evaluating an employee’s performance based
on pre-defined competencies and their behavioral indicators, in turn is called
competency based assessment.
The competencies are like the Pole Star for the organization especially the senior
leaders and act as a constant check for the employees to direct and redirect their
efforts to yield specific results. Since, competencies can be assessed and analyzed;
it gives a fairly objective evaluation of an employee’s performance. The best part
about the competencies is that they can be learnt unlike personality traits which are
20
characteristic to an individual. If the organization identifies criteria critical to its
success in the form of core competencies, they can be cascaded down to the whole
organization.
Within the competency based assessment approach the organizations can find a
structured model to integrate their management practices as well. A systematic
approach like this can help define organizational priorities clearly and align the
human resource strategies to create and build on key behaviors which are desired
and would be rewarded. It also creates individual employees accountable and
responsible for their performance and learning and development while creating a
culture of transparency
Formative assessment
Summative assessment
21
data for selection (to the next stage in education or to employment). For all these
reasons the validity and reliability of summative assessment are of the greatest
importance. Summative assessment can provide information that has
formative/diagnostic value.
Integrated assessment
Diagnostic assessment
22
Collaborative contracts must include provisions for program oversight and protocols
to address concerns or issues that arise among partners and within the team.
Program oversight must be defined in the contracts between the community
program and the county board of education. These include, but are not limited to,
staff evaluation and discipline, health, and safety concerns, communication process,
and impasse procedures. Protocols should be specified within collaborative
agreements of each participating Pre- program. Teams must consider the roles of
each partner and classroom administrators when developing and implementing a
program oversight protocol. Collaborative Head Start classrooms are monitored for
compliance with the Head Start Performance Standards in partnership with local
and federal Head Start personnel. Childcare centers must be licensed under
Childcare Center Licensing. Child care centers are reviewed by child care licensing
specialists as part of child care licensing regulations. Information provided by
reviewing agencies should be made available to all partners in collaborative
classrooms. Provisions for program oversight should include a process for reviewing
of monitoring results from federal Head Start, and child care licensing, as
applicable.
Assessments should reveal how well students have learned what we want them to
learn while instruction ensures that they learn it. For this to occur, assessments,
learning objectives, and instructional strategies need to be closely aligned so that
they reinforce one another.
Our objective is for students to learn to apply analytical skills, but your assessment
measures only factual recall. Consequently, students hone their analytical skills and
are frustrated that the exam does not measure what they learned.
Your assessment measures students’ ability to compare and critique the arguments
of different authors, but your instructional strategies focus entirely on summarizing
the arguments of different authors. Consequently, students do not learn or practice
the skills of comparison and evaluation that will be assessed.
23
The demonstrated ability of understanding and applying the knowledge and
thinking which underpins the action required and taken in a specific context.
Vast volumes of South African educational documents highlight the following four
competencies (paraphrased from http://www.citehr.com/14370-competency-
mapping-urgent.html) of which two mention reflexive/reflex
TOPIC QUIZ
24
KM-05-KT02: Evidence collection and recording concepts and principles
Sources of evidence
Finding the best sources also depends on the type of question you have, and the
type of research study that is most appropriate for that question. Evidence is
published in a wide range of sources including journals, books, research reports,
and increasingly directly onto websites. Sources may contain different types of
information, such as clinical guidelines, systematic reviews, controlled trials or
qualitative research. Some sources will provide access to a range of information
types, i.e. PubMed, while others will specialize in one particular type, i.e. the
Cochrane Database of Systematic Reviews. This following presentation provides an
overview of key sources available to help guide your search for evidence.
Deciding which sources of evidence are the best ones to use will depend on many
factors, including the purpose of your search, the time available, and whether you
have access to the resource itself. We all belong to extensive information networks
and use lots of different sources to find answers to our questions. Often the sources
we use are based on our previous behavior patterns, or are the ones that we are
most familiar with or have the easiest access to. Think about the last time you
needed to find some information to help you make a health care decision.
25
As noted above, a direct source differs from an indirect source. Indirect sources
involve the reports and analysis of direct information by other people. Reports,
articles, and books by scholars are the main category of secondary sources that we
tend to use, especially when we are beginning to develop our knowledge about an
issue. Once again, the key to using such information is always to think about the
context in which it was produced. In other words, information from indirect sources
is only as good as our understanding of that source itself. For example, as a result
of changes over the past two decades, an academic commentary such as the one
by James on advertising in the 1970s may not be precisely relevant to
contemporary concerns. There is no general rule to apply to such analysis of
sources, except that we must always think about the context (who obtained the
information; when, where, and how the information was arrived at) as well as the
text (what the information is).
For example, imagine you are watching a television program on advertising. The
host makes some comments on nationalistic commercials, saying that they always
produce an emotional reaction and that is why they are effective. Is this source
useful for an academic investigation? If you answered no, then you would, in some
circumstances, be correct. But the important question to ask is 'Why?'. Let us
contrast this hypothetical television program with a more usual source: academic
writing. The trustworthiness of academic writing is based on the idea that the
person doing the writing is an expert in that area, through their close study of the
topic, their skills as a researcher, their careful, long-term analysis, and their
involvement in a system in which articles and books are published only after the
scrutiny of other qualified academics to determine if they are 'right' or not. In other
words, the claims are trustworthy because an institutionalized method makes them
trustworthy. It is a social convention that academic work is regarded as being more
'sound' (if often more remote) than 'popular' work; it is also a worthwhile social
convention because there are good reasons to accept this distinction in soundness.
Learning Domains
26
Learning is not an event. It is a process. It is the continual growth and change in the
brain's architecture that results from the many ways we take in information,
process it, connect it, catalogue it, and use it (and sometimes get rid of it).
Learning can generally be categorized into three domains: cognitive, affective, and
psychomotor. Within each domain are multiple levels of learning that progress from
more basic, surface-level learning to more complex, deeper-level learning. The level
of learning we strive to impact will vary across learning experiences depending on
1) the nature of the experience, 2) the developmental levels of the participating
students, and 3) the duration and intensity of the experience.
Routine assessment is carried out during the course, and is designed to organize
students' independent work and a systematic monitoring of their academic
knowledge.
27
Students receive grades for exams after passing final assessments. A cumulative
grade and an exam grade form a final grade. The final grade is a weighted sum of
the cumulative grade, midterm grade and a grade received on the exam. The grade
composition for each discipline is defined in the course syllabus, for example:
The final grade is a grade for the course that is included in the transcript of
records/diploma. In extraordinary occasions, a student can be exempted from an
exam and granted a resulting grade equal to the cumulative grade.
Please keep in mind that the HSE’s academic rules are the same for all students
(Russian, foreign, exchange and non-exchange). To succeed on a course, we advise
international students to attend all lectures and seminars, prepare their home
assignments and meet the deadlines for all essays, reports, etc.
Your POE is marked together with your exam. The marks you get for your POE forms
30% of your fiats you study a subject (or learning area) all your assessments are
kept in a folder. So all these evaluations, plus your final exam paper makes up your
complete Portfolio of Evidence. When you get your POE from the ICB, it will consist
of the Administration Book, which informs you exactly how to complete your POE.
You hand all of it in at your final exam, and the assessors (examiners) marks all of it
to identify if you pass or if you have to repeat the work.
Keep in mind that you must get a POE for every single subject you do. So if you sign
up for three subject exams, then the ICB will send your three Portfolios of Evidence.
28
If your POE is incomplete, that can mean that you fail that subject. So make it as
complete as possible. Likewise make certain that you make a copy of your POE
before you hand it in. Put it in a safe place, so that you understand if something
occurs to your POE, you have a back-up strategy.
Your assessor may ask you to put together a portfolio of evidence, which is a formal
collection of documentary evidence that you have obtained during your learning,
your career experience, unpaid work, relevant personal interests, and professional
development. The more current the evidence, the more useful it is for the purposes
of your application for Skills Recognition. However, there is no hard and fast rule
about currency; it depends on the competencies that you are applying for. Your
assessor will be able to advise you how recent your evidence needs to be.
Your assessor will probably provide you with a list of suggested evidence for your
portfolio. You may also want to consider the ideas listed here: evidence of courses
or training completed in or outside work. These could be certificates, diplomas or
29
degrees, diplomas or certified transcripts of grades and subjects/courses
undertaken (certified statements of results). These may have been gained through
formal training, Skills Recognition or internal training within an organization..
Workplace documents:
Validity: Defined
The term validity has varied meanings depending on the context in which it is being
used. Validity generally refers to how accurately a conclusion, measurement, or
concept corresponds to what is being tested. For this lesson, we will focus on
validity in assessments.
30
Factors That Impact Validity
Before discussing how validity is measured and differentiating between the different
types of validity, it is important to understand how external and internal factors
impact validity
A student's reading ability can have an impact on the validity of an assessment. For
example, if a student has a hard time comprehending what a question is asking, a
test will not be an accurate assessment of what the student truly knows about a
subject. Educators should ensure that an assessment is at the correct reading level
of the student.
Student self-efficacy can also impact validity of an assessment. If students have low
self-efficacy, or beliefs about their abilities in the particular area they are being
tested in, they will typically perform lower. Their own doubts hinder their ability to
accurately demonstrate knowledge and comprehension.
Student test anxiety level is also a factor to be aware of. Students with high test
anxiety will underperform due to emotional and physiological factors, such as upset
stomach, sweating, and increased heart rate, which leads to a misrepresentation of
student knowledge.
Measurement of Validity
Validity is measured using a coefficient. Typically, two scores from two assessments
or measures are calculated to determine a number between 0 and 1. Higher
coefficients indicate higher validity. Generally, assessments with a coefficient of .60
and above are considered acceptable or highly valid.
Types of Validity
There are three types of validity that we should consider: content, predictive, and
construct validity. Content validity refers to the extent to which an assessment
represents all facets of tasks within the domain being assessed. Content validity
31
answers the question: Does the assessment cover a representative sample of the
content that should be assessed?
For example, if you gave your students an end-of-the-year cumulative exam but the
test only covered material presented in the last three weeks of class, the exam
would have low content validity. The entire semester worth of material would not be
represented on the exam.
Educators should strive for high content validity, especially for summative
assessment purposes. Summative assessments are used to determine the
knowledge students have gained during a specific time period.
It is a near-consensus that the way we produce and consume goods and services in
modern economies is not sustainable. We systematically (and knowingly) overuse
natural resources, ignore the social cost of Nature’s services… So, the diagnosis is
more or less uncontroversial and agreed upon. However, the question about the
right therapy is still unsettled: how can we do it all in a better way? Generally, there
are three strategies that are often named as necessary to achieve true
sustainability: efficiency, consistency and sufficiency
Let us begin with the least controversial sustainability strategy: efficiency. If you ask
an economist, he likely will tell you efficiency is the simple most important thing.
Actually, modern economics is mainly about various aspects of efficiency. However,
in the context discussed here efficiency has a slightly different meaning. First of all,
traditionally understood “economic” efficiency is a notion confined to things that
can be easily expressed in quantitative terms (at best in money). Moreover, it is
viewed through the lens of rational choice theory that underlies modern economics,
which means that efficiency from the point of view of an individual (firm, household
etc.) is focused on. E.g., it is efficient to exploit coal reserves fastly if it is expected
that the coal price will decline in the future. What I mean by “efficiency” here is
something different. It is thought as an imperative to use (natural) resources
efficiently, i.e. not to waste them. Efficiency in this context is to be understood with
regard to social costs that are to be economized upon, not individual/market costs.
32
Moreover, it is not at all confined to quantifiable aspects of human economic
activity – there are many ecosystem services that have no “price tag” attached,
which does not mean that they need not be used efficiently.
A less well-known but still crucial strategy on the path toward sustainability is called
consistency (or, by some, industrial ecology In general terms, consistency is the
trial to integrate material and production cycles into each other as well as into
natural cycles. It is more than recycling, because recycling is only an ex-post
utilization of parts of once-used materials, often linked to unpleasant side-effects
like pollution and toxic waste. Consistency is a much more holistic and
encompassing integrated approach, within which re-cycling is planned already
before production actually begins.
TOPIC QUIZ
I. Evidence collection and recording concepts are described and the most
suitable selection motivated for specific scenarios
33
KM-05-KT03: Evidence evaluation concepts and principles
Rules of Evidence
It is not good enough to just collect any evidence. Just as the way we collect
evidence is guided by the principles of assessment, the way we collect evidence is
guided by the rules of evidence.
Valid
Current
Sufficient
34
Demonstrate competence that is able to be repeated
Comply with language, literacy and numeracy levels which match those
required by the work task (not beyond)
Authentic
Validity
The assessor is assured that the learner has the skills, knowledge and attributes as
described in the module or unit of competency and associated assessment
requirements. For example, the evidence will not be valid if you instruct a candidate
to solve printer problems by simply asking: Type a standard office memo on a word
processor. Validity is assured when the performance required matches the
performance described in a competency standard.
Currency
Sufficiency
The assessor is assured that the quality, quantity and relevance of the assessment
evidence enables a judgement to be made of a learner’s competency.
A judgement has to be made concerning how much evidence to call for. How much
is required for the assessor to accept the performance as competent? Too little
evidence risks the assessment not being reliable; too much leads to waste of time
and effort.
35
Authenticity
The assessor is assured that the evidence presented for assessment is the learner’s
own work.
It is very easy to get too much evidence. It is also very easy to get too much
evidence that doesn’t really help us to make good decisions. Because of this, it is in
everyone’s interests to guide our candidates through the selection, organization
and submission of evidence.
The first thing we need to do, however, is work out what makes quality evidence.
The answer to this is quite simple. It is evidence that lets us make decisions about
whether someone can do what it is that they are meant to be able to do, i.e., it will
help us to recognize competency.
36
“accurate” and “fair”. This presentation will first review some of the concerns and
challenges in using teacher professional judgment in large scale school based
assessment – both theoretical and practical issues - and then demonstrate how the
development of judgment can be supported and enhanced by online professional
interaction, including the sharing of student work samples, and benchmarking of
assessments. The implications for assessment criteria always interpreted differently
by teachers according to personal background, previous experience, unconscious
expectations, internalized and personalized preferences regarding the relative
importance of different criteria and ideological orientation
"We are social beings who construe the world according to our values and
perceptions" individual schools as assessment communities will also be discussed.
Effective feedback
Feedback is most effective when it is given at the time of the learning so that
students can make improvements as they go. However, written feedback can be
beneficial to learning if the following points are taken into consideration:
37
Some students have difficulty understanding and processing written
feedback.
When students are presented with grades and comments, the grades can
cancel the beneficial effects of the comments.
Teachers often give too much feedback, which students find overwhelming
and difficult to understand.
As a result, the tool measures compliance or ‘must dos’ for all public offices as well
as additional performance criteria, which will help determine whether your public
38
office is performing above the minimum requirements at a higher level of maturity,
capability and capacity.
The tool is currently under re-development due to the issue of the new Standard on
records management. The following three modules are still available for use:
Each module contains a questionnaire which will assist your organization to assess
its recordkeeping processes and practices. You can use the tool in a modular style
(i.e. choose to complete a module and evaluate the results before moving onto the
next module) or you can use the whole tool and evaluate all results together.
For large organizations or organizations located at different sites, the tool can be
used to examine separately these different segments of your organization for
conformity with organizational recordkeeping policy, standards and procedures.
Assessment of standards
Standards specify what a candidate for assessment needs to know, do, and
understand. Whether it is by task or evidence, assessment can use a variety of
methods and approaches (appropriate to the learner and the context) that give the
learner the opportunity to show evidence of achievement/competence.
39
These standards represent assessment competencies in the administrative contexts
in which assessment is relevant to educational administrators. These contexts
include: (a) assisting teachers in creating and using assessment effectively; (b)
providing leadership in the creation and implementation of building- or district-level
assessments policies; and (c) using assessment results in their capacity as
administrators in making decisions about students, teachers, and instruction, and in
reporting on assessment results to a variety of stakeholders and constituencies.
40
Grading of results in vocational education and training is a controversial topic.
Training Packages do not specify whether or not units should be graded. However,
many units delivered in TAFE NSW have been set up so that student results show a
graded outcome an argument in favor of graded outcomes is that they can provide
additional information to students, employers and educational institutions about the
level of performance demonstrated by a student.
The traditional method of assigning marks as a percentage does not sit comfortably
within a competency based system. Where a unit requires graded results, the
challenge for Institute teachers is to develop and implement a suitable approach to
the assessment process.
41
Determine the Purpose of the Activity then develop the learning
objectives to “closing the gap. Permission to reproduce this work for
educational purpose is granted, provided proper citation is given."
Remedial action
Rework and repair are generally the remedial actions taken on products, while
services usually require additional services to be performed to ensure satisfaction
Remedial Action is often enacted through ‘Remedial Action Plans’ (RAPs) of three or
more stages. For example, one North American river protection scheme drew up a
plan that identified environmental problems and sources of pollution, evaluated and
carried out actions to restore the area, and confirmed that these actions had been
effective
These actions normally have negative effects on a company’s image and profits. A
fine balance is often struck between denying the existence of problems or
42
wrongdoing and choosing to publicize this existence before being exposed removes
the company’s ability to control the way any scandal is seen.
The secondary school endeavors to monitor and support students with difficulties.
This role includes:
43
TOPIC QUIZ
I. The rules and criteria of evidence evaluation are listed and explained
II. Evidence evaluation and judgement concepts are explained in terms of
competency levels, purpose, impact and mitigation
III. Feedback practices are explained for various stakeholder needs
IV.
44
KM-05-KT04: Assessment administration and regulatory practices
Management of Assessment
In the Construction Education and Training Authority, the Education and Training
Providers are responsible for the management of assessment processes being
conducted by assessors registered to assess unit standards and/or qualifications
within the primary focus of Catamite is important that Education and Training
Providers to have high quality assessment practices to ensure that the success and
credibility of the outcome-based education and training system.
Assessment events are designed at the element level as a minimum square Bullet
lite
Assessors has considered all matters contained in performance criteria in the unit
Assessment use the least amount of evidence that is needed for the assessor to
make a valid professional judgement that competency has been achieved
45
employees (under our direction) and placed in specially designed storage boxes for
safe keeping. Details of each file are captured on our secure database, according to
your company’s requirements.
Regular reports are sent to our customers detailing all the archived files. Any
movement of files is recorded and a monthly statement is available upon request
which lists all activity, revealing all movement.
Document destruction
Each box is allocated a destruction date according to the laws that govern them and
our customers’ requirements. Prior to this date, you will be provided with a printout
of all the files flagged to be destroyed. Only with your written authorization will we
proceed to eliminate these files.
Document Security
Fire – Our fire prevention facilities include pumps, tanks, sprinklers, hydrants, hose
reels, hand held extinguishers and dense smoke detectors which are connected via
control panels to our security office. Theft – Multiple physical deterrents have been
installed. In addition, all facilities are patrolled on a 24-hour basis and have a direct
link to the security company’s control room.Unauthorised access – Only individuals
that have been authorized by our customers and our security-cleared personnel
have access.
We will issue a Certified Statement of Results if you have lost or did not collect your
certificate from your centre and they no longer have the origin
46
Replacement Certificate
We can only issue an original replacement certificate if the originals were damaged
or defaced. Please return your original certificate(s) with a completed application.
This service is only available for certificates issued from 2001 onwards.
All candidates receive a Statement of Results, and if they are successful in the
exam they will also receive a certificate.
Statements of Results are released online, approximately four to six weeks after the
exam for paper-based exams, and two to three weeks after the exam for computer-
based exams. The majority of your students’ results will be available at the start of
the release window, with any outstanding results being released in the days
following.
National Senior Certificate or NSC is a high school diploma and is the main school-
leaving certificate in South Africa. This certificate is commonly known as the
matriculation (matric) certificate, as grade 12 is the matriculation grade. The NSC,
previously known as the Further Education and Training Certificate or FETC,
replaced the Senior Certificate with effect from 2008 and was phased in starting
with grade 10 in 2006.
Pupils study at least seven subjects, including two compulsory official South African
languages, Mathematics or Mathematical Literacy, Life Orientation and three
elective subjects. Students have the option of either taking the NSC (National Senior
Certificate) or the NCV (National Certificate Vocational).
Subjects are all taken on the same level - there is no higher or standard grade as in
the past. There are three pass levels, each with different pass requirements. First is
a pass with higher certificate, and one must obtain 40% in home language, 40% in
two other subjects and must obtain 30% in three other subjects. A learner, who
passed matric with higher certificate level, cannot enroll for a University Degree or
a Diploma at an institute of higher learning. The mean mark in any subject is usually
about 55. Only a small proportion of candidates score an 'A' in any subject (from as
47
little as 2% to a maximum of about 10% in subjects taken by highly select groups.)
A further 8 – 15% is likely to gain a 'B' and about 20 – 25% achieve a 'C' grade. The
National Senior Certificate is a group certificate and records an aggregate mark.
The Department of Basic Education has responsibility for general educational policy
to be implemented by nine provincial education departments and private providers
such as the Independent Examinations Board (IEB). There are nine provincial
examination boards and three independent boards, of which the IEB is the biggest.
The IEB operates on a national level catering primarily for independent schools.
48
TOPIC QUIZ
References
Crooks, T. (2001). "The Validity of Formative Assessments". British
Educational Research Association Annual Conference, University of Leeds,
September 13–15, 2001.
Huhta, Ari (2010). "Diagnostic and Formative Assessment". In Spolsky,
Bernard; Hult, Francis M. The Handbook of Educational Linguistics. Oxford,
UK: Blackwell. pp. 469–482.
Shepard, Lorrie A. (2005). "Formative assessment: Caveat emptor" (PDF).
ETS Invitational Conference The Future of Assessment: Shaping Teaching
and Learning, New York, October
Black, Paul; Wiliam, Dylan (2009). "Developing the theory of formative
assessment". Educational Assessment, Evaluation and Accountability. 21
(1): 5–31. doi:10.1007/s11092-008-9068-5.
Scriven, Michael (1967). "The methodology of evaluation". In Stake, R. E.
Curriculum evaluation. Chicago: Rand McNally. American Educational
Research Association (monograph series on evaluation, no. 1.
Wiliam, Dylan (2006). "Formative assessment: getting the focus right".
Educational Assessment.
Bloom, Benjamin S. (1968). Learning for mastery. Los Angeles, USA:
University of California press.
Bloom,, Benjamin S.; Hasting, Thomas; Madaus, George (1971). Handbook
of formative and summative evaluation of student learning. New York,
USA: McGraw-Hill.
Black, Paul; Wiliam, Dylan (2003). "'In praise of educational research':
formative assessment". British Educational Research Journal.
49
R. Pr ́ egent, Charting your course: How to prepare to teach more
effectively, Atwood, 2000.
Cauley, K, M.; McMillan, J. H. (2010). "Formative Assessment
Techniques"Paid subscription required. The Clearing House. 83 (1) – via
Proquest.
Harlen, Wynne; James, Mary (1997). "Assessment and Learning:
differences and relationships between formative and summative
assessment". Assessment in Education: Principles, Policy & Practice.
Nicol, David; Macfarlane-Dick, Debra (2005). Rethinking Formative
Assessment in HE: a theoretical model and seven principles of good
feedback practice. Quality Assurance Agency for Higher Education.
Nicol, David J.; Macfarlane-Dick, Debra (2006). "Formative assessment
and self-regulated learning: a model and seven principles of good
practice". Studies in Higher Education.
Cowie, Bronwen; Bell, Beverley (1999). "A model of formative assessment
in science education". Assessment in Education.
50