Program Assessment Process
Program Assessment Process
Outcomes Assessment)
Program Assessment Getting Started
The teaching/learning/assessment cycle (the left side of the figure below) depicts instructor activities at the
course level. The instructor begins by determining student learning objectives. Next, she provides students with
opportunities to learn what they need to know to meet those objectives. Following the learning opportunities,
tests or assignments are administered to determine the extent to which students have met the objectives.
Finally, results are used to make improvements in the course.
The program assessment process is depicted on the right of the figure, and is very similar to the teachinglearning-assessment cycle. The program assessment process involves almost the same steps at the program
level. It differs because of the focus on the entire program rather than a single course. The process also
provides the opportunity for faculty members to come together to discuss student learning within the program
and to see how courses fit together.
The formal program assessment process begins with the determination of the mission and goals of the
program, and is ideally linked to the mission and goals of the institution. However, in practical terms, the
process often begins with the program objectives. The following pages describe the basic program assessment
process, which involves the following six steps.
1.
2.
3.
4.
5.
6.
Although these are the typical meanings of these terms, many people use them synonymously and/or
interchangeably. Goals are typically more general, objectives and outcomes more specific. Objectives look
toward the future, outcomes reflect the current state. Thus, objectives are developed at the beginning of the
assessment or instruction process, whereas outcomes are the results of tests or assignments that address the
objectives. These terms are also scalable. Academic programs have goals, objectives, and outcomes, but
courses, course modules, class sessions, and even topics within a single class session can have goals,
objectives, and outcomes. More information on writing goals and objectives. [These should link to revised
pages]
Please note: Some accrediting organizations use or define these terms differently. If your program is accredited
by a disciplinary or professional organization, use the terminology as it is used by that organization.
What are the needs of our graduates upon completion of a degree in our discipline?
Are there specific accreditation or certification requirements for our department and/or college?
Are there any recommendations for goals that have been developed by professional organizations in
our field that are aligned the goals we want graduates to achieve?
Are there any recommendations made by business and industry that could translate into goals for our
program?
Have peer institutions published goals that might be appropriate for our graduates?
It is not necessary to achieve complete consensus by all program faculty, but most groups should be able to
achieve a functional consensus (i.e., sufficient agreement to prevent paralysis and allow progress). All
important goals should be included, even if it could be difficult to document students' progress and
achievement. Programs offered at different Penn State locations should have similar goals, but faculty may
want to include a limited number of campus-specific goals.
Communications
Graduates will demonstrate an understanding of the diversity of groups in a global society and their
relationship to the field of communications.
Comparative Literature
Graduates will obtain a global awareness through a wide range of international literary and language
experiences that are critical in todays world.
Engineering
Graduates will have the ability to design and conduct experiments as well as analyze and interpret
data.
Physics
Graduates will acquire the theoretical and experimental background for work in areas such as
acoustics, astrophysics, biophysics, chemical physics, computer science, mathematical physics, and
engineering.
Psychology
Graduates will understand and generate applications of psychology to individual, social, and
organizational issues.
When writing learning objectives, it helps to think in terms of the level of knowledge, skill or attitude you expect
your graduates to attain. Bloom (1956) developed a taxonomy of learning objectives that is useful for
developing learning objectives. Using a verb to describe the student actions makes the statement measurable
and helps you later define the type of assessments needed to show the extent to which the objectives were
achieved. Sample Verbs for Learning Objectives (pdf) provides a list of verbs associated with each level.
See examples of learning objectives below. See also Program Assessment: Options for Getting Started (pdf)
for suggestions about how to approach the development of learning objectives at the program level.
Learning objective 1: Students will be able to analyze a film for its inclusion or exclusion of certain
diverse groups.
Learning objective 2: Students will be able to write an unbiased news article about a two-sided issue.
Learning objective 1: Students will be able to analyze a text set outside of the United States and infer
the social and economic conditions based on the authors description.
Learning objective 2: Students will be able to compare a non-U.S. publication with a U.S. publication
and show the similarities and differences between the two settings.
Engineering goal
Graduates will have the ability to design and conduct experiments as well as analyze and interpret data.
Learning objective 1: Students will be able to choose the appropriate method or procedure when faced
with an engineering related problem.
Learning objective 2: Students will be able to create a hypothesis and determine the methods
necessary to test their hypothesis.
Learning objective 1: Students will be able to distinguish between different types of aerobic
conditioning exercises.
Learning objective 2: Students will be able to correctly measure standard vital signs.
Physics goal
Graduates will acquire the theoretical and experimental background for work in areas such as acoustics,
astrophysics, biophysics, chemical physics, computer science, mathematical physics, and engineering.
Learning objective 1: Students will be able to evaluate an article from a discipline-specific text and
write a critique using theoretical principles.
Learning objective 2: Students will be able to employ the correct procedure for a given experiment.
Psychology goal
Graduates will understand and generate applications of psychology to individual, social, and organizational
issues.
Learning objective 1: Students will be able to distinguish between several different psychological
issues.
Learning objective 2: Students will be able to recommend a course of treatment for an individual based
on the issues presented.
Engineering (pdf)
Fine, Communication, and Performing Arts (pdf)
Health Sciences (pdf)
Humanities (pdf)
Law Education and Business (pdf)
Physical and Natural Sciences (pdf)
Social Sciences (pdf)
Course 1
Course 2
Course 3
Course x
Objective 1
Objective 2
Objective 3
Objective x
Begin by creating a table that lists the program objectives in the left column and the courses in your program
across the top. The courses should be listed in the prescribed order that students take them, with your elective
courses last. Once you have the chart completed, give a copy to each faculty member in your department and
have them indicate which courses are contributing to the attainment of each objective and the depth of
coverage using the following scale (or one like it):
This strategy helps you determine which courses address your objectives and helps identify the courses to
target for data collection, which is the next step in the program assessment process. The courses that list the
amount of coverage at a level 3 are usually the ones where you can focus some of your data collection efforts.
This alignment activity also allows program faculty to demonstrate the role and importance of their courses in
the curriculum and helps to verify that the curriculum is appropriately structured and balanced to attain the
program objectives. The process offers faculty an opportunity to identify gaps and redundancies and to make
deliberate decisions about whether those are acceptable.
Each course in the curriculum should be linked to at least one program objective. Some courses will be
associated with more than one objective. Since objectives are directly linked to goals, this alignment will, by
default, ensure that each course is linked to at least one program goal.
If specific general education courses are required as pre-requisites for courses in the degree program, those
should be identified and included in the alignment.
This information should be presented in a way that reflects the traditions of the field. Some degree programs
will present the alignment in a table or grid format [hyperlink to Objective_Course Linking Template.docx] or as
a list of courses under each objective, but your program is not limited to these two formats. See an example
from the Elementary and Kindergarten Education program at Penn State Berks.
Provide evidence of how well students are meeting your goals and
objectives
After goals and objectives are mapped to the courses in the curriculum, it is time to select the evidence that will
best demonstrate the extent to which students have met the stated goals/objectives. The curriculum map will
help you determine which courses are likely to provide the most appropriate evidence. Evidence may be drawn
from a wide variety of sources, including answers to specific test questions, student writing samples, team
project reports, and survey questionnaires. Once this array of material, information, and evidence is gathered,
the program will have a better sense of the programs outcomes, i.e., the overall level of achievement.
Together the program faculty will need to decide:
Assessment experts recommend a mix of direct and indirect evidence of student learning (see below). If
assignments will be used, it is not necessary to collect them from all students. It is acceptable to sample
student work for the purposes of program assessment. However, it is important for faculty to determine a
benchmark level of performance that is deemed acceptable. Evidence of student learning and achievement
should be presented in aggregate form for the program. Evidence should be anonymous. Students' names
should be removed and course level data should not be linked to or used in evaluations of individual faculty
members.
Be careful not to collect more information than you will be able to analyze, because doing so can divert
attention from critical areas and result in overwork. Over time, each program will refine its process and more
accurately target areas of concern or interest.
Direct evidence
Direct evidence typically includes samples of students' work from courses or other scholarly activities, such as
papers, essays or presentations. If faculty members use specific criteria such as performance on professional,
national, or state exams, or rubrics to evaluate students at the program level, these instruments and outcomes
can also provide direct evidence of student achievement. You may have heard that student course grades are
not considered useful evidence for program assessment. See Can grades be used for assessment? (pdf) for
reasons why and for appropriate ways to use course assignment scores for evidence of student learning.
Setting Benchmarks
The next decision to be made is to set benchmarks. What score on the assignment or set of MC questions
indicates that a student has met the objective? Would you be satisfied with an individual student score of
75%, or would you expect a 90%? According to Linda Suskie (2012), the benchmark depends on the objective.
In some cases, 75% might be enough. In other cases, for example nurses learning to do injections, you
probably want your benchmark closer to 100%.
The next question to address is the level of overall performance you will accept as evidence that your students,
collectively, have met the objective. That is, what performance level will you accept as evidence of success
and what level will lead you to want to make changes. Suskie suggests that you do some further investigation
on any rubric criteria or test questions for which fewer than 50% of students reached your benchmark.
Indirect evidence
Perceptions and viewpoints that can be obtained through focus groups or surveys of students, alumni, and
employers serve as indirect evidence. Retention, graduation, and job placement statistics, as well as data on
participation in university or department programs that can be linked to program goals (e.g., internships, study
abroad, undergraduate research participation rates) are also considered indirect evidence.
For example, you may determine that the program objective was not written well, or perhaps isnt really
important after all. In this case, your strategy would be to re-write, replace or delete the objective. Or, you may
hypothesize that the objective isnt being properly addressed in the course(s). Perhaps students arent getting
enough practice, for example, or there isnt enough emphasis by the instructor. Alternatively, it may be that the
assessment itself is not constructed in a way that best addresses the objective. Perhaps the assignment
directions and/or rubric need to be revised. For multiple choice questions, it isnt uncommon for the test
question itself to be worded in a way that results in many students missing it even if they know the material.
Analyzing test questions can be done using a statistical procedure called item analysis.
After you have determined what is underlying the students poor performance, you can make the necessary
changes. Be sure to repeat the assessment so that you will know if your changes made a difference.
Accreditation Cycles
If a program is accredited by a disciplinary accreditation organization, the plan should reflect the organizations
review cycle. Other programs should develop a five-year plan and rotating schedule for assessing student
learning consistent with the Middle States cycle; Penn State was reaccredited in 2005 and submitted a periodic
review report in 2010.
Frequency
Each program determines the frequency with which the program objectives are assessed and how often
evidence will be gathered from each source. Not every program objective needs to be assessed every time an
associated course is taught. Neither does evidence need to be collected from every student in a program or
course. A program may decide to collect information from all courses or only some of its courses. Faculty will
want to consider a variety of data collection strategies, including those that permit comparisons across
locations, but also those that permit particularly successful approaches to be recognized.
How frequently each program objective is assessed depends on a variety of factors. Faculty might want to use
the following questions to guide their development of the assessment plan and schedule:
Is one particular objective associated with lower than expected student achievement?
Does one objective reflect a known (and easily verified) program strength?
Are some objectives associated with fundamental or crucial program goals?
How many courses contribute to (are associated with) foundational objectives?
Which courses are linked to core objectives and are they required for all students?
How often are courses taught?
Does the manner in which an objective is met vary frequently through time, such as with advances in
the field, or is there only one way to meet a particular objective?
How often is the degree program reviewed, either internally, externally, or by an accrediting
organization?
Resources
Reference
Suskie, L. (2012) Summarizing, Understanding and Using Assessment Results, presented at Penn
State Harrisburg on May 10, 2012.