Sie sind auf Seite 1von 53

Assessing The Effectiveness Of

Your Academic Advising


Program
Tom Grites
Assistant Provost
The Richard Stockton College
Tom.Grites@stockton.edu

Audience Poll
What are your expected learning outcomes
for this webinar?
A.
B.
C.
D.
E.
F.

Why do we assess?
What do we assess?
Where do we begin?
What tools do I use?
Who defines our success?
What I need to do tomorrow.

Overview

Terminology and Rationale


Assessment as a Process
Focus on the Contexts
Non-Assessment
Summary

Terminology
Assessment
Evaluation
Measurement (the tools)

Assessment
(re: academic advising)
Assessment is the process through which
we gather evidence about the claims we are
making with regard to student learning and
the process/delivery of academic advising in
order to inform and support improvement
(Campbell, 2008)

Uses/Contexts
Assessment tends to be more related to
programmatic issues and outcomes
Evaluation tends to be more related to
people (advisor) skills, performance, and
outcomes
Its OK to use evaluation as part of the
assessment process

Intentions

(related to both)

Formative more associated with


assessment; includes a wider range of
efforts; requires more analysis; provides a
broader perspective; focus on
improvement
Summative more associated with
evaluation; more limited effort; focus on
Does it work? or How well was job
performed?

The Rationale
a lack of assessment data can
sometimes lead to policies and practices
based on intuition, prejudice, preconceived
notions, or personal proclivities none of
them desirable bases for making
decisions
(Upcraft and Schuh, 2002, p. 20)

More Rationale
In God we trust; all others bring data.

An ounce of data is worth a pound of


opinion.
(Magoon, c. 1975)

Other Reasons

Accountability
Effectiveness
Accreditation
Trustees/Regents
Legislators

Program Improvement (to monitor and


improve student success) the most
important reason

The Assessment Process:


A Cycle
Resources:
Assessment of Academic Advising Package
(3 CDs available from NACADA via www.nacada.ksu.edu)

Assessment of Academic Advising Institute


(Feb 12-14, 2014

Albuquerque, NM)

Getting Started: Identify


Stakeholders
Complete set of advising constituents (students,
staff and faculty advisors)
Broad range of key offices (Registrar, Enrollment
Management, similar advising units, certain
campus referral resources, IR office)
Critics, Antagonists, and Naysayers
FYIs Faculty Senate, Deans Council, Retention
Committee, others as appropriate

The Advising Hub

What Do We want To Know or Demonstrate


as a Result of Academic Advising?
Focus on student learning
Connect learning to mission, vision, values,
goals in your advising program
How will your program contribute to
student learning?
Who, what, where, when, how will learning
take place?
Define measures of student learning
Gather evidence, set levels of expected
performance

The Assessment Process/Cycle


Alignment with institutional and unit
missions
Specify goals and/or objectives
Identify the outcomes expected (student
learning and/or programmatic)
Gather evidence (the measurements)
Share findings, interpretations, and
recommendations
Begin implementation and re-start the cycle

Mission/Purpose

A working model

Academic advising is integral to fulfilling the teaching and


learning mission of higher education. Through academic
advising, students learn to become members of their higher
education community, to think critically about their roles and
responsibilities as students, and to prepare to be educated
citizens of a democratic society and a global community.
Academic advising engages students beyond their own world
views, while acknowledging their individual characteristics,
values, and motivations as they enter, move through, and
exit the institution.

(Preamble, Concept of Academic Advising, NACADA, 2006)

Goals/Objectives
(how we intend to achieve our mission)

These need to emanate from and reflect the nature of the unit to
be assessed (total institution, Advising Center and its
clientele, College Deans Office, etc)
Examples:
To assist students to become independent and lifelong
learners
To assist students in understanding the relevance of the total
curriculum
To assist students in making good decisions based on their
own evidence (e.g., selecting a major)

Identify Outcomes
Student Learning Outcomes examples

All students will select an appropriate major by the end


of their third semester.

All students will become engaged in at least one cocurricular activity each semester.

All students will be able to identify and will select courses


that enhance their human capital.

At least 30% of the students will choose to participate in a


service learning course.

All (CC) students will be able to distinguish among the


A.A., A.S., and A.A.S. degree programs

A Task For Youre: Course


Selection
How many courses are in your Catalog?
(A)
How many courses are required to earn a
degree from your institution? (B)
What percentage of what your institution
offers do students actually take in order to
earn a degree? (B/A)
Now, for each course a student takes, how
many are eliminated?

Outcomes (continued)
Programmatic/Process Outcomes examples
As a result of our advising services, the
retention/persistence rate of first-year students will
increase by 10% in the next 3 years.
As a result of our intervention strategies, the percentage
of students who are removed from academic probation
will increase by 10% in the next academic year.
After two advising sessions, all students will come to
their future sessions with a degree audit already run and
with a plan for meeting outstanding requirements

Everybodys Favorite
All students will be able to
understand, appreciate, and
articulate the value of general
education.

Gather Evidence

Mapping the Experience (Maki, 2004)*

Not all outcomes will necessarily occur as a


direct result of what we do as advisors, so we
need to know what other learning
opportunities exist in order for the students to
meet our stated goals/objectives.
WHAT learning is to occur?
WHERE might it be learned?
By WHEN should it be learned?
*This process can also inform the kinds of evidence that
need to be gathered for appropriate assessment.

The Advising Hub

Types of Measurement and Data


Qualitative open-ended survey questions; focus groups;
in-depth responses, but small N

Quantitative descriptive, structured, numbers and


statistics from surveys, demographics, etc; limited content
responses, but large N

Direct observations; recorded data; pre-post


information

Indirect perceptions, inferences, even inclinations


Use Multiple Measures!!!

Gather (Multiple) Evidence


Satisfaction Surveys

(OK, but not enough)

Institutional Data (changes of major, drop/add


transactions, grades in gateway courses, retention and
graduation rates, use of services provided elsewhere, advisor :
advisee ratios, NSSE, etc)
Office Data (number of appointments vs. walk-ins, nature of
sessions, results of sessions, transcript analyses, other advisor
tasks/activities; What did you learn?)
Focus groups (of clients, of faculty advisors, others a
qualitative measure)
The Advising Syllabus* can inform what evidence should be
collected
*http://www.nacada.ksu.edu/Clearinghouse/AdvisingIssues/syllabus101.htm
http://intraweb.stockton.edu/eyos/page.cfm?siteID=123&pageID=42#syllabus

Share the Results


Tips

Be sure that the stakeholders you identified


earlier are informed throughout the process in
order to enable their support in the
decision-making for implementation of your
recommendations.
Academics have a preferred method of
review, so it makes sense to conform to their
expectations.

Sharing the Results


(Format and Content)
These elements are often best provided in a
standard research report or journal format
Purpose of the assessment project
Method of data collection
Results found
Interpretation of the results
Recommendations with timetable for and
anticipated cost of implementation
Executive Summary or Abstract

How Results Will Inform Decision-Making


Revise pedagogy or curriculum or policy/procedure
Develop/revise advisor training programs
Design more effective programming
advising, orientation, mentoring, etc.
Increase out-of-class learning opportunities
Shape institutional decision making
planning, resource allocation

Sample Implementation Recommendations

Redesign the advising effort in the


Orientation Program
Develop a peer advising/mentoring
program
Streamline office procedures
Initiate proposals for policy changes
Improve communication with other
service offices and personnel
Request/Reallocate resources (human,
fiscal, and/or physical)

You Did It!!


This will complete the assessment cycle,
which provides the evidence for change and
improvement.
Completion of the cycle may also provide
new goals and objectives, new assessment
strategies and tools, and other aspects that
will be need to be included in beginning the
next cycle.
(See Darling, 2005 handout)

Youve Earned a Break


Please take a few minutes to submit any
questions you may have at this point via the
chat function.

Back to the Original


Contexts
People

Academic advising, as a teaching and learning


process, requires a pedagogy that incorporates
the preparation, facilitation, documentation, and
assessment of advising interactions. Although the
specific methods, strategies, and techniques may
vary, the relationship between advisors and
students is fundamental and is characterized by
mutual respect, trust, and ethical behavior.
(Concept of Academic Advising, NACADA, 2006)

NACADA Core Values


Academic Advisors are responsible
to the individuals they advise
for involving others, when appropriate, in
the advising process
to their institutions
to higher education in general
to their educational community
for their professional practices and for
themselves personally

Assessment (Evaluation) of Advisors


SELECTION
TRAINING
EVALUATION
RECOGNITION/REWARD

Selection of Academic
Advisors

Use the best


Add from other resources/units
Target specific populations
Cross disciplinary lines
Develop mentors
Use other skills/expertise

Potential Pitfalls
Making a distinction
Faculty Advising (Programmatic;
Assessment)
Faculty Advisors (Personal; Evaluation)
Inappropriate Comparisons
Professional Academic Advisors
Peer Advisors
No Improvement Plan
Training

Faculty vs. Professional


Staff
Advisors
Too often all are expected or
required to advise, but also
teach, publish, seek grants,
etc no selection
Training ranges from near
nothing to perhaps a day or
2, but usually only a few
hours

They are hired via a search


process and have specific
job descriptions they are
selected

Their training is systematic,


intentional, and ongoing;
staff development is
expected

They are evaluated through


annual performance reviews

They are rewarded with


salary and benefits

Evaluation is not systematic


Recognition/Reward is very
limited in the tenure and
promotion process; mostly
intrinsic; can also be a
reverse structure (better =
more)

ASSESSMENT (Evaluation)
37 %

OF ALL INSTITUTIONS HAD NO


PERFORMANCE EVALUATION MEASURES FOR
FACULTY IN THEIR ACADEMIC ADVISING ROLE

44 %

in 2 yr public institutions

25 %

in 4 yr public institutions

39 %

in 4 yr private institutions
(Habley, 2004)

PARAMETERS (faculty advisors)


Faculty Contract
List of Responsibilities
Availability of Resources
Assignment of Advisees
Recognition/Reward

Tools for Assessment (and/or Evaluation)


Of Advisors
Self evaluation
Student surveys (locally designed)
Survey of Academic Advising (ACT)
Academic Advising Inventory (NACADA)
Student Satisfaction Inventory (Noel-Levitz)
NACADA Clearinghouse

Back to the Original


Contexts
Program
a lack of assessment data can
sometimes lead to policies and practices
based on intuition, prejudice, preconceived
notions, or personal proclivities none of
them desirable bases for making
decisions

(Upcraft and Schuh, 2002, p. 20)

Other Tools and Strategies

Satisfaction Surveys
Institutional Data
Office Data
Focus groups
The Advising Syllabus
External Reviews
CAS Standards
Others

CAS Assessment Worksheet

An Economic Model
Though not an outcomes-based model per se, this
approach to assessment is a functional analysis
based on the premise that every task an advisor
performs and every operation that an advising unit
conducts has some monetary value related to it.
The analysis results in a comparison of the fiscal
expenditures required to perform the tasks to the
cost benefits as results.
The model operates from the perspective of a
threat to the existence of an advising unit,
function or personnel. A quick example

Determining Your Worth


Identify every function the unit performs
Identify all possible alternatives for each
function, if the unit was dissolved
Determine the cost of those functions that
cannot be replaced and who would perform
them; estimates will sometimes be required
Determine the cost of those functions that could
be eliminated

(In Markee and Joslin, 2011)

Where are the data?


Bill Gates colleges today know more
about how many kids attend basketball
games and which alumni give money than
how many students showed up for
economics class during the week
(jn review of Academically Adrift).

Where are the Data?


Jeff Selingo Think about it. Before we buy
a car, we can find various measures on
everything from gas mileage to results of
safety tests. We can turn to objective sources
to check comparisons of similar vehicles and
see which cars hold their value over time. But
when it becomes to potentially one of the most
expensive purchases in a lifetime, the attitude
from colleges has always been that we should
just trust them on the quality of their product.
(p. 25)

What Are We Not AssessingAnd Should We Be?


Student expectations, intentions
Whether advising strategies actually can be
attributed to different types of student
success (removed from probation, successful
choice of major, overcome a skills deficiency
or harmful social habit, etc)
Retention and graduation rates of transfer

students

Expectations vs. Experience


Expect
Be Undecided
7%
Change Majors
12
Fail a course
1
Extra time to complete degree
8
Drop out
1
Transfer institutions
12
Work while in school
36
Seek personal counseling
6
Need tutoring
15
Seek career guidance
5
(Habley 2011)

Experience
20%
65-85
16
60
40
28
60
27
20
25

Non-Assessment

(continued)

Use and value of articulation agreements


number of students who use them, are they
updated
Currency of academic policies, e.g., course
repeats, course pre-requisite criteria,
drop/add/withdrawal processes, academic
warning, probation, and suspension policies
Does advisor training result in better
advising?

Summary
Assessment is a process, not an event
Collaboration and cooperation are necessary for
productive assessment to occur
An ounce of data is worth a pound of opinion
(Magoon, c. 1975)
Avoid the N of 1 syndrome
The purpose and results of assessment should
always be used for program and/or advisor
improvement in order to realize maximum
student development and success

Questions?

Tom.Grites@stockton.edu

Das könnte Ihnen auch gefallen