Hi Tips folks!.. Happy New Year!

   Below is an assortment of comments contributed by members of faculty to
   the well known (and somewhat celebrated!) Chronicle of Higher Education:
   Colloquy--1998 -- cleaned-up and internet friendly. I suspect that they will
   provide members of faculty with a rather useful resource when dealing with
   and/or debating some student ratings issues.

   - All the very best!
     -----------

    The January 16 Chronicle article on "Student
    Evaluations" quoted one researcher as
    saying, "With student ratings, you've got
    someone who's sat through 40 hours of a
    course." Similarly, among these Colloquy
    responses, another writer opined that "It is
    ironic that so many individuals believe that
    students are not capable of providing
    meaningful evaluative feedback or make
    accurate reflections on their experiences."
    These statements seem to imply that students
    give accurate ratings. Relevant to this
    proposition is the following empirical
    demonstration provided by L. E. Stanfel
    (1995, Journal of Instructional Psychology,
    22, pp. 117-125):

    "For Questions 1 ["The instructor's]
    objectives for the course have been made
    clear."] and 4 ["The instructor explains
    clearly to students how they are
    evaluated."] the vehicle of [documentation]
    was an open-book, multiple-choice quiz ...
    every student answered both quiz parts
    correctly and thereby proved that they knew
    both the course objectives and the procedure
    by which they would be evaluated. ... For
    Question 5 ["Tests and written assignments
    are graded and returned in a reasonable
    period of time."] a dated sign-up sheet was
    circulated at each occasion when graded
    documents were returned ... The original
    documents' being dated, the student
    signatures proved the graded documents had
    been returned to class at the earliest
    possible opportunity" (Stanfel, 1995).

     For the first course reported by Stanfel
    (1995), Question 1 received: 1 strongly
    agree; 6 agree; 7 uncertain; 19 disagree;
    and 6 strongly disagree. The data reported
    allow the following calculations: on a
    5-point scale, if 5 were the best score, the
    average response to Question 1 should have
    been 5.00; instead, the average response to
    Question 1 was 2.41; the averages for
    Questions 4 and 5, respectively, were 2.64
    and 3.79; for the second course, the
    averages for Questions 1, 4, and 5 were
    1.88, 2.00, and 3.75, respectively. As
    Stanfel notes, "since [responses to three
    questions] have been proven patently
    incorrect, it is impossible to suppose the
    remainder [of responses to the
    questionnaire] is any less erroneous"
    (Stanfel, 1995).

    This negative halo effect demonstrated by
    Stanfel probably reflects the discriminant
    invalidity which Greenwald and Gillmore
    document (1997, American Psychologist, 52,
    p. 1214). Potentially making matters worse,
    Marsh and Roche (1997, American
    Psychologist, 52, p. 1188), two enthusiastic
    defenders of student evaluations of
    instruction, speculate that "Global or
    overall ratings ... may be more susceptible
    to context, mood, and other potential biases
    than are specific items that are more
    closely tied to actual teaching
    behaviors..." If that is the case, then
    student responses to any global items on
    Stanfel's (1995) questionnaire would have
    been even more untruthful than those to the
    specific items described above. So there you
    have it: students are in a position to
    provide accurate reports about their
    classroom experiences, but the fact is that
    in many cases they do not provide accurate
    reports.

    I would also like to comment on what one of
    the writers in this colloquy refers to as
    "forty years of very consistent research
    demonstrating the underlying validity and
    usefulness of student-provided data." Not
    even counting the recent work by Greenwald
    and others, consider the negative evidence
    provided by several actual classroom
    experiments such as that of Chacko (1983,
    Educational Research Quarterly, 8 (2),
    19-25):

    "... Pre- and posttreatment measures of Ss'
    perceptions of instruction were obtained
    from an experimental and a control group of
    Ss using a Likert-type scale consisting of
    items describing commonly shown teaching
    traits and behaviors. The 2 groups did not
    vary in their premeasure scores. A midterm
    examination was administered, then evaluated
    in such a way that Ss in the experimental
    group were more harshly graded than Ss in
    the control group. Results show that the
    ratings of Ss in the experimental group
    changed to reflect more negative perceptions
    of instruction. These shifts were
    statistically significant in 7 of the 10
    rating items. Such changes were not seen
    in the control group" (Chacko, 1983).

    Findings such as these, bearing negatively
    on the validity of student ratings, have
    been dismissed on spurious grounds by
    unabashed advocates of student evaluations.
    These advocates would gain a lot more
    credibility if they would acknowledge that
    student ratings have both validity and
    invalidity and if they would constructively
    address the invalidity problem. In the
    meantime, the rest of the academic community
    must be vigilant and resolved in the face of
    Pollyannaish claims about student ratings
    that can then be used to justify misuse of
    ratings by administrators who are so
    inclined.

         -- James M. Puckett, Associate
         Professor of Psychology, Texas A&M
         University, Kingsville (posted 1/26, 12
      --------------------------------------------

    I have just two simple points with respect
    to student ratings:



    (1) Students are neither objective nor fully
    informed. They can have axes to grind. They
    do not know what is good teaching in the
    discipline or whether the content is being
    handled thoroughly or appropriately. That is
    why they are not teaching the class.

    (2) Higher education is the only part of the
    "consumer economy" where the "customer"
    expects less for his or her money than they
    are paying. If education is merely
    credentialization, then the student may
    reasonably expect that any professor who
    impedes or disturbs progress towards the
    credential is doing him or her a disservice.
    However, if learning is truly the student's
    highest priority (instead of the 35 hours a
    week job for minimum wage to pay for the car
    and apartment and cable TV), the student
    needs to assume primary responsibility for
    learning (or not doing so). All the
    professor can do is provide a suitable
    structure to assist the student. That
    structure must include sufficient rigor and
    demands on the student's time and effort,
    even if discomfort is produced.

         --Anthony Duben, Assoc. Dean, Col. of
         Science and Tech., Southeast Missouri
         State Univ. (posted 1/30, 12:20 p.m.,
         E.S.T.)
        _______________________________________

      It is hardly surprising that individuals
    representing the evaluation industry within
    academe (both those who produce instruments
    and correlate the results, and those
    administrators who use the products to
    determine promotion and/or tenure) would
    defend the validity, integrity, and goals
    of their work.

    Part of the problem is cultural, and part is
    a combination of epistemology and ethics.
    The cultural problem is that no matter how
    much we may seek to produce value-neutral
    means for measuring performance, inevitably
    evaluation rankings confer actual or
    perceived social status along with "neutral"
    measurement. An "A" student has higher
    status than a "C" student, and the same is
    true with professorial evaluations. This
    means that no matter how much individuals or
    systems may seek objectivity, cultural value
    systems inhibit these aims.

    The ethical and epistemological issue
    concerns the debate between process and
    outcome as aims of education. In a tradition
    extending from Aristotle to Dewey, it has
    been held that the ultimate test of a
    successful practice can come only at the end
    of one's life. It is only in retrospect,
    surveying the narrative of a whole life,
    that one can truly know whether one has
    lived a good life. Meanwhile, one immerses
    oneself in the process of seeking the mean.
    The measure of such is the "person of
    practical wisdom"; a person and her
    character, not some feeling state or
    measurable performance level in knowledge or
    skills.

    Those who advocate outcome based assessment
    of student learning as a major criterion for
    teacher evaluation stand within the
    utilitarian tradition of the greatest good
    for the greatest number. They tend to
    believe that quantitative evaluation of
    outcome, in terms of goods such as pleasure,
    is both possible and worthwhile; indeed,
    they often believe that the only "real"
    results are those that can be measured by
    whatever criteria they have devised. They
    are concerned with product rather than
    process, with how to measure the
    satisfaction of consumer desire rather than
    the character of learners.

    Utilitarianism is the default position of
    middle management in industry for reasons
    that are self-evident. It is not clear to me
    that such ought to be the case with
    professions in general and with teaching in
    particular. But then, I believe that there
    are rational, other-regarding, valid reasons
    to defend the importance of developing
    inter-personal character in cultural,
    social, and political contexts, and some of
    my colleagues in the social sciences do not.

         -- David H. Fisher, Ph.D., Chair &
         Professor of Philosophy (posted 1/16,
         10:45 a.m., E.S.T.)
        ----------------------------------------

    Those who advocate outcome based assessment
    of student learning as a major criterion for
    teacher evaluation stand within the
    utilitarian tradition of the greatest good
    for the greatest number. They tend to
    believe that quantitative evaluation of
    outcome, in terms of goods such as pleasure,
    is both possible and worthwhile; indeed,
    they often believe that the only "real"
    results are those that can be measured by
    whatever criteria they have devised. They
    are concerned with product rather than
    process, with how to measure the
    satisfaction of consumer desire rather than
    the character of learners.

    Utilitarianism is the default position of
    middle management in industry for reasons
    that are self-evident. It is not clear to me
    that such ought to be the case with
    professions in general and with teaching in
    particular. But then, I believe that there
    are rational, other-regarding, valid reasons
    to defend the importance of developing
    inter-personal character in cultural,
    social, and political contexts, and some of
    my colleagues in the social sciences do not.

         -- David H. Fisher, Ph.D., Chair &
         Professor of Philosophy (posted 1/16,
         10:45 a.m., E.S.T.)
        ______________________________________

    Student evaluations of university professors
    are of very limited validity. If, on the
    evaluation form, the student were simply
    asked to list their anticipated course
    grade, an extremely high correlation between
    the evaluation and the expected grade would
    be observed.

    This is well known and clearly invalidates
    the entire exercise. The obvious result is
    extreme grade inflation. It also results in
    exorcising the best teachers from the
    classroom -- the persons who refuse to
    succumb to the resultant pressure for grade
    inflation.

    The pressure comes from the "Administrative
    Class" that operates the universities. For
    the administrator, a good teacher is, by
    definition, someone who gives high grades.
    Their logic is childishly simple. If the
    students get poor grades, then the teacher
    either cannot teach, or the teacher cannot
    motivate the students to learn, or both.

         -- Charles A. Weatherford, professor
         and Chairman of Physics, Florida A&M
         University (posted 1/12, 12:35 p.m.,
         E.S.T.)
        -----------------------------------------

     There is currently a debate simmering (it's
    not quite raging, yet) among health-care
    providers and evaluators concerning
    "consumer satisfaction" surveys. At times it
    seems that, while "everybody's doing them,"
    no one is really certain what they measure
    or what their worth is in evaluating
    providers (individuals or organizations) of
    health care.

    In many ways, this parallels the situation
    in academe. It is important, for a number of
    reasons, to allow consumers to provide an
    assessment of their experiences -- in the
    emergency room or the classroom. Such
    assessments should be one element of an
    "evaluation" of the service or experience.
    To allow them to become the major type of
    information -- or worse, the only
    information -- to judge the worth or value
    of the enterprise is tempting because
    questionnaire-type surveys are so easy to
    administer. But this focus on only one
    source or type of information should be
    resisted, if possible.

    To quote a famous oracle of program
    evaluation (or was it Yogi Berra?), we
    should seek out simplicity and them distrust
    it!
        -- John Myers, Administrator of
         Planning & Evaluation, Trumbull County
         Alcohol, Drug, and Mental Health Board
         (posted 1/16, 11:30 a.m., E.S.T.)

       -------------------------------------------

        The deleterious effects of student
    evaluations are diverse and affect all
    professors who aspire to maintain high
    standards in their evaluation of student
    performance. What I find particularly
    distressing is the general attitude of many
    students who "know the game." That is to
    say, they are well aware of the impact of
    their assessments and will often focus on
    mundane aspects of a professor's performance
    simply because they have been awarded a
    grade which did not fit their expectations
    before they entered the course.

    If a professor is to believe or place any
    credence to these, often not relevant,
    comments she/he will in effect consciously
    or subconsciously construct course
    requirements based not on expertise gained
    from present and past research, but; on what
    is needed to satisfy the wishes of students
    who can hardly be expected to promote a
    course which requires "good old fashioned
    hard work."

         -- Al Figone, professor, Humboldt State
         University (posted 1/12, 12:30 p.m.,
         E.S.T.)
        ---------------------------------------



Modern man has developed a kind of Gallup-poll mentality,
relying on quantity instead of quality and yielding
to expediency.
>>Walter Gropius

Successful teachers are effective in spite of
the psychological theories they suffer under.
>>Old Educational proverb
>>
---------------------------------->>>
[[(#:-) >>>>
---------------------------------------->>>
                              +
....John C Damron, PhD        *
....Douglas College, DLC   *  *  *
....P.O Box 2503         *    *    *
....New Westminster, British Columbia
....Canada V3L 5B2  FAX: (604) 527-5969
....e-mail: [EMAIL PROTECTED]

                http://www.douglas.bc.ca/

        http://www.douglas.bc.ca/psychd/index.html


 Student Ratings Critique:

 http://www.mankato.msus.edu/dept/psych/Damron_politics.html



Reply via email to