Energy Resources. ( B. Pharmacy, 1st Year, Sem-II) Natural Resources
Anatomy of course redesign tamu presentation (2)
1. Anatomy of Course Redesign:
How to know what works by
removing the guesswork.
Dr. Mike Simmons | Dr. Ron Carriveau 1
2. INTRODUCTION
This session focuses on the procedures and data
collection used to provide meaningful
information for use by faculty in making valid
instructional and course redesign decisions.
2
3. Why We Did What We Did:
• Course redesign goals and challenges with
which institutions have difficulty
– Linking student learning to Student Learning
Outcomes
– Instruction and assessment providing evidence of
student outcome attainment
– Demonstrating how assessment results are used to
make instructional changes
3
4. Background
• Background
– Five years of QEP implementation
– Providing evidence of student outcome
attainment (NOT GRADES)
– UNT process and methodology used to
address these goals and challenges
4
6. NextGen Outcome Based Model
• Outcome attainment measures
• Instructional methods
• Foundation
– Assessment
– Intended student learning outcomes
– Instructional methods used
6
7. The NextGen Outcome based assessment model
Develop outcome statements that tell what
students are to achieve.
Align instruction to Use outcome statements to
outcome statements develop test items
Use assessment
results to inform
and improve
instruction, asse
ssment, and
outcomes
Develop instructional Develop assessments to
strategies that tell how measure the degree to which
opportunities will be Align instruction
to students are achieving and
provided to help students have achieved.
assessment
achieve.
Source: Carriveau, R.S. (2011). Connecting The Dots: Writing Student Learning Outcomes and Outcome Based Assessments. Fancy Fox
Publications, Denton, TX.
8. Item
1. Specific The Three Level Model
2. Outcome
3. sLO 1.1.1
4.
5. Specific
General Outcome
Outcome
6. GLO 1.2
sLO 1.1.2
7.
8. Specific
9. Outcome Goal
10. sLO 1.1.3 1
11.
12.
13.
Specific
14.
Outcome
15. sLO 1.2.1 General Outcome
16.
GLO 1.2
17. Specific
18. Outcome
19. sLO 1.2.2
Source: Carriveau, R.S. (2011). Connecting The Dots: Writing
20. Student Learning Outcomes and Outcome Based
Assessments. Fancy Fox Publications, Denton, TX.
9. Anatomy of Course Redesign: How to know what works by removing
the guesswork.
RATIONALE FOR ATTAINMENT
9
10. Why Use Attainment Values?
• Valid measure of student attainment of learning
outcomes as the basis for instructional changes and
course redesign
• Provides superior measures to the use of grade
distributions, percentages, and other traditional
achievement measures which lack the ability to
address student learning outcomes at the course
level and provide little to no evidence for course
improvement
10
13. Applying Attainment Measure to a Marketing Class Learning Activity
Experiential Activity: OBSERVING MARKETING AND GENDER STEREOTYPING
The purpose of this assignment is to examine the marketing of toys and sports equipment as well
as advertising images of boys and girls in play and sports contexts. The focus will be on memory
capabilities of adults compared to the memory capabilities of children. Class will be divided into
three groups. Each group will be assigned a specific task to research and will post their findings
online.
Marketing Experiential Activity
Students will be able to detect Item Item sLO GLO
1.2.3 differences between sub‐cultural
market segments’ attitudes toward n=10 % sLO AV GLO AV
brands. 23 83 1.2.3 1.2
18 67 1.2.3 1.2
1.3.1 Students will be able to recognize 22 88 1.2.3 1.2
consumer sub‐cultural market 4 38 1.2.3 69 1.2 69
segments’ VALs.
2 94 1.3.1 92 1.3
13 98 1.3.2 1.3
Students will be able to detect 14 86 1.3.2 1.3
1.3.2 differences between consumer
sub‐cultural market segments’ 16 85 1.3.2 91 1.3 91
attitudes toward brands. 42 56 2.1.2 2.1
32 50 2.1.2 53 2.1 53
2.1.2 Students will be able to relate how
self‐identity may impact consumers
on consumption choices.
14. Examples of Statements That Can Be Made Using
the Three Level Model
• The class as a whole met the criterion on four out of five specific learning
outcomes (sLOs).
• The class met the criterion on both general outcomes (GLO level) and the
associated Goal.
• An improvement goal is to raise the general outcome (GLO level) criteria to
82%.
• Our improvement goal is to increase Goal 1 attainment by 5 points within a
year.
15. Anatomy of Course Redesign: How to Know what works by removing
the guesswork.
INFORMATION FOR
INSTRUCTIONAL DECISIONS
15
16. Information for instructional
decisions
• Need An Analytics (Many Sources of Data)
Approach To Make Good Instructional and
Course Redesign Decisions
– “Analytics is quickly becoming a term that gets
slapped onto any existing product.” G. Siemans
• This is academic (course level analytics) - key
difference from learning (SIS or system
analytics)
16
17. Learning Analytics
Penetrating the Fog: Analytics in Learning and
Education – Educause
“learning analytics is the measurement, collection,
analysis and reporting of data about learners and their
contexts, for purposes of understanding and optimising
learning and the environments in which it occurs.”
1st International Conference on Learning
17
Analytics and Knowledge
19. Obtaining Information About Students
Formative information
• Monitoring learning
• Measuring achievement on the construct scale
• Determining the degree to which student is learning
• Determining the degree to which student meets outcome expectations
Pre-Post Information
• Measuring gains
• Measuring growth
• How far did individual move
• How far did group move
Summative Information
• Making judgments
• Assign proficiency level
20. PRECOURSE DEMOGRAPHIC PROFILE
1. Gender 4. Hours Worked per Week Decision Questions
Male Currently do not work
Female Less than 10 hrs per week
10 to 20 hrs per week
Does this year differ significantly
2.Race 21 to 30 hrs per week
African American greater than 30 hrs per week from last year?
Asian
Caucasian 5. Residence Would seeing this demographic
Hispanic Resident Hall information prior to starting the
Indian City of Denton class cause you to want to make
Non-Resident Alien Denton County changes to the instruction and
Outside of Denton County course. design?
3. Classification
Freshman 6. Age
Sophomore Below 19 How might you make changes in
Junior 19 to 20 instructional design to
Senior 21 to 22 accommodate the profile of the
Post Bac 23 to 25 class you are getting?
Masters 26 to 30
Doctoral 31 to 40
41 to 50
Greater than 50
25. NextGen Information
• Attitude Toward Course Survey
• Learning Environment Preference Survey
• Course Format Preference Survey
• Student Evaluation of Teaching Effectiveness
• Tests
• Engagement
25
26. Student Preferences for N-Gen Course Format
Versus Traditional FTF Format for a COMM course
Table 1 shows the results of the Student Format Preference surveys administered at the
end of each semester. Provided are the counts and percentages of student preferences
for course format plus the preferences by the categories of Successful (grade of A,B,C)
and Unsuccessful (D,F,W,I). The values in parentheses are percentages.
Total Un- Un-
Total Number Success Success success success
Preferred Preferred Number Un- preferred preferred preferred preferred
COMM 1010 N-gen FTF Success successful N-Gen FTF N-Gen FTF
2009 Spring
(n=595) 334(.56) 261(.44) 564(.95) 31(.05) 318(.56) 246(.44) 16(.52) 15 (.48)
2009 Fall (n=507) 232(.46) 275(.54) 472(.93) 35(.07) 216(.46) 256(.54) 16(.46) 19(.54)
2010 Spring
(n=386) 181(.47) 205(.53) 380(.98) 6(.02) 177(.47) 203(.53) 4(.67) 2(.33)
Student Comments
Student comments as to why they preferred an NextGenformat versus an
FTF format were also collected, and the current semester comments are
sent to you in a separate email.
27. Student Preferences for NextGen Course Format Versus Traditional FTF Format
Student Comments
Student comments as to why they preferred an NextGen format versus an FTF format were also
collected. After conducting a study of student responses over a two year period, it was found that
the following categories emerged.
Categories with descriptions for reasons why students chose NextGen or Traditional FTF.
Format Reason Category Description
Students liked that they could control the rate at which they absorbed information.
Pace
NextGen
Flexibility
Students liked that they could do assignments whenever and wherever they wanted.
Learning Students found it easier to learn content when it is internet based.
Practice Students liked that there were more opportunities to practice and learn
Manage Students needed a structure so that they wouldn’t procrastinate.
FTF
Learning Students found it easier to learn content when format is FTF.
People Students felt that they needed the face to face (lecture) interaction.
Technical Students had difficulties with computers , network, and technology used
28. SURVEY RESULTS
STUDENT ATTITUDE TOWARD COURSE SUBJECT (SATCS)
N = 61 Class enrollment = 99 Diff (* is sig at
+/- Pre (SD) Post (SD) .05) (SD)
1 This subject is worth knowing. + 4.29 (0.56) 4.33 (0.66) +0.05 (0.67)
2 I like this subject. + 4.00 (0.55) 4.14 (0.73) +0.14 (0.57)*
3 Knowing this subject makes me more employable. + 3.90 (0.70) 3.81 (0.98) -0.10 (1.09)
4 This subject is easy to learn. + 3.71 (0.64) 3.95 (0.67) +0.24 (0.70)
5 This subject should be required for all students. + 3.29 (1.00) 3.48 (0.93) +0.19 (0.87)*
6 This is a difficult subject for me. - 2.35 (0.67) 1.95 (0.76) -0.40 (0.82)
7 Learning this subject requires a lot of hard work. - 2.86 (0.79) 2.33 (0.86) -0.52 (1.08)
8 Knowing this subject is valuable to me. + 4.14 (0.66) 3.81 (0.81) -0.33 (0.91)
9 This subject makes me feel anxious or uncomfortable. - 1.81 (1.81) 1.52 (0.51) -0.29 (0.64)
10 This subject does not fit into my overall educational needs. - 1.90 (0.77) 2.10 (0.89) +0.19 (0.51)*
11 This subject is interesting. + 4.25 (0.55) 4.05 (0.61) -0.20 (0.70)
12 This subject is difficult to understand. - 2.24 (0.70) 2.00 (0.89) -0.24 (0.89)
13 This is a complicated subject. - 2.45 (0.89) 2.15 (0.88) -0.30 (1.08)
14 I know a lot about this subject. + 3.00 (0.80) 3.35 (0.81) +0.35 (0.81)*
15 This subject is relevant to my personal goals. + 3.80 (0.83) 3.75 (0.97) -0.05 (0.89)*
16 I can learn this subject. + 4.25 (0.44) 4.25 (0.55) 0.00 (0.65)
17 This subject is useful to my everyday life. + 4.00 (0.55) 3.90 (0.77) -0.10 (0.70)*
18 I will have no application of this subject in my profession. - 2.05 (0.87) 1.95 (0.87) -0.10 (1.00)
19 I am scared by this subject. - 1.48 (0.51) 1.52 (0.60) +0.05 (0.59)*
20 I want to learn more about this subject. + 3.95 (0.74) 3.90 (0.70) -0.05 (0.74)*
21 This is a fun subject. + 3.90 (0.77) 4.05 (0.67) +0.14 (0.66)*
29. Learning Environment Preferences Survey (LEP) Report
Course:
Semester:
1. What to Learn
2. How to Learn
3. How to Think
4. How to Judge
31. Engagement – Course Level
• CLASSE is a pair of survey instruments that enable one
to compare what engagement practices faculty
particularly value and perceive important in a
designated class with how frequently students report
these practices occurring in that class.
• CLASSEStudent is the survey instrument completed by
each student enrolled in the designated class, while
CLASSEFaculty is the survey instrument completed by
the faculty instructor of the designated class.
http://assessment.ua.edu/CLASSE/Overview.htm
31
32. Exploring Other Analytics
• Big Data
– McKinsey Global Institute defines big data as
“datasets whose size is beyond the ability of
typical database software tools to capture, store,
manage and analyze.” (James Manyika, “Big Data: The Next
Frontier for Innovation, Competition, and Productivity,” Executive Summary,
McKinsey Global Institute, May 2011)
• Learning management system analytics
32
36. Scenario Which of the following
statements best represents your
You just got your end of course
results on student attainment of conclusion about the particular
outcomes. online activity?
•For a particular online activity that
was designed to help students A. I created a great learning
achieve a high degree of attainment
on a particular sLOs, five test items environment but I need to fix
were used. that dang 45% correct item that
•Four of the test items showed
above 80% percent correct and one didn’t work last year either.
was at 45% correct, which of course
lowered considerably the average B. I need to consider
the attainment value for the sLO.
•Your attitude toward the course
redesigning the online activity
topic survey showed a slight positive so that students will do better
result
•70% of the students who took the
on attaining the sLO.
format preference survey said they
prefer NextGen to traditional FTF. C. I will make sure that next
•There was also high positive semester I spend more review
agreement on the engagement
survey between what you thought time on what the 45% test item
was important and what students
saw happening in the class.
was covering.
36
37. Information and Contacts
• Mike Simmons, Ph.D.
– mike.simmons@unt.edu
• Ron Carriveau, Ph.D.
– ronald.carriveau@unt.edu
http://nextgen.unt.edu
http://clear.unt.edu/
Twitter: @nextgeneducate
http://www.slideshare.net/simmonsweb
37
Editor's Notes
Introduce Ron and MikeMike Simmons, Senior Associate Director in CLEAR - the Center for Learning Enhancement, Assessment and Redesign at the University of North TexasRonald Carriveau, Outcomes, Assessment, and Measurement Specialist. QEP Assistant Director. Also a member of the CLEAR team at the University of North TexasOften when you submit a conference presentation proposal it seems brilliant in your mind. Then you show up at the conference and you realize both how much you don’t know, and sometimes just how much you do know. So far the sessions at this conference have proven more of the former than the latter (not speaking for Ron, just for me). Hopefully though, the value we add will be in leading a discussion/conversation by planting ideas in your head and giving you sufficient time for feedback and commentary.
We are committed to quickly moving through the material so that there is plenty of time for discussion and questions at the end of our session. As a result, we ask you to save your questions for the end of the presentation. Slides are numbered so if you want to come back to any specific slide, just make a note of it.What is your primary responsibility? Instructor, Administrator, Staff, Student, OtherMany roles in the room and in the end probably the best way to look at this is to view us as fellow beggars looking for food.Our conversation is about our efforts to provide information and tools which allow faculty to make valid instructional and course redesign decisions. Our specificfocus is on providing information that is of value to the faculty without overwhelming them. We expect that most of you have similar aspirations or may be ahead of us, but there’s not doubt we are all moving in this same direction.The graphic you see is one of UNT’sNextGeneration course redesign posters that market these courses to students. We’ll use our experience with NextGen as the example for our conversation today.
We need to take just a minute to give you some context and background so that our current directions and ideas will make some sense. First, let me share why we did what we did. At UNT, we found ourselves in the same spot as most other institutions who endeavor to redesign courses. There are many reasons to systematically redesign courses. In our case there were three main concerns that we wanted to try and address:Linking student learning to Student Learning OutcomesInstruction and assessment providing evidence of student outcome attainment Demonstrating how assessment results are used to make instructional changesJane Wellman, George Kuh morning sessions – “intentionality matters”
Here’s what we did:We made course redesign the focus of our QEP. We did not redesign for redesign’s sake, but we chose the redesign as a vehicle by which we could get to the concerns I just listed. Many of you have chosen other vehicles to reach similar destinations.We are at year 5 – we’re just about to submit our SACS 5th year report. We saw early on, and continue to confirm, that we needed something other than grades to provide the evidence of student outcome attainmentSo, we builtNextGenLarge undergraduate core courses.Making Big Classes BetterEngaged learning to create higher level learning experiences.
Outcome attainment measures are used to evaluate how well the class as a whole did on achieve specific learning outcomes and how well particular instructional formats worked in terms of outcome attainment.The instructional methods include combinations of in-class lecture, online instruction, and small group experiential activities.The foundation of this approach is the relationship of the item to the intended student learning outcomes and to the instructional methods used.
I’d like to show you the model, but as you will see, this is something that all of you are doing in one way or another – whether formally or informally.The real focus for the rest of our conversation today is to discuss how we use assessment results to inform and improve instruction.
In order to talk about the information needed to improve instruction, we must first spend a little time setting the scene for how we created our ruler or yardstick. We use a three level model and Ron will be discussing this in more detail at his session on Tuesday. For our purposes today, we just want to show you how it works in terms of the courseYou might be saying to yourself “wait, there are four levels on that diagram”. And you’d be right, but as Ron will soon discuss, the item column simply allows the specific assessments to be fully mapped to the goal levels. More in this in a moment. But for now, please notice that the arrows point in different directions because as you all know, we work in different directions at various times. Sometimes we have external or higher level goals that are prescribed, suggested, mandated for us, so that’s where we begin. Other times, the instructor has specific outcomes that are needed for the course. Most times, this process is a combination of the two scenarios – some goals mandated, and some sLO’s in place. It’s the point of the three level model to provide a framework for this to happen in an organized manner.Dr. Carriveau will now spend a few minutes talking about how the three level model allows us to calculate the attainment values that serve as our yardstick in determining what can actually be improved in instruction.Dr. Kuh challenged us this morning to “show the data”…so that is where we started…with a yardstick.
We now have the ruler…attainment values.We can begin to understand how changes in instruction have an impact on learningWe are a campus with lots of high impact practicesNextGen contains a number of the practices…and integrative learningWe are here to explore
Let’s start with what we all have in common. The big ideas are great, but most of us need to pick a simple starting point:Formative Monitoring learningMeasuring achievement on the construct scaleDetermining the degree to which student is learningDetermining the degree to which student meets outcome expectationsPre-Post informationMeasuring gainsMeasuring growthOnly those with pre and post scoresHow far did individual moveHow far did group moveSummative informationMaking judgments Assign proficiency levelAssign gradDetermine if retake is allowed
What are some other types of information that institutions will typically have? Demographic information.We are also finding that other course demographics, like size of course, time of day, room type, and student demographics have an impact on the dataKuh…said today that when we control for variables these types of information have less impact than we think…or do they? Perhaps it’s impolite to disagree with the plenary speaker, so I won’t. As Dr. Kuh said, it’s more important what we do in the instruction.
Dashboarding and visual presentation is more user friendly for quicker decisions
Source is institutional data, in our case Instituitional Research and EffectivenessAt this point, institutions are not likely to produce dashboard
Nextgen v. non
Additional, specific information that we collected and use for Next Gen – again these are not terribly unusual.
Pre and post
What to Learn.Student who prefers this approach to learning focuses on facts and primarily wants only to be given “right” answers to specific questions. The teacher is viewed as an Authority and the only source of information. Learning is viewed as an exchange of information and content. A clear-cut, objective testing/grading method is preferred. A straight-lecture format is preferred with the teacher being in complete control of the classroom environment. Uncertainty and fuzziness in the learning process is not acceptable.2. How to Learn. Student who prefers this approach to learning focuses on the methods and processes of learning, like problem solving. This person needs to have a variety of activities/methods used in class. Learning is seen as the quantity of facts, subjects, and/or methods that can be learned. Class discussions in which the opinions of others are expressed are desirable. The teacher needs to be more than just a source of facts. Challenges are enjoyable. Hard work is viewed as being the primary factor in success. 3. How to Think. Student who prefers this approach to learning focuses on independent thinking as the primary purpose of learning. The tendency is to integrate the process of learning with content. Making connections between classes and across disciplines is important. The need for presenting reasoned arguments is understood. There is a tendency to reject rote learning and memorization. Essay tests are preferred and viewed as opportunities to demonstrate thinking. Learning about self is viewed as an important part of education. 4. How to Judge. Student who prefers this approach to learning focuses on the synthesis of different ideas and viewpoints in specific disciplines or areas. Peers in class are viewed as being genuine sources of learning (in addition to the teacher). The teacher is valued as the expert, but the preference is that the teacher assume the role of facilitator and co-participant in the learning. Judging ideas and arguments begins to be based on the quality of the evidence presented. There is a tendency to be a self-directed learner, seeking new challenges on one’s own. Evaluation is viewed as constructive criticism and as an opportunity for learning.
CategoriesDifferences over time in course
As discussed in the plenary session earlier today, we are all chasing the holy grail of of “engagement”…but at the classroom level seems to be where we need to know more in order to make valid instructional decisions.Working in an active partnership with Dr. Bob Smallwood and the CLASSE team to move this conversation forward. Lots of questions…but this is as good a place to start as any.
Student academic behaviors and other performance data that is increasingly available to institutions
So what? Well I don’t know yet. But I want faculty to consider things like this regularly within the context of their own discipline. And I want the information to be available easily – without faculty overhead.
You just got your end of course results on student attainment of outcomes. For a particular online activity that was designed to help students achieve a high degree of attainment on a particular sLOs, five test items were used. Four of the test items showed above 80% percent correct and one was at 45% correct, which of course lowered considerably the average the attainment value for the sLO. Your attitude toward the course topic survey showed a slight positive result, and 70% of the students who took the format preference survey said they prefer NextGen to traditional FTF. There was also high positive agreement on the engagement survey between what you thought was important and what students saw happening in the class. Which of the following statement best represent your conclusion about the particular online activity. A. I created a great learning environment but I need to fix that dang 45% correct item that didn’t work last year either. B. I need to consider redesigning the online activity so that students will do better on attaining the sLO. C. I will make sure that next semester I spend more review time on what the 45% test item was covering.
Mike: Implementation of this metrics driven approach includes constant exploration of what form data should take to be most useful - especially in terms of the many workload and administrative challenges that a faculty member faces in an education system that is ever-increasing in demands on instructors.Our overall intent is to find the best ways to support faculty as they endeavor to make good instructional decisions.