thinking broadly about educational technology

54
Thinking broadly about educa3onal technology Edward Price [email protected] Department of Physics

Upload: others

Post on 15-Nov-2021

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Thinking broadly about educational technology

Thinking'broadly'about'educa3onal'technology'

'Edward'Price'[email protected]'

''''

Department'of'Physics'

Page 2: Thinking broadly about educational technology

Acknowledgements''

Noah'Finkelstein,'Fred'Goldberg,''Steve'Robinson,'Beth'Simon'

Charles'De'Leone,'Stephen'Tsui,''Clarisa'Bercovich,'Robin'Marion''

Chase'Keene,'Kornel'Czarnocki,'Alicia'Hart,''Lydia'Saucedo,'John'Saunders'

CSUSM'IITS''

HewleP'Packard'Technology'for'Teaching'MicrosoQ'External'Research'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 3: Thinking broadly about educational technology

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 4: Thinking broadly about educational technology

I'believe'that'the'mo3on'picture'is'des3ned'to'revolu3onize'our'educa3onal'system'and'that'in'a'few'years'it'will'supplant'largely,'if'not'en3rely,'the'use'of'textbooks.''The'educa3on'of'the'future,'as'I'see'it,'will'be'conducted'through'the'medium'of'the'mo3on'picture…'where'it'should'be'possible'to'obtain'100%'efficiency.'

Thomas'Edison,'1922'

Cuban,'Larry.'Teachers'and'Machines:'The'Classroom'Use'of'Technology'Since'1920.''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 5: Thinking broadly about educational technology

Typical(classroom-(today(

Earliest(known(example(of(a(school-room(from(Sumer,(circa(3000(BC(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 6: Thinking broadly about educational technology

Will'computers'‘fix’'educa3on?'

Knowledge'Comprehension'Applica3on'Analysis'Synthesis'Evalua3on'

Problem'solving'Communica3on'Collabora3on'

Management'of'complex'tasks'Nature'of'science'

What(do(we(mean(by(learning?(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 7: Thinking broadly about educational technology

•  Before'thinking'about'technology'specifically,'what'do'we'know'about'teaching/learning?'

•  What'are'our'goals?'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 8: Thinking broadly about educational technology

Scien>fic'teaching'involves'ac>ve'learning'strategies'to'engage'students'in'the'process'of'science'and'teaching'methods'that'have'been'systema>cally'tested'and'shown'to'reach'diverse'students''

521

Since publication of the AAAS 1989 re-port “Science for all Americans” (1),commissions, panels, and working

groups have agreed that reform in scienceeducation should be founded on “scientificteaching,” in which teaching is approachedwith the same rigor as science at its best (2).Scientific teaching involves active learningstrategies to engage students in the processof science and teaching methods that havebeen systematically tested and shown toreach diverse students (3).

Given the widespread agreement, it mayseem surprising that change has not pro-gressed rapidly nor been driven by the re-search universities as a collective force.Instead, reform has been initiated by a few pi-oneers, while many other scientists have ac-tively resisted changing their teaching. Sowhy do outstanding scientists who demandrigorous proof for scientific assertions intheir research continue to use and, indeed, de-fend on the basis of the intuition alone, teach-ing methods that are not the most effective?Many scientists are still unaware of the dataand analyses that demonstrate the effective-ness of active learning techniques. Othersmay distrust the data because they see scien-tists who have flourished in the current edu-cational system. Still others feel intimidatedby the challenge of learning new teachingmethods or may fear that identification asteachers will reduce their credibility as re-searchers (3).

This Policy Forum is needed becausemost scientists don’t read reports but theydo read Science. In addition, reports gener-ally do not offer a guide to learning how to

do scientific teaching, as we do with sup-porting online material (SOM) (3) and table(see page 522). We also present recommen-dations for moving the revolution forward.

Implementing Change in Lectures Active participation in lectures and discovery-based laboratories helps students develop thehabits of mind that drive science. However,most introductory courses rely on “transmis-sion-of-information” lecturesand “cookbook” laboratory ex-ercises—techniques that are nothighly effective in fostering con-ceptual understanding or scien-tific reasoning. There is mount-ing evidence that supplementingor replacing lectures with activelearning strategies and engagingstudents in discovery and scien-tific process improves learning and knowl-edge retention (3).

Introductory classes often have high en-rollments, frequently approaching 1000students in biology courses. This need notbe an impediment to scientific teaching.Many exercises that depart from traditionalmethods are now readily accessible on theWeb, which makes it unnecessary for teach-ers to develop and test their own (3).Quantitative assessment indicates that theseinteractive approaches to lecturing signifi-cantly enhance learning, and although timeallocated to inquiry-based activities re-duces coverage of specific content, it doesnot reduce knowledge acquisition as meas-ured by standardized exams (4).

Faculty are also using computer sys-tems to engage students, assess learning,and shape teaching. Students can be askedto read and solve problems on a Web site,and their answers can be analyzed beforeclass to guide the design of lectures (3).

Some scientists have replaced lectures al-most entirely. Laws’s course “Calculus-BasedPhysics Without Lectures” at DickinsonUniversity (5) and Beichner’s program,SCALE-UP, at North Carolina State Uni-versity (see figure, this page) rely on a prob-lem-based format in which students work col-laboratively to make observations and to ana-lyze experimental results. Students wholearned physics in the SCALE-UP format at a

wide range of institutions demonstrated betterproblem-solving ability, conceptual under-standing, and success in subsequent coursescompared with students who had learned intraditional, passive formats (3).

These results are neither isolated nordiscipline-specific. At the University ofOregon, Udovic showed dramatic differ-ences between students taught biology in atraditional lecture and those taught “Work-shop Biology,” a series of active, inquiry-based learning modules (6). Similarly im-pressive results were achieved by Wright ina comparison of active and passive learningstrategies in chemistry (7). Others havetaught cross-disciplinary problem-basedcourses that integrate across scientific dis-ciplines, such as Trempy’s, “The World

According to Microbes,” atOregon State University,which integrates science,math, and engineering. Thecourse serves science ma-

jors and nonmajors, and outcome assess-ments indicate high content retention andstudent satisfaction (8).

Students as ScientistsScientists of all disciplines have developedinquiry-based labs that require students todevelop hypotheses, design and conduct ex-periments, collect and interpret data, andwrite about their results (9). Many of theseinvolve simple, inexpensive materials con-figured so that they invite students to asktheir own questions. In addition to labs thathave already been tested in the classroom,resources are available to help teachers con-vert cookbook labs into open-ended, in-quiry-based labs (3). Some schools provideintroductory-level students with the opportu-nity to conduct original research in a profes-sor’s research lab rather than take a tradition-

E D U C AT I O N

Scientific TeachingJo Handelsman,1* Diane Ebert-May,2 Robert Beichner,3 Peter Bruns,4

Amy Chang,5 Robert DeHaan,6,† Jim Gentile,7 Sarah Lauffer,1

James Stewart,8 Shirley M. Tilghman,9 William B. Wood10

1Howard Hughes Medical Institute Professor, Departmentof Plant Pathology, University of Wisconsin–Madison;2Department of Plant Biology, Michigan StateUniversity; 3Department of Physics, North CarolinaState University; 4Howard Hughes Medical Institute;5American Society for Microbiology; 6National ResearchCouncil; 7Dean of Natural Sciences, Hope College;8Department of Curriculum and Instruction, Universityof Wisconsin–Madison; 9President, Princeton Uni-versity; 10Department of Molecular Cellular, andDevelopmental Biology, University of Colorado atBoulder. For complete addresses, see SOM.

*To whom correspondence should be addressed.E-mail: [email protected].†Present address: Division of Educational Studies,Emory University, Atlanta, GA, 30322, USA.

POLICY FORUM

A physics classroom at North Carolina StateUniversity arranged for traditional lectures (in-set) and redesigned for group problem-solvingin the SCALE-UP program.

www.sciencemag.org SCIENCE VOL 304 23 APRIL 2004

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 9: Thinking broadly about educational technology

Learning'principles'

1.'Learning'builds'on'prior'knowledge'2.'Learning'is'a'complex'process'requiring'scaffolding''3.'Learning'is'facilitated'through'interac3on'with'tools''4.'Learning'is'facilitated'through'peer'interac3ons'5.'Learning'is'facilitated'through'establishment'of'

'norms'and'expecta3ons'

J.'D.'Bransford,'A.'L.'Brown,'and'R.'R.'Cocking,'How'People'Learn:'Brain,'Mind,'Experience,'and'School'(NAP,'Washington,'D.C.,'2000).'E.'F.'Redish,'Implica3ons'of'Cogni3ve'Studies'for'Teaching'Physics,'American'Journal'of'Physics'62,'796'(1994).'D.'Hammer,'et'al.,'in'Transfer'of'Learning:'Research'and'Perspec>ves,'edited'by'J.'Mestre'(Informa3on'Age'Publishing,'2004).'Y.'Engeström,'Learning'by'Expanding:'An'Ac>vityMtheore>cal'Approach'to'Developmental'Research'(OrientalKonsul3t'Oy,'1987).'M.'Cole,'Cultural'Psychology:'A'Once'and'Future'Discipline'(Harvard'University'Press,'Cambridge,'MA,'1998).'L.'S.'Vygotsky,'Thought'and'Language'(MIT'Press,'Cambridge,'MA,'1986).'E.'G.'Cohen,'Designing'Groupwork:'Strategies'for'the'Heterogeneous'Classroom'(Teachers'College'Press,'1994),'p.'203.'R.'Driver,'P.'Newton,'and'J.'Osborne,'Establishing'the'Norms'of'Scien3fic'Argumenta3on'in'Classrooms,'Science'Educa3on'84,'287l312'(2000).'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 10: Thinking broadly about educational technology

•  Where'does'technology'come'in?'•  How'can'we'think'about'how'technology'gets'used'and'changes'the'classroom?'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 11: Thinking broadly about educational technology

Clickers or Flashcards: Is There Really a Difference?N. Lasry, John Abbott College, Montreal Canada, and Harvard School of Engineering & Applied Sciences, Cambridge MA

A growing number of physics teachers are cur-rently turning to instructional technologies such as wireless handheld response systems

—colloquially called clickers. Two possible rationales may explain the growing interest in these devices. The first is the presumption that clickers are more effective instructional instruments. The second rationale is somewhat reminiscent of Martin Davis’ declaration when purchasing the Oakland Athletics: “As men get older, the toys get more expensive.” Although personally motivated by both of these rationales, the effectiveness of clickers over inex-pensive low-tech flashcards remains questionable. Thus, the first half of this paper presents findings of a classroom study comparing the differences in stu-dent learning between a Peer Instruction group using clickers and a Peer Instruction group using flash-cards. Having assessed student learning differences, the second half of the paper describes differences in teaching effectiveness between clickers and flashcards.

About Peer InstructionPeer Instruction (PI) is a student-centered instruc-

tional approach developed at Harvard by Eric Mazur.1 The method has been welcomed by the science com-munity and adopted by a large number of colleges and universities due, among other reasons, to its common sense approach and its documented effectiveness.1,2 A schematic description of the PI method used in this study is shown in Fig. 1.

In PI, the progression of any given class depends on the outcome of real-time student feedback to Con-

cepTests: multiple-choice conceptual questions. In the early 1990s, students displayed their answer to Con-cepTests using a show of hands and later flashcards. Instructors would then count or estimate the number of students holding each alternative conception. Due to the tediousness of counting flashcards in large en-rolment courses, flashcards were replaced with wired classroom communication systems3 and later with wireless clickers.

Study Description First-semester students in a two-year Canadian

public community college were randomly assigned by the registrar to one of two sections of an algebra-based mechanics course. Instruction in the first section consisted of PI with clickers (n = 41) while the other

ConcepTest Students Vote

correct ans <30%

Revisit Concept

correct ans: 30%-80%

Peer discussion (2-3min) students try to convince

each other

Students revote

Brief lecture ( 10min)

correct ans >80%

Remaining misconception

explained

Next Topic

Fig. 1. A Peer Instruction Implementation Algorithm.

242 DOI: 10.1119/1.2895678 THE PHYSICS TEACHER Vol. 46, April 2008

What'do'we'do'with'clickers?'Technology'≠'pedagogy'

Beaty'&'Gerace'2009;'Lasry,'2008;''Mazur,'Peer'Instruc>on'(1997)'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 12: Thinking broadly about educational technology

Clickers'vs'how'we'use'them'

Clickers(as(a(tool(•  Fast,'easy,'private''•  Limited'answer'choices'•  Response'from'all'students'•  Formalize'par3cipa3on'•  Automate'sharing'•  Provide'referent'for'

discussion'•  Save'data'for'review,'

grading,'research'

Pedagogies(featuring(class(response(

•  Reading'quizzes'•  In'class'conceptual'ques3ons'•  Peer'Instruc3on'(Mazur)'•  Ques3on'sequences'(Bao)'•  Ques3on'driven'instruc3on'

(BeaPy)'

Beaty'&'Gerace'2009;'Lasry,'2008;'Mazur,'1997;'Reay,'Li,'&'Bao,'2008'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 13: Thinking broadly about educational technology

Thinking'about'tools'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 14: Thinking broadly about educational technology

Thinking'about'tools'

•  Affordances'•  Constraints'•  Tools'shape'what'we'do'•  Enable'new'possibili3es'•  Not'determinis3c'

Finkelstein,'et'al.,'2005;'Lasry,'2008;'Norman,'1988;'Thornton'&'Sokoloff,'1990'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 15: Thinking broadly about educational technology

Tools'&'pedagogy…'is'that'it?'•  Norms''

–  ' sense'making'–  ' responsibility'for'genera3ng'ideas'–  ' responsibility'for'evalua3ng'ideas'

•  Roles'–  Who'does'what'

•  Instructor'ac3ons,'grading'prac3ces'lead'to'norms,'perceived'by'students'

•  Classrooms/instructors'have'varia3on'in'norms'and'prac3ces''•  Implica3ons'for'feedback'and'how'it'is'used'

James'&'Willoughby,'2011;'Turpen'&'Finkelstein,'2010'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 16: Thinking broadly about educational technology

Small'Group'Discussion'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 17: Thinking broadly about educational technology

A'framework'for'thinking'about'the'physics'classroom'

(Other'students,'instructor)'

(How'do'things'work'here?)'

(Who'does'what?)'

(“Technology”'but'also'

representa3ons,'language,'etc)'

Cole,'1996;'Engeström,'1987;'Kaptelinin'&'Nardi,'2006;'Nardi,'1996'

A'student'learning'physics'is'engaged'in'an'ac3vity'as…'

part'of'a'community…'

with'rules/norms…'

and'roles…'

using'tools…''

In'a'broader'context'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 18: Thinking broadly about educational technology

YOUR(TURN(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 19: Thinking broadly about educational technology

For'MBL'and/or'simula3ons…'

•  How'does'the'use'of'this'tool'relate'to'your'goals?'

•  For'the'ways'you'(might)''use'them,'how'does'the'tool'–  Reorganize'who'does'what?'–  Change'par3cipa3on?'–  Allow'new/different'norms?'–  Reinforce/support'exis3ng'norms?''

Finkelstein,'et'al.,'2005;'Lasry,'2008;'Norman,'1988;'Thornton'&'Sokoloff,'1990'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 20: Thinking broadly about educational technology

A'whiteboardlintensive'physics'class'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 21: Thinking broadly about educational technology

Whiteboards'provide'good'collabora3ve'space,'but'are'vola3le'and'fixed'in'size'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 22: Thinking broadly about educational technology

Tablets'as'digital'whiteboards'

•  Student'groups'use'Tablet'PCs'and'work'on'slides'prepared'by'the'instructor''

•  Student'slides'can'be'projected'for'whole'class'discussion'and'are'archived'on'the'web'

Price'&'Simon,'2009'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 23: Thinking broadly about educational technology

Projec3ng'student'work'during'whole'class'discussion'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 24: Thinking broadly about educational technology

Archiving'student'work''

Students'found'online'archive'very'useful,'many'page'views,'including'student'solu3ons'We'had'lots'of'ideas'for'how'to'use'this'to'close'the'loop,'but…'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 25: Thinking broadly about educational technology

Tablets'and'collabora3on'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 26: Thinking broadly about educational technology

Digital'photographs'and'Flickr.com'

Whiteboards'+'archiving'•  Cameras'w/'wirelesslenabled'SD'cards''•  Courselspecific'Flickr'account'•  Images'can'be'organized'•  Tags'and'comments'

Price,'2011'Ed'Price,'APS'Experienced'Faculty'

Workshop,'April'2013'

Page 27: Thinking broadly about educational technology

Students'find'online'archive'very'useful,'many'hits'before'quizzes,'emphasis'on'student'work'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 28: Thinking broadly about educational technology

Feedback'and'revising'thinking'

Before'Flickr'was'used'– Student'work'was'seldom'edited'aQer'the'class'discussions.'

With'Flickr'archive'– Students'began'to'edit'whiteboards'aQer'whole'class'presenta3ons…'“fixing”'mistakes'

– The'photo'now'captured'an'edited'solu3on'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 29: Thinking broadly about educational technology

“Closing'the'loop”'

•  Photographing'whiteboards'mo3vated'a'final'round'of'instructor'feedback'and'student'revision.'

•  Unintended'and'unexpected'•  Students'now'take'responsibility'for'evalua3ng,'correc3ng'their'work'

•  Arose'from'student'interest'in'their'work'as'a'resource,'facilitated'by'the'technology'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 30: Thinking broadly about educational technology

YOUR(TURN(AGAIN(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 31: Thinking broadly about educational technology

For'online'homework'systems'

•  How'does'the'use'of'this'tool'relate'to'your'goals?'

•  For'the'ways'you'(might)''use'them,'how'does'the'tool'–  Reorganize'who'does'what?'–  Change'par3cipa3on?'–  Allow'new/different'norms?'–  Reinforce/support'exis3ng'norms?''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 32: Thinking broadly about educational technology

STUDENTS’(WRITING(OF(SCIENTIFIC(EXPLANATIONS(IN(A(LARGE(CLASS(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 33: Thinking broadly about educational technology

Adap3ng'a'small,'discussionllab'course'to'large,'lecture'format'

Can'we'do'this?'What'does'it'look'like?'

Does'it'work?'

Development'supported'by'NSF'ESIl0096856'and'DUEl0717791'

Learning(Physical(Science((LEPS)(

!!

Fred Goldberg Stephen Robinson

Edward Price Rebecca Kruse

Danielle Boyd Harlow Michael McKean

Supported by National Science Foundation Grant 0717791 © 2011 San Diego State University Research Foundation

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 34: Thinking broadly about educational technology

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 35: Thinking broadly about educational technology

Calibrated'Peer'Review'

CPR'was'developed'at'UCLA'hPp://cpr.molsci.ucla.edu/Home.aspx'''

A'weblbased'tool'that'supports'students’'construc3on'and'evalua3on'of'explana3ons.'

3'stages:'1.  Text'entry''2.  Calibra3on'3.  Peer'review'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 36: Thinking broadly about educational technology

Calibrated'Peer'Review'Text'entry'stage'View'background'material'&'prompt'

Enter'text'/'upload'images' Other'

students’'texts'

Peer/self'review'stage'

Evaluate'&'score'3'peers’'texts'

Evaluate'&'score'own'text'

Review'results'&'feedback'

Calibra3on'texts'

Evalua3on'ques3ons'and'scoring'rubric'

Calibra3on'stage'

Evaluate'&'score'calibra3on'texts'

Receive'feedback'on'calibra3ons'

Price,'Goldberg,'et'al.'2012'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 37: Thinking broadly about educational technology

CPR'Task'Example'

www.postersession.com

Doing experiments with hands-on materials and computer simulations

Developing a Large-Enrollment, Guided-Inquiry, Conceptual Physics Course1

Fred Goldberga, Stephen Robinsonb, Edward Pricec, Danielle-Boyd Harlowd and Michael McKeana aSan Diego State University; bTennessee Technological University; cCSU San Marcos; dUC-Santa Barbara

References (1)  Supported by NSF Grant 1044172

(2)  Goldberg, F., Robinson, S., and Otero, V. (2008). Physics and Everyday Thinking. It’s About Time, Mount Kisco, NY 10549.

(3)  Goldberg, F., Otero, V. and Robinson, S. (2010). Design principles for effective physics instruction: A case study from Physics and Everyday Thinking. Am. J. Phys. 78 (12), 1265-1277.

(4)  Goldberg, F., Robinson, S., Price, E., Harlow, D., and McKean, M. (2012). Learning Physical Science. It’s About Time, Mount Kisco, New York, 10549.

(5)  F. Goldberg, E. Price, S. Robinson, D. Boyd-Harlow, and M. McKean. (2012)Developing the learning physical science curriculum: Adapting a small enrollment, laboratory and discussion based physical science course for large enrollments. Phys. Rev. ST Phys. Educ. Res. 8, 010121.

(6)  We mainly use the PhET simulations, available at phet.colorado.edu/

(7)  Price, E., Goldberg, F., Patterson, S. and Heft, P. (2012). Supporting scientific writing and evaluation in a conceptual physics course with Calibrated Peer Review. Physics Education Research Conference, Philadelphia, PA .

(8)  http://cpr.molsci.ucla.edu/Home.aspx

(9)  W. K. Adams, K. K. Perkins, N. Podolefsky, M. Dubson, N. D. Finkelstein, and C. E. Wieman, "A new instrument for measuring student beliefs about physics and learning physics: The Colorado Learning Attitudes about Science Survey," Phys. Rev. ST PER, 2 (1), 010101 (2006).

(10)  V. Otero and K. Gray (2008). Attitudinal gains across multiple universities using the Physics and Everyday Thinking curriculum. Phys. Rev. ST PER, 4, 020104.

•  One-semester physics course for prospective elementary teachers and general education science students, intended for large enrollment, lecture-style settings (45 hours class time).

•  Adapted from Physics and Everyday Thinking (PET)2,3, a small enrollment, lab and discussion centered course, with extensive white-boarding to share ideas (75 hours class time). Also adapted from Learning Physical Science,4,5 a large enrollment curriculum in physical science.

Learning Physical Science (LEP)

•  LEP focuses on core ideas and practices of science.

•  Specific activities focusing on the nature of learning and nature of science

•  Kits in Ziploc bags are distributed to groups.

•  Students use the online Calibrated Peer Review system7, developed at UCLA,8 to construct their own explanations and evaluate their peers’ explanations.

•  During course students write and evaluate explanations involving: (1) describing a chain of interactions in terms of energy; (2) explaining how prisms in eyeglasses correct for ‘seeing double;’ (3) applying Newton’s Second Law; and (4) using model to explain a new (for them) magnetism phenomenon.

•  Students use ‘clickers’ to make predictions, interpret representations, and share experimental observations and conclusions. When ‘clicker’ results are inconclusive, students share their thinking with whole class.

•  Clicker question from 2nd lesson of semester: person gives cart a quick shove and lets go. The cart continues to move along track with very little slowing down. Students are shown speed-time graph for motion. Clicker results are shown below in red. Two students share their reasoning for choices B and C.

Supporting claims with evidence and reasoning

•  Students see video clip of a nail being rubbed by a magnet (to magnetize it). The nail is then hammered, which demagnetizes it. Students use the class consensus alignment model of magnetism to explain what happened.

•  Student uploads diagrams and text to the CPR system. (Partial text is below.)

Example: Explanation task on magnetism

Key Findings: Comparing LEP and PET

Developing and using models: Example from magnetism

•  Majority of class initially proposed a ‘separation’ model to explain previous observations: dragging magnet across nail to magnetize it causes the entities inside nail (+/- or N/S) to separate to the two ends.

•  Separation model, however, could not explain why cutting a magnetized nail anywhere along its length would result in two pieces that are each magnetized. Through experimentation and discussion, class reached consensus on an ‘alignment’ model.

•  What do you gain and what do you loose when adapting a curriculum designed for a small, highly interactive class to a large enrollment class?

•  Staff developed 12-question conceptual assessment (multiple-choice with paired reasoning question) covering common content of LEP and PET. Administered pre/post to 3 LEP classes and 12 PET classes in Fall 2012. Findings: no significant difference in normalized gains between LEP and PET classes.

•  Administered the Colorado Learning Attitudes about Science Survey (CLASS)8 to 3 LEP classes in Fall 2012; compared results to 9 PET classes in Spring 20079. Findings: average pre/post shifts in LEP classes was about one-half the shift in PET classes.

Writing and evaluating explanations

CQ 2-1: At what point on the graph do you think the finger

lost contact with the cart?

•  After doing experiment with hands-on materials, students model and extend phenomena with computer simulations.6

Acknowledgements The authors would like to acknowledge Melissa Dancy, Heath Kirkwood, Scott Patterson, Paul Heft and CSUSM’s ITS.

www.postersession.com

Doing experiments with hands-on materials and computer simulations

Developing a Large-Enrollment, Guided-Inquiry, Conceptual Physics Course1

Fred Goldberga, Stephen Robinsonb, Edward Pricec, Danielle-Boyd Harlowd and Michael McKeana aSan Diego State University; bTennessee Technological University; cCSU San Marcos; dUC-Santa Barbara

References (1)  Supported by NSF Grant 1044172

(2)  Goldberg, F., Robinson, S., and Otero, V. (2008). Physics and Everyday Thinking. It’s About Time, Mount Kisco, NY 10549.

(3)  Goldberg, F., Otero, V. and Robinson, S. (2010). Design principles for effective physics instruction: A case study from Physics and Everyday Thinking. Am. J. Phys. 78 (12), 1265-1277.

(4)  Goldberg, F., Robinson, S., Price, E., Harlow, D., and McKean, M. (2012). Learning Physical Science. It’s About Time, Mount Kisco, New York, 10549.

(5)  F. Goldberg, E. Price, S. Robinson, D. Boyd-Harlow, and M. McKean. (2012)Developing the learning physical science curriculum: Adapting a small enrollment, laboratory and discussion based physical science course for large enrollments. Phys. Rev. ST Phys. Educ. Res. 8, 010121.

(6)  We mainly use the PhET simulations, available at phet.colorado.edu/

(7)  Price, E., Goldberg, F., Patterson, S. and Heft, P. (2012). Supporting scientific writing and evaluation in a conceptual physics course with Calibrated Peer Review. Physics Education Research Conference, Philadelphia, PA .

(8)  http://cpr.molsci.ucla.edu/Home.aspx

(9)  W. K. Adams, K. K. Perkins, N. Podolefsky, M. Dubson, N. D. Finkelstein, and C. E. Wieman, "A new instrument for measuring student beliefs about physics and learning physics: The Colorado Learning Attitudes about Science Survey," Phys. Rev. ST PER, 2 (1), 010101 (2006).

(10)  V. Otero and K. Gray (2008). Attitudinal gains across multiple universities using the Physics and Everyday Thinking curriculum. Phys. Rev. ST PER, 4, 020104.

•  One-semester physics course for prospective elementary teachers and general education science students, intended for large enrollment, lecture-style settings (45 hours class time).

•  Adapted from Physics and Everyday Thinking (PET)2,3, a small enrollment, lab and discussion centered course, with extensive white-boarding to share ideas (75 hours class time). Also adapted from Learning Physical Science,4,5 a large enrollment curriculum in physical science.

Learning Physical Science (LEP)

•  LEP focuses on core ideas and practices of science.

•  Specific activities focusing on the nature of learning and nature of science

•  Kits in Ziploc bags are distributed to groups.

•  Students use the online Calibrated Peer Review system7, developed at UCLA,8 to construct their own explanations and evaluate their peers’ explanations.

•  During course students write and evaluate explanations involving: (1) describing a chain of interactions in terms of energy; (2) explaining how prisms in eyeglasses correct for ‘seeing double;’ (3) applying Newton’s Second Law; and (4) using model to explain a new (for them) magnetism phenomenon.

•  Students use ‘clickers’ to make predictions, interpret representations, and share experimental observations and conclusions. When ‘clicker’ results are inconclusive, students share their thinking with whole class.

•  Clicker question from 2nd lesson of semester: person gives cart a quick shove and lets go. The cart continues to move along track with very little slowing down. Students are shown speed-time graph for motion. Clicker results are shown below in red. Two students share their reasoning for choices B and C.

Supporting claims with evidence and reasoning

•  Students see video clip of a nail being rubbed by a magnet (to magnetize it). The nail is then hammered, which demagnetizes it. Students use the class consensus alignment model of magnetism to explain what happened.

•  Student uploads diagrams and text to the CPR system. (Partial text is below.)

Example: Explanation task on magnetism

Key Findings: Comparing LEP and PET

Developing and using models: Example from magnetism

•  Majority of class initially proposed a ‘separation’ model to explain previous observations: dragging magnet across nail to magnetize it causes the entities inside nail (+/- or N/S) to separate to the two ends.

•  Separation model, however, could not explain why cutting a magnetized nail anywhere along its length would result in two pieces that are each magnetized. Through experimentation and discussion, class reached consensus on an ‘alignment’ model.

•  What do you gain and what do you loose when adapting a curriculum designed for a small, highly interactive class to a large enrollment class?

•  Staff developed 12-question conceptual assessment (multiple-choice with paired reasoning question) covering common content of LEP and PET. Administered pre/post to 3 LEP classes and 12 PET classes in Fall 2012. Findings: no significant difference in normalized gains between LEP and PET classes.

•  Administered the Colorado Learning Attitudes about Science Survey (CLASS)8 to 3 LEP classes in Fall 2012; compared results to 9 PET classes in Spring 20079. Findings: average pre/post shifts in LEP classes was about one-half the shift in PET classes.

Writing and evaluating explanations

CQ 2-1: At what point on the graph do you think the finger

lost contact with the cart?

•  After doing experiment with hands-on materials, students model and extend phenomena with computer simulations.6

Acknowledgements The authors would like to acknowledge Melissa Dancy, Heath Kirkwood, Scott Patterson, Paul Heft and CSUSM’s ITS.

CPR(Task(''Students'use'an'alignment'model'of'magne3sm'to'explain'a'nail’s'being'magne3zed'by'a'magnet,'and'demagne3zed'aQer'being'hit'with'a'hammer.''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 38: Thinking broadly about educational technology

CPR'Task'Example'

www.postersession.com

Doing experiments with hands-on materials and computer simulations

Developing a Large-Enrollment, Guided-Inquiry, Conceptual Physics Course1

Fred Goldberga, Stephen Robinsonb, Edward Pricec, Danielle-Boyd Harlowd and Michael McKeana aSan Diego State University; bTennessee Technological University; cCSU San Marcos; dUC-Santa Barbara

References (1)  Supported by NSF Grant 1044172

(2)  Goldberg, F., Robinson, S., and Otero, V. (2008). Physics and Everyday Thinking. It’s About Time, Mount Kisco, NY 10549.

(3)  Goldberg, F., Otero, V. and Robinson, S. (2010). Design principles for effective physics instruction: A case study from Physics and Everyday Thinking. Am. J. Phys. 78 (12), 1265-1277.

(4)  Goldberg, F., Robinson, S., Price, E., Harlow, D., and McKean, M. (2012). Learning Physical Science. It’s About Time, Mount Kisco, New York, 10549.

(5)  F. Goldberg, E. Price, S. Robinson, D. Boyd-Harlow, and M. McKean. (2012)Developing the learning physical science curriculum: Adapting a small enrollment, laboratory and discussion based physical science course for large enrollments. Phys. Rev. ST Phys. Educ. Res. 8, 010121.

(6)  We mainly use the PhET simulations, available at phet.colorado.edu/

(7)  Price, E., Goldberg, F., Patterson, S. and Heft, P. (2012). Supporting scientific writing and evaluation in a conceptual physics course with Calibrated Peer Review. Physics Education Research Conference, Philadelphia, PA .

(8)  http://cpr.molsci.ucla.edu/Home.aspx

(9)  W. K. Adams, K. K. Perkins, N. Podolefsky, M. Dubson, N. D. Finkelstein, and C. E. Wieman, "A new instrument for measuring student beliefs about physics and learning physics: The Colorado Learning Attitudes about Science Survey," Phys. Rev. ST PER, 2 (1), 010101 (2006).

(10)  V. Otero and K. Gray (2008). Attitudinal gains across multiple universities using the Physics and Everyday Thinking curriculum. Phys. Rev. ST PER, 4, 020104.

•  One-semester physics course for prospective elementary teachers and general education science students, intended for large enrollment, lecture-style settings (45 hours class time).

•  Adapted from Physics and Everyday Thinking (PET)2,3, a small enrollment, lab and discussion centered course, with extensive white-boarding to share ideas (75 hours class time). Also adapted from Learning Physical Science,4,5 a large enrollment curriculum in physical science.

Learning Physical Science (LEP)

•  LEP focuses on core ideas and practices of science.

•  Specific activities focusing on the nature of learning and nature of science

•  Kits in Ziploc bags are distributed to groups.

•  Students use the online Calibrated Peer Review system7, developed at UCLA,8 to construct their own explanations and evaluate their peers’ explanations.

•  During course students write and evaluate explanations involving: (1) describing a chain of interactions in terms of energy; (2) explaining how prisms in eyeglasses correct for ‘seeing double;’ (3) applying Newton’s Second Law; and (4) using model to explain a new (for them) magnetism phenomenon.

•  Students use ‘clickers’ to make predictions, interpret representations, and share experimental observations and conclusions. When ‘clicker’ results are inconclusive, students share their thinking with whole class.

•  Clicker question from 2nd lesson of semester: person gives cart a quick shove and lets go. The cart continues to move along track with very little slowing down. Students are shown speed-time graph for motion. Clicker results are shown below in red. Two students share their reasoning for choices B and C.

Supporting claims with evidence and reasoning

•  Students see video clip of a nail being rubbed by a magnet (to magnetize it). The nail is then hammered, which demagnetizes it. Students use the class consensus alignment model of magnetism to explain what happened.

•  Student uploads diagrams and text to the CPR system. (Partial text is below.)

Example: Explanation task on magnetism

Key Findings: Comparing LEP and PET

Developing and using models: Example from magnetism

•  Majority of class initially proposed a ‘separation’ model to explain previous observations: dragging magnet across nail to magnetize it causes the entities inside nail (+/- or N/S) to separate to the two ends.

•  Separation model, however, could not explain why cutting a magnetized nail anywhere along its length would result in two pieces that are each magnetized. Through experimentation and discussion, class reached consensus on an ‘alignment’ model.

•  What do you gain and what do you loose when adapting a curriculum designed for a small, highly interactive class to a large enrollment class?

•  Staff developed 12-question conceptual assessment (multiple-choice with paired reasoning question) covering common content of LEP and PET. Administered pre/post to 3 LEP classes and 12 PET classes in Fall 2012. Findings: no significant difference in normalized gains between LEP and PET classes.

•  Administered the Colorado Learning Attitudes about Science Survey (CLASS)8 to 3 LEP classes in Fall 2012; compared results to 9 PET classes in Spring 20079. Findings: average pre/post shifts in LEP classes was about one-half the shift in PET classes.

Writing and evaluating explanations

CQ 2-1: At what point on the graph do you think the finger

lost contact with the cart?

•  After doing experiment with hands-on materials, students model and extend phenomena with computer simulations.6

Acknowledgements The authors would like to acknowledge Melissa Dancy, Heath Kirkwood, Scott Patterson, Paul Heft and CSUSM’s ITS.

www.postersession.com

Doing experiments with hands-on materials and computer simulations

Developing a Large-Enrollment, Guided-Inquiry, Conceptual Physics Course1

Fred Goldberga, Stephen Robinsonb, Edward Pricec, Danielle-Boyd Harlowd and Michael McKeana aSan Diego State University; bTennessee Technological University; cCSU San Marcos; dUC-Santa Barbara

References (1)  Supported by NSF Grant 1044172

(2)  Goldberg, F., Robinson, S., and Otero, V. (2008). Physics and Everyday Thinking. It’s About Time, Mount Kisco, NY 10549.

(3)  Goldberg, F., Otero, V. and Robinson, S. (2010). Design principles for effective physics instruction: A case study from Physics and Everyday Thinking. Am. J. Phys. 78 (12), 1265-1277.

(4)  Goldberg, F., Robinson, S., Price, E., Harlow, D., and McKean, M. (2012). Learning Physical Science. It’s About Time, Mount Kisco, New York, 10549.

(5)  F. Goldberg, E. Price, S. Robinson, D. Boyd-Harlow, and M. McKean. (2012)Developing the learning physical science curriculum: Adapting a small enrollment, laboratory and discussion based physical science course for large enrollments. Phys. Rev. ST Phys. Educ. Res. 8, 010121.

(6)  We mainly use the PhET simulations, available at phet.colorado.edu/

(7)  Price, E., Goldberg, F., Patterson, S. and Heft, P. (2012). Supporting scientific writing and evaluation in a conceptual physics course with Calibrated Peer Review. Physics Education Research Conference, Philadelphia, PA .

(8)  http://cpr.molsci.ucla.edu/Home.aspx

(9)  W. K. Adams, K. K. Perkins, N. Podolefsky, M. Dubson, N. D. Finkelstein, and C. E. Wieman, "A new instrument for measuring student beliefs about physics and learning physics: The Colorado Learning Attitudes about Science Survey," Phys. Rev. ST PER, 2 (1), 010101 (2006).

(10)  V. Otero and K. Gray (2008). Attitudinal gains across multiple universities using the Physics and Everyday Thinking curriculum. Phys. Rev. ST PER, 4, 020104.

•  One-semester physics course for prospective elementary teachers and general education science students, intended for large enrollment, lecture-style settings (45 hours class time).

•  Adapted from Physics and Everyday Thinking (PET)2,3, a small enrollment, lab and discussion centered course, with extensive white-boarding to share ideas (75 hours class time). Also adapted from Learning Physical Science,4,5 a large enrollment curriculum in physical science.

Learning Physical Science (LEP)

•  LEP focuses on core ideas and practices of science.

•  Specific activities focusing on the nature of learning and nature of science

•  Kits in Ziploc bags are distributed to groups.

•  Students use the online Calibrated Peer Review system7, developed at UCLA,8 to construct their own explanations and evaluate their peers’ explanations.

•  During course students write and evaluate explanations involving: (1) describing a chain of interactions in terms of energy; (2) explaining how prisms in eyeglasses correct for ‘seeing double;’ (3) applying Newton’s Second Law; and (4) using model to explain a new (for them) magnetism phenomenon.

•  Students use ‘clickers’ to make predictions, interpret representations, and share experimental observations and conclusions. When ‘clicker’ results are inconclusive, students share their thinking with whole class.

•  Clicker question from 2nd lesson of semester: person gives cart a quick shove and lets go. The cart continues to move along track with very little slowing down. Students are shown speed-time graph for motion. Clicker results are shown below in red. Two students share their reasoning for choices B and C.

Supporting claims with evidence and reasoning

•  Students see video clip of a nail being rubbed by a magnet (to magnetize it). The nail is then hammered, which demagnetizes it. Students use the class consensus alignment model of magnetism to explain what happened.

•  Student uploads diagrams and text to the CPR system. (Partial text is below.)

Example: Explanation task on magnetism

Key Findings: Comparing LEP and PET

Developing and using models: Example from magnetism

•  Majority of class initially proposed a ‘separation’ model to explain previous observations: dragging magnet across nail to magnetize it causes the entities inside nail (+/- or N/S) to separate to the two ends.

•  Separation model, however, could not explain why cutting a magnetized nail anywhere along its length would result in two pieces that are each magnetized. Through experimentation and discussion, class reached consensus on an ‘alignment’ model.

•  What do you gain and what do you loose when adapting a curriculum designed for a small, highly interactive class to a large enrollment class?

•  Staff developed 12-question conceptual assessment (multiple-choice with paired reasoning question) covering common content of LEP and PET. Administered pre/post to 3 LEP classes and 12 PET classes in Fall 2012. Findings: no significant difference in normalized gains between LEP and PET classes.

•  Administered the Colorado Learning Attitudes about Science Survey (CLASS)8 to 3 LEP classes in Fall 2012; compared results to 9 PET classes in Spring 20079. Findings: average pre/post shifts in LEP classes was about one-half the shift in PET classes.

Writing and evaluating explanations

CQ 2-1: At what point on the graph do you think the finger

lost contact with the cart?

•  After doing experiment with hands-on materials, students model and extend phenomena with computer simulations.6

Acknowledgements The authors would like to acknowledge Melissa Dancy, Heath Kirkwood, Scott Patterson, Paul Heft and CSUSM’s ITS.

CPR(Task(''Students'use'an'alignment'model'of'magne3sm'to'explain'a'nail’s'being'magne3zed'by'a'magnet,'and'demagne3zed'aQer'being'hit'with'a'hammer.''

EvaluaVon(quesVons:''“Does'the'first'paragraph'correctly'describe'that'within'the'unmagne3zed'nail'there'are'(many)'3ny'magnets'that'are'randomly'oriented;'that'is,'their'NPs'(or'SPs)'point'in'different'direc3ons,'or'something'similar?”'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 39: Thinking broadly about educational technology

WriVng(scienVfic(explanaVons(–(average(final(exam(performance(of(students(in(two(types(of(courses(

training,!the!interrater!reliability!was!90%.!Each!rater!scored!half!of!the!students!in!each!curriculum,!thus!minimizing!rating!artifacts.!Figure!2!shows!the!average!score!for!the!students!in!each!curriculum.!The!difference!between!the!averages!is!1.1,!which!is!statistically!significant.!!!

!Figure$2.$Average!post!test!explanation!question!performance!for!LEP!(N=203)!and!PET!(N=316)!students.!Error!bars!are!standard!error!of!the!mean.!The!difference!is!statistically!significant!(p<!0.0001).!!The!Colorado!Learning!Attitudes!About!Science!Survey![Adams!et.!al.,!2006]!was!used!to!gather!information!on!students’!views!about!science!and!learning.!CLASS!was!administered!as!a!voluntary!onWline!preW!and!postW!assessment!in!the!three!LEP!classes.!CLASS!survey!results!are!analyzed!by!comparing!students’!responses!to!an!“expert”!response.!A!postWpre!difference!is!reported!in!terms!of!percent!shift!towards!or!away!from!the!expert!response;!a!positive!shift!indicates!more!expert!like!responses!at!the!end!of!the!semester.!Figure!3!indicates!the!average!shift!for!155!LEP!students!who!completed!pre!and!post!CLASS!surveys!in!fall!2012.!Data!from!PET!and!PSET!classrooms!(9!courses,!N=368),!reported!by!Otero!and!Gray![2008],!are!included!for!comparison.!The!PET/PSET!classes!showed!increased!attitudinal!shifts!compared!to!the!LEP!classes.!The!difference!between!the!averages!is!4.8,!which!is!statistically!significant!and!represents!nearly!2!items!on!the!survey.!However,!the!LEP!classes!compare!favorably!to!large,!traditional!introductory!physics!courses,!where!negative!shifts!or!shifts!of!+1W2%!have!been!reported!in!large!algebraW!and!calculusWbased!physics!courses,!and!courses!for!nonWscience!majors![Perkins!et.!al.,!2005].!

$Figure$3.!Average!shift!towards!expertWlike!views!on!CLASS!survey!for!LEP!(N=155)!and!PET/PSET!(N=368)!students!as!reported!by!Otero!and!Gray![2008].!Error!bars!are!standard!error!of!the!mean.!The!difference!is!statistically!significant!(p=!0.01)!

post test explanations

LEPPET

0

1

2

3

4

5

expl

anat

ion

scor

e

curriculum

Mann Whitney testP valueExact or approximate P value?P value summarySignificantly different? (P < 0.05)One- or two-tailed P value?Sum of ranks in column A,BMann-Whitney U

< 0.0001Approximate****YesTwo-tailed64810 , 7013120045

MeanStd. DeviationStd. Error of Mean

LEP2.7271.5500.1088

PET1.6381.6360.09204

error bars are SEM

CLASS

Shi

ft to

war

ds e

xper

t (%

)

LEP

PET/PSET

0

5

10

15

Students'in'courses'that'included'5'CPR'tasks'(LEP)'outperformed'students'in'courses'with'tradi3onal'assignments'(PET)'

Price,'Goldberg,'et'al.,'in'prepara3on'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 40: Thinking broadly about educational technology

CPR'as'a'tool'

•  Supports'goal'of'students'being'able'to'construct,'cri3cally'evaluate'explana3ons'

•  With'CPR,''–  Instructor'as'developer,'but'students'as'graders/evaluators'

– Task'development'is'intensive,'but'grading/administra3on'is'minimal'

–  Implicit'sugges3on'that'students'can'develop'(some)'exper3se'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 41: Thinking broadly about educational technology

FLIPPING(THE(CLASSROOM(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 42: Thinking broadly about educational technology

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 43: Thinking broadly about educational technology

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 44: Thinking broadly about educational technology

For'screencasts'(and'books)'and'flipped'classrooms?'

•  How'does'the'use'of'this'tool'relate'to'your'goals?'•  For'the'ways'you'(might)'use'them,'how'do'these'tools'–  Reorganize'who'does'what?'–  Change'par3cipa3on?'–  Allow'new/different'norms?'–  Reinforce/support'exis3ng'norms?''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 45: Thinking broadly about educational technology

MOOCS,(ONLINE(COURSES,(&(THE(WHOLE(FUTURE(OF(EDUCATION(

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 46: Thinking broadly about educational technology

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 47: Thinking broadly about educational technology

Learning'principles'

1.'Learning'builds'on'prior'knowledge'2.'Learning'is'a'complex'process'requiring'scaffolding''3.'Learning'is'facilitated'through'interac3on'with'tools''4.'Learning'is'facilitated'through'peer'interac3ons'5.'Learning'is'facilitated'through'establishment'of'

'norms'and'expecta3ons'

J.'D.'Bransford,'A.'L.'Brown,'and'R.'R.'Cocking,'How'People'Learn:'Brain,'Mind,'Experience,'and'School'(NAP,'Washington,'D.C.,'2000).'E.'F.'Redish,'Implica3ons'of'Cogni3ve'Studies'for'Teaching'Physics,'American'Journal'of'Physics'62,'796'(1994).'D.'Hammer,'et'al.,'in'Transfer'of'Learning:'Research'and'Perspec>ves,'edited'by'J.'Mestre'(Informa3on'Age'Publishing,'2004).'Y.'Engeström,'Learning'by'Expanding:'An'Ac>vityMtheore>cal'Approach'to'Developmental'Research'(OrientalKonsul3t'Oy,'1987).'M.'Cole,'Cultural'Psychology:'A'Once'and'Future'Discipline'(Harvard'University'Press,'Cambridge,'MA,'1998).'L.'S.'Vygotsky,'Thought'and'Language'(MIT'Press,'Cambridge,'MA,'1986).'E.'G.'Cohen,'Designing'Groupwork:'Strategies'for'the'Heterogeneous'Classroom'(Teachers'College'Press,'1994),'p.'203.'R.'Driver,'P.'Newton,'and'J.'Osborne,'Establishing'the'Norms'of'Scien3fic'Argumenta3on'in'Classrooms,'Science'Educa3on'84,'287l312'(2000).'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 48: Thinking broadly about educational technology

Evaluation of Evidence-Based Practices in Online Learning

A Meta-Analysis and Review of Online Learning Studies

Narrative Synthesis

In addition to the meta-analysis comparing online learning conditions with face-to-face

instruction, analysts reviewed and summarized experimental and quasi-experimental studies

contrasting different versions of online learning. Some of these studies contrasted purely online

learning conditions with classes that combined online and face-to-face interactions. Others

explored online learning with and without elements such as video, online quizzes, assigned

groups, or guidance for online activities. Five of these studies involved K–12 learners.

Key Findings

The main finding from the literature review was that

! Few rigorous research studies of the effectiveness of online learning for K–12 students

have been published. A systematic search of the research literature from 1994 through

2006 found no experimental or controlled quasi-experimental studies comparing the

learning effects of online versus face-to-face instruction for K–12 students that provide

sufficient data to compute an effect size. A subsequent search that expanded the time

frame through July 2008 identified just five published studies meeting meta-analysis

criteria.

The meta-analysis of 51 study effects, 44 of which were drawn from research with older learners,

found that2

! Students who took all or part of their class online performed better, on average, than

those taking the same course through traditional face-to-face instruction. Learning

outcomes for students who engaged in online learning exceeded those of students

receiving face-to-face instruction, with an average effect size of +0.24 favoring online

conditions.3 The mean difference between online and face-to-face conditions across the

51 contrasts is statistically significant at the p < .01 level.4 Interpretations of this result,

however, should take into consideration the fact that online and face-to-face conditions

generally differed on multiple dimensions, including the amount of time that learners

spent on task. The advantages observed for online learning conditions therefore may be

the product of aspects of those treatment conditions other than the instructional delivery

medium per se.

2 The meta-analysis was run also with just the 44 studies with older learners. Results were very similar to those for

the meta-analysis including all 51 contrasts. Variations in findings when K-12 studies are removed are described

in footnotes. 3 The + sign indicates that the outcome for the treatment condition was larger than that for the control condition. A

– sign before an effect estimate would indicate that students in the control condition had stronger outcomes than

those in the treatment condition. Cohen (1992) suggests that effect sizes of .20 can be considered “small,” those of

approximately .50 “medium,” and those of .80 or greater “large.” 4 The p-value represents the likelihood that an effect of this size or larger will be found by chance if the two

populations under comparison do not differ. A p-value of less than .05 indicates that there is less than 1 chance in

20 that a difference of the observed size would be found for samples drawn from populations that do not differ.

xiv

[differences]'may'be'the'product'of'aspects'of'those'treatment'condi3ons'other'than'the'instruc3onal'delivery'medium'per'se.'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 49: Thinking broadly about educational technology

The'Cow'Tipping'Point'

Ehrenfeld,'David.'“Unethical'Contexts'for'Ethical'Ques3ons,”'Harper’s'Magazine,'October,'2002'

I'propose'to'widen'the'context'gradually'so'that'we'always'know'the'vantage'point'from'which'we'are'viewing'the'[physics'educa3on]'landscape.'Eventually,'the'basic'truths'of'the'maPer'should'be'fairly'clear,'if'they'aren’t'already;'and'the'conclusions'we'ought'to'reach'about'the'technology'will'be'obvious.''…'As'we'move'further'and'further'from'the'original'narrow'context,'we'gradually'leave'the'realm'of'science'and'medicine'and'we'enter'the'territory'of'ethics,'economics,'and'social'welllbeing.''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 50: Thinking broadly about educational technology

MOOCs,'online'courses,'&''the'future'of'educa3on'

•  In'these'models,'what'are'the'implicit'(or'explicit)'theories'of'learning?'Are'they'consistent'with'research'on'learning?'Compared'to'what?'

•  Roles'of'faculty,'instruc3onal'developers,'teachers,'students''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 51: Thinking broadly about educational technology

MOOCs,'online'courses,'&''the'future'of'educa3on'

•  In'these'models,'what'are'the'implicit'(or'explicit)'views'about'the'purposes'and'mechanisms'of'educa3on?'

•  Providing'access'–'of'what'sort,'for'whom?'•  Who'profits?'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 52: Thinking broadly about educational technology

Technology'in'the'classroom'

A'classroom'is'a'community,'learning'is'a'social'process.'Technology'should'be'designed'and'used'to'support'this.'''Clickers,'videolbased'experiments,'and'online'archives'can'extend'and'enrich'the'classroom,'and'support/structure'interac3ons.'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 53: Thinking broadly about educational technology

Technology'in'the'classroom'

Let'pedagogical'goals'drive'the'use'of'technology.'Technology'≠'pedagogy.'What'you'do'is'more'important'than'the'tools'you'use.'But'tools'can'reorganize'ac3vi3es,'roles,'and'norms.'Keep'an'eye'on'the'broader'context'in'which'we'work.''

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'

Page 54: Thinking broadly about educational technology

References'Atkin,'J.'Myron;'Black,'Paul;'Coffey,'Janet'eds.'(2001)'Classroom'Assessment'and'the'Na>onal'Science'Educa>on'Standards,(NRC'BeaPy,'I.D.;'Gerace,'W.J.;'Leonard,'W.J.;'&'Dufresne,'R.J.'(2006)'Designing'effec3ve'ques3ons'for'classroom'response'system'teaching,'AJP'74:1,'31l39'BeaPy,'I.D.'&'Gerace,'W.J.'(2009)'TechnologylEnhanced'Forma3ve'Assessment:'A'ResearchlBased'Pedagogy'for'Teaching'Science'with'Classroom'Response'Technology,'J'Sci'Educ'Technol'18:146–162''Black,'Paul'&'Wiliam,'Dylan''(1998)'Assessment'and'Classroom'Learning,'Assessment'in'Educa>on:'Principles,'Policy'&'Prac>ce,'5:1,'7l74'Bransford,'J.D.;'Brown,'A.L.'&'Cocking,'R.R.'(1999)'How'People'Learn:'Brain,'Mind,'Experience,'and'School'Na3onal'Academy'Press'Cole,'M.'(1996)'Cultural'psychology:'A'once'and'future'discipline,'Cambridge,'MA:'Harvard'University'Press.'Cowie,'Bronwen'&'Bell,'Beverley'(1999)'A'Model'of'Forma3ve'Assessment'in'Science'Educa3on,'Assessment'in'Educa>on:'Principles,'Policy'&'Prac>ce,'6:1,'101l116'Engeström,'Y.'(1987)'Learning'by'Expanding:'An'Ac>vityMTheore>cal'Approach'to'Developmental'Research,'Helsinki:'OrientalKonsul3t.'Fakcharoenphol,'W.,'PoPer,'E.'&'Stelzer,'T.'(2011)'What'students'learn'when'studying'physics'prac3ce'exam'problems,'PRSTMPER,'7,'010107'Harlen,'Wynne'(1999)'Purposes'and'Procedures'for'Assessing'Science'Process'Skills,'Assessment'in'Educa>on:'Principles,'Policy'&'Prac>ce,'6:1,'129l14'Henderson,'C.'&'Harper,'K.'(2009)'Quiz'Correc3ons:'Improving'Learning'by'Encouraging'Students'to'Reflect'on'Their'Mistakes,'TPT'47:9,'581l586'Kaptelinin,'V.,'Nardi,'B.'A.'(2006)'Ac>ng'with'Technology:'Ac>vity'Theory'and'Interac>on'Design,'Cambridge:'MIT'Press.'Lasry,'N.'(2008)'Clickers'or'Flashcards:'Is'There'Really'a'Difference?'The'Physics'Teacher'46,(242l244'Mason,'A.'&'Singh,'C.'(2010)'Do'advanced'physics'students'learn'from'their'mistakes'without'explicit'interven3on?,'Am.'J.'Phys.'78:7,'760l767'Nardi,'B.'A.'(1996)'Context'and'Consciousness:'Ac>vity'Theory'and'HumanMcomputer'Interac>on,'Cambridge,'MA:'MIT'Press.'Norman'(1988)'The'Psychology'of'Everyday'Things'Otero,'Valerie'&'Nathan,'Mitchell'(2008)'Preservice'Elementary'Teachers’'Views'of'Their'Students’'Prior'Knowledge'of'Science,'JRST,'45:4,'497–523'Price,'E.'&'Simon,'B.'(2009)'Ubiquitous'Presenter:'A'Tablet'PClbased'system'to'support'instructors'and'students'in'the'classroom,''TPT,'47(570'Price,'E.'et'al.'(2011)'Don’t'erase'that'whiteboard!'TPT,'in'press.'Reay'NW,'Li'P,'Bao'L'(2008)'Tes3ng'a'new'vo3ng'machine'ques3on'methodology.'Am'J'Phys'76(2):171–178'Sadler,'D.'Royce'(1998)'Forma3ve'assessment:'Revisi3ng'the'territory,'Assessment'in'Educa>on,'5:1,'77l84'Sadler,'D.'Royce'(1989)'Forma3ve'assessment'and'the'design'of'instruc3onal'systems,'Instruc>onal'Science'18,'119l144''Turpen,'C.'&'Finkelstein,'N.D.'(2010)'The'construc3on'of'different'classroom'norms'during'Peer'Instruc3on:'Students'perceive'differences,'PRSTMPER'6,'020123'

Ed'Price,'APS'Experienced'Faculty'Workshop,'April'2013'