Preview only show first 10 pages with watermark. For full document please download

One-dimensional Regression In High School

   EMBED


Share

Transcript

INSTITUT FOR NATURFAGENES DIDAKTIK KØBENHAVNS UNIVERSITET One-dimensional regression in high school Jeanette Kjølbæk Kandidatspeciale Juni 2015 IND’s studenterserie nr. 39 INSTITUT FOR NATURFAGENES DIDAKTIK, www.ind.ku.dk Alle publikationer fra IND er tilgængelige via hjemmesiden. IND’s studenterserie 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. Ellen Berg Jensen: 15-åriges viden om klimaforskelle (2007) Martin Sonnenborg: The Didactic Potential of CAS (2007) 15. 16. 17. 18. 19. 20. 21. 22. Philipp Lorenzen: Hvem er de nye studenter? Baggrund, interesse & uddannelsesstrategi (2010) 23. 24. 25. 26. 27. 28. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. Karina Søgaard og Sarah Kyhn Buskbjerg: Galoisteori i Gymnasiet (2007) Ana Hesselbart: Mathematical reasoning and semiosis (2007) Julian Tosev: Forskningslignende situationer (2007) Niels Nørskov Laursen: En Covarians-tilgang til Variabelssammenhænge i gymnasiet (2007) Katja Vinding Petersen: Lyd og Liv (2007) Jesper Bruun: Krop og computer i fysikundervisning (2008) Jakob Svendsen: Matematiklærerens forberedelse (2009) Britta Hansen: Didaktik på tværs af matematik og historie (2009) Nadja Ussingkær: En didaktisk undersøgelse af brudte lineære funktioner i rammerne af tysk fritidsmatematik (2009) Thomas Thrane: Design og test af RSC-forløb om vektorfunktioner og bevægelse Flemming Munch Hansen: Samspil omkring differentialregningens elementer i gymnasiets matematik og fysik (2009) Hasan Ademovski og Hatice Ademovski: Proportionalitet på mellemtrinnet - Design af didaktiske situationer baseret på stofdidaktisk analyse (2009) Signe Ougaard: Logiske strukturer i matematisk analyse på gymnasieniveau. Et forløb om kvantorer og ε-δ-definition af grænseværdi (2010) Jesper Winther Sørensen: Abstrakt algebra i gymnasiet - design, udførelse og analyse af undervisning i gruppeteori (2010) Sofie Stoustrup: En analyse af differentialligninger på A-niveau i STX ud fra den antropologiske didaktiske teori (2010) Jan Henrik Egballe Heinze: Eksponentialfunktioner i STX (2010) Mette Beier Jensen: Virtuelgalathea3.dk i biologiundervisningen i gymnasiet (2010) Servet Dönmez: Tosprogede elever og matematik i gymnasiet (2011) Therese Røndum Frederiksen: Designing and implementing an engaging learning experience about the electric sense of sharks for the visitors at Danmarks Akvarium (2011) Sarah Elisabeth Klein: Using touch-tables and inquiry methods to attract and engage visitors (2011) Line Kabell Nissen: Matematik for Sjov – Studie- og forskningsforløb i en eksperimentel kontekst (2011) Jonathan Barrett: Trigonometriske Funktioner i en Gymnasial Kontekst – en epistemologisk referencemodel (2011) Rune Skalborg Hansen: Et studie i læringsopfattelse og -udbytte – Om fysik C kursister på Københavns VUC (2011) Astrid Camilus: Valideringssituationer i undervisningsforløb om differentialligninger og radioaktivitet (2012) Niven Adel Atie: Didaktiske situationer for fuldstændiggørelse af kvadratet i andengradsligningen (2013) Morten C. B. Persson: Kvantekemi i gymnasiet - Tilrettelæggelse, udførelse og evaluering af et undervisningsforløb (2013) Sofie Birch Jensen: Køn, evaluering og The Force Concept Inventory (2013) Simone Gravlund Nielsen: Når børn forsker i matematik (2013) Henrik Egholm Wessel: Smartphones as Scientific Instruments in Inquiry Based Science Education (2013) Nicole Koefoed: Et didaktisk design om definition, eksistens og eksakt værdi af bestemt integral (2013) Trine Louise Brøndt Nielsen: From Master’s programme to labour market – A study on physics graduates’ experience of the transition to the labour market (2013) Rie Hjørnegaard Malm: Becoming a Geologist – Identity negotiations among first year geology students (2013) Mariam Babrakzai Zadran: Gymnasiealgebra I et historisk perspektiv – Matematiske organisationer I gymnasiealgebra (2014) Marie Lohmann-Jensen: Flipped Classroom – andet end blot en strukturel ændring af undervisningen? (2014) Jeppe Willads Petersen: Talent – Why do we do it? (2014) Jeanette Kjølbæk: One-dimensional regression in high school (2015) Abstract Development of calculators and IT-tools has caused that students can easily solve mathematical tasks without knowing intermediate calculations and the mathematics behind the technique. Unfortunately a consequence of this is that the mathematics theory behind these techniques has been given a lower priority. This is the case with teaching in one dimensional regression in high school, where many students learn the instrumented techniques to make regression without knowing what to find or how to find it. In this thesis, the theory about one-dimensional regression at the mathematical community and the possibilities to work more theoretically with one-dimensional regression in high school is investigated. IND’s studenterserie består af kandidatspecialer og bachelorprojekter skrevet ved eller i tilknytning til Institut for Naturfagenes Didaktik. Disse drejer sig ofte om uddannelsesfaglige problemstillinger, der har interesse også uden for universitetets mure. De publiceres derfor i elektronisk form, naturligvis under forudsætning af samtykke fra forfatterne. Det er tale om studenterarbejder, og ikke endelige forskningspublikationer. Se hele serien på: www.ind.ku.dk/publikationer/studenterserien/ ONE-DIMENSIONAL REGRESSION IN HIGH SCHOOL jeanette kjølbæk A Master Thesis in Mathematics Department of Science Education, University of Copenhagen Submitted: 1 June 2015 name: Jeanette Kjølbæk title: One-dimensional regression in high school A Master Thesis in Mathematics department: Department of Science Education, University of Copenhagen supervisor: Carl Winsløw time frame: December 2014 - June 2015 ABSTRACT Development of calculators and IT-tools has caused that students can easily solve mathematical tasks without knowing intermediate calculations and the mathematics behind the technique. Unfortunately a consequence of this is that the mathematics theory behind these techniques has been given a lower priority. This is the case with teaching in one-dimensional regression in high school, where many students learn the instrumented techniques to make regression without knowing what to find or how to find it. In this thesis, the theory about one-dimensional regression at the mathematical community and the possibilities to work more theoretically with one-dimensional regression in high school is investigated. Initial, the external didactic transposition was analyzed. This was done by analyze and present the theory of regression in the mathematical community and describe how regression is included in the curricula, written exams and textbooks. The presentation and descriptions were based on the Anthropological Theory of Didactic (ATD). Based on the analysis four main questions were highlighted and an epistemological reference model (ERM) was developed. The second part of the thesis concerns the internal didactic transposition and focus on the design and evaluation of the teaching sequence. The planned and the realized didactic process was analyzed using ATD and ERM. Finally, the possibilities and obstacles to work more theoretically with regression were presented and discussed. The analysis of the external didactic transposition showed that one technological-theoretical discourse of linear regression is based on mathematics well-known to students in high school. This discourse were applied in the design of the teaching sequence. Further it was found that the students only are presented for a minimum of technological-theoretical discourse and do not learn how to determinate the best class of functions. The technological-theoretical discourse of linear regression and the question of best class of functions were tested in a high school class. It turned out that the students had a sensible basis to reach the technological discourse. The technical work to reach the theoretical level was difficult, especially the algebraic notation and the rules of sum were found to be challenging. The determination of the best class of functions was found to function in practice. iii ACKNOWLEDGEMENTS This thesis marks the end of my education at the Department of Mathematical Sciences and the Department of Science Education at the University of Copenhagen. The thesis combines my interest of didactic and mathematics. The thesis is a thesis in mathematical didactics, where the theory of didactic is used to design and evaluate a teaching sequence. First of all I would like to thank Anne Tune Svendsen for the opportunity to test my teaching sequence and for executed the teaching sequence in her class. A great thank to my supervisor, Carl Winsløw, for competent and thorough supervision throughout the process. Thanks for you always have been disposal with excellent advice. A big thank to my sister, Louise Kjølbæk, for proofreading the thesis. Finally thanks to the students and employees at the Department of Science Education for advices and to make the thesis office a great place to be. v CONTENTS 1 introduction . . . . . . . . . . . . . . . . . . . . . . . . . 1.1 Objectives . . . . . . . . . . . . . . . . . . . . . . . . . 1.2 Structure of the thesis . . . . . . . . . . . . . . . . . . 2 the anthropological theory of didactics . . . . 2.1 Praxeologies . . . . . . . . . . . . . . . . . . . . . . . . 2.1.1 Mathematical praxeologies and organizations 2.1.2 Didactic praxeologies and moments . . . . . . 2.2 The didactic transposition . . . . . . . . . . . . . . . . 2.2.1 Level of codetermination . . . . . . . . . . . . I 3 4 5 6 . . . . . . . . . the external didactic transposition and the epistemological reference model . . . . . . . . . . . . . 1 st research question . . . . . . . . . . . . . . . . . . . the scholar knowledge . . . . . . . . . . . . . . . . . . 4.1 Historical perspective . . . . . . . . . . . . . . . . . . . . 4.2 Regression . . . . . . . . . . . . . . . . . . . . . . . . . . 4.2.1 Types of regression . . . . . . . . . . . . . . . . . 4.2.2 Definition of the best linear model . . . . . . . . 4.2.3 Definition of the best model . . . . . . . . . . . . 4.3 The method of least squares . . . . . . . . . . . . . . . . 4.3.1 Elementary algebra . . . . . . . . . . . . . . . . . 4.3.2 Multivariable calculus . . . . . . . . . . . . . . . 4.3.3 Linear algebra . . . . . . . . . . . . . . . . . . . . 4.4 Statistical approach . . . . . . . . . . . . . . . . . . . . . 4.4.1 Maximum likelihood estimation . . . . . . . . . 4.4.2 Random variable and orthogonal projection . . 4.5 The quality of a model . . . . . . . . . . . . . . . . . . . 4.5.1 Visual examination . . . . . . . . . . . . . . . . . 4.5.2 The correlation coefficient . . . . . . . . . . . . . 4.5.3 The coefficient of determination . . . . . . . . . 4.6 Non-linear regression . . . . . . . . . . . . . . . . . . . . the knowledge to be taught . . . . . . . . . . . . . . . 5.1 Curricula . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.2 Written exams . . . . . . . . . . . . . . . . . . . . . . . . 5.2.1 The type of tasks . . . . . . . . . . . . . . . . . . 5.2.2 The methods and techniques . . . . . . . . . . . 5.3 Textbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . 5.3.1 Gyldendals Gymnasiematematik . . . . . . . . . 5.3.2 Lærebog i matematik . . . . . . . . . . . . . . . . 5.3.3 Hvad er matematik? . . . . . . . . . . . . . . . . epistemological reference model . . . . . . . . . . . 6.1 Q1 : The type of model (F i ) . . . . . . . . . . . . . . . . 1 1 2 5 5 6 8 10 12 15 17 19 19 20 21 22 25 26 27 30 33 38 38 39 41 41 42 44 45 49 49 51 51 54 56 56 56 57 63 65 vii viii contents 6.2 6.3 6.4 II 7 8 9 10 11 12 Q2 : The best model (f 2 F i ) . . . . . . . . . . . . . . . Q3 : The best linear model (f 2 FL ) . . . . . . . . . . . . Q4 : Evaluate the model . . . . . . . . . . . . . . . . . . . 66 67 69 the internal didactic transposition: the teaching sequence . . . . . . . . . . . . . . . . . . . . . . . . . 2 nd research question . . . . . . . . . . . . . . . . . . . design of the teaching sequence . . . . . . . . . . . 8.1 The class . . . . . . . . . . . . . . . . . . . . . . . . . . . 8.2 Mathematical organizations . . . . . . . . . . . . . . . . 8.2.1 MO belonging to Q3 . . . . . . . . . . . . . . . . 8.2.2 MO belonging to Q1 . . . . . . . . . . . . . . . . 8.3 The specific aims . . . . . . . . . . . . . . . . . . . . . . 8.4 Didactic principles . . . . . . . . . . . . . . . . . . . . . 8.5 Description of the planned didactic process . . . . . . . 8.5.1 Module 1: Definition of best linear model . . . . 8.5.2 Module 2: The proof of best linear model . . . . 8.5.3 Module 3: The type of model . . . . . . . . . . . a priori analysis . . . . . . . . . . . . . . . . . . . . . . . 9.1 Module 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.2 Module 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . 9.3 Module 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . a posteriori analysis . . . . . . . . . . . . . . . . . . . . 10.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . 10.2 Analysis of the realized teaching sequence . . . . . . . 10.2.1 Q3 : The best linear model f 2 F L . . . . . . . . 10.2.2 Q1 : The type of model F i . . . . . . . . . . . . . discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.1 Redesign . . . . . . . . . . . . . . . . . . . . . . . . . . . 11.2 The rationale for teaching more theoretically . . . . . . conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 73 75 75 76 76 78 78 79 81 81 83 85 87 87 89 92 99 99 101 101 110 115 115 117 119 list of references . . . . . . . . . . . . . . . . . . . . . . . . 121 III appendix . . . . . . . . . . . . . a appendix . . . . . . . . . . . . . . a.1 The tasks . . . . . . . . . . . . a.2 The tasks with solutions . . . a.3 Tables of the didactic process a.4 The planned didactic process a.5 The realized didactic process a.6 The students’ solutions . . . . a.7 The transcriptions in danish . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125 127 127 137 162 166 183 188 202 LIST OF FIGURES Figure 1 Figure 2 Figure 3 Figure 4 Figure 5 Figure 6 Figure 7 Figure 8 Figure 9 Figure 10 Figure 11 Figure 12 Figure 13 Figure 14 Figure 15 Figure 16 Figure 17 Figure 18 Figure 19 Figure 20 Figure 21 Figure 22 Figure 23 Figure 24 Figure 25 Figure 26 Figure 27 Figure 28 The research methodology and structure of the thesis . . . . . . . . . . . . . . . . . . . . . . . . A mathematical praxeology . . . . . . . . . . . Connection between MOs . . . . . . . . . . . . The didactic transposition . . . . . . . . . . . . Level of didactic codetermination . . . . . . . . Examples of vertical, horizontal or perpendicular errors . . . . . . . . . . . . . . . . . . . . . Projection of v to W . . . . . . . . . . . . . . . . Projection of ~b to C(A) . . . . . . . . . . . . . . Examples of scatter plot and residual plot . . . Anscombe’s quartet . . . . . . . . . . . . . . . . Exam exercise from 1980 . . . . . . . . . . . . . Exam exercise from 2013 . . . . . . . . . . . . . Task of type TD and TQD from 1998 . . . . . . Illustration of the LSQ method in "Hvad er matematik? - C" . . . . . . . . . . . . . . . . . . Illustration of projection in "What are mathematics? A" . . . . . . . . . . . . . . . . . . . . . ERM . . . . . . . . . . . . . . . . . . . . . . . . . Data set in a example . . . . . . . . . . . . . . . Scatter plot and residual plot of the linear and exponential model . . . . . . . . . . . . . . . . . Illustration of MOs to Q3 . . . . . . . . . . . . . Scatterplot of the data set in exercise 1 and 2 . Illustration of displacement of a data set . . . . The representation of the data set in exercise 9 Outline of the planned and realized didactic process . . . . . . . . . . . . . . . . . . . . . . . The plots of the transformed data set . . . . . . The rules of sum written at the black board . . A student’s solution to task 1a . . . . . . . . . A student’s solution to task 2c + 2d) . . . . . . The linear model in task 11a . . . . . . . . . . . 2 6 8 10 12 23 33 35 42 43 53 53 54 58 59 63 64 65 68 82 84 93 101 104 106 107 109 111 ix Figure 29 Plots of the displacement, when the graph window is set . . . . . . . . . . . . . . . . . . . . . 116 L I S T O F TA B L E S Table 1 Table 2 The properties of the three methods to linear regression . . . . . . . . . . . . . . Exam exercises regarding regression in 1999 + 2009-2013 . . . . . . . . . . . . . . make . . . . 1975. . . . 25 52 ACRONYMS ATD Anthropological Theory of Didactics MO Mathematical organization ERM Epistemological reference model LSQ Least squares D Data set D = {(x1 , y1 ), . . . , (xn , yn )} FL The class of linear functions FL = {f(x) = ax + b|a, b 2 R} FE The class of exponential functions FE = {f(x) = bax |a 2 R+ , b 2 R} FP The class of power functions FP = {f(x) = bxa |a, b 2 R} Fi The collection of three class of functions (linear, exponential and power functions). Fi with i 2 {L, E, P} x 1 INTRODUCTION 1.1 objectives Development of calculators and IT-tools in the recent 10-15 years has had a great influence on mathematics in high school, not only on the techniques to solve exercises, but also on the mathematical tasks and content. This has enabled an inclusion of new mathematical themes in the curricula and one of these themes is one-dimensional regression, which is mandatory for all students in high school. Using the calculators and IT-tools, students can now easily solve several tasks without knowing the intermediate calculations to solve the tasks and furthermore solve tasks that would have be very painstaking by hand. Unfortunately, a consequence is that the mathematics behind the techniques has been given a lower priority. The motivation for investigating one-dimensional regression is the frequently imprecise approach to the theme, where the teacher focus on the instrumented techniques to determine the best linear, exponential or power model, without presenting the students to what is meant by best or how the models are determined. This approach has the unfortunate consequence that the instrumented techniques became mysterious processes the students cannot explain or justify. Winsløw (2003) denotes this a black box effect. This effect is of course undesirable and recently in the magazine of LMFK it has been discussed how the theory of regression can be presented to avoid the black box effect. In the articles1 the possibilities and obstacles to teach linear regression is discussed and four different proofs of the formula of the best linear model are presented. Thus, it is interesting to consider the didactic problem of teaching onedimensional regression in high school and investigate the possibilities to design a teaching sequence containing theoretical work with onedimensional regression to avoid the black box effect. An important part of the didactic problem lies in the elaboration of a mathematical knowledge complete enough to allow the development of theoretical insight in regression. To elaborate the mathematical knowledge, the knowledge in the mathematical community and high school will be 1 Nr. 1/Januar: Fodgængerversion af lineær regression. Nr. 2/Marts: Regression med mindste kvadraters metode, Mere om linear regression. Nr. 3/Maj: De mindste kvadraters metode. 1 The magazine of LMFK is a member’s magazine for the union of teachers in mathematics, physics and chemistry 2 introduction analyzed and the relevant knowledge will be transformed to design the teaching sequence. The teaching sequence will be tested in a high school class. 1.2 structure of the thesis Figure 1: The research methodology and structure of the thesis Figure 1 illustrate the research method used and the structure of the thesis. Initial, the main elements of the Anthropological Theory of Didactics (ATD) are described (Chapter 2). The elements constitute the theoretical basis of the thesis. This is followed by part I, which concerns the external transposition and the epistemological reference model (ERM). Firstly, the first research question about the body of knowledge and the transposition is presented (chapter 3). In chapter 4, the scholar knowledge about onedimensional regression at the mathematical community is presented, afterward the knowledge to be taught is described by analyzing curricula, exams and textbooks, in section 5.1, 5.2, 5.3. The analysis of the external transposition constitutes the basis of the ERM. The ERM is presented in chapter 6. The first part is basis for the second part, which describes the internal transposition, hence the design and the evaluation of the teaching sequence. Part II prefaces with the second research question about the possibilities and obstacles to work more theoretically with regression (chapter 7). This is followed by a description of the design in chapter 8 and the a priori analysis in chapter 9. These chapters describe the reflections about the design and contain detailed descriptions of the planned didactic process. The realized didactic process and the students’ work is evaluated in the a posteriori analysis in chapter 10. 1.2 structure of the thesis Finally, the teaching sequence is discussed are described in chapter 11. The thesis is finalized with a conclusion in chapter 12. Navigation Reference in the thesis is marked with different colors. The brown color marks references to the list of reference. Click on the reference and you will jump to the reference. The same procedure, but with green reference will guide to a homepage. The blue reference will guide to other parts of the thesis, for examples chapter, section, subsection, theorems and definitions. 3 T H E A N T H R O P O L O G I C A L T H E O RY O F D I D A C T I C S Didactics of mathematics is the science of study and aid to study mathematics. Its aim is to describe and characterize the study processes (or didactic processes) in order to provide explanations and solid responses to the difficulties of people (students, teachers, parents, professionals, etc.) studying or helping others to study mathematics. (Bosch and Gascón, 2006, p. 60) The theoretical basis of thesis is based on ATD, a research programme evolved by Yves Chevallard, studying the phenomena related to teaching and learning. ATD postulates that mathematics is a human activity, produced, taught, learnt and practiced in educational institutions. One of the main point in ATD is that what is taught and learnt in an educational institution are influenced by the nature and origin of the mathematical knowledge and the institutional setting. Thus to study didactic phenomena the knowledge outside the classrooms and the specific institution have to be considered. In ATD, the notion of didactic transposition and the level of codetermination is used to study the conditions and constraints coming from the society and institutions. To describe mathematical activity ATD proposes a general epistemological model, where mathematics is described by praxeologies. According to ATD, a mathematical activity be comprised of two related aspects; the study process in which mathematical knowledge is constructed and the result of the constructions, the mathematical knowledge. These two components are described by respectively didactic and mathematical praxeologies. 2.1 praxeologies The didactic and mathematical praxeologies are structured in two interrelated components, praxis and logos. The praxis or "know how" is the way of doing to solve a problem or task, while logos or "knowledge" refers to the thinking and justification of the way of doing. In every praxeology the two components occur simultaneously and affect each other, since no praxis can be executed without being justified or explained (logos) and the justification or explanation (logos) influence the way of doing (praxis) (Bosch and Gascón, 2006, p. 59). 5 2 6 the anthropological theory of didactics 2.1.1 Mathematical praxeologies and organizations Like any other praxeology, a mathematical praxeology is composed of a practical and theoretical block. The practical block is made of types of problems or tasks (T) to be studied as well as the technique (⌧) to solve them. The types of task can be anything from solving an equation, determining coefficients in a model, evaluate a model or making a proof. Solving a task requires work with the task and the technique refers to the way of working. The technique characterizes the type of tasks in the practical block. The type of tasks and the technique constitute the practical block [T / ⌧] of a praxeology. This means that a specific task (t) can belong to more practical blocks, depending on which technique used to solve the task. For instance, there are different techniques to solve the task t: Given a function f(x) = 2x + 3 and a value of f(x0 ) = 5. Determine x0 : ⌧A : Isolating x0 . x0 = f(x0a)-b ⌧G : Draw f(x) as a graph. Read off x0 when f(x) = f(x0 ). ⌧I : Use solve at a computer algebra system (CAS). As the technique characterizes the type of task and the technique ⌧I can be used to determine an unknown variable in all classes of functions, several tasks belong to the practical block [T / ⌧I ]. In contrast the technique ⌧A is more specific and fewer tasks belong to the practical block characterized by this technique. The technique used to solve a task is, according to ATD, influenced by the institutional and historical settings. The logos or theoretical block includes the discourses necessary to describe, explain and justify the technique and is formed by the two levels: technology (✓) and theory (⇥). The technology explains and justifies the technique, while the theory is an argument which explain and justify the technology. The technology and theory constitute the theoretical block [✓ / ⇥] of a praxeology. The union of the two blocks (figure 2) constitutes the four basis components of a mathematical praxeology. A praxeology is usually described by a four-tuble [T / ⌧ / ✓ / ⇥]. Type of tasks T Technique ⌧ Technology ✓ Theory ⇥ Praxis Logos Figure 2: A mathematical praxeology is the union of the two main components: Praxis and logos The two blocks are inseparable and developed together. Solving a task 2.1 praxeologies consists in the construction of praxeologies bringing up ways of doing (praxis) and ways of justifying (logos). Development of the practical block raises new technological needs that influence the theoretical block. Similarly, the developments in the theoretical block influence the praxis, as description and justification of tasks can lead to new types of tasks or techniques. According to ATD, mathematical activities consist in activating praxeologies (old knowledge) to construct or reconstruct mathematical praxeologies to solve problems and tasks (Rodríguez et al., 2008, p. 291). Solving problems or tasks will result in construction of mathematical praxeologies and which praxeologies constructed or reconstructed depends on the task and the technique used to solve the task. The mathematical praxeologies and particular the collection of mathematical praxeologies can be classified by the notion of mathematical organization (MO), which is a collection of praxeologies arranged into a sequence of increasing complexity. The classification is based on the technique, technology or theory and forms a sequence of MOs with increasing complexity. The simplest MO is the punctual, generated by a unique type of tasks with the related technique. The punctual MO is the same as a mathematical praxeology. In punctual MO the theoretical block is often ignored or implicit assumed (Garcia et al., 2006, p. 227). The collection of praxeologies sharing technological discourse forms a local MO. The local MO is characterized by its technology and is made up of practical blocks sharing technology and theory. Like a task can belong to several mathematical praxeologies, a punctual MO can belong to different local MOs depending on the technology used to describe and justify the technique. Going a step further a regional MO is obtained by connecting several local MOs having the theoretical discourse in common. The most complex MO is the global MO, which is a collection of regional MOs. The connection between the MOs is shown in figure 3. The notion of mathematical praxeologies and MOs are used to describe and analyze mathematical knowledge, but as mathematical knowledge never is a definitive construction, the praxeologies and MOs are either. According to Bosch and Gascón (2006), it is possible to find knowledge (logos) that are not used to solve problems, since one did not know or has forgotten what type of tasks it can help to solve. Further techniques (praxis) which is not described do also exist. Praxeologies and MOs are result of a complex and ongoing process, where new tasks and mathematics are studied in a study process. The study is by itself a human activity and thus it can be modelling in term of praxeologies, called didactic praxeologies. 7 8 the anthropological theory of didactics Figure 3: The connection of punctual, local and regional MOs. The figure illustrates how punctual MOs can be collected in local MOs and how these local MOs can be collected in a regional MO. In fact, there is no mathematical praxeology without a study process that engenders it but, at the same time, there is no study process without a mathematical praxeology in construction. Process and product are the two sides of the same coin. (Garcia et al., 2006, p. 227) 2.1.2 Didactic praxeologies and moments The study process or didactic process constitutes together with mathematical praxeologies the aspects of mathematical activity. The didactic process is, in ATD, modeled in terms of praxeologies, called didactic praxeologies. The didactic praxeologies refer to any activity related to setting up or develop praxeologies and so can be used to model didactic processes for students and teachers. Students’ didactic praxeology is a model of how students study and learn mathematical praxiologies, whereas teachers’ didactic praxeology is a model of how teacher helps students to acquire the mathematical praxeologies, for instance by setting up the teaching (Barbé et al., 2005, p. 239). The study process can be organized into six moments, each of which characterize different dimensions of the studied MO: The moment of first encounter is the first meeting with the type of tasks or questions of the process. The first encounter should moti- 2.1 praxeologies vate the development of the MO at stake. The second moment, the moment of exploratory concerns the exploration of the type of tasks and elaboration of techniques to solve the tasks. The third moment, the technological-theoretical moment, consists of the construction of the technological-theoretical environment related to the technique. In this moment the students describe and justify the techniques, which can generate a need for new techniques. The fourth moment, the technical moment concerns the technical work with the task, in which a technique is improved or new techniques are developed to make the technique more reliable. In the firth moment, the moment of institutionalization, the elaborated and praxeology is identified and finally in the the moment of evaluation the praxeology is evaluated by investigate the limitations of the technique (Barbé et al., 2005, p. 238-239). In a study process the six moments do not have a chronological or linear structure and some moments can occur simultaneously as the moments are closely interrelated. Each moment has a specific function in the study process and a study process able to carry out all the moments leads to the construction of a local MO generated by the initial question or task (Barbé et al., 2005, 239). The teachers’ task is to construct a didactic process that enables the students to carry out all the moments to let the students establish the MO in stake. The possible ways of constructing the didactic process can be described by use of didactic praxeologies, since didactic praxeologies, as every praxeology, include a set of problematic didactic tasks, didactic techniques and didactic technologies and theories. A classic example of a teacher’s didactic task is how to set up and teach a MO under the certain conditions given to the teacher. In this thesis, the didactic task: "How to establish a MO around linear regression" is considered. The technique to a didactic task is the way the teacher solve the task, for instance the tools the teacher applied (ex: textbooks and tasks) and the way the teacher teach. ATD postulated that didactic tasks and the corresponding techniques can be set up and described by consider the six didactic moments, for instance; How to manage the first encounter?, Which specific tasks can help the students explore techniques?, How to introduce the technologicaltheoretical environment? (Rodríguez et al., 2008, p. 292). The didactic technologies and theories are the discourses that explain and justify the techniques. For instance, classroom teaching can be justified by the perspective that"students learn what the teacher explains clearly" (Garcia et al., 2006, p. 228), while the discourse of applying inquiry based teaching is that students learn most, if they by themselves construct the knowledge by experiments. It is important to notice that the didactic techniques and the techno- 9 10 the anthropological theory of didactics logical - theoretical discourse used in mathematical activity is mainly determined by institutions and historical tradition, for which the didactic research have to go beyond what identifiable in classrooms. The teachers’ practice (and the students) is strongly conditioned by restrictions of both mathematical and didactical origin. Thus, according to ATD, analyze and describe the MOs and the didactic praxeologies, the conditions and constraints from the society, scholarly and educational institutions have to be considered by examine the didactical transposition and the level of codetermination (Bosch and Gascón, 2006). 2.2 the didactic transposition ATD postulates that the mathematical and didactic praxeologies occurring in educational institutions are mainly determined by conditions and constraints that cannot be reduced to a specific classroom. To analyze the knowledge of regression and the conditions and constraints to teach in regression, it is important to study the origin of the theory and the rationale of the knowledge. To this the didactic transformation and the levels of determination can help. The process of didactic transposition refers to the transformation through which a piece of mathematical knowledge is transposed from the mathematical community until it is actually learned in a given institution. The knowledge taught in school is a production generated outside school that is made teachable through transposition. The transposition is a complex process consisting of three steps. Figure 4 illustrates the steps of the didactic process and shows how the knowledge is distinguish in four components; the scholarly knowledge, the knowledge to be taught, the taught knowledge and the learned knowledge (Bosch and Gascón, 2006, p. 56). Figure 4: The didactic transposition The first component is the knowledge produced in the mathematical community and this knowledge takes place at universities and occupation. The process of transposition starts with a selection of bodies of scholarly knowledge to be transmitted. This knowledge then undergoes a process of deconstruction and reconstruction to make the knowledge teachable in school (Bosch and Gascón, 2006, p. 53). This process is executed by the noospheres, who are politicians, mathematicians and members of the teaching system. The products of the 2.2 the didactic transposition process are the curricula, official programmes, exams, textbooks and recommendations to teacher etc. These products constitute the knowledge to be taught. The knowledge to be taught set up the mathematical content taught in the teaching institutions. Based on the knowledge to be taught, the teacher construct and teach concrete knowledge in the classroom, which constitute the taught knowledge. The last transformation takes place in the classroom when the students transform the taught knowledge to learned knowledge. The taught knowledge will be identical for students in a class whereas the learned knowledge will be distinct since students learn differently. The didactic transposition can also be divided into two parts; the external and internal transposition to differentiate the three institutions; the mathematical community, the educational system and the classroom (Winsløw, 2011, p. 537). The external transposition is the first transformation going from the mathematical community to the educational system, where the noospheres constitute the knowledge to be taught. The internal transposition is the second transposition from the educational system to the classroom. This transposition is called internal, since the teacher can influence this step by the proposed activities, test, resources and the teacher’s discourse. The didactic transposition underlines that mathematical praxeologies always arise in institutional setting and that the knowledge differ in the institutions (Bosch and Gascón, 2014, p. 70). This means that the knowledge in a specific institution has to be considered when doing research. For this reason, it makes no sense to apply the scholar knowledge as reference knowledge when consider knowledge in high school, as the transformation has imposes a lot of limitations of the knowledge found in the class. As the researchers always will belong to at least one of the three institutions, the ATD postulates the necessity to make an ERM such that the researchers avoid to be subject to the institution examined (Bosch and Gascón, 2006, p. 57). The ERM has to be elaborated from empirical data of all three institutions to secure that no institutions became more dominant than other. The collection of knowledge in an ERM should let the researchers answer questions occurring when analyzing didactic problems. Some relevant questions in this thesis are: What is the object called "regression"?, Where does it come from?, How does it vary from one institution to another, both in time and in institutional space?, What kind of praxeologies is regression made of in the different institutional setting?. These questions are part of studying what is taught and learnt in the classrooms about regression. In the thesis these questions will be answered to illuminate the bodies 11 12 the anthropological theory of didactics of knowledge existing about regression at the mathematical community and in high school. As in any act of modeling, models are used to simplify and structure data (Winsløw, 2011, p. 8) and in this thesis the data relevant for high school will be collected in a ERM, that will be used in the design and evaluation of the teaching sequence. 2.2.1 Level of codetermination ATD points out that the taught knowledge in school is determined through the didactic transposition and the frames in the educational institutions. Consequently, the conditions and constraints affecting the mathematical and didactic transposition occurring in the classroom have to be considered. Chevallard suggests a hierarchy of determination that can help identifying these restrictions at the different levels. In the hierarchy two kinds of restrictions are identified. The generic restrictions are the restrictions coming from the levels above the discipline and the specific restrictions arise from the level below the discipline (knowledge to be taught). The two kinds of restrictions may help to identify conditions and constraints that influence and define the knowledge taught, but go beyond what identifiable in the classroom (Barbé et al., 2005, p. 239). The hierarchy is structured into a sequence from the most generic level, (the civilisation) to the specific task (the subject), see figure 5. The highest level is the civilisation, followed by the society, the school and the pedagogy. These levels restrict the possible ways of organizing the didactic process and the kind of MOs that can be studied. For instance, restrictions at the school level Figure 5: Level of didactic are; the structure of time and class, the textcodetermination books used and CAS available. The generic levels influence the lower levels. The specific restrictions are arranged in the discipline, domain, sector, theme and subjects. The last two levels, subject and theme, are according to Garcia et al. (2006) the only levels the teachers can influence, as all the above levels are determined by the curricula and educational authorities. This didactic phenomenon is called the thematic autism of the teacher (Barbé et al., 2.2 the didactic transposition 2005, p. 257) and refers to that the teachers only have to decide how to organize the study of a MO at the theoretical level. The thematic autism of the teacher has the unfortunate influence that school mathematics rarely works at levels higher than the themes, as the teachers’ didactic problem is restricted to the selection of tasks to be studied. This means that teachers in a specific institution have not to decide, which domain a mathematical subject or theme belongs to, because the decision already is taken. For instance, regression belongs to two different domains depending on the institutions investigated. In high school the theme belongs to the domain of functions, while it at university is taught in the domain of statistics. Restrictions on the lower levels, especially from the curricula, influence that the themes, sectors and domains are only slightly connected in high school (Barbé et al., 2005, p. 257). As a result, the students do not form regional or global MOs, as the four lowest levels in the hierarchy correspond to the four different MOs (Barbé et al., 2005, p. 257). The lowest level in the hierarchy (the subject) corresponds to a mathematical praxeology, as a type of tasks with a technique is examined. The theme is formed by the types of tasks and techniques sharing technology, i.e. a local MO. The sector is formed by a collection of themes sharing theory, thus similar to a regional MO. The last level below the discipline is constituted by regional MOs. The restrictions coming from the different levels influence the teachers’ didactic process and consequently the students’ MO. In this thesis, ATD will be used to investigate the MOs about regression and to analyze the conditions and constraints related to regression in high school. 13 Part I THE EXTERNAL DIDACTIC TRANSPOSITION AND THE EPISTEMOLOGICAL REFERENCE MODEL To the didactician [the concept of didactic transposition] is a tool that allows to stand back, question the evidence, erode simple ideas, get rid of the tricky familiarity on the object of study. It is one of the instruments of the rupture didactics should make to establish itself as a proper field; it is the reason why the "entrance through the knowledge" into didactic problems passes from power to action: because the "knowledge" becomes problematic through it and, from now on, can figure as a term in the formulation of problems (either new or reformulated ones), and in their solution. (Bosch and Gascón, 2006, p. 55) 3 1st R E S E A R C H Q U E S T I O N According to ATD every didactic problem has to consider the way the mathematical knowledge is constructed and taking into account the conditions and restrictions imposed by the institutional setting, since what students actually learn in an institution depends of the transposition. The transposition lead to conditions and restrictions that influence what teachers teach and students learn in high school, therefore before designing the teaching sequence about regression the mathematical knowledge in different institutions will be analyzed. In this part the external transposition of regression will be analyzed and based on the analysis an ERM will be elaborated. The ERM will be basis of the design and the evaluation of the teaching sequence in part two. This part will investigate: • Which bodies of knowledge regarding regression exist in the mathematical community? • Which elements of scholar knowledge are relevant in relation to high school? • Which bodies of scholar knowledge are transposed to high school and how are they described in the knowledge to be taught? 17 4 THE SCHOLAR KNOWLEDGE This chapter concerns the first step in the didactic transposition, the scholar knowledge. In the chapter the bodies of knowledge about onedimensional regression existing in the mathematical community will be presented. 4.1 historical perspective: the derivation of the least squares method and regression The least squares (LSQ) method was one of the earliest statistical methods and became a very important statistical method in the nineteenth century. Adrien-Marie Legendre published the method first time in 1805. The reason for introducing the method was the problem of determining the relationship between objects (mostly the orbits of the comets) based on various measurements, when more data points than unknowns parameters were given (the problem of finding the best approximate solution to a system of m linear equations in n unknowns, m > n). The method was first published in an appendix to a work about the orbits of comets and immediately, the publication had a widespread effect. Within ten years the method became a standard method to solve astronomical and geodetical problems. Few years later (18081809), Robert Adrain and Carl Friedrich Gauss also published the method and since Gauss went far beyond Legendre, the development and discovery is mostly assigned to him (Katz, 2009, p. 820). In Legendre’s treatment of the method was a lack of formal consideration of probability, for which reason the accuracy of the method could not be determined. In contrast, Gauss justified in his publication Theory of the Motion of Heavenly Bodies from 1809 the theoretical basis of the method by linking the method to probability. In addition, Gauss proved algorithms to calculate the estimates, the method of Gaussian elimination. In the publication, the assumptions that were basis of his arguments and the methods can be found. The three principles are (Waterhouse, 1990, p. 43-45): 1) If several direct measurement of a quantity, all equally reliable are given, then the arithmetic mean of the measured value should be used as the estimate of the actual value of the quantity. 2) The parameter values that seem most likely, given the measurements, should be chosen (what Gauss mean by "most likely" is described below) 19 20 the scholar knowledge 3) The parameter values should be chosen such that the errors are distributed over all the equations rather than just a few of them. We should suppose that larger and smaller errors are equally possible in all of the equations without distinction. With principle 3) Gauss remarked that it makes no sense to throw some of the observation away, but rather the errors have to balance. In principle 2) the phrase "most likely" refers to greatest probability. Using the principles, Gauss showed that the parameter values are most probable, when the sum of the LSQ are minimized. Gauss worked with a function (x), which describe the probability of an error of magnitude x. Gauss proved by assuming 1) that the function (x) must be normal distributed, if the function satisfies; that it is symmetric both R1 about zero, asymptotic to the real axish in-h 2 x2 p directions and -1 (x)dx = 1. The function was (x) = ⇡ e (Katz, 2009, p. 821-822). Denote that (x) is equivalence with the standard normal distribution when h= p12 . Gauss claimed that the errors (ei ) of m various observations are independent if the observations are independent. Consequently the probability of all the errors was given by the product of the m joint functions (ei ): m Y i=1 (ei ) = hm ⇡-(1/2)m e-h 2 (e2 +...+e2 ) m 1 Q Pm 2 Gauss proved that m i=1 (ei ) is maximal (principle 2), when i=1 ei is minimal (Katz, 2009). Using the three principles, Gauss justified the method of LSQ. Initial, Galton used the word reversion (Katz, 2009, p. 826) Even through the LSQ method is the earliest form of regression, Legendre or Gauss did not use the term regression to describe the work. First around 1870-1880 the term was used to describe relation between biological phenomena, studied by Francis Galton. Analyses of Galton’s work show that he did not use the LSQ method to determine the relation between the empirical data. Galton simply plotted the points and estimated the line by eye. In the last decade of the nineteenth century George Udny Yule showed how to calculate the regression line using the LSQ method (Katz, 2009, p. 826). 4.2 regression The problem which regression is used to solve is the problem of determine the underlying relation of some measurements, where the measurements reflect the reality with some errors. The idea is that any phenomenon can be describe by a relation or a theory and that 4.2 regression the measurements are reflections of some ideal underlying relation. In real life the measurements will be random spread around the model that describe the relation. Regression is used to model the underlying relation based on the measurements. This way to consider regression agrees with the etymology of the word. Regression is compound of the verb regress and the suffix ion. To regress means to move backward or go back and the suffix -ion denote an action (Dictionary.com, 2015). Thus, regression can be interpreted as the method to go back from the measurements and determine the underlying relation, which produces the measurements. Regression analysis is the determination of a best mathematical model by identifying the relationship between a dependent and one or more independent variables. Making regression analysis two problems have to be considered: The determination of the class of functions that has to be used to model the data set (the type of model) and estimation of the parameters given the class of functions (the specific model). The two problems are interrelated and influence each other, however the techniques to solve the problems are quite different. Determine the specific model require construction of mathematical praxeologies, while determination of the class of functions requires construction of mathematical praxeologies mixed with non-mathematical object. 4.2.1 Types of regression There are four types of regression: Simple linear regression, simple non-linear regression, multiple linear regression and multiple nonlinear regression. The type of regression is classified by the number of independent variables and the relation between the variables. If the data set D = {(x1 , y1 ), . . . , (xn , yn )} is simple, the function is of one independent variable. In the thesis the notation D will refer to a simple data set. Simple linear regression refers to determination of the best line f 2 FL = {f(x) = ax + b|a, b 2 R} for a given data set D. Simple non-linear regression refers to determination of the best model of one independent variable given a non-linear class of functions. In this thesis the class of exponential functions FE = {f(x) = bax |a 2 R+ , b 2 R} and power functions FP = {f(x) = bxa |a, b 2 R} will be considered. The notation Fi , with i 2 {L, E, P} will refer to the collection of the three classes of functions. The multiple regression models are functions of two or more independent variables. The models can either be linear or non-linear in the parameters. The class of functions determines the type of regression done, for example exponential regression refer to determine the best f 2 FE . 21 22 the scholar knowledge Later, it will be described how exponential regression is converted to a problem of simple linear regression. This thesis will primary consider simple linear regression. As the method of simple linear regression is used to determine other models than the simple linear models, the connection and application to multiple linear models and simple non-linear models will be illustrated and reviewed. 4.2.2 Definition of the best linear model By far, no clear definition of "best model" has been stated, since no unique definition exist. Study the mathematical community, several definitions and justifications of "best model" can be found and in the following some of the definitions and the related methods will be presented to show the complexity of interpreting the "best model". The complexity of interpreting "best" is illustrated in the case of linear models. It seems that the definition of best model depends on the disciplines examined. In statistics the best model is define as the most probable given a data set, whereas the best model minimize some errors (vertical, perpendicular etc.) in the discipline of mathematics. In mathematics two very different approach to linear regression can be found. The geometric approach is an inaccurate approach, where the model is determined by eye and ruler. The technological-theoretical discourse of the technique is defective, because no common criterion or description of how to make the line can be made. As it is natural for one’s mind to consider the orthogonal distances, most common the length of the perpendicular distances will be minimized using this method. The method is vague, as it is based on an individual preference. This causes that several best models in a given class of functions can be determined. The algebraic approach is a more rigorous approach, because the sum to minimize can be explicit stated and consequently the technologicaltheoretical discourses can be described. In the algebraic approach different quantities to sum can be found. The quantities are characterized by the type of error and by the size of the error. What quantity is minimized affect which method to use to minimize the sum. The size of the errors To avoid positive errors cancel negative errors out, the errors have to be positive. This can be obtained by use of absolute values of the errors (|ek |) or squaring the errors (e2k ). Using the absolute values the errors are weighted by their distances, which means that an er- 4.2 regression ror twice as large is weighted twice. Squaring the errors, large errors are weighted more, for instance an error twice as large is weighted by four. Using the absolute size of the errors, outliers in data set influence the model to a lesser extent than using the squared errors (Weisstein, 2015). The type of the errors Figure 6 illustrates that three types of errors; the vertical, the horizontal and the perpendicular exist and that the type of errors used had a influence on the line determined. Figure 6: The best linear model calculated by minimize the sum of the squares of the (A): The vertical errors, (B): The horizontal errors, (C): The perpendicular errors. The figure are retrieved 23/2 2015 from http://elifesciences.org/content/2/ e00638 The figure illustrates that the choice of dependent and independent variable is important when making regression, as the model differ depending on whether x is the independent or dependent variable. However, often the relation of the variable is clear when modeling data. For example, consider the variable pulse and workload. Workload has an effect on pulse, while the reverse makes no sense, thus pulse has to be the dependent variable. Determine the best line the errors have to minimal, why it makes sense to minimize the errors in the variables by the perpendicular distance. Using the perpendicular distance, the shortest distance from the point to the line is used, which means that the errors in both the independent and dependent variables be taken into account. Because both variables are taken into account, the determination of the distance can be difficult if the variables are measured in different units (for instance pulse and workload), since calculations of perpendicular distances require that the units of the variables are equal. Thus one variable has to be rescaled, before the distances can be determined. For that reason using the perpendicular distances require lot of calculation. Usually, when consider measurements the errors in the independent variable is few or zero, why it makes sense only to minimize the er- 23 24 the scholar knowledge rors in the dependent variable (vertical errors) and not the errors in the independent variable (horizontal errors). The method The size of the errors and the type of errors that should be minimized affect the method used to minimize the sum. In the following three different methods, which can be used to determine the best line will shortly be commented. The least absolute deviation method: Minimize the sum of the absolute values of the vertical errors n X |yk - (axk + b)| k=1 The linear LSQ method: Minimize the sum of the squared vertical errors n X yk - (axk + b) 2 k=1 The total LSQ method: Minimize the sum of squared perpendicular errors. In the formula the variables are assumed to be measured in the same units. n X (yk - (axk - b))2 1 + b2 k=1 The least absolute deviation method can be helpful to used when all observations have to be given equal emphasis, as the method is resistant to outliers in data (Siemsen and Bollen, 2007, p. 228). The method has the disadvantaged that no unique solution to a data set exists, as infinitely many lines to one data set can occur1 . In addition, the method is not simple to use, because the sum is non differentiable. The linear LSQ method is simple to use and the method will always give one unique solution. The unique solution and the method will be described in 4.3. The total LSQ method is easy to use when the variables are measured in the same units, but in the cases were they are not, the method is not straightforward, as firstly the variables have to be rescaled. Because of the rescaling different lines can be found. If the errors in the independent variable is zero, the method is equivalence with the linear LSQ method. Table 1 sum up the properties of the three methods. 1 Consider the three points (0,0), (3,2), (3,6). The linear model y = kx with k 2 [ 23 ; 2] minimize the sum of the absolute values of the vertical errors 4.2 regression Linear LSQ Least absolute deviation Total LSQ Not robust Robust Not robust Simple method Not simple method Not simple method Always one unique solution Possibly more solutions More solutions (depends of the rescaling) Table 1: The properties of the three methods to make linear regression Finally, the statistic interpretation of the best model is commented. In statistics, the best linear model is defined as the model that maximizes the likelihood function given some distribution of the errors (like Gauss did). The maximum likelihood estimation can be performed when the distribution of the errors belongs to a parametric family of probability distribution. Assuming the errors are independent and normal distributed, the linear model determined by the estimation will be similar to the model determined by the linear LSQ method. If the errors instead follow a Laplace distribution, the linear model determined by maximum likelihood estimation will be equivalence to the linear model determined by the least absolute deviation method. Analyzing the distribution of the errors, the relevant method to use in a given situations can be determined and justified. The most applied method used to determine the best linear model found in the mathematical community is the linear LSQ method that minimize the vertical errors. The rationales for using this method are: • Determination of the linear model by linear LSQ method is simple • Assuming the errors are independent and normal distributed, the model determined by the linear LSQ method is also the most probable model. • In many applications, an error twice as large as another error is more than twice as bad. Squaring the errors accounts for this discrepancy. In the rest of the thesis, the best linear model refers to the model f 2 FL determined by the linear LSQ method, unless otherwise mention. 4.2.3 Definition of the best model The criterion for the best linear model can be generalized to the nonlinear and multiple models, but in contrast to the linear case, the LSQ method are not simple to use when the models are non-linear. For instance, consider the simple non-linear function f(x, ), where is a 25 26 the scholar knowledge vector of m parameters. Using the LSQ method the sum S( ) that are minimized is given by: S( ) = n X yk - f(xk , ) 2 k=1 The minimum of S( ) occurs in a stationary point, i.e. rS( ) = 0. The derivatives are given by n X @S @f(xk , ) =2 (yk - f(xk , )) @ j @ j k=1 k, ) Since @f(x is a function of both the independent variable and the @ j parameters, the parameters cannot be directly determined. This implies that no direct method to determine the parameters (i.e. the model) exists. Consequently the non-linear model is found by approximation of the model by linear models. This problem is mostly solved by iterative refinement. As the technique of iterative refinement is demanding and complex, if possible, other techniques are used to model non-linear functions. One technique is linearization, where a non-linear data set is transformed into a linear domain and then the squared errors (in the linear domain) are minimized by the linear LSQ method. This technique can be applied to regression models that can be transformed to be linear in the parameters, for instance exponential and power functions. The technique is discussed in section 4.6. Other non-linear models can be considered as linear in the parameters and can be easily solved by either simple or multiple linear LSQ method (The technique to solve multiple linear models is shortly described in section 4.3.3). Examples of non-linear models are polynomials, f(x) = 0 + 1 · log(x) and f(x) = 0 + 1 · sin(x) + 2 · sin(2x). 4.3 the method of least squares The method of LSQ is a standard approach to the approximate solution of overdetermined systems. Given a data set D = {(xk , yk ), k = 1, . . . , n} and a class of functions, the task is to find the model f(x) which minimize the sum S(f) = n X yk - f(xk ) 2 k=1 This means to find a model f(x) which satisfy that {(xk , f(xk ), k = 1, . . . , n} is closest to the data set D, where closest is measured as the sum of squared vertical distances from the model to the data points. The LSQ method can be applied to all classes of functions, but as described in section 4.2 the method is only simple, when the class of 4.3 the method of least squares functions is linear. Thus the method will be described for the linear class of functions FL = {f(x) = ax + b|a, b 2 R}. Using the method of linear LSQ a unique linear model can be determined by solve task tM : Minimize Pn 2 k=1 (yk - (axk + b)) . In the following, the theorem of the linear model and several techniques to solve task tM (prove the theorem) will be presented. To solve the task many techniques has to be used, however one will be essential to solve the task. This technique will be highlighted. The essential techniques belong to the mathematical domain of; elementary algebra, multivariable calculus and linear algebra. For shortness, the following notations will be used: n 1 X x= xk n k=1 Sx = Sy = Sxy = Tx2 = n X (xk - x) k=1 n X (yk - y) k=1 n X n 1 X y= yk n k=1 S2x = S2y = n X (xk - x)2 k=1 n X (yk - y)2 k=1 (xk - x)(yk - y) k=1 n X x2k k=1 Txy = n X xk yk k=1 Theorem 4.3.0.1 (Linear LSQ). Let D with n > 2 be given. Assume not all xk are equal. The linear model y = ax + b, which minimizes the sum P 2 S(a, b) = n is given by: k=1 yk - (axk + b) Pn Sxy (x - x)(yk - y) Pn k a = 2 = k=1 , b = y - ax 2 Sx k=1 (xk - x) 4.3.1 Elementary algebra The following proof is based on elementary algebra and the technique is: completing the square. The technique is well-known to students at second and third year in high school, where the technique is taught in relation with quadratic polynomial2 . 2 The technique is used to convert a quadratic polynomial of the form ax2 + bx + c to the form a(x + h)2 + k, with h, k constants 27 28 the scholar knowledge In this proof and the subsequent the following rules of summation is used: n X k=1 ak ± n X bk = k=1 n X k=1 n X n X (ak ± bk ) k=1 c · ak = c · n X ak k=1 c = nc k=1 The proof is based on Winsløw (2015). 4.3.1.1 1st proof of Theorem 4.3.0.1 Proof. Given D with n > 2 and not all xk equal. Determine a, b that minimize the sum S(a, b) = n X yk - (axk + b) 2 k=1 Pn P P First denote that Sx = k=1 (xk - x) = n xk - n k=1 k=1 x = x · n Pn x · n = 0 and similar Sy = k=1 (yk - y) = 0. Since not all xk are equal, S2x > 0. Let D = y - b - ax. S(a, b) = = = = n X k=1 n X k=1 n X yk - axk - b 2 y + (yk - y) - a(x + xk - x) - b D + (yk - y) - a(xk - x) k=1 n ✓ X 2 2 D2 + 2D(yk - y) - 2aD(xk - x) + (yk - y)2 - k=1 2a(xk - x)(yk - y) + a2 (xk - x)2 ◆ = nD2 + S2y - 2aSxy + a2 S2x ✓ ✓ ◆ ✓ ◆ ◆ aSxy Sxy 2 Sxy 2 2 2 2 = nD + Sx a - 2 2 + + S2y Sx S2x S2x ✓ ◆ Sxy 2 (Sxy )2 2 2 = nD + Sx a - 2 + S2y Sx S2x The first two terms depend on the parameters a, b, while the last two terms are constants. 4.3 the method of least squares 29 The sum S(a, b) is minimal, when the two terms that depend on the parameters a, b are equal zero. ✓ ◆ Sxy 2 Sxy Sxy 2 2 Since Sx > 0, Sx a - 2 = 0 , a- 2 = 0 , a = 2 Sx Sx Sx Since n 6= 0, nD2 = 0 , D2 = 0 , 0 = D = y - b - ax The sum S(a, b) is proved to be minimal when Pn Sxy (x - x)(yk - y) Pn k a = 2 = k=1 2 Sx k=1 (xk - x) b = y - ax In the domain of elementary algebra many different techniques to make the proof can be found. The techniques used influence the didactic process and the mathematical praxeology constructed. In the following proof another technique in the domain of elementary algebra is used. The technique is to constrain the problem to a centralized data set and used that min (f(a, b))=min ( (a)) + min( (b)), when f(a, b) = (a) + (b). The proof is designed by me based on Key (2005) and Nielsen (2015). 4.3.1.2 2nd proof of Theorem 4.3.0.1 Proof. Given D with n > 2 and not all xk equal. Determine a, b that minimize the sum S(a, b) = n X yk - (axk + b) 2 k=1 Let {(hk , vk )|k = 1, . . . , n} be a data set with hk = xk - x and vk = yk - y. The data set is centralized since n X hk = k=1 n X (xk - x) = k=1 and similar n X xk - k=1 n X vk = k=1 n X n X k=1 x = x·n-x·n = 0 (yk - y) = 0 k=1 The best line v = ah + b of the centralized data set is f(a, b) = = = = n X k=1 n X k=1 n X k=1 n X k=1 vk - ahk - b 2 v2k - 2ahk vk - 2bvk + a2 h2k + 2abhk + b2 v2k - 2a v2k - 2a n X k=1 n X k=1 hk vk - 2b h k vk + a 2 n X k=1 n X k=1 vk + a 2 n X k=1 h2k + nb2 h2k + 2ab n X k=1 hk + nb2 30 the scholar knowledge The sum f(a, b) can be written as two second degree polynomials (a) + (b) in respectively a and b. Using min(f(a, b))=min( (a)) + min( (b)), (a) og (b) are minimized. P 2 Since n k=1 hk > 0 the polynomial (a) = a 2 n X h2k - 2a k=1 is minimal for n X hk vk + k=1 n X v2k k=1 Pn k=1 hk vk a= P n 2 k=1 hk (b) = nb2 is minimal for b = 0. The sum f(a, b) is minimized when Pn k=1 hk vk a= P n 2 k=1 hk b=0 Since a displacement do not change the slope of the line, the best line y = ax + b for the data set D is given by: Pn Pn (x - x)(yk - y) k=1 hk vk Pn k a = Pn = k=1 2 2 k=1 (xk - x) k=1 hk The intersection b is given by: v = ah , y - y = a(x - x) , y = a(x - x) + y. 4.3.2 Thus b = y - ax Multivariable calculus The following proof is based on multivariable calculus. The technique is to determine the global minimum of a function of two variables. The technique requires knowledge of function of two variables, more precise: how to differentiate a function of two variables, both the first and second derivative, how to calculate extreme values, i.e. the second derivative test and a method to determine if the points are local/global. Before the proof, a definition and a theorem from first year calculus at university is stated (Kro, 2003). The proof is based on the definition and the theorem. Definition 4.3.2.1 (Stationary point). Let f : Rn ! R be a function of n variables. A point p is called stationary if rf(p) = 0. Theorem 4.3.2.1 (Second derivative test). Let f(x, y) be a C2 -function and assumes that (a, b) is a stationary point of f. Let Hf(a, b) be the Hessian of f in (a, b). Let D be the determinant of the Hessian in (a, b), so D =det(Hf(a, b)). 4.3 the method of least squares • If D < 0, then (a, b) is a saddle point • If D > 0 and @2 f (a, b) @x2 > 0, then is (a, b) a local minimum • If D > 0 and @2 f (a, b) @x2 < 0, then is (a, b) a local maximum • If D = 0, then the test gives no conclusion The following proof is mainly based on Miller (2006). 4.3.2.1 3rd proof of theorem Theorem 4.3.0.1 Proof. Given D with n > 2 and not all xk equal. Determine a, b that minimize the sum S(a, b) = n X yk - (axk + b) 2 k=1 Firstly the stationary points of the function S(a, b) are determined. The stationary points are given by: rS(a, b) = 0 , @S @S (a, b) = 0 ^ (a, b) = 0 @a @b The derivatives of the function are: ◆ ◆ n ✓ n ✓ X X @S 2 = 2(yk - axk - b) · -xk = -2 yk · xk - axk - bxk @a k=1 k=1 ✓ ◆ 2 = -2 · Txy - aTx - b · n · x ◆ ◆ n ✓ n ✓ X X @S = 2(yk - axk - b) · -1 = -2 yk - axk - b @b k=1 k=1 ✓ ◆ = -2 · n · y - a · x - b Setting the derivatives equal to 0 and dividing with -2 gives: Txy - aTx2 - b · n · x = 0 n· y-a·x-b = 0 Rearrange the terms gives a linear system with two equations and two unknowns (a, b). Txy = aTx2 + b · n · x n·y = n· a·x+b 31 32 the scholar knowledge The solution is: b = y-a·x ^ Txy = aTx2 + b · n · x , Txy = aTx2 + (y - a · x) · n · x Txy - y · n · x = a(Tx2 , 2 - x · n) , Pn Txy - y · n · x yk · xk - y · n · x a= = k=1 Pn 2 2 2 2 Tx - x · n k=1 xk - n · x Pn Pn Using Sxy = - x)(yk - y) = k=1 (xkP k=1 yk · xk - y · n · x and P n 2 = 2 - n · x2 gives: S2x = n (x x) x k=1 k k=1 k a= Sxy S2x The function has one stationary point. The point is: a= Sxy S2x and b = y-a·x Since the function only has one stationary point and S(a, b) > 0, a local minimum is also a global minimum. Thus it is enough to show that the stationary point is a local minimum. By theorem 4.3.2.1 the stationary point is a local minimum if @2 S det(HS(a, b)) > 0 and @a 2 (a, b) > 0. The second derivatives are: n X @2 S (a, b) = 2 x2k = 2Tx2 2 @a k=1 @2 S (a, b) = 2n @b2 n X @2 S (a, b) = 2 xk = 2nx @b@a k=1 The determinant of the Hessian: det(HS(a, b)) = = @2 S (a, b) @a2 @2 S @b@a (a, b) 2Tx2 2nx @2 S @a@b (a, b) @2 S (a, b) @b2 = 4nTx2 - 2nx 2 2nx 2n ✓ ◆ ✓X ◆ n 2 2 2 2 = 4n · Tx - nx = 4n · xk - nx k=1 = 4n · n X (xk - x)2 > 0 k=1 4.3 the method of least squares Pn @2 S 2 Since @a 2 (a, b) = 2 k=1 xk > 0, the stationary point corresponds to a local and thus a global minimum of S(a, b). It is proved that S(a, b) has global minimum when a = b = y-a·x 4.3.3 Sxy S2x and Linear algebra The last technique to prove theorem 4.3.0.1 is based on linear algebra. Using the technique the data is treated as vectors. The technique is to use orthogonal projection to minimize the length of the residual vector. Before the proof, a definition and two theorems that the proof is based on, are given. Definition 4.3.3.1 (Orthogonal projection). Let V be an Euclidian space with a vector v. Let W be a linear subspace of V. A vector w 2 W is the orthogonal projection of v onto W if • w2W • (v - w) ? W. The vector w is denoted projW v Theorem 4.3.3.1 (The approximation theorem). Let V be an Euclidian space with a vector v. Let W be a linear subspace of V. projW v is the vector in W closest to v 2 V and is the only vector with this property. kv - projW vk 6 kv - wk 8w 2 W Proof. Let w be a arbitrary vector in W. Figure 7: Projection of v to W 33 34 the scholar knowledge kv - wk2 = kv - projW v + projW v - wk2 = kv - projW vk2 + kprojW v - wk2 + 2hv - projW v, projW v - wi = kv - projW vk2 + kprojW v - wk2 using that projW v - w 2 W (see figure 7) and consequently by definition 4.3.3.1 v - projW v ? projW v - w , hv - projW v, projW v - wi = 0. Since kprojW v - wk2 > 0 we have kv - projW vk 6 kv - wk8w 2 W Next we show that projW v is unique. Let v = projW v + m for a vector m 2 W ? . Assume projW v is not unique. Let v = p + n with p 2 W, n 2 W ? . Then projW v + m = p + n , projW v - p = n - m Since projW v - p 2 W and n - m 2 W ? the projW v is unique because hprojW v - p, n - mi = hprojW v - p, projw v - pi = 0 so projW v = p. The next theorem uses theorem 4.3.3.1 to prove the best approximation solution of an overdetermined system. In simple linear regression with n data points we have a system of n equations with two unknowns (a, b). The proof of the best linear model(theorem 4.3.0.1) is a special case of the following theorem. Theorem 4.3.3.2 (The best approximation solution to an overdetermined system). Let A~x = ~b, with A a n ⇥ k matrix, ~x 2 Rk , ~b 2 Rn and A 6= 0, ~x 6= ~0, ~b 6= ~0. The best solution x~⇤ to the system is given by x~⇤ = (AT A)-1 AT ~b The proof of theorem 4.3.3.2 is mainly based on Khanacademy. k ~ n Proof. Let A~x = ~b, with A a n ⇥ k2matrix, 3 ~x 2 R 2, b 23R . 2 3 x b 6 17 6 17 6 7 6 6 7 x2 7 b2 7 7 7 , ~x = 6 ~b = 6 Let A = 6 and 6 7 6 7 a ~ a ~ . . . a ~ 2 k5 . . 4 1 6 .. 7 6 .. 7 4 5 4 5 xk bn ~ If a solution to A~x = b exist the solution fulfills a~1 x1 + a~2 x2 + . . . + a~k xk = ~b This means that there must exist a linear combination of the column vector of A equal to ~b, i.e. ~b must be in the column space of A, denoted C(A). 4.3 the method of least squares If ~b 2 / C(A), A~x 6= ~b. Thus the system cannot be solved. In this case a solution x~⇤ that gets Ax~⇤ as close as possible to ~b is determined. This means a solution x~⇤ such that kAx~⇤ - ~bk 6 kA~x - ~bk 8~x 2 Rk Since Ax~⇤ 2 C(A), ~b 2 Rn and C(A) is a linear subspace of Rn , projC(A)~b is the vector in C(A) closest to ~b (theorem 4.3.3.1). Hence kAx~⇤ - ~bk is minimized when Ax~⇤ = projC(A)~b, see figure 8. Figure 8: Projection of ~b to C(A) By definition 4.3.3.1 Ax~⇤ - ~b ? C(A) , Ax~⇤ - ~b 2 C(A)? , Ax~⇤ - ~b 2 N(AT ) where N(AT ) denote the null space of AT . The best solution is given by: Ax~⇤ - ~b 2 N(AT ) , AT · (Ax~⇤ - ~b) = 0 AT Ax~⇤ = AT ~b , x~⇤ = (AT A)-1 AT ~b , In the last equation is used that the column vectors of A are linear independent, which implies that A has full column rank. Thus (AT A)-1 exists. In the proof the best approximate solution is obtain by minimize kAx~⇤ - ~bk. Let Ax~⇤ = v, then p kAx~⇤ - ~bk = kv - ~bk = (v1 - b1 )2 + (v2 - b2 )2 + . . . + (vn - bn )2 . For that reason the technique is equivalent with the linear LSQ method and can be used to prove the best linear model(4.3.0.1). 35 36 the scholar knowledge 4.3.3.1 4th proof of Theorem 4.3.0.1 The proof is a special case of theorem 4.3.3.2 with A a n ⇥ 2 matrix, ~x 2 Rn and ~b 2 R2 . Proof. Given D with n > 2. Let 2 3 1 x1 6 7 61 x 7 27 6 X = 6. . 7 6 .. .. 7 4 5 1 xn ~p = 2 y1 3 6 7 6y 7 6 27 ~y = 6 . 7 6 .. 7 4 5 yn " # b a From theorem 4.3.3.2 the best solution to X~p = ~y is given by ~p = (XT X)-1 XT ~y Want to determine ~p: " # 1 1 ... 1 T X = x1 x2 . . . xn XT X = " 1 = " n (XT X)-1 = (XT X)-1 XT = (XT X)-1 XT ~y = = 1 x1 x2 # n¯x 1 x1 3 7 " # 6 # Pn 61 x 7 n x 27 6 k k=1 · 6 . . 7 = Pn Pn 2 .. .. 7 . . . xn 6 x k=1 k k=1 xk 4 5 1 xn ... 1 n¯x Tx2 1 nTx2 - n¯x 2 · 2 · 1 nTx2 - n¯x 1 nTx2 - n¯x 2 2 2 " Tx2 # -n¯x -n¯x n " # Tx2 - x1 n¯x . . . Tx2 - xn n¯x nx1 - n¯x . . . nxn - n¯x ✓ ◆ ✓ ◆3 2 2 6y1 Tx - x1 n¯x + . . . + yn Tx - xn n¯x 7 ✓ ◆ ✓ ◆ 7 ·6 4 5 y1 nx1 - n¯x + . . . + yn nxn - n¯x 3 2 nyT ¯ x2 -Txy nx¯ 2 nT ¯ 2 5 x -(nx) 4 nTxy -n2 x¯ y ¯ nTx2 -(nx) ¯ 2 This parameters a, b is proved to be: 3 2 3 " # 2 nyT ¯ x2 -Txy nx¯ y xa 2 2 b nTx -(nx) ¯ 5=4 5 ~p = = 4 nT 2 ¯y Sxy ¯ xy -n x a 2 2 2 Sx nT -(nx) ¯ x 4.3 the method of least squares In the rewriting is used that b= nyT ¯ x2 - Txy nx yT ¯ x2 - Txy x = nTx2 - nx)2 Tx2 - nx2 yT ¯ x2 - ynx2 - Txy x + ynx2 Tx2 - nx2 Txy - ynx y(T ¯ x2 - nx2 ) = -x 2 2 2 Tx - nx Tx - nx2 = = y - xa a= nTxy - n2 xy nTx2 - nx 2 = Txy - xny Tx2 - nx2 = Sxy S2x The proof of simple linear regression is a special case of theorem 4.3.3.2. The theorem (4.3.3.2) can be used to determine the best multiple linear model (multiple linear regression). Let f be a multiple linear function of k variables, st. f(x1 , . . . , xk ) = a0 + a1 x1 + . . . ak xk . Let n data points in Rk+1 be given: (x11 , . . . , x1k , y1 ), . . . , (xn1 , . . . , xnk , yn ) . Let X be a n ⇥ (k + 1) matrix, ~y 2 Rn and ~p 2 Rk+1 . 2 3 2 3 2 3 1 x11 . . . x1k a0 y 6 7 6 7 6 17 61 x 7 6 7 6 7 21 . . . x2k 7 6 6 a1 7 6 y2 7 X = 6. ~ p = ~ y = 7 6 7 6 7 .. .. 7 6 ... 7 6 ... 7 6 .. . . 4 5 4 5 4 5 1 xn1 . . . xnk ak yn The parameters ~p is given by ~p = (XT X)-1 XT ~y (theorem 4.3.3.2). The four proofs of theorem 4.3.0.1 differ in complexity. The first two proofs are based on mathematical knowledge that can be found in high school. The other two proofs are based on mathematics that are not taught in high school. Making the two last proofs teachable in high school while keeping its power require implementation of new mathematical domain (linear algebra, function of two variable) in high school. This shows that the proofs cannot be implemented in high school without reduce its power or justification. 37 38 the scholar knowledge 4.4 statistical approach 4.4.1 Maximum likelihood estimation The statistical approach to make simple linear regression is based on normal distributed stochastic variables and their probability density functions. Let {y = + ↵(x - x¯ )|↵, 2 R} be the class of linear functions describing the linear relation between the variables. Using the maximum likelihood estimation, the parameters that make the observed values (D) most probable is estimated (Ditlevsen and Sørensen, 2009). The basis of the statistical linear model y(x) = + ↵(x - x) is independent stochastic variables Y1 , . . . , Yn each normal distributed with mean + ↵(xk - x) and variance 2 , i.e. Yk ⇠ N( + ↵(xk - x), 2 ). Assuming the following assumptions, the model determined will be the most probable. • The mean of Yi is a linear function of the xi • The Yi are independent and normal distributed • The variance of the residuals is constant across observations (homoscedasticity) (Ditlevsen and Sørensen, 2009, p. 105) Assuming Yk to be independent normal distributed stochastic variables, the most probable values of the parameters are found by maximize the likelihood function. The likelihood function is the joint density of yk regarded as a function of the parameters ↵, , 2 . The likelihood function is: ◆ n 1 1 X 2 2 Ly (↵, , ) = · exp - 2 (yk - - ↵(xk - x)) 2 (2⇡ 2 )n/2 k=1 The estimates, which maximize the likelihood function is denoted ↵, ˆ ˆ , ˆ 2 and fulfill that Ly (↵, ˆ ˆ , ˆ 2 ) > Ly (↵, , 2 ) 8↵, , 2 2 R ⇥ R ⇥ [0, 1[. Maximize the likelihood function is simplified to a problem of maximize the log-likelihood function. As log(x) is a strictly monotonically increasing function, the transformation does not affect the optimality. The log-likelihood function is: n log(Ly (↵, ˆ ˆ , ˆ 2 )) = - log(2⇡ 2 2 )- 2 n 1 X 2 (yk - - ↵(xk - x))2 k=1 The log-likelihood function log(Ly (↵, ˆ ˆ , ˆ2 )) is maximal, when Pn 2 k=1 (yk - - ↵(xk - x)) is minimal. Thus the problem is reduced to the LSQ method and the method to determine ↵, is also found 4.4 statistical approach to be similar to the method described in section 4.3 (Ditlevsen and Sørensen, 2009). After determine ↵, the variance can be determined, since the function is reduced to be of one variable. The estimate of the parameters is given by: ↵ˆ = Sxy S2x ˆ =y ˆ2 = n 1 X (yk - y - ↵(x ˆ k - x))2 n k=1 The estimated variables are independent and their distributions are: ↵ˆ ⇠ N(↵, 2 /S2x ), ˆ ⇠ N( , 2 /n), nˆ2 ⇠ 2 2 n-2 Using the maximum likelihood function the linear model is justified as the most probable model. This justification has a high epistemological value when consider and argue for the definition of the best linear model, as it seems rational that the best linear model is the most probable. However, the technique of maximum likelihood function is difficult to transform to be simple while keeping its functional character, since both the technique of maximum likelihood function and the LSQ method has to be used. 4.4.2 Random variable and orthogonal projection In the following, another approach to simple linear regression in statistics is shown. The approach is based on vector space of random variables and orthogonal projection. The best linear model (Theorem 4.3.0.1) is proved by considering a linear space of random variables and use orthogonal projection to show the best predictor of Y given X is the conditional expectation E(Y|X). The proof of the best linear model is a special case of a more general theorem, which first is stated. The theorems and proof are mainly based on Kachapova and Kachapov (2010) and Grimmett and Stirzaker (2001) Theorem 4.4.2.1 (Projection theorem). Let W be a Euclidian space of random variables with scalar product given by (X, Y) = E(X · Y). Let Y satisfy E(Y 2 ) < 1. Let M 2 W. The following two statements are equivalent: E (Y - M)V = 0 8V 2 W kY - Mk 6 kY - Vk 8V 2 W Theorem 4.4.2.2. Let X and Y be random variables and suppose E(Y 2 ) < 1. Let W = {f(X)|f : R ! R, E(f(X)2 ) < 1}. The best predictor of Y given X is the conditional expectation E(Y|X). 39 40 the scholar knowledge ˆ is minimal. Proof. The best predictor of Y has to fulfill that kY - Yk Have to prove that E(Y|X) 2 W and E (Y - E(Y|X))V = 0 8V 2 W (theorem 4.4.2.1). First E(Y|X) 2 W is proved: Using Cauchy-Schwarz inequality and E(Y 2 |X) = E(Y 2 ): E E(Y|X)2 6 E(E(Y 2 |X)) = E(Y 2 ) Since E(Y 2 ) < 1 ) E E(Y|X)2 < 1 and so E(Y|X) 2 W. Now E (Y - E(Y|X))V = 0 8V 2 W is proved Let V = f(X) 2 W. E (Y - E(Y|X))V = E(Y · f(X)) - E E(Y|X) · f(X) = E(Y · f(X)) - E(Y · f(X)) =0 using the property: E(r(X) · E(Y|X)) = E(r(X) · Y) 8r : X ! R. By theorem 4.4.2.1 kY - E(Y|X)k < kY - Vk8V 2 W. Theorem 4.4.2.2 gives that the best predictor is given by the conditional expectation. In the following the theorem is used in the special case when W = {f(X) = aX + b|a, b 2 R}. Proof. Let W = {aX + b|a, b 2 R}. Let the vector in W closest to Y be + ↵X for ↵, 2 R (theorem 4.4.2.1) Let " = Y - - ↵X. Then E(" · 1) = 0 and E(" · X) = 0, as " 2 W ? and 1, X 2 W. This implies: E((Y - ") · 1) = E(Y · 1) - E(" · 1) , E((Y - ") · 1) = E(Y · 1) and E((Y - ") · X) = E(Y · X) - E(" · X) , E(Y - ") · X) = E(Y · X) This leads to a system of two linear equations. Using Y - " = gives: E( + ↵X) = E(Y) E( X + ↵X2 ) = E(Y · X) , , + ↵X + ↵E(X) = E(Y) E(X) + ↵E(X2 ) = E(Y · X) 4.5 the quality of a model The solution of the linear system is given by: + ↵E(X) = E(Y) , = E(Y) - ↵E(X) = µY - ↵µX E(X) + ↵E(X2 ) = E(Y · X) , (E(Y) - ↵E(X))E(X) + ↵E(X2 ) = E(Y · X) 2 , 2 ↵(E(X ) - E(X) ) = E(Y · X) - E(Y)E(X) ↵= ↵= E(Y · X) - E(Y)E(X) E(X2 ) - E(X)2 Cov(X, Y) 4.5 2 X , 2 X where µY , µX is the mean, Cov(X, Y) the covariance and ance. Substitute µY , µX , Cov(X, Y), be similar to 4.3.0.1. , 2 X the vari- with y, x, Syx , S2x , the linear model the quality of a model After a model is determined, a natural question is; How well did the model represent the set of data?. To answer this questions, the model has to be evaluated. As several ways to estimate the quality of the model exist, only the simplest methods will be presented. The methods presented are equivalent with the methods in the high school textbooks. 4.5.1 Visual examination A simple, though not qualitative precise way to evaluate models is to conduct visual examination of the data and error distribution by use of plots. Scatter plot can be used to study the relationship between two variables, since the plot provides a good indication of the nature of the relationship. Using scatter plot, a linear relation is easy to recognize, whereas non-linear relations are more difficult to recognize because some non-linear relations look like each other (figure 9). As some non-linear functions can be transformed to be linear, plotting the transformed data can help recognize the relation. For instance, exponential and power functions became linear in the variables with respectively semi-log and log-log transformation (see section 4.6). The distribution of errors can be investigated on residual plots. Residual plot is used to investigate obviously deviations from randomness between a data set and a model. Making models the errors are assumed to be constant across observations. Figure 9 shows examples of scatter and residuals plot. 41 42 the scholar knowledge (a) Exponential model (b) Power model Figure 9: Examples of scatter plot and residual plot. At figure a) the errors are random, while the variances of the errors in b) is systematic 4.5.2 Two strongly related definitions of the Pearson correlation coefficients appear; The population and the sample. The correlation coefficient The most familiar correlation coefficient is the Pearson correlation coefficient, which measures the correlation of a linear relation between two variables. Definition 4.5.2.1 (Sample Pearson correlation coefficient). The correlation coefficient r 2 [-1, 1] is given by Pn Sxy Sxy (xk - x¯ )(yk - y) ¯ q r = pPn k=1 =q =p Pn 2 2 ¯) ¯ S2x · S2y S2x · S2y k=1 (xk - x k=1 (yk - y) The correlation coefficient is a measure of the strength and direction of a simple linear relation between two variables x and y. If r < 0 the slope of the linear model is negative, while r > 0 denote a positive slope. A value of 0 implies that there are no linear correlation of the variables, while |r| = 1 implies that the linear model describes the relationship perfectly. When |r| is close to 1 the variables are strongly related. The formula of the correlation coefficient is complex and from the formula it is not intuitively, why the coefficient is a measure of the strength and direction. Using vectors can help make sense to the formula. Consider the data set D as vectors x = (x1 , . . . , xn ), y = (y1 , . . . , yn ). Then the correlation coefficient is the angle ✓ between the normalized vectors. Normalize the vectors yield z = (x1 - 4.5 the quality of a model x, . . . , xn - x) and w = (y1 - y, . . . , yn - y). The angle between the normalized vectors are given by: Pn Sxy z·w (xk - x¯ )(yk - y) ¯ p pPn cos(✓) = = Pn k=1 =p q 2 2 kzk · kwk ¯) ¯ S2x S2y k=1 (xk - x k=1 (yk - y) The knowledge of vectors and linear algebra can be used to justify why the coefficient measures the strength and direction. Two vectors that are linear independent are orthogonal, i.e. r = cos(✓) = 0. Two vectors pointing roughly in the same direction are closely related and the angle between them is small. Thus cos(✓) will be about 1. The correlation coefficient can be usable if data has a linear relation, but the coefficient cannot be used blindly, as the value does not characterize the relationship. Anscombe’s quartet (figure 10) illustrates how four very different distributed data set can have equal correlation coefficient. For instance is Set 3 perfectly linear, except for one outlier, while Set 2 is non-linear. The example confirms that the correlation coefficient have to be combined with visual examination before a model can be evaluated. Figure 10: Anscombe’s quartet: Four sets of data with same correlation coefficient r = 0.816. Retrieved 18/2 2015 from http://2.bp.blogspot.com/_IFzDPHUxHI0/ SG0ocfCh01I/AAAAAAAAADI/VAqSLJd0dLc/s400/anscombe_quartet.gif 43 44 the scholar knowledge 4.5.3 The coefficient of determination The coefficient of determination measures (like the correlation coefficient) the strength of a model. Definition 4.5.3.1 (The coefficient of determination of simple linear models). The coefficient of determination r2 2 [0, 1] is given by Pn (ax + b - y) ¯ 2 2 Pn k r = k=1 ¯ 2 k=1 (yk - y) In the case of linear models, the coefficient of determination is the square of the correlation coefficient and the formula of r2 can be inferred from the correlation coefficient: ✓ ◆ 2 (Sxy )2 (Sxy )2 S2x Sxy 2 S2x 2 2 Sx r = 2 2 = = · = a · Sx Sy (S2x )2 S2y S2x S2y S2y P P 2 2 2 a2 · n a2 · n k=1 (xk + x - 2xk x) k=1 (xk - x) = = S2y S2y Pn Pn (axk - ax)2 (axk - ax + ax + b - y)2 = k=1 2 = k=1 Sy S2y Pn (ax + b - y)2 Pn k = k=1 2 k=1 (yk - y) (Winsløw, 2015) The domain [0, 1] of the coefficient of determination can be proved by using that S(a, b) > 0 and S(a, b) = S2y minimized (a result from 1st proof) (Sxy )2 r2 = 2 2 = Sx Sy (Sxy )2 S2x S2y = (Sxy )2 , S2x (Sxy )2 S2x (S )2 S(a, b) + Sxy2 x when S(a, b) is 61 (Winsløw, 2015). Generalize the formula of coefficient of determination to all classes of functions yields: Pn 2 k=1 (f(xk ) - y) r2 = P n 2 k=1 (yk - y) The sum in the numerator is called the explained sum, as it explain the squared distance between the model and the mean, while the sum in the denominator is called the total sum, as it is the sum of the distance between the observed value and the mean. Thus, the coefficient of determination can be described as the measure of the proportion of the variation in the dependent variable y, which the independent variable x can explain. If r2 = 0.8, the x-variable can explain 80% of the variation in y, while the rest 20% must be explained from other factors. A r2 value close to 1 indicate a strong relation. 4.6 non-linear regression The coefficient of determination is a useful tool to explain how certain predictions can be made, however using the quantity to choose and justify the class of functions used to model a data set requires caution. Firstly, a perfect model can be found to all data set using the class of polynomials. A model perfectly fit a data set does not imply that the type of model (the class of functions) is useful to describe the actual relation. Further it is not possible to decide the class of functions based on the r2 -values, which the Anscombe’s quartet was an example of. Consequently to choose and justify the class of functions used to model a data set, the phenomenon behind the data has to be considered. Thus the determination of the class of functions requires construction of mathematical praxeologies mixed up with non- mathematical objects. As non-mathematical objects have to be considered to evaluate the class of functions, the criteria of a good model differ and no general criteria can be set up. If the purpose is to demonstrate theoretical laws, r2 values of 0.9 - 0.95 are interpreted as small values, while values of 0.6-0.7 are high, when verifying relations in social science, as many other factors influence the relation. 4.6 non-linear regression As demonstrated earlier, it is different to use the LSQ method to make non-linear regression, because no direct method to minimize the sum exist. Because of that the determination of non-linear models are reduced to tasks, which are simpler to solve. The non-linear functions, which can be transformed to be linear in the parameters are solved by linear LSQ method on the transformed data set. Two non-linear functions, which can be transformed to be linear are the exponential and power functions and therefore the best exponential and power model is found by linearization and linear LSQ on the transformed data. Exponential regression An exponential function is linearized by use of the properties of logarithm: y = b · ax , log(y) = log(b · ax ) , log(y) = log(b) + x · log(a) If D is the original data set, the transformed data set will be { xk , log(yk ) |k = 1, . . . , n}. Linear regression on the transformed data set gives a linear model log(y) = Ax + B. Using that log(x) 45 46 the scholar knowledge and 10x are inverse function the model is transformed back to an exponential model by log(y) = Ax + B , y = (10A )x · 10B = ax · b with a = 10A and b = 10B The exponential model y = b · ax minimizes n ✓ X k=1 log(yk ) - log(b) + x · log(a) ◆2 The exponential model is given by: log(a) = Pn k=1 log(yk ) - log(y) 2 k=1 (xk - x) xk - x Pn log(b) = log(y) - log(a) · x Power regression Similarly, the best power model f 2 FP is determined by use of logarithm transformation and linear regression. The power function is transformed by: y = b · xa , log(y) = log(b · xa ) , log(y) = log(b) + a · log(x) The last expression is a linear function log(y) of log(x). The data set D is transformed to { log(xk ), log(yk ) |k = 1, . . . , n }. Linear regression on the transformed data set gives a linear model log(y) = A · log(x) + B. The linear model is transformed back to a power model by: log(y) = A · log(x) + B , y = xA · 10B = xa · b with a = A and b = 10B The power model y = b · xa minimizes n ✓ X log(yk ) - (log(b) + log(x) · a) k=1 ◆2 The power model is given by: a= Pn k=1 log(xk ) - log(x) log(yk ) - log(y) Pn 2 k=1 log(xk ) - log(x) log(b) = log(y) - a · log(x) The linearization of the functions correspond to work with straight lines on logarithmic paper, since the exponential and power functions form straight lines at respectively semi-log and log-log paper. The use of logarithm paper to make exponential and power regression is described in the next chapter. 4.6 non-linear regression The technique of linearization are further applied to other types of non-linear regression, for instance logarithm or reciprocal functions. Another type of transformation is to transform a simple non-linear function to a multiple linear function and use multiple linear regression. This transformation is applied in polynomial regression. The polynomial f(x) = a0 + a1 x + . . . + ak xk is treated as a linear function in k variable by consider x, x2 , . . . , xk as independent variables. The problem of minimize LSQ for non-linear functions are, if possible simplified to problems, which are easier to solve. However, transformation of non-linear data requires caution, since the model solved by the simpler method is not the best model in the "strict" sense, (minP 2 imize n k=1 (yk - f(xk ))P). Making exponential regression with the xk 2 LSQ method, the sum n k=1 (yk - B · A ) is minimized and this sum differs from the sum minimized by use of linearization and linear LSQ method. The two models will differ, because the points are not weighted equally, e.g. small errors are weighted greater using linearization. In the rest of the thesis, LSQ method will refer to the method used to determine the best exponential and power model. The best exponential model is defined as the function f 2 FE which minimize Pn 2 k=1 log(yk ) - (log(b) + x · log(a)) . The best exponential model is defined as the function f 2 FP which P 2 minimize n k=1 log(yk ) - (log(b) + log(x) · a) . 47 5 T H E K N O W L E D G E T O B E TA U G H T In this chapter, the knowledge to be taught about simple regression in high school will be examined. The chapter highlights the conditions and constraints crucial for teachers’ didactic processes and the mathematical praxeologies taught. The knowledge to be taught is composed of many sources, some of them are common for high schools, e.g. the curricula and exams, whereas sources as textbooks and materials differ between high schools. The common sources, the curricula and exams are analyzed to illustrate the generic (purpose, aim, didactic principle) and specific (specific task, themes) restrictions. In the analysis, old exams and curricula are included to highlight how the MOs about regression have evolved over time. Afterwards the content in three series of textbooks is presented to illustrate the MOs taught in high school. The examples and citations from the sources are translated from danish. 5.1 curricula The curricula state the reasons to teach a discipline, what should be taught, what should be learned and didactic principles. These elements set up constraints for teachers and influence what and how MOs be taught. The conditions and constraints in high school is analyzed in the curricula of STX B-level from 1999 and 2013 . The analysis focuses mainly on the curriculum from 2013, since the curriculum influences the taught knowledge nowadays. In the curriculum from 2013 the application of mathematics in real life is emphasized as being very important when teaching and learning mathematics and are also highlighted as a reason to learn mathematics. Mathematics is based on abstraction and logical thinking and comprise a lot of method to modeling and problem treatment. Mathematics is essential in many professions, in science and technology, in medicine and ecology, in economy and social sciences and as a basis for policy making [...] Mathematics as a science has evolved in a continuous interaction between application and theory building. (STXBekendtgørelsen, 2013, Section 1.1) In the curriculum the necessity and application of mathematics are clearly pointed out, as well as the theoretical part of mathematics, 49 50 the knowledge to be taught where the students should learn to "practice simple reasons and proofs" (STXBekendtgørelsen, 2013, Section 2.1) to justify the methods and mathematics used. Some aims of the teaching in mathematics is that the students "know about how mathematics can contribute to understand, formulate and solve problems within different disciplines" (STXBekendtgørelsen, 2013, Section 1.2) and further learn "the important aspects of mathematics interaction with culture, science and technology" (STXBekendtgørelsen, 2013, Section 1.2). These aims demonstrate that the application of mathematics has to be central when doing mathematics. The specific domains, sectors, themes and subjects that have to be taught can be found in the core substance. In the curricula from 1999 and 2013 regression belongs to the domain of functions. - [...] and the use of semi log paper have to be mentioned. In relation to the treatment of power functions the use of log-log paper has to be mentioned. The facilities of the calculator to do regression must be commented. (STXBekendtgørelsen, 1999, Section 11.2) - the notation f(x), the characteristics of the following functions: Linear function, polynomial, exponential-, power and logarithm functions and the characteristic properties of these functions graph, application of regression. (STXBekendtgørelsen, 2013, Section 2.2) The two extracts from the curricula indicate that the MOs about regression have evolved in time. In 1999, regression is mentioned in relation with logarithm paper, which could be an indicator of that regression is taught in relation with this theme. Nowadays logarithm paper is no longer a part of the core substance. Furthermore, it seems that the technique has change, as the calculator only has to be commented in 1999, which indicate that the students had other techniques than the instrumented to make regression. In the students’ study process of mathematical knowledge the students have to develop academic aims. The most relevant in relation to regression are: - Apply simple functions in modeling of given data, make simulation and extrapolation and reflect about the idealization and range of the models - Practice simple reasons and proofs - Demonstrate knowledge about the use of mathematics in selected areas, including knowledge of use in the treatment of a more complex problem - Use of IT-tools to solve mathematical problems 5.2 written exams (STXBekendtgørelsen, 2013, Section 2.1) Finally, it is worth mentioned how the curricula restrict the generic levels (in the level of determination) of the teacher. In the curriculum from 2013 it is clearly that the application of mathematics has to be in focus, which means that the teacher has to "consider emphasis on mathematical applications" (STXBekendtgørelsen, 2013, Section 3.1), and that "mathematics has to be mixed up with other disciplines" (STXBekendtgørelsen, 2013, Section 3.4). Furthermore, the teacher has to set up the students’ study process in a way that let the students learn mathematics in an interaction between "by hand" and "by CAS", such that "the calculator and IT become essential tools in the students acquisition of notation and problem solving" (STXBekendtgørelsen, 2013, Section 3.3). 5.2 written exams The written exam in mathematics is the final written examination that test students’ knowledge for the purpose of grading the students’ abilities in written mathematics. The exam exercises be an indicator of what students have to learn and consequently teachers make use of the exercises as guideline of what to teach1 . Since only "application of regression" is part of the curricula, the mathematical knowledge of regression is rarely examined in the oral exam. Thus the knowledge of regression examined can be found in the written exams. In the following, exam exercises and evaluations reports are analyzed and the types of tasks, methods and techniques (related to regression) appearing in the exam exercises are presented. The analysis is based on all levels of exam exercises from 1975-1999, 2009-2013 and evaluation reports from 2002-2011. 5.2.1 The type of tasks The first exercise in regression appeared in an exam set from august 1980. The exercise was remarkable because until 1985 no other exercises regarding regression were found. After 1985 exercises in regression became more frequently, but not in the same extent as today, see table 2. The lay out of the exercises from 1980-1999 and 2009-2013 are very similar (see figure 11 and 12), with a text explaining the context (phenomenon) of the data set D, a table with data and two-four tasks. In the following I have listed the types of tasks found in the exams without take into account the technique used to solve the tasks, since 1 Studies have shown that exam exercises play an important role in designing teaching (Svendsen, 2009, p. 74) 51 52 the knowledge to be taught Year Number of exercises/ Linear Exponential Power Number of exam set regression regression regression 1975 - 1979 0/30 0 0 0 1980 - 1984 1/30 0 1 0 1985 - 1989 6/33 0 4 2 1990 - 1994 5/24 0 3 2 1995 - 1999 7/38 0 5 2 2009 - 2013 37/38 7 17 13 Table 2: Exam exercises regarding regression in 1975-1999 and 2009-2013. The second column shows the ratio of exercise and number of exam set. The last three columns describe the type of regression. I cannot deduce the techniques from the exams. The types of task in the exams are: TRL : Given a data set D. Determine the best model f(x) within a given class FL = {f(x) = ax + b |a, b 2 R} TRE : Given a data set D. Determine the best model f(x) within a given class FE = {f(x) = bax |a 2 R+ , b 2 R} TRP : Given a data set D. Determine the best model f(x) within a given class FP = {f(x) = bxa |a, b 2 R} TPL : Given a data set D and a class of functions (FE or FP ). Draw the data set D in a suitable coordinate system (semi-log or log-log) TQD : Given a data set D and a class of functions Fi (FL or FE or FP ). Evaluate the model f(x) for a data set D TQP : Given a point P = (x0 , y0 ) 2 / D and a model f(x). Comment the range of the model. TXM : Given a model f(x) and a value of f at x0 . Determine x0 TXD : Given a data set D, one class of functions Fi (FL or FE or FP ) and a value of f at x0 . Determine x0 TYM : Given a model f(x) and x0 . Determine f(x0 ) TYD : Given a data set D, one class of functions Fi (FL or FE or FP ) and x0 . Determine f(x0 ) In figure 11, examples of tasks belonging to TQD , TYD , TXM are presented. In the exercise from 1980 (figure 11) the class of functions has to be examined (task TQD ) before the model is determined. The model is determined in relation to the second task (TYD ). The last task is of type TXM . Figure 12 illustrates examples of tasks of types TRL , TYM , TXM . 5.2 written exams The valuation prices of a specific danish stamp appear in the following table. Year 1972 1974 1976 1978 The valuation price in kr. 175 225 300 400 a) Explain that the price trend approximately can be describe by a function of the type f(x) = a · bx It is assumed that the price trend continues. b) What is the valuation price in 1990? c) When is a valuation at 1000 kr. obtain? Figure 11: Exercise from 1980 containing tasks of types TQD , TYD , TXM (Petersen and Vagner, 2003) The table show the assessment of the number of traffic casualty in the first half-year of the years 2007-2012 from the Roads Directorate. Years 2007 2008 2009 2010 2011 2012 Traffic casualty 195 190 161 110 107 82 In a model, is assumed that the development in the number of traffic casualty can be described by a function of the type y = ax + b where y is the number of traffic casualty in the first half-year and x is the time (measured in years after 2007). a) Use the data in the table to determine the coefficients a and b b) Use the model to determine the number of traffic casualty in the first half-year of 2013. c) Determine that year, where the number of traffic casualty in the first half-year are 50 people in accordance to the model. Source: FDM Figure 12: Exercise from 2013 (Undervisningsministeriet, 2013) containing tasks of type TRL , TYM , TXM The problem in the first three types of tasks (TRL , TRE , TRP ) is to determine the model given a class of functions, i.e. respectively linear, exponential and power regression. The listed types of tasks appear in the exam from 1980-1999 + 20092013, however not all in the whole period, since the type and frequency of tasks have changed. The types of tasks TPL , TQD , TXD , TYD occur only in the exams before 1999, whereas linear regression (TRL ) only occurs in the recent exams, see table 2. In all exercises at least one task of types TXM , TXD , TYM , TYD is in- 53 54 the knowledge to be taught cluded. The four types of tasks have in common that a variable is estimated by intra- or extrapolation. Besides prediction of values and determination of models, the exercises contain also types of tasks (TQD , TQP , TPL ) about the quality of the model. In exams from 1980-1999 the quality of the class of functions is considered in several exercise (task of type TQD , TPL ) and mostly the class of functions is evaluated before the specific model is determined (figure 13). In the exams from 2009-2013, the task of type TQP is the only task of type concerning the quality of the model. At waterworks the operating costs D (measure in 1000) by treatment of water depends on the treated volume of water x (measured in m3 /hour). The scheme shows some corresponding values of x and D. Volume of water x(m3 /hour) 30 50 80 140 200 Operating costs 55 91 143 248 350 Draw the information from the scheme in a suitable coordinate system and explain hereby that the operating costs D approximately can be described by a function of the volume of water x of the type D = bxa Figure 13: Types of tasks TPL and TQD from 1998 (Petersen and Vagner, 2003) 5.2.2 An instrumented technique is a set of rules and methods in a technological environment that is used for solving a specific type of problem. (Drijvers and Gravemeijer, 2005, p. 169) The methods and techniques In rest of the thesis three kind of techniques is used to distinguish between the way of solving a task. Graphic techniques: Indicate that a task is solved by use of graphs, for instance determine the best model by eye and ruler, make scatter plot or residual plot Algebraic techniques: Refer to solving tasks by use of algebra, for instance determine the best model by use of theorem 4.3.0.1 or calculate the quantities r, r2 by the formulae in definition 4.5.2.1 or 4.5.3.1. Instrumented techniques: Refer to solving tasks by use of tools with the intermediate calculations invisible for the students, for instance commandoes to do regression or calculation of the coefficient of determination. It is not possible to deduce the techniques from the exam, but examination of other exercises and the evaluation reports indicates the techniques used. In exams from 1980-1999 students got tasks where the formula of exponential and power functions had to be determined from straight lines on logarithm paper. It seems reasonable to infer that the technique also had been applied to make regression. This indicates that students made regression by the following technique: 5.2 written exams 1) Plot the points at logarithmic paper (semi-log or log) (Task TPL ) 2) Draw the best line by eye 3) Determine the formula of the linear model using two points 1 (a = yx22 -y -x1 , b = y1 - ax1 ) 4) Transform the linear model back to a model in the variable x and y (cf. section 4.6). Using the graphic technique the approach is geometric (section 4.2). Step 1) and 4) can be explained by the theory of logarithm, while step 2) and 3) cannot be justified by mathematics. The evaluation report supported that the technique described was the technique used in the exams from 1980-1999. In the evaluation report is written: "the students should use regression or calculate from reading of two points on a line drawn in technical paper" (Grøn, 2004). By the accessibility and development of calculators in the end of the 20th century, the first signs of the use of linear LSQ method and instrumented techniques are found in high school. The development caused that an instrumented technique was mentioned in the curricula (STXBekendtgørelsen, 1999, section 11.2), however the instrumented technique did not replace the graphic technique immediately (Grøn, 2004). Since the instrumented technique to make regression is based on the LSQ method, not only two techniques, but also two methods (LSQ and the geometric) were actually taught in a period. The two methods were both accepted until 2007 where the first exam, after the new reform in 2005, was hold. With the reform in 2005 the LSQ method became mandatory and is nowadays the only acceptable method. The analysis of the evaluation reports shows that an evolution in methods and techniques had happened in the period 1999-2005. New techniques to solve regression occur and consequently new MO was taught. The tasks were now solved by other techniques and so new practical blocks were taught. The LSQ method has been implemented in the curriculum from 2005 and exercises regarding regression have since been a stable and permanent part of the written exam. The exercises related to regression is only solved by the instrumented technique nowadays and thus the tasks can be solved without knowledge of the intermediate calculations or the method behind. 55 56 the knowledge to be taught 5.3 textbooks In this section, the content of regression in textbooks is presented. The following three points will be presented: • How is the problem described and what tasks are related to the problem • Which techniques are presented to solve the problem • How is the solution justified or proved 5.3.1 Gyldendals Gymnasiematematik The serie of Gyldendals Gymnasiematematik consists of four books, of which only the first book (c-level) describes regression (Schomacker et al., 2010). The book presents regression in the domain of functions and mainly through examples. Regression is introduced by problematize how a relation between a data set can be describe, when the class of functions is unknown. The problem is TR : Given a data set D. Determine the best model f(x) within a class of functions Fi . The technique to solve tasks of type TR is shown by five examples (all linear models). The technique presented is: Initial make scatter plot. Solve TRL at CAS. Plot the data together with the model. In the book, linear regression is described as "the method to determine the formula of the straight line that describe the location of the points best possible" (Schomacker et al., 2010, p. 54). No further description or explanation about the method or what best means is included. Besides task of type TR , TRL , tasks of type TXM , TYM are presented, but not solved. Exponential and power regression are presented and described by use of an exam exercise of respectively type TRE and TRP . The technique presented is the instrumented. In the presentation of task TR for non-linear data it is described how the phenomenon have to be considered when modeling the data. The book presents the types of tasks TR , TRL , TRE , TRP , TXM , and TYM and the instrumented technique without presenting the technological - theoretical discourse. 5.3.2 Lærebog i matematik Lærebog i matematik is a serie of four books at A level. Regression is presented in the first book (Brydensholt and Ebbesen, 2010) in the context of functions and the linear LSQ method is described and proved in the second book (Brydensholt and Ebbesen, 2011) in relation to differentiation. The first book presents the problem of how measured data has to be 5.3 textbooks modeled by approximate a class of functions (TRL , TRE , TRP ) to the data. The technique, which are instrumented, is shown through an example. In the example, the data is plotted and a figure of the data with the model are presented, but the visual examination is not described. The technique is described as: "Find the constants a and b by a statistical method, called linear regression" (Brydensholt and Ebbesen, 2010, p. 181). However, the book did not explain the statistical method or make further explanation of the method. The exponential and power regression are presented through examples of tasks of type TRE and TRP without explaining or justifying the technique. The technique to solve TRE is described as: "We use all the data in the table and determine the constants a and b by exponential regression" (Brydensholt and Ebbesen, 2010, p. 194). After presenting exponential regression the book presents that exponential functions became straight lines in semi-log paper, but no connection to exponential regression is made. The presentation of power regression and power function at log-log paper is similarly. In the second book, the method of linear LSQ is described and the best linear model is defined. The formula of the parameters in the linear model f(x) = ax + b is presented in a theorem (Brydensholt and Ebbesen, 2011, p. 106) Given the points (x1 , y1 ), (x2 , y2 ), . . . , (xN , yN ). The best straight line describing the points is the equation y = ax + b with P P P P P N xy - ( x)( y) y-a x P P a= , b= N x2 - ( x)2 N The theorem is proved by multivariable calculus. The parameters are proved by partial derivatives (like 3th proof), but the proof lacks to show that the parameters minimize the function. The authors state that the last part is omitted, since it "requires knowledge of function with more reel variables to determine the extremum" (Brydensholt and Ebbesen, 2011, p. 106). This shows that the proof in the transformation to be teachable in high school lose its power and rational, since the students’ knowledge limits what can be done. 5.3.3 Hvad er matematik? The serie of Hvad er matematik? consists of three books (A, B, C). In the serie, regression is presented at C-level in the domain of functions and at A-level in a supplementary section, named Regressionsmodeller. The C level book has a long introduction to the problem and describes what regression is used to. The book presents the problem of finding 57 58 the knowledge to be taught an ideal model that describes data and describes that the way to solve the problem is to go back (the etymology) and find the model that the data set reflect. The book describes how data points in real life are spread randomly around an ideal line (describing the phenomenon) and the problem is to determine the formula of the ideal line, called the regression line. It is explained that the technique (instrumented) is based on a method to calculate the best model and this method depends on a criterion of best model. For a linear model the regression line and the method are defined in a separate box. Definition: Regression line The line that best fit the data points is called the regression line and is obtained by making linear regression. The best is determined by the least squares method. (Grøn et al., 2014, p. 59) In the book the linear LSQ method is illustrated by figure 14 and it is described that the line that minimize the sum of the squares is the regression line. Figure 14: The illustration of the LSQ method in Hvad er matematik? - C (Grøn et al., 2014, p. 60) After linear regression and the linear LSQ method is presented, methods to evaluate models are presented and discussed. The techniques; coefficient of determination and residual plot are presented and further, it is described how other disciplines have to be considered, since "even through r2 is close to 1, [..], we cannot with certainty know that there is a genuine casual relationship" (Grøn et al., 2014, p. 60). The problem presented in the sections about exponential and power regression is similar with the linear, however the definition for best model is not specified. For instance, the best power model is described as: "Best is based as linear and exponential regression on a decision about how we measure this deviation. The measure of the deviation is thus more complicated compared to linear regression" (Grøn et al., 2014, p. 188). In the section about logarithm an exercise2 which should let the students study the connection between non-linear regression (exponen2 The exercise consists of four tasks, where the first task is of type TRE . In the second task the transformed data has to be calculated. In the third task the linear model of 5.3 textbooks tial and power) and linear regression is included. In the A level book (Grøn et al., 2013) the formula of the the best line is proved and the coefficient of determination, hypothesis test, multiple linear regression, non-linear regression is presented. The linear LSQ method is proved by simplify the sum and use the results of the simplified problem to prove the linear LSQ method. The linear LSQ method is proved by elementary algebra. The linear LSQ method is further explained using linear algebra and projection. Pn 2 Firstly, the t that minimize i=1 (xi - t) is determined (theorem 1) (Grøn et al., 2013, p. 402). Theorem 1: The mean is the number with the smallest variation n Given data {x1 , . . . , xn } the mean x¯ = x1 +...+x gives the lowest n 1 2 2 variation, Var(t) = n (x1 - t) + (x2 - t) + . . . + (xn - t)2 Theorem 1 is proved in two ways: using the vertex of second degree polynomial and by projection. The technique using projection is: Let x be the n-dimensional vector of {x1 , . . . , xn } and d the n-dimensional vector given by d = {1, 1, . . . , 1}. The sum is equal (x - t · d)2 . Since x and d are vectors, t · d is the position vector and e = x - t · d the vector that links the vectors x and t · d, see figure 15. Figure 15: Illustration of projection in What are mathematics? A (Grøn et al., 2013, p. 404) As (x - t · d)2 = e2 , the sum is minimal when the length of e is minimized, thus t is determined by minimize the length of e. The projection of x at d minimizes the length of e, hence t is determined by: xd = t · d = x·d x·d x1 + x2 . . . xn x1 + x2 + . . . xn ·d , t = 2 = 2 = = x¯ 2 2 1 +...+1 n d d (Grøn et al., 2013, p. 403-404) the transformed data are determined and in the last task the exponential model is compared with the linear model. 59 60 the knowledge to be taught After this the best proportional model y = kx is presented (theorem 2). The proof of the theorem is left to the students. Theorem 2: The best proportionality Given data {x1 , . . . , xn }. The value of k that minimize the variation of the model y = kx is k = x·y . x2 The variation is given by: 1 var(k) = n (y1 - kx1 )2 + (y2 - kx2 )2 + . . . (yn - kxn )2 The two theorems are used to prove theorem 3 (the best linear model) (Grøn et al., 2013, p. 409). Theorem 3: The best straight line For the data (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ) the best straight line y = ax + b is given by a= (y - y)(x - x) (x - x)2 = 1 n ((y1 - y)(x1 - x) + . . . + (yn - y)(xn - x)) 1 n ((x1 - x)2 + . . . + (xn - x)2 ) b = y - ax The values of a, b minimize the variation of the straight line Var(a, b) = 1 (y1 - ax1 - b)2 + . . . (yn - axn - b)2 n The proof is divided into two steps, based on respectively theorem 1 and 2. First, a is treated as a constant a0 and b is found by minimize var(a0 , b). Then, var(a, b) is minimized by substitute b into var(a, b). In summarized form the proof is: 1 (y1 - a0 x1 - b)2 + . . . + (yn - a0 xn - b)2 n 1 = (z1 - b)2 + . . . + (zn - b)2 with zi = yi - a0 xi . n Var(a0 , b) = Using theorem 1 implies b = z = y - a0 x = y - a0 x. Rewrite the straight line: y = ax + b = ax + y - ax = a(x - x) - y ✓ ◆ 1 2 2 Var(a, b) = y1 - a(x1 - x) - y + . . . + yn - a(xn - x) - y n ✓ ◆ 1 2 2 = (y1 - y) - a(x1 - x) + . . . + (yn - y) - a(xn - x) n 1 = (v1 - au1 )2 + . . . + (vn - aun )2 n with vi = yi - y¯ and ui = xi - x¯ . Using theorem 2 implies a= v·u u2 = (y-y)(x-x) . (x-x)2 (Grøn et al., 2013, p. 408-409) 5.3 textbooks 61 The algebraic proof is followed by an explanation of how the best line can be interpreted geometric using vectors. Let x = {x1 , x2 , . . . , xn }, y = {y1 , y2 , . . . , yn } and d = {1, 1, . . . , 1}. 1 1 2 Then var(a, b) = n (y - ax - bd)2 = n e with e = y - (ax + bd). e is the vector that links the vector ax + bd with y. The smallest length of e is the orthogonal, i.e. the projection of y on a normal vector to the plane defined by x and d. Because it requires knowledge to orthogonal projection in a n-dimensional space to determine e, the proof is not completed. 6 EPISTEMOLOGICAL REFERENCE MODEL In this chapter an ERM will be elaborated based on the scholar knowledge and the knowledge to be taught. The knowledge will be structured in four main questions, clearly connected and interdependent. Common to the four questions is that a data set D is given. The four questions are: Q1 : What class of functions in Fi is suitable to model the data? Q2 : Given a class of functions in Fi . What do we mean by best model f 2 Fi ? Q3 : Special case of Q2 . What do we mean by best linear model f 2 FL ? Q4 : Given a model f. What is the quality of the model? The connection between the questions is illustrated by arrows in figure 16. The figure shows how answering a question may requires a movement to another question. For instance, question Q2 may lead to Q3 and a movement back to Q2 as exponential and power regression is reduced to a problem of linear regression by linearization. Figure 16: ERM: The four questions in the ERM. At the figure the connection between the questions are illustrated by arrows Before the MOs belonging to the four questions is described, an example of the tasks and techniques to the different questions will be presented and the connection between the questions will be illustrated. 63 64 epistemological reference model Example of tasks, techniques and the connection between Q1 -Q4 Based on the number of road victims in Denmark in 1979-1999, predict the number of road victims in Denmark in 2010. The data is: Year Victims Year Victims Year Victims 1979 9267 1986 8301 1993 8427 1980 8477 1987 7357 1994 8105 1981 8546 1988 7321 1995 8523 1982 8427 1989 7266 1996 8672 1983 8105 1990 9267 1997 8301 1984 8523 1991 8477 1998 7357 1985 8672 1992 8546 1999 7321 Source: Faktalink.dk To solve the task, several techniques is used. First, the relation of the variables (Q1 ) can be investigated by use of a scatter plot (figure 17). The scatter plot shows a clearly decreasing Figure 17: The relation between year (after 1979) and number of road accident victims tendency, but whether the relation is linear or non-linear is difficult to conclude from the plot. To determine the class of functions (Q1 ), the classes of functions have to be evaluated (Q4 ), which require that the specific models f (Q2 ) are determined. The linear model f 2 FL is determined (task of type TRL ) by techniques belonging to MOs in Q3 . Using the algebraic techniques, the exponential model (task of type TRE belonging to Q2 ) is determined by linearization and linear LSQ method (technique belonging to Q3 ). The models can be evaluated (Q4 ) using the coefficient of determination, plots of data and model and residual plots. The coefficient of determination (0.94 and 0.92) shows that the mathematical relation of the data is good for both models. Examining the scatter and residual plots do either helps to reject one of the models, since the residuals 6.1 q 1 : the type of model (F i ) look similar for both models (figure 18). To determine the class of functions (Q1 ) and consequently the best (a) Linear model (b) Exponential model Figure 18: Scatter plot and residual plot of the linear and exponential model model of the data, the class of functions that is most reasonable relative to predict the phenomenon is selected. Predicting road victims by the linear model means that no road accidents will occur in the end of 2015, while a more reasonable number of 2308 road victims will be predicted at the same time using the exponential model. Thus the exponential model f 2 FE is used to make predictions. 6.1 q 1 : the type of model (F i ) The question: What class of functions in F i is suitable to model the data? is the starting point in modeling, when the class of functions F i is unknown. The task related to the question is T R : Given D. Determine best model f(x) within classes of functions F i . Answering the question requires that MOs is mixed with non - mathematical objects, since a mathematical relation does not imply causal relation. The mix of mathematical and non-mathematical objects caused that no unique solution and technique to the task exist and consequently no general MO belonging to the task can be made. Even though no unique solution exists, some classes of functions are more reasonable than others. To determine a reasonable class (or classes) of functions several techniques exist and which to use depend on the specific data set. Some techniques are: • Make plot (scatter plot, logarithm plot or residual plot) • Consider the quality of the model f 2 F i • Use of the properties of the class of functions (growth, differentiation) 65 66 epistemological reference model • Consider the purpose of the model • Consider the phenomenon behind the data One technique is to make plots, since a linear relation is easy to recognize. If data follow a linear model (at either scatter plot, logarithm paper) and the relations seems feasible with the phenomenon the data comes from, the class of functions can be determined from graphs. If it is not possible to determine the relation between the variables from graphs, the possible classes of functions can be evaluated by consider the quality of the models. This requires that one or more models have to be determined (Q2 ) and afterwards evaluated (Q4 ). Using growth or differentiation of functions can help determine the model (examples are given in the a priori analysis). Furthermore the purpose with the model has to be considered to solve Q1 , since a class of functions can describe data very well in a restricted range, but not be useful for predictions (cf. the example). Independent of the technique or techniques used the phenomenon behind the data has always to be considered. 6.2 q 2 : the best model (f 2 F i ) The question What do we mean by best model f 2 F i ? deals with the interpretation and definition of the best model given a class of functions in F i . The type of tasks related to the question are T RL , T RE and T RP (section 5.2.1). The three tasks can be solved by three kinds of techniques: • The graphic ⌧ G by plotting the data and draw the best line (plot at respectively semi-log and log-log paper for the non-linear models) • The algebraic ⌧ A using theorem of the best model (in the linear case theorem 4.3.0.1. In exponential and power regression similar theorem exist) • The instrumented ⌧ I using a specific push button technique (linear regression, exponential regression and power regression) As the technique determines the mathematical praxeology, each type of task belongs to three different mathematical praxeologies. The technological - theoretical discourse is constituted of the knowledge that explains and justifies the technique. The practical block [T RE /⌧ A RE ] is justified by the definition of the best exponential model, which is justified by linearization (the properties of logarithm) and the linear LSQ method. The practical block [T RP /⌧ A RP ] is justified by the definition of the best power model, which is justified by linearization (the properties of logarithm) and the linear LSQ method. 6.3 q 3 : the best linear model (f 2 F L ) The theoretical blocks of [T RE /⌧ IRE ] and [T RP /⌧ IRP ] is lacking, since the intermediate calculations are hidden and users do not know about the IT-software. G The theoretical blocks belonging to [T RE /⌧ G RE ] and [T RP /⌧ RP ] is justified by linearization and the imprecise definition of the best linear model (best line by eye). If the connection between the mathematical praxeologies is not learnt, the MOs developed will be punctual. Because the theoretical block of punctual MO often is ignored or implicit assumed, the definition of best non-linear model and the connection between Q2 and Q3 is not learnt. 6.3 q 3 : the best linear model (f 2 F L ) This question is a special case of Q2 and the type of tasks related to the question is T RL . As described in Q2 , three techniques to the type of tasks occurs and so three different practical blocks can be made. Since the graphic technique is explained and justified by visualization and an individual sense perception of small errors, no formal mathematics can be used to justify the technique. Hence the technologicaltheoretical discourse to [T RL /⌧ G RL ] cannot be justified by formal mathematics. Common for the algebraic and instrumented technique is the formal mathematics of the linear LSQ method, since both techniques can be justified by this method. But even though the formal mathematics for the techniques are equivalence, the technological-theoretical discourses differ, since the instrumented technique is not justified by mathematics. The algebraic technique using theorem 4.3.0.1 is based on formal mathematics and since the theorem can be proved in more domains several technological-theoretical discourses exist to the practical block [T RL / ⌧ A RL ]. Figure 19 illustrates the different technological and theoretical discourses. The definition of the best linear model constitutes the technology and the domain in which the theorem is proved constitutes the theory. The technological discourse ✓M concerns minimization and the best linear model in this discourse is defined as the model f 2 FL which Pn 2 minimize S(f) = k=1 (yk - f(xk )) . To ✓M three theoretical discourses exist, because the minimum of S(f) can be proved in the domain of elementary algebra, multivariable calculus or linear algebra. 67 68 epistemological reference model Figure 19: Illustration of techniques to task of type TRL and the technological-theoretical discourses to the techniques. If no discourse is illustrated the technological-theoretical discourse is not justified by formal mathematics The technological discourse ✓S is the statistic approach, where the best model is defined as the most probable model f 2 FL given D. In this discourse the theorem is proved by the same theoretical discourses as belonging to ✓M , however the notations differ. For instance, is the conditional expectation used to denote the projection (the proof in section 4.4.2), however the proof is based on projection of random variable, i.e. the theory of projection in the domain of linear algebra. The technological-theoretical discourses for the algebraic technique can be used to justify and explain the calculation of the best linear model in the instrumented technique. Thus one of the technologicaltheoretical discourses to [TRL , ⌧A RL ] have to be developed to let the students know the definition of the best model, which can be used to justify the instrumented technique. In the c-level textbooks only the practical blocks [TRL /⌧IRL ], [TRE /⌧IRE ], [TRP /⌧IRP ] of the MOs were presented. In the textbook Lærebog i matematik-Bind 2 the MO presented is [TRL /⌧A RL /✓M /⇥MC ], while Hvad er matematik?-A presents the MO A [TRL /⌧A RL /✓M /⇥EA ] and [TRL /⌧RL /✓M /⇥LA ]. The practical blocks from Q2 and Q3 can be collected in local and regional MOs depending on the technology-technological discourse. A A For instance, can the practical blocks [TRL /⌧A RL ], [TRE /⌧RE ] and [TRP /⌧RP ] be integrated in a local MO with technology: minimize S(f) for f 2 FL , since the transformed functions belong to FL . Further the punctual G G MOs of [TRL /⌧G RL ], [TRE /⌧RE ] and [TRP /⌧RP ] share the imprecise definition of best linear model by eye and hence the punctual MOs can be integrated in a local MO sharing the technology. 6.4 q 4 : evaluate the model 6.4 q 4 : evaluate the model This question deals with the evaluation of a model. The quality of a model is evaluated in connection with predictions, to know how certain the predictions are and in relation to Q1 to compare models in different classes of functions and justifying the class of functions. To estimate the quality of a model, the best model f 2 F i has to determined (solve Q2 or Q3 ). As described in section 4.5 several techniques to evaluate models exist and more techniques have to be combined to evaluate a model, since statistics quantities do not characterize the relationship (cf. Anscrombe’s quartet) and visual examination does not measure the quality. The grand type of tasks: Given D and a class of functions F i . Evaluate the model, can be solved by several techniques. The techniques relevant for high school are: ⌧PS : Draw the data set D and the model f(x) in a coordinate system ⌧PL : Draw the data set D in a suitable coordinate system (semi-log or log-log) ⌧PR : Make a plot of the residuals ⌧Qr : Determine the correlation coefficient ⌧Qr2 : Determine the coefficient of determination ⌧QP : Given a actual point P = (x0 , y0 ) about the phenomenon, where (x0 , y0 ) 2 / D. Calculate f(x0 ) and compare f(x0 ) and y0 to evaluate the model (cf. 5.2.1) Besides the concrete mathematical techniques to evaluate a model, the phenomenon the data comes from has to be considered, since mathematical relation does not imply causal relation. For instance, using the linear model to predict road victims (example 6), the number be negative after 2015, which is inconsistent with the reality. In the ERM proper mathematical technological-theoretical discourses first appear in the arrow from Q2 to Q3 , and subsequently in Q3 and Q4 . This is due to the fact that Q2 is justified by the theory belonging to Q3 and Q1 is mainly justified from knowledge of the phenomenon. Linearization (Q2 $ Q3 ), the linear LSQ method (Q3 ) and statistics quantities (Q4 ) advanced mathematical theory constitutes the theoretical block of the MOs. 69 Part II THE INTERNAL DIDACTIC TRANSPOSITION: THE TEACHING SEQUENCE How can we reorganize or modernize the teaching in one-dimensional regression in ways that make use of the affordance of technology while presenting regression in a more complete and satisfactory way than as a set of modeling tools? 7 2nd R E S E A R C H Q U E S T I O N In part I the relevant scholarly knowledge and the result of the transformation of the scholarly knowledge to be teachable in high school were presented. The process of transposition simplifies the knowledge and one can argue that it is over-simplified. The students learn the techniques to determine the best model in a given class of functions without actually knowing what is meant by the best model. Very few of the c-level textbooks define what is meant by the best model or include a description of the technique to find the best model, the technological discourse ✓ M to Q 3 . This imply that students get a minimum of technological-theoretical discourse, which is needed to explain and justify the praxis. If students only have a minimum of technological-theoretical discourse, the techniques, which are instrumented, appear as black boxes, with even the intermediate calculations being invisible for the students. Based on the ERM, I will design a teaching sequence introducing high school students to the technological-theoretical discourse of regression with the purpose to investigate the possibilities to work more theoretically with one-dimensional regression in high school. The teaching sequence will be observed and the data collected will be used to answer the following questions: 1. What are the possibilities and obstacles to establish the MO [T R L / ⌧ A R L / ✓ M / ⇥ E A ] about the best linear model (Q 3 in ERM)? 2. What are the possibilities and obstacles to establish a MO around how to choose a suitable class of functions (Q 1 in ERM)? 73 8 DESIGN OF THE TEACHING SEQUENCE Based on the ERM and the knowledge of the external didactic transposition, I designed a teaching sequence. This chapter will contain description of the class, the aims, the MO, the didactic principles and the considerations and choice I had in the process of making the design. Before reading this and the following chapters, I will recommend that you have look at the content and exercises of the teaching sequence in appendix A.1, A.3 and A.4. 8.1 the class The teaching was conducted in a second year class at Gladsaxe Gymnasium. In the class, the mathematical skills are very unlike with about an equal number of weak and strong students. Based on their mathematical skills their usual teacher has divided the class in two groups. The teaching was only carried out in the group, where the skills were highest, hence the teaching was taught to 15 students. The decision was made by the teacher, because the other group needed more time to exam repetition. The teaching was placed in the beginning of a repetition sequence, hence it should be natural for the students to work with regression again. The students have learned about regression at first year, so the teaching is based on that. The students have been taught in the instrumented techniques to make regression and the technique to determine the coefficient of determination. The students have briefly been introduced to the linear LSQ method, but have not worked with the technological-theoretical discourses related to Q2 and Q3 . Because the students have learnt about linear regression, the encounter with Q3 will not be a real first encounter. However, in the design it is characterized as a first encounter, since the students have not considered the initial question of the process. The students have not been taught in the questions related to Q1 . The students are taught in logarithm paper and are familiar with that exponential and power functions became straight line at respectively semi-log and log-log paper. The time frame for the teaching sequence was three modules, each of 90 minutes. The teaching was performed by the usual teacher to let the situation be as natural as possible. Using the usual teacher the psychological aspect of a new teacher was avoided, since the students 75 76 design of the teaching sequence know the teacher and reverse. Further I could observe the class. On the contrary, I could not control the didactic process in detail. To let teaching be as close to the design, I met with the teacher three times before the teaching. 8.2 mathematical organizations The design is divided into two parts. The first part (module 1 and 2) dealing with the question "What do we mean by best linear model f 2 FL ?" (Q3 ) and the second part (module 3) treating the question "What class of functions in Fi is suitable to model the data?" (Q1 ). The reason that Q3 was taught before Q1 is that answering Q1 requires that Q2 and Q3 are answered. Thus, it is natural to teach Q3 before Q1 to let the students actually know how the best model f 2 Fi is determined. Another reason was to let the students realize in the simple case of linear models, that determine the best model f 2 Fi requires a definition of best and this definition can be formulated by mathematics. This should let the students search for a definition of best class of functions Fi and let them realize that no definition of best Fi can be made. 8.2.1 MO belonging to Q3 In the first part the students should develop a MO belonging to Q3 . Due to the fact that regression belongs to the domain of functions in the curriculum and that students in high school do not know about the mathematics belonging to the statistics discourse, the technological discourse taught in the design will be ✓M . I have decided that the students should work with the theoretical discourse ✓EA , because this discourse can be taught without a previous introduction to new mathematical domains. The theoretical discourse has to be reached by work with task tM : P 2 Minimize S(a, b) = n k=1 (yk - (axk + b)) . In the design I have decided to let the students deduce the formula of a and b (theorem 4.3.0.1) in a way similar to the second proof described in the scholar knowledge, however some smaller changes in the proof are made to let the study process be simpler. The specific changes are described later. I have chose to use the second proof, since the technique to minimize the sum to a certain extent are related to the way the students usually minimize functions. This means that the problem of minimize a functions of two variables can be linked to the problem of minimize 8.2 mathematical organizations a function of one variable. This should let the students reconstruct old knowledge about optimization. The technique to solve the task has the disadvantage that the transformation of the data (centralized) make the proof longer. Furthermore the proof requires knowledge of displacement. Thus the simplification of the task is not free of charge, since new knowledge has to built, however the displacement also has advantages in relation with other MO. The displacement can explain the transformation of the x-variable1 and furthermore it can be linked to the transformation of exponential and power functions to linear functions. In contrast to the second proof, the first proof described in the scholar knowledge is much shorter, however I have decide not to use this proof, because the technique of completing the square are not simple. I suppose that the students will have trouble to justify why (y - y + x - x) should be adding. Further the technique will be difficult to link to other MOs. Didactic process to reach the theoretical level In the design, I have made an exercise, which should help the students to solve task tM . The exercise is divided into two step, which again is divided into subtasks. The first part is to rewrite the sum S(a, b) to let the term only depends on one of the variable, i.e. S(a, b) = (a) + (b). The second part is to minimize the sum by minimize each of the functions, min(S(a, b))= min( (a)) + min( (b)) . The second part is the essential part, since the problem can be linked to knowledge belonging to the students’ MOs. The reason for the division in two step is to let the students knowing the current functions of the subtasks and highlight the chain of reasoning to solve tM . The division of task tM into subtasks should help the students solve the task by themselves, since the subtasks can be solved by applying known techniques and draw on established praxeologies. Some techniques will belong to old knowledge, while others techniques have to be established in the previous teaching. In the exercise the students are guided through the different subtasks. For instance, I have chose that the students have to explain 2 why yk - (axk + b) = y2k - 2axk yk - 2byk + a2 x2k + 2abxk + b2 instead of letting them expand the expression without knowing the result. Knowing the result are supposed to help the students to expand the expression, since they know what to find. In other subtasks I have make hints that should help the students what techniques to use. To let the students’ study process of the proof be as simple as possible while keeping the accuracy, I have made some changes in relation to the second proof described in the scholar knowledge. 1 In several tasks the students have to transform the x-variable, but the students are not presented to the theory behind the transformation. 77 78 design of the teaching sequence In the second part of the proof the sums are substituted with letters to simplify the expression. This has the advantage that the expression do not contain the notation of xk , yk . In the exercise, I have chose to use the notation xk , yk instead of hk , vk in the case of centralized data, since the students normally indicate points with (x1 , y1 ), (x2 , y2 ) and describe linear functions with the symbolic expression y = ax + b. Since the abstract calculation and algebra already are difficult, I think changing the letters will only caused further confusion. I am aware that using xk and yk can cause problems when generalize the formula, as similar notation (xk , yk ) will be used to describe the centralized data and a general data set. 8.2.2 MO belonging to Q1 In module 3 the students have to consider how to model a data set D without knowing the class of functions Fi . As described in ERM no specific MO can be set up to this question. Because of time, I have restricted the classes of functions to the three classes of functions the students are familiar with. Even through the classes of functions is restricted, I have designed exercises that should let the students discover that the phenomenon of the data has to be considered to make models and that more models can be reasonable to model a data set. In the design, the exercises are designed such that the students have to solve tasks related to Q2 , Q3 , Q4 to answer Q1 . The phenomena included in the exercises are selected, such that the students should be able to relate to them. This should ensure that the phenomena of the data can be included in the solutions of the tasks. Because of the time frame, I have decided not to present the theoretical block belonging to Q2 and Q4 . The students are supposed to use the techniques belonging to the questions, but the techniques will not be justified. It is planned that the connection between Q2 and Q3 should be shortly institutionalized. 8.3 the specific aims The aim with the design is to investigate the possibilities to work more theoretically with regression in high school, especially Q1 and Q3 . The purpose is to give the students a precise description of the best linear model f 2 FL and let the students construct the theoretical discourses through the work with theorem 4.3.0.1. The first two modules should let the students develop MOs related to Q3 . In module 3 the students have to work with questions related to Q1 and Q4 . The intention is that the students learn to justify their choice of model by including the quality of the model (Q4 ) and the phe- 8.4 didactic principles nomenon. Below is listed the specific aims of the design: 1. Define the best line (the technological discourse ✓M of Q3 ) 2. Explain the chain of reasoning to the task tM : P 2 Minimize n (theoretical discourse ✓EA ) k=1 yk - (axk + b) 3. Determine Fi and explain the choice of Fi (Q1 ) 4. Demonstrate knowledge about modeling and reflect about the idealization and range of the models (Q4 ) The last aim the students have become acquainted with, since it is part of the curriculum (cf. 5.1). 8.4 didactic principles In this section three general didactic principles used in design and execution of the teaching are presented. Inquiry instruction The teaching sequence is designed such that the students through work with concrete data and numeric examples should be able to generalize from numeric and graphic examples to general abstract definition and notation. The inquiry instruction should allow the students to examine prior mathematical objects and generate assumption based on the examples. I will use this approach, because it might let the students reflect and interpret their result (Lawson, 2009). Before the students can solve task tM , the techniques to solve the task have to be part of the students praxeological equipment. In the design I have chose that the students have to develop the techniques and knowledge by working with concrete examples, that should be generalized. In contract the knowledge could have been institutionalized, but since definition and techniques are easier to apply after having experience with the use of them (Winsløw, 2013, p. 157), I have included exercises that not belong to the theme of regression. The teaching is designed such that the students have to work in a numeric or graphical settings and from these settings generalize to algebraic definition and techniques. This should entail that the students will work with the praxis before the technological-theoretical discourse is developed (Lawson, 2009). Using the inquiry instruction, the students should be able to link the new mathematical objects to discoveries, ideas and patterns that they had made and so the comprehension and retention should become bigger (Lawson, 2009). 79 80 design of the teaching sequence Method of working The teaching is designed to be a mixture between class teaching and group work. It is planned that the students have to develop techniques and construct knowledge in group work and in class teaching the knowledge should be institutionalized. A considerable part of the teaching is planned to be in groups in order to develop the mathematical objects through discussions. In the students’ first encounter with Q3 and in the work with Q1 the group work is essential, as the students should discuss the meaning of best f 2 FL and best Fi . Class teaching and class discussions are used to validate exercises and institutionalize mathematical knowledge. The teaching is planned to include the students most possible, such that the students participate effectively in almost all activities. Thus the teacher should include the students in the validations and institutionalizations. In the summaries of the exercises, it is planned that the students should be the most active, while the teacher should hold a humble position controlling the summaries and answers questions. IT-tool Pragmatic value; the efficiency for solving tasks Epistemic value; the insight the IT-tool provides into the mathematical objects and theories to be studied (Gyöngyösi et al., 2011, p. 2) In the teaching, the IT-tool will be TI-Nspire, because the students normally used that. I have planned that TI-Nspire should be an essential tools in the students acquisition of problem solving, because of its high pragmatic value when making graphs and regression. The use of TI-Nspire should make it possible for the students to focus on the principles and theory, such that the best linear model f 2 FL , the transformation and the best class of functions Fi can be explored at a higher theoretical level (Gyöngyösi et al., 2011, p. 2-3). Applying TI-Nspire, the techniques and mathematics are invisible for the students and therefore calculations have to be done by hand in exercise 3 (about sums) to ensure that the students can discover and justify the rules of sums. I have planned that the students in their first encounter with Q3 should make the best line using a graphic technique. Since it is possible to insert a line and move it in TI-Nspire, TI-Nspire is used to investigate the best line. Both the pragmatic and epistemic value of this technique is high, since the students by themselves has to reflect where the line should, but at the same time they can make several lines. Furthermore the symbolic expression is determined without the need of calculations by hand. In accordance with the curricula I have designed the teaching such that IT became an important tool in the students work with conceptualization and problem solving (STXBekendtgørelsen, 2013, section 3.3). 8.5 description of the planned didactic process 8.5 description of the planned didactic process In the following, ATD and the ERM is used to describe the didactic process in the teaching. The exercises and specific schemes of each module are included in appendix A.1, A.3 and A.4. In appendix A.3 the problem/task, the techniques, the technology/theory and the didactic moments of each episode are summaries in schemes, while the time, activity, representation, method of working and the role of the teacher and the students can be viewed in appendix A.4. 8.5.1 Module 1: Definition of best linear model and introduction to sums and centralized data The goal of the module is to let the students explore techniques to solve TRL and let the students specify the question "What do we meant by best linear model?". This should give rise to a definition of the best linear model (✓M ). Further the students should work with sums, means and displacement. According to ATD, any didactic process requires a first encounter with the MO in question and the question "What do we mean by best linear model?" should make up the first encounter with the MO. In the beginning of the module, the teacher will introduce the aim of the modules. Exercise 1 +2 The first two exercises are a attempt to let the students work with the question of Q3 by solving the exercises. The exercises will constitute the first and an explanatory moment and should motivate the work with regression in the teaching. In the teaching the students have to elaborate techniques to determine the best line based on the data sets (figure 20) and make criteria of the best line. Afterwards the students have to discuss the criteria of the line and specify what they mean by best line. The purpose of the exercises is that the students realize that the question has to be specified, since their techniques are insufficient (more best line can be found). The specification should made basis for the technological discourse of Q3 . After the students have make criteria, the criteria should be discussed in class and based on the students’ criteria the teacher should introduce the best line f 2 FL . The institutionalization should let the question be specified to the task tM . The technological discourse ✓M should give rise to the technical moment, where a technique to minimize the sum has to be developed. In the introduction to the technological discourse, the teacher should motivate the work with sums. 81 82 design of the teaching sequence (a) Scatterplot of the data(b) Scatterplot of the data set in exercise 1 set in exercise 2 Figure 20: Scatterplot of the data set in exercise 1 and 2 The following exercises (Exercise 3-5) concern sums, mean and displacement. The aim of the exercises is to let the students develop and construct the techniques that they have to apply in the technical moment of Q3 . Exercise 3 The intention with exercise 3 is to let the students deduce the rules of sums by letting the students calculate sums and compare the sums. In the exercise, the students have to elaborate techniques to calculate sums in a numeric setting and afterwards discover and explain the connection between the calculated sums. The connection should let the students develop new techniques (rules of sum) that are used to solve 3j. The teacher can start the institutionalization of the rules of sum before the students have finish the exercise, if the teacher evaluates that they do not discover the connections by themselves. In the institutionalization the results of the sums have to be filled in the table at the blackboard and the connection between the sums should be validated. The connection between the sums has to be generalized and the rules of sums institutionalized. Exercise 4 + 5 In exercise 4 the students will reencounter mean and the techniques to calculate mean. Based on the numeric calculations (task 4a + b) the students have to generalize the technique and write a definition of mean by using the symbol of summation. The exercise is constructed, since the students have to apply the definition to solve task tM . Exercise 5 concerns displacement of data. The exercise is included to let the students learn that displacement of data does not change the interconnection of the data and so not the slope of the best line. The students have not work with displacement before, however the 8.5 description of the planned didactic process knowledge should be accessible, because of the visual representation. Since the representation is graphic, the students should be able to get an intuitive idea about data as a configuration of points that can be moved. The last two tasks (5j-k) are included if some are finished before time is up. After the students have worked with the two exercises the students’ work will be validated. The definition of mean and the technique to make displacement should be institutionalized. This should be followed by definition of centralized data.The teacher should finish by proving that every data set can be centralized (task 5j). If time, the teacher makes a summary of the mathematical objects in the module: The definition of the best line, the rules of sums and centralized data. The objects are written on a slide. In the establishment of the praxeology to displace data, vectors will not be introduced, as it will be a new mathematical object. Instead I have chose to let the students work with the displacement in a graphic setting in TI-Nspire. The graphic representation should justify that changing data with a constants does not change the interconnection between the points and so not the slope of the best line. 8.5.2 Module 2: The proof of best linear model In the second module the students should produce a new reliable technique to tasks of type TRL and establish a technological-theoretical discourse that can be used to justify and validate the instrumented technique ⌧IRL . In the module the students should develop the formula in theorem 4.3.0.1 by solving task tM . The module will begin by reencounter the essential points from module 1. Then the teacher will introduce task P 2 tM : Determine aand b that minimize n k=1 yk - (axk + b) . Using figure 21 the teacher should recall the technique to centralize data (subtracting (x, y)) and emphasize that centralize data does not change the slope of the best line. The students should solve task tM by assuming x = y = 0 This work should constitute the technical moment of Q3 . The students should work in groups with the different subtasks, which should let the students develop a technique to determine the best line. As the essential step occur in the second part, the teacher has to ensure that the students do not use long time at the first part. If many problems arise in the first part the teacher can institutionalize the sub tasks (1a1c), after which the students may continue with the second part. Since the techniques should be well-known to the students, the role of the teacher is to help the students by recalling techniques and explain the algebraic notation. 83 84 design of the teaching sequence Figure 21: The graphic representation used to illustrate and justify the displacement of a data set. The figure shows how the interconnection between the points are similar. After the students have work with task tM , the teacher will go through the task and generalize the technique. The teacher will use figure 21 to show and justify how the technique can be generalized. Exercise 6 Given their is sufficient time, exercise 6 can be done after the presentation of the theorems. The exercise consists of a task of type TRL which has to be solved by the algebraic and instrumented technique. The exercise should let the students get insight into the calculations in the instrumented technique and let the students validate the instrumented technique. Introduction to best exponential and power model If time, the module can be ended with this introduction. Otherwise it is moved to module 3. The part concerning best line should be ended by introducing the best exponential and power model. The teacher has to draw parallel to the linear case and explain that the best exponential and power model can be calculated by formulas similar to the linear case. The teacher has to illustrate graphic in TI-Nspire how data can be transformed to be in line and subsequently the relation between the linear model and the exponential or power model. Homework The students have to solve exercise 7 - 9 at home. The purpose of the exercises is to let the students recall the techniques to solve the tasks, since knowing the technique should make it possible to work 8.5 description of the planned didactic process with technological-theoretical questions about Q1 in the last module. In the exercises technological questions related to both Q1 and Q4 appear. The students have to consider the range and utility of the models in tasks 7d and 9b, while task 9a (task of type TR ) should constitute the students’ first encounter with Q1 . The data set in exercise 9 is represented graphic in stead of in a table, to let the students visualize the development of the data. 8.5.3 Module 3: The type of model The intention with the module is that the students learn that no criterion exist to determine the best class of functions Fi , however some classes of functions are more useful than other and further that the phenomenon has to be considered. In the module the students have to bring old technique into play to evaluate the models and use them to justify their choice of model. The module should start with a follow-up on the homework, where two students have to present exercise 8 + 9. If the students not explain their choice of model in exercise 9, the teacher should ask about their choice. For the students that have not made their homework, this will constitute the first encounter with Q1 . Exercise 9 should be used as motivation to consider the best classes of functions in the following exercises. Exercise 10 + 11 The purpose with exercise 10 +11 is that the students should realize that it is far from simple to determine the class of functions to model data and that the phenomenon has to be taken into account when determine the class of functions. The activity are supposed to constitute the technological-theoretical and technical moment in the students didactic process, since the students both have to determine specific models f 2 Fi and work with technological-theoretical questions, as the choice of model has to be explained. Exercise 10 contains tasks of type TR , TXM , TQP . The purpose is to let the students discover that data can fit more models and that models have a restricted range. I hope that the students choose different classes of functions, such that a discussion about which classes of functions to used can be started in the group. Otherwise the teacher has to ask for the class of functions that are not investigated. In exercise 11 the students have to solve tasks of types TRL , TRP , TYM . The exercise should illustrate that the model with highest coefficient of determination is not necessary the best model. In this exercise it should be clear that it is unrealistic to use a linear model when the phenomenon is taken into account. When the students have discussed the exercises in groups, the exer- 85 86 design of the teaching sequence cises are presented in class. In the presentation of the exercises, the teacher have to draw attention to the arguments of the choice. It is important that the teacher emphasizes the real world complications and consequently that no model can describe the real life perfect. Exercise 12+13 The aim with exercise 12 is that the students learn that answering Q1 the purpose (interpolation/extrapolation) of the model has to be considered. In the exercise two different models have to be used to predict the number of victims in respectively 1995 and 2016. I have chose that the students should extrapolate the year 2016, since the number of victims is negative. This clearly show that the linear model is unrealistic in domains far from the data set. Given their is sufficient time, the students will work with exercise 13 is quite different from the previous, since the class of functions used to model the data does not belong to Fi . This exercise should constitute a moment of evaluation, since the students should realize that not all data can be modeled by the three class of functions they have at one’s disposal. The module has to be finished by final remarks on the techniques to choice the class of functions. Finally, the teacher should underline how mathematics can be used to justify the best model f 2 Fi given the class of functions, but not the best class of functions Fi . 9 A P R I O R I A N A LY S I S This chapter will contain an analysis of the exercises in the teaching. The techniques, solutions and explanation that can occur are described. The a priori analysis make up the focus points in the observation and in the a posteriori analysis. To each exercise a reference to the specific exercise (in the appendix) is included in the headline. 9.1 module 1 First encounter and the explanatory moment: Exercise 1 + 2 (A.1) In the first exercises the students have to make the best line and discuss their techniques. The reason that the lines are based on intuition and visual impression, it is difficult to make explicit the technique used. The work with the lines should let the students make criteria of best line and in that way give their first answer to the question at stake. The arguments in task 1c + 2c and the criteria of best line in task 2e can be: 1. The best line minimizes 2. The best line minimizes 3. The best line minimizes 4. The best line minimizes Pn k=1 |yk Pn - (axk + b)| yk -b a | Pn |yk -(axk -b)| k=1 1+b2 k=1 |xk Pn k=1 - yk - (axk + b) 2 5. The best line has the same number of points below and above the line 6. The best line goes through the points (xi , yi ) and (xj , yj ), where yi 6 yk 6 yj for k 2 {1, . . . , n} 7. The best line goes through the mean (x, y) 8. The best line goes through two points from the data set D 9. If we know the exact value of (0, y) the best line has to go through this point. My hypothesis is that criteria 5 and 8 will be the most frequent, as these criteria seems reasonable and furthermore can be viewed from the plots. The criteria about sums (1-4) are supposed to be rare, as they are not visible or intuitive. However, it is possible that some students will suggest the criteria, because they have been introduced to the LSQ method at first year. The institutionalization of the technological argument ✓M may give rise to questions about the criterion, for instance; "Why do we minimize the squared vertical distances or not just the numeric distance", since the criterion is not intuitive or logical. 87 88 a priori analysis The rules of sums: Exercise 3 (A.1) In this exercise the students have to develop techniques to calculate sums and afterwards generalize the techniques. The tasks (a-h) can be solved directly by summing the terms. However, which terms to sum P will not be obvious, as the symbol n k=1 only have been visited in 2 relation with -test. In the first two tasks (summing a constant), the students can get trouble, since no variable k is included. The students have to realize what the index of summation k = 1 to n means for the sum and which terms to sum. Once it has been established how to deal with the index to sum the terms, the students can apply the same technique in the following tasks. In task 3i the students have to recognize the connection between the P P sums. Some should be simple, for instance 4k=1 5k = 5 4k=1 k, as the equivalence results are side by side in the table, while the conP nection of the sum of two terms, for instance 4k=1 (2 + 5k) = 58 P would be more difficult to recognize, as two results ( 4k=1 2 = 8 and P4 k=1 5k = 50) have to be added to discover the connection. The last task (3j) can only be solved if the relations (rules of sums) are discovered, since the sum will be painstaking to solve by hand. Using the rules of sums, the technique is to split the sum up and move the constant 2 in front of the sum. The formula of the mean: Exercise 4 (A.1) The first two tasks can be solved by techniques (calculate mean) that the students already have. After this the students have to generalize the technique to set up the definition of the formula of the mean for the values (x1 , . . . , xn ). This will be far from simple and both cognitive and didactical obstacles are supposed to occur. Firstly the algebra can be an obstacles, because the numbers have to be represented by x1 , . . . , xn and the total number by n. Secondly the symbols of summation can give troubles if the knowledge about the index and summation are not established in the previous exercise. Examples Pn P 1 Pn of definitions that can occur is: n x/n and k=1 x, k=1 xn /n, P n 1 k=1 xk . n Displacement of data: Exercise 5 (A.1) In this exercise the students have to investigate the connection between displacement of data and the best line through graphical representation of the displacement and comparison of the symbolic expression of the three linear models. The techniques to solve the exercise require knowledge of the commands in TI-Nspire to subtract a constants from data set, make plots and linear regression, which are commands the students are supposed to know. After plotting data, 9.2 module 2 the students have to visualize that subtracting (x, y) from the data displaces the data (x to the left and y down) without changing the relative location of points. The activity should let the students observe that; "It is like taking the whole configuration of points and moving it together" or "The relative location of the points is similar, but the points are moved in the neighbourhood of (0, 0)". After having plotted the data and calculated the linear models it should be explicit from a comparison of the linear models that the slopes are equal, but the intersection changes. The students should be able to make arguments as; "The slope of the three models are equal, but the intersection change" or "Displacement of the data does not change the slope, as the relative location of the points are similar however the intersection changes as the points are displace vertical and horizontal". P The last two tasks (j-k) are to prove that n k=1 (xk - x) = 0 and similar Pn k=1 (yk - y) = 0. The tasks require that the techniques developed in the previous exercises (3-4) are recalled and collected, which is far from simple because of the algebraic setting. The techniques that have to be used and collected to solve the tasks are: Pn Pn Pn 1. k=1 (xk - x) = k=1 xk k=1 x. Pn 2. k=1 x = nx. The students have to realize that x represents a constant. As x normally represents a variable, it can be a hurdle to identify x as a constant. Pn 1 Pn 3. The formula of the mean. n k=1 xk = x , k=1 xk = xn 9.2 module 2 The proof (A.1) In the second module the students have to reinvest the techniques learnt in the first module as well as former knowledge to prove the theorem. Because of the algebraic setting and the representation of constants and variables, the mathematical objects and techniques have to be considered from a different angle than the students usual do. P 2 The task tM : Minimize s(a, b) = n k=1 (yk - (axk + b)) is divided into the following subtasks: Assume x = y = 0 2 1a) Explain yk - (axk + b) = y2k - 2axk yk - 2byk + a2 x2k + 2abxk + b2 1b) Use the rules of sum to split up the sum, put a and b in front of the sum and sum over b 1c) Simplify the expression by using the assumptions The expression is rewritten to a2 A - 2aB + C + nb2 2a) Explain that A, B, C are constant 2b) Divide the expression into two second degree polynomials 2c) Minimize the expression by minimize each polynomial 2d) Write the values of a and b 89 90 a priori analysis In class: Generalize by substitute xk and yk with xk - x and yk - y. Step 1 The first task (1a) can be solved using either the algebraic technique of quadratic identities1 or the algebraic technique of multiplying term by term. 2 1. yk - (axk + b) = y2k + (axk + b)2 - 2yk (axk + b)) = y2k + (ax2k + b2 + 2axk b) - 2yk (axk + b) 2. yk - (axk + b) 2 = (yk - (axk + b)) · (yk - (axk + b)) 3. yk - (axk + b) 2 = (yk - axk - b) · (yk - axk - b) The algebraic technique of quadratic identities (1.) requires that the students first recognize the two terms; yk and (axk + b) and next use the identity at (axk + b)2 . As the students normally use the quadratic identity at squared bracket with two terms, some problems can occur to identify that (axk + b) has to be treated as one term. Using the algebraic technique of multiplication term by term, the expression can be expanded directly. The students have to be careful with the signs of the terms and the notation, when entering the expression. The expanded form has nine terms, which can be simplified to six terms. In the next task (1b) the students have to reinvent the rules of sums to rewrite the expression. Firstly, the techniques n X k=1 yk ± xk = n X k=1 yk ± n X xk k=1 has to be applied to divide the expression into six sums. Then the techniques n n n X X X cxk = c xk and c = c·n k=1 k=1 k=1 have to be applied. Using the techniques require that xk and yk are treated as indexed variables, such that 2, a and b can be put outside the sums. A problem that can occur is to identify a and b as constants for the sums, since they are variables in the expression. A specific P 2 problem can be to rewrite n k=1 b . The students have to know that Pn P n 2 2 2 k=1 b = b k=1 1 to simplify the sum to nb . In the last task (1c) the students have to realize that x = 0 is equivalent P with n k=1 xk = 0 (and similar with y = 0). This requires that the students know: x=0, n n X 1 X xk = 0 , xk = 0 n k=1 k=1 1 Quadratic identities means: (a + b)2 = a2 + b2 + 2ab or (a - b)2 = a2 + b2 - 2ab 9.2 module 2 Step 2 In this step the students have to minimize the expression from step 1. The step can pose a series of problems, since the techniques and mathematical objects have to be used in another settings than usual; first of all the symbols used will be complicated since xk and yk are constants and a and b variables, totally reverse to the formula y = ax + b. Secondly, the constants a, b and c in the standard notation of second degrees polynomials are not the same as the variables a and b in the expression. Thus the a and b occurring in the formula of the vertex are not equivalence with the a and b in the expression. To minimize the obstacles the expression a2 n X x2k - 2a k=1 n X k=1 yk xk + n X y2k + nb2 k=1 is rewritten to a2 A - 2aB + C + nb2 in the task. The rewriting avoids the use of xk and yk and simplifies the expression. The relation of P Pn Pn 2 A = n k=1 xk , B = k=1 xk yk and C = k=1 yk can be observed by comparing the expressions. Because of the notation (xk and yk ) it would be challenging for the students to explain why A, B and C are constants. The students have to identify (xk , yk ) as given points (i.e. constants) to realize that they are constants and furthermore that a sum of constants are a constants. The task should let the students make explanations as; "The sums are constants, since xk , yk are given data points, and summing constants gives a constant." or "The xk and yk are constants, thus the sum of these is a constant". In the next task (2b), the students have to realize that the sum S(a, b) can be divided into two functions (a) + (b). Due to the standard notation of second degree polynomials, the use of a and b as variables and A, B, C as constants will be a cognitive obstacle for the students, since a, b and c normally represents constants. Because C is independent of the variables, the expression can be divided in two ways. 1. (a) = a2 A - 2aB + C and 2. (a) = a2 A - 2aB and (b) = nb2 (b) = C + nb2 It does not matter that two ways to divide the expression S(a, b) exist, actually it can be an advantage, because it can give occasion for a discussion, where the essential (split up the polynomial) in the task can be highlighted. This step is crucial for solve the task, because the problem of minimize the function S(a, b) is reduced to a problem of minimize functions of one variable. The division of the expression in two ways also has a mathematical quality, since the work do not appear as one unique long screed, which the students have to discover or guess. The students will be presented for that a proof can have small variations without making any difference. 91 92 a priori analysis In the following task (2c) the two polynomials have to be minimized. Firstly, the students have to realize min(S(a, b)) = min( (a) + (b)) = min( (a)) + min( (b)) The polynomials can be minimized either by use of differential calculus or by the formula of vertex. Using differential calculus the students have to solve 0 (a) = 0 and 0 (b) = 0. Using the formula of vertex -b the symbols can be an 2a obstacles, since a and b in the formula is not equivalence to the a and b in the polynomials. For instance, if (a) = a2 A - 2aB + C, the students should recognize that A is the constant belonging to the squared variable and -2B is the constant belonging to the variable of B power one and so set up -(-2B) = A . After having determined the 2A values, the students have to justify the values minimize the polynomials. The value of b should be clear. The value of a can be explained using second degree polynomials or monotony. The argument using polynomials is: P 2 Since x2 > 0, the sum A = n k=1 xk > 0. Thus the value of a minimize the polynomial. Using monotony is much more complicated, since 0 (a) = 2aA - 2B, which causes that the monotony has to be checked for both B < 0 and B > 0. The argument is: B B B If B > 0, then 12 A 0. B B 1B If B < 0, then 2 A < A < 2 A This gives B B 2 · 2A A - 2B = 4B - 2B = 2B < 0 and 2 12 A A - 2B = B - 2B = -B > 0. B Thus a = A is a minimum. Exercise 6 (A.1) In the exercise the students have to solve TRL using the technique P developed in task tM . The calculation of n k=1 (yk - y)(xk - x) and Pn 2 are hard to do by hand, but using commands in TI(x x) k=1 k Nspire the sums can be calculated and insert in the formula. The students should realize the equivalence of the instrumented and algebraic technique and consequently links the techniques. 9.3 module 3 The last module constitutes a didactic process about the class of functions (Fi ), where the students should work with Q1 (and to some extent Q4 ). The students’ expectations to what happens in mathematics can be an epistemological obstacles in the work with Q1 , since the exercises 9.3 module 3 have not one unique solution that is correct with logical necessity, which the students are accustomed to. Exercise 7 + 8 (A.1) The exercises contain tasks of types TRL , TYM , TQP , TRP and TXM . The techniques to make regression and extrapolation should be well-known to the students. However, task 7a is not simple, since a transformation of the independent variable has to be made before linear regression should be made. This transformation is not straightforward, because the years of the world records are not continuous, hence not all years from 1998-2007 are included in the table. Subtraction of (1998, 0) is the technique to transform the data set. If the students lacks the transformation, the following tasks can still be correct. If the students lacks to transform the data, this can give occasion to discuss transformation (displacement) and make connection to the previous work (ex: 5 and task tM ). Exercise 9 (A.1) The exercise constitutes the students first encounter with task TR : Determine the best model f(x) within classes of functions. At the graphic representation of the data, the absolute change in the dependent variable ( y) can be read to increase with the year. For that reason it should be clear that the data cannot follow a linear model. Whether the class of functions in Fi has to be exponential or power cannot be determined from the representation (figure 22). One technique (⌧Qr2 ) Figure 22: The representation of the data set in exercise 9 is to determine the coefficient of determination. Using exponential regression the students will get a good model (r2 = 0.998) whether they use x =year or x =year-1996. Making power regression different scenario can be found: 93 94 a priori analysis 1. Using x =year, the model will be f(x) = 0 · x491.409 , i.e. f(x) = 0. The reason is that x491.409 is enormous for x > 1, so the other parameter is round down to 0. 2. Using x = year - 1996 (Transforming data by subtracting (1996, 0)) will give an error message "Domain error. An argument must be in a specified domain", since ln(0) is undefined. 3. Using x=year-1995 (Transforming data by subtracting (1995, 0)) will give a reasonable model with r2 = 0.899 The first two scenarios can give rise to a discussion about the technique to determine f 2 FP , where the practical impact of data can be emphasized. Either the original data can be used as it is far away from the center (0, 0) or the transformed data, since ln(0) is undefined. To determine f 2 FP , the data has to be transformed to be close to the centrum (0, 0) (technique 3). Evaluating the power model, it can be found that the coefficient of determination is lower than in the exponential model, which could let the students make argument as; "The best model is the exponential (f(x) = 1.59 · 1.28x ) since the coefficient of determination is higher than the power model". Using the properties of growth of exponential function (f(x + k) = b · ax · ak ), the students can argue that the size in which the capacity changes depends on the size of the capacity. It seems reasonable that setting up windmills effect other to do it in the domain of the data set, since it describe the period (1996-2011) where many set up windmill. Thus it seems reasonable, when considering the properties of growth together with the phenomenon that an exponential model is suitable to describe the data set. Another technique to argue for the exponen0 (x) x ·ln(a) tial model is using differential equation ff(x) = b·ab·a = ln(a) , x but the students do not have the didactical conditions for using this technique. Exercise 10 (A.1) In exercise 10 the students have to solve tasks of types TR , TYM and TQP . In the first task (10a) the students can use scatter plot to reject a linear model. The following techniques can be used to investigate the best f 2 FE and g 2 FP . 1. ⌧Qr2 : Using the coefficient of determination the students will found similar values for both models 2. ⌧PS : Plotting the data and the models, both models will be found to fit the data 9.3 module 3 3. Properties of the function. f(x + k) = b · ax · ak = f(x) · ak g(x · k) = b · xa · ka = f(x) · ka The techniques can let students make arguments as: A1 "The exponential model is best, since the coefficient of determination is slightly higher" A2 "Both models can be used, as the both fit the data. The coefficient of determination is similar, so I cannot select which are best" A3 "I will chose the exponential model. It seems reasonable that the number of infant mortality decrease with the same per cent, since the contribution to minimize infant mortality will decrease when the number get smaller." In the last task (10d) the students will get (whether they use exponential or power model) that the predicted value is 7.5 lower than the actual value (predicted value is 2.4, the actual value is 9.9). This should give rise to a discussion of the model. Answering the task requires that the phenomenon of infant mortality is considered. The task should let the students set up arguments as: "The model is underestimated with 7.5, hence the model cannot be used to predict the number of infant mortality in a domain far from the original data" and "We see that the model underestimate the infant mortality by 7.5. This is due to the fact that it is assumed that the infant mortality per mile continue to decrease with about 10%, which is unrealistic. The drop in infant mortality in the period 1933-1973 was a result of development in healthcare and disease." Exercise 11 (A.1) The first three tasks can be solved by use of techniques well-known to the students. In the last task (11d) the relation of distance and time have to be discussed. To justify the class of functions different qualitative arguments, depending on the technique, can be suggested: A1 Coefficient of determination: Both models have r2 ⇡ 1 (Linear model: r2 = 0.9998, power model: r2 = 0.9993). This shows that both models fit the data well and so can be used to describe the data. A2 The average speed at marathon will be 297sek/km= 12.12 km/h using the linear model and 328.9 sek/km= 10.95 km/h using the power model. Using the data point (25, 7413) the runner’s average speed at 25km is 296.5sek/km. This means that the runner has to run with the same average 95 96 a priori analysis speed at marathon as for 25km, if the linear model is used. Using the power model the average speed at marathon will be about 1.2km/h slower than the average speed at 25km. Since the average speed decreases, when the distance increases, the power model is selected. A3 A table of time per km shows that the speed is not constant. The longer the distance, the more time used per km, i.e. the slower speed. Distance (km) 0.5 1 3 5 7 10 15 20 25 Time per km (sek/km) 196 215 259.7 272.8 284.1 286.9 289.4 295.2 296.5 Since the time per km is not constant the relation between the distance and time is not linear, hence the time at marathon can not be predicted by a linear model. A4 The maximal speed is equal (independent of the distance) using the linear model. Using a power model the time per km f(x) a-1 (assuming a > 1), increases when the distance x = b·x increases, so the maximal speed decreases. Therefore the maximal speed depends on the distance. Since a runner’s maximal speed decreases when the distance increases, the linear model is unrealistic. A5 The maximal speed for runners at a given distance decreases when the distance increases, so the power model seems most reasonable. Using the knowledge about functions and race, the students may realize that it seems inappropriate to use the linear model. The students should realize that the class of functions is determined in an interplay between the phenomenon of the data and mathematics. Exercise 12 (A.1) Like exercise 10, the students can get a good idea of how the data behaves, when making a scatter plot. Using the coefficient of determination a linear model seems reasonable (r2 = 0.947) to apply to make interpolation (task 12a). Extrapolate the linear model, the students will determine a negative number of victims. Once the students have found out that the linear model is unreasonable, the students are supposed to investigate the classes of exponential and power functions. The techniques that can be used are: 1. ⌧Qr2 : The coefficient of determination. The coefficient of determination are similar (r2 = 0.924, r2 = 0.923) 2. Properties of the functions: Both classes of functions tend to zero (limx!1 bax = 0 for 0 < a < 1 and limx!1 bxa = 0 for 9.3 module 3 a < 0) without being negative, when the independent variable increases, in accordance with the phenomenon 3. Properties of growth: Since the effort and action made to minimize road victims will decrease concurrently with the number of road victims, it is likely that the number of road victims decrease with the same per cent each year Extrapolate the two models, the students will found respectively 2262 and 2278 road victims in 2016, which will let the students do arguments as: "The number of road victims in 2016 will be about 2270. Using the exponential model the number will be 2262 and using the power the number will be 2278." or "The number will be 2262, as the exponential model fit the data best. However the power model seems also reasonable to model the data. The number only differ with 16 in the two models. " Exercise 13 (A.1) The first task (13a) will be difficult for the students, since they have not learn about logistic growth. First, when the students have classified the type of growth as logistic, logistic regression can be made. Subsequently the logistic model can be used to solve task b + c. The exercise can also be solved by techniques belonging to the students’ equipment, however the technique is not as simple as logistic regression. The technique is to divide the data into two data sets described by two different models. The number of rabbits in the first eight years (0 6 x 6 8) can be described by an exponential model, while the number of rabbits after 12 years (x > 12) is almost constant, hence can be modeled by a constant model. 97 10 A P O S T E R I O R I A N A LY S I S 10.1 methodology In this section, the methodology for data collection and the data material are briefly described. In the following section the data material will be analyzed corresponding to the focus points set up in the 2nd research question and based on the a priori analysis. The methodology for data collection and the type of data collected was based on the 2nd research question, since the data and the subsequently analysis of the data should make it possible to answer the 2nd research question. To investigate the possibilities and obstacles for the students’ theoretical work with regression, data about the students’ techniques, their explanations of the techniques and their arguments for the class of functions (Fi )chose were needed. To answer the first question in the 2nd research question, data about the students’ techniques and solutions to the subtasks in tM was collected. The techniques were used to clarify which praxeologies the students applied, while the calculations and the solutions illustrated the constrains and obstacles the students had when using the technique. Since the techniques and solutions to each exercise appear in the students’ handouts1 , the handouts were collected. In the handouts, boxes to the calculations were included and the students were told to write their solutions in these boxes. To answer the second question in the 2nd research question, the students’ techniques, considerations and arguments to determine the class of functions and their approach to tasks of type TR was collected. The arguments can be found in the students’ solutions and in their oral discussions. Due to that, each group was sound-recorded to collect data about their arguments and explanations. This gave insight into discussions, where the explanations were made and different arguments were discussed. The sound recordings further gave data about the obstacles occurring and the students’ interaction with the teacher on the different exercises. The sound recordings made up an important part of the data material. The sound recordings were supported by observations from the classroom. During the group work, I walked around between the there 1 The students were given the exercises at paper, which were collected in a handout 99 100 a posteriori analysis groups and took notes about the students’ approach to the different exercises, the obstacles occurring and noted the techniques used in TI-Nspire. During class teaching I noted how many students participated and took pictures of the blackboard. For instance, I observed which class of functions the students start modeling and which technique they used to decide the class of functions. In the observation I focused on the technique and the justification occurring. To organize the observation, a table with the following general heading was used: Activity Didactic Tasks and Technological/ Observed moment techniques theoretical elements activity The following data was collected about the teaching sequence: - Sound recordings of the group work - Sound recordings of the classroom teaching - Students’ solution to the exercises and their notes - Notes made in the teaching by observing Sound recording of three groups in three modules, each of length 95 min with a break of 5 min. After the teaching approximately 13.5 hours of sound recordings, observation notes and 14 sheet of the students’ solutions were collected. In the analysis of the data I listened to the sound recordings and transcribed the points worth discussing. The points worth discussing are for instance: an interesting explanation, a dialogue about the techniques, spontaneous citations about the study process etc. In the transcript the students have been anonymized by using letters A to E in the groups instead of their names. Similar the students are named with letters in the sound recordings from the class teaching. Therefore students A in one dialogue may not be the same as student A in another dialogue. The transcripts used in the thesis have been translated. The transcript in danish can be found in appendix A.7. Translation of the transcript posed challenges, because it was difficult to translate spoken danish to spoken english. Because of that the dialogue (and so words) presented in the thesis are literally not the students own anymore. In the translation, the meaning of the words are interpreted in such a way that it be intelligibly what the students meant. (Van Nes et al., 2010) The students’ solutions are examined and the different techniques, solutions and explanations to each task are draw together, see appendix A.6. The collection of the techniques, solutions and explanations made it possible to compare what actually happened in each exercise with what could have happen, as described in the a priori analysis. 10.2 analysis of the realized teaching sequence 10.2 analysis of the realized teaching sequence In this section the students’ didactic process as it occurred in the teaching will be analyzed. Since the activities planned took longer time than expected, the designed teaching was altered in the process. An outline of the planned and realized didactic process is sketch in figure 23. A detailed described of the realized didactic process and the figure of the outline can be found in appendix A.5. Figure 23: Outline of the planned and realized didactic process. Each string illustrates a module and each module is divided into the time spent on the activity. The color illustrates the didactic process. The numbers refer to the exercises. Figure 23 illustrates that the technical work with the techniques to the technical moment and the institutionalization constituted a greater part in the work with Q3 than planned, which was because the technical work was difficult. This caused that the students did not do all the technical work by themselves, thus the theoretical discourse of Q3 was mainly institutionalized. Focus of the analysis was on the students’ didactic process of respectively Q3 and Q1 . In the analysis the students’ solutions and quotations are italicized. In the analysis, a reference to the students’ solution and the dialogue (before translation) is made in rectangular brackets. 10.2.1 Q3 : The best linear model f 2 FL First encounter and the exploratory moment The students’ first encounter and the exploratory moment with Q3 worked smoothly. The students elaborated techniques to make the 101 102 a posteriori analysis best line and had intuitive and qualitative criteria of the line they made. It turned out that the students had difficulties explaining why the line they had made was the best, which comments as; "It is actually difficult to explain"[1] and "It is indeed very good. But I don’t know why"[1] shows. The students’ criteria of the best line turned out to be very similar. The most frequent criteria were[A.6]: • I have placed the line in the middle of all the points • I choose this line, since it hit as many points as possible. Further I put the beginning of the line at the intersection of y and x axis • I have counted the number of dots on each side of the line and they are equal The first and last criteria are related to criterion 5 in the a priori analysis, while the second criterion can be related to criterion 9 in the a priori analysis. It is encouraging that the students were on track to make mathematical criteria of the best line. The students ideas and intention of the best line, called the students’ protomathematics2 , indicate that the students had a sensible basis, which can be used to specify the question of best linear model. In the class, the students discussed the techniques they had elaborated and their solutions to the task. This led the students to adjust their line and state more exactly criteria. The following dialogue is from a discussion, where student C adjusted the line after discussing the techniques. Student B: But what are your reasons for this line? Student C: That it should hit as many points as possible Student A: Hit, so hit them Student C: Yes, like this [Student C showed what she meant by hit on the screen] Student A: So it does not matter, it just had to hit Student B: What about the one that sits in the middle Student D: There must be some which are outside Student C: I might as well have put it that way a little more down here. It would probably be more right Student A: Well, I have written that there has to be an equal number of points the same side. So in same distance from the line Student C: But. If you place it here. Then it is really far away from this one. From this point. [2] Few students were rather close to the definition of best line, as they include mean, sum or distances in their criteria. One student argued: "Since there are many points in east and west I make the line based on 2 Protomathematics is a set of skills and knowledge that a group or an individual knows and uses in life. Protomathematics offering fertile sources for mathematization without itself being mathematics. (Thomas, 1996, p. 11-13) 10.2 analysis of the realized teaching sequence what I think is the mean of the points"[A.6]. From the criterion it is not clear if the student refered to the point of gravity or if the criteria was similar to criteria 7 in the a priori analysis. Another student proposed that "All the points on the one side, their total distance has to be roughly the same as the other side."[3], perhaps an idea of summing the distances, like criteria 1-3 in the a priori analysis. Even though the criteria was loosely formulated, the inclusion of "sum" and "distances" indicate that the student had a rather strong intuition about summing a distance, and consequently useful ideas that could be used in the technological-theoretical moment. The technological-theoretical moment The discussions in the groups (when solving exercise 1+2) indicate that the students found out that the question had to be defined precisely to be answered. For instance one student said: "So you know you can not use the eye. It is [ makes disappointed sound]"[4]. The soundrecordings and the observation indicate that the students had realized the problem with using graphic techniques and the need of a more precise technique to determine the best line. It is however not possible to say if the students realized the need of a definition and a technique because of the exercises or if it is due to their previous work with regression (they know one "best" line exists). Because the students had relevant answers to the question, the specification of the question was based on these answers. The teacher institutionalized the definition of best linear model and the technology ✓M . The experimental approach to encountering linear regression seems reasonable to include to let the students construct the technological discourse (Aim 1 of the design). The technical moment The technical moment consists of developing a technique to minimize tM . As described in the a priori analysis the students had to solve task tM and reach the theoretical level ⇥EA through different subtasks. It was intended that the subtasks were based on old knowledge, in form of well-known techniques and praxeologies, but it turned out that the techniques were not to well establish. The students did not develop the algebraic techniques of sums and mean and they did not realize the connection between displacement and the symbolic expression of the best linear model. The last should have been the basis for generalizing the result of the technical work to a technique. Displacement of data In exercise 5, the students did not succeed to make a connection be- 103 104 a posteriori analysis tween displacement and the symbolic expression of the best linear model. It appeared that the students spent long time and effort to calculate the transformed data (subtract (x, y)), even though the teacher had recalled the technique. One reason, that the technological-theoretical level of displacement was not reached, was that the displacement of the data was invisible in Nspire. Nspire determines the values of the axis based on the data set, which meant that the points in all three cases were plotted the (a) Original data set (b) Vertical displacement (c) Centralized data set Figure 24: The scatterplots of the three data set. The scatterplots are similar except the axis, which did it difficult to discover the displacement. same place and only the axis were changed, see figure 24. I must admit that the problem was not something I had foreseen would be an obstacle to the students’ work, since otherwise I would have solved it (a redesign of the exercise is described in the discussion). Based on observation and the sound recordings it is clear that the students had difficulty explaining what they saw, as nothing happened with the location of the points, but only with the axis. The teacher had to help the students by asking about the coordinates before the students realized that the coordinates had changed. The following conversation shows how the teacher helped the students to state what happened. Student A: In c. Then you had to make those calculation we have made and insert them in a plot. But on our plots it is the same as, you know the same points The teacher: When you have subtracted the mean Student A: Yes, like this. We plot them The teacher: So then you plot right now Student A: The points are the same now. Is that the point? The teacher. Yes [ interruption of student B] Students B: But see the number here 10.2 analysis of the realized teaching sequence 105 The teacher: But they are not identical. What is the difference? Student B: It goes in minus Student A: Yes yes. The y and the x values The teacher: Yes, so you say it is the same. What do you mean? Student A: I just think that the points are in the same location, but Student B: Just on different, they are located to Student C: If you set it up in a different coordinate system so would they not be located the same places [8] The dialogue indicates that the students were aware that the relative location of the points was similar, but the students missed to realize that subtracting (x, y) implies that the points were moved in the neighbourhood of (0, 0) (A priori analysis ex:5). Because of time constraints, the students did not compare the symbolic expression of the three linear models and so they lack to discover that transforming the data did not change the slope of the linear model. The knowledge was institutionalized by the teacher. The obstacles described caused that the current function with the activity was not linked to questions about regression (monumentalism), hence the purpose of the exercise was missing. The students also said that the exercise was about "writing a lot of numbers in Nspire"[9] and "calculating the mean and difference" [9]. The subtasks in the proof The subtasks to solve tM turned out to be very difficult for the students. For that reason, the students only did some of the subtasks, by themselves, while the teacher institutionalized the others. In the following, the essential possibilities and obstacles that occurred in the subtasks are described. In the first module the technique and theory of summation and mean were handled. The students’ solutions and the observations show that the students elaborated techniques to calculate sums in the numeric setting. The technique to sum a constant was found to be the only obstacle. One group found the technique with help from the teacher, while the other two groups elaborated the technique by discussing the examples in the handout. Student A: Student C: Student A: Student C: Student A: Student B: Student A: Student D: I do not quite understand. What is k. k equal 1. The index This is what it starts from No, it is not Or it increases by one No. Try to look at the examples Is it not just that you have to put two up four times Yes - I also think that So it is 2 + 2 + 2 + 2 [5] In the exercise about mean, it became clear that the theory of sum- Monumentalism refer to the construction of praxeologies without knowing its current functions or why it was once built (Rodríguez et al., 2008, p. 292) 106 a posteriori analysis mation was not to well established and the students had challenges using the notation in the algebraic setting. Student B: We must make. The number five which are above the summation have to be n. And then we have to say Student A: No Student B: Yes Student A: What is written below is n Student B: Yes, it also comes to be called n, but this above will also be called n. [6] The praxeologies developed about mean and sum appeared to be insufficient when the techniques had to be used in the proof. The students were only able to apply the rule Pn Pn Pn k=1 (xk ± yk ) = k=1 xk ± k=1 yk , but not the other two rules, which the following is an example of: Student B: Can we use number one? [ The students refer to the rules of sums written on the blackboard (figure 25)] The teacher: Yes, where will you use number one? Student B: Is it not that where it says k? The teacher: k? Student C: No Student A: Also, I will use, yes number one as he said. But there were we have a constant. And this are for instance there where two stand before the letter The teacher: Yes, that is correct. Two is a constant. Are there other letters, I know it is letters, but are there other letters that are constants up here? If we try to think. What are our variables and what are our constants? Student A: Those which are raised to the second. So that is a to the second power [7] Figure 25: The rules of sum written at the black board As expected, the students could not explain why a and b could be put outside the sum. This illustrates that the students’ praxeologies were not stable to variation in the notation, since they were unable to 10.2 analysis of the realized teaching sequence use the algebraic rules of sum when the ostensive changes3 . The more stable theory, the less sensitive is a technique for variation, which indicate a lack of the technological-theoretical discourse in the students praxeologies. In short, the algebraization of the rules and the technological questions did not lead to an establishment of a fluid interrelation between the technical and technological-theoretical level of sums and mean. This implies that the students did not have the conditions to use the algebraic techniques (rules of sums, definition of mean) in the proof. 2 The expansion of yk - (axk + b) was found to be based on wellknown techniques. The students’ solutions showed that all three techniques described in the a priori analysis were used. Some students made small errors when using the techniques (figure 26), but the task (1a) was manageable. Figure 26: A student’s solution to task 1a). The technique is multiplication of the brackets. Note the student wrote axn · axn = ax2n [A.6] In addition to the problems caused by the new technologies of sum and mean, the students also had difficulties with the algebraic symbols used. In task 2a) the students had difficulty realizing that the sums were constants. It was obvious that the notation of xk and yk was in conflict with the way the students usually use the notation x and y. The representation of x and y as variables and a, b and c as constants is linked to the students’ application of them in the various mathematical context of functions (their ostensives), but here the students had to reverse the representation, which they did not manage to do. Due to the fact that no students recognized xk and yk as constants, the 3 Any praxeology is activated through the manipulation of ostensives and the evocation of non-ostensives. The ostensives are symbols, words, gestures and representations connected to the abstract mathematical concepts. For instance ax2 + bx + c is an ostensive connected to second degree polynomials. (Arzarello et al., 2008) 107 108 a posteriori analysis teacher institutionalized that the points were constants. The discussion after the teacher’s help and the students’ solutions indicate that the students realized that xk and yk represented the data and so be constants. However, no students made arguments like the arguments in the a priori analysis, since they not argued for that a sum of constants are a constant. In task 2b, the students did not master the definition of second degree polynomials in an algebraic setting, where the symbols differ from the standard notation. Student E: Can the a, b, c be the same as x. So ax2 + bx + c. Boom The teacher: Yes. What could then be the x? Student E: It could be a, b, c. Big a, b, c The teacher: The big a, b, c were what? Student B + E (in chorus): They were constants The teacher: Yes, so what should vary here? Student E: The little a, b, c Student B: The a, b have to vary [10] Firstly it is worth mentioning that the students did not understand the formulation of the task (2b). Listening to the students’ discussions, it seems that the formula; sum of two second degree polynomials caused the barrier. Two variables in the same function and furthermore that the variables were named a and b were a general problem for all students. The dialogue shows how student E first suggested to let A, B and C (the constants) be the variable and later the a, b and c (from the standard notation) be the variables. It was obvious that all the students were confused about the use of symbols and that they needed help from the teacher to identify the variables (see dialogue 11). It turned out that the students had hard time using their praxeologies about second degree polynomials when the ostensives changes. In the students’ solutions both ways to divide the expression S(a, b) can be found, thus the potential to discuss the step S(a, b) = (a) + (b) was available, but unfortunately neither the students nor the teacher did. Because the division was not discussed, the mathematical quality of the proof was missing (cf. a priori analysis). Because of time constraints, only one group solved task 2c. Based on that group and the observation made in the institutionalization, I infer that the task was manageable. During the institutionalization of task 2c), it was surprising that several students were able to participate in the discussion and that the students suggested both techniques (differential calculus, vertex) to solve the task. Using the formula of vertex, the students succeed to link the formal definition of vertex to the specific case, see figure 27 (cf. a priori analysis). In addition, one student explained that the 10.2 analysis of the realized teaching sequence values minimize the polynomials (using polynomials cf. a priori analysis), which illustrates that the students was able to link the subtask P 2 to the previous tasks, as the sum (A = n k=1 xk ) was applied. The fact that a great part of the students’ work was based on alge- Figure 27: A student’s solution to task 2c + 2d).[A.6] braic techniques, which were not sufficiently stable, it did not succeed to let the students by themselves solve task tM . For that reason a part of the technical work with the subtasks was skipped and was only institutionalized. Judging from the observation, the students only mastered to focus on the techniques in the subtasks and consequently they did not link the subtasks in a didactic process about the task tM . This caused that the students did not realize that the technique to solve tM was to split up the expression in two expressions with each one variable and minimize each of these (Aim 2 of the design). 10.2.1.1 Institutionalization and evaluation In the first two modules the institutionalization occurred continuously in the teaching and it turned out that the institutionalization constituted for a larger part than planned (figure 23) because the technological moment proved to be difficult. In relation to the MO of Q3 , the students did not reach the theoretical discourse by themselves, but only in an interplay between the technical work with the subtasks and institutionalization from the teacher. Based on the comments made after institutionalization of the theorem, it seems that at least few of the students had realized that they had developed a general technique to make linear regression and were able to establish a connection between the instrumented techniques to make linear regression and the technological-theoretical discourse [✓M /⇥EA ]. Student A: So this is not the way we have to calculate it. It is thus a proof The teacher: Fortunately, Nspire does it for you 109 110 a posteriori analysis Several students: Yes please Students A: But this is the reasons for it Student B (to student A): So now you know what Nspire does The teacher: So each time you make regression in Nspire, linear regression in Nspire, then it calculate the value of a in this way and similarly with the value of b [12] In the teaching no evaluation of the MO was made. In the redesign, examples of tasks to evaluate the MO are suggested. 10.2.2 Q1 : The type of model Fi 10.2.2.1 The first encounter with Q1 The first encounter with Q1 , turned out to be in the class, since only one student (student B) had made the homework. Student B had only made exponential regression hence the three scenarios about power regression (described in the a priori analysis) did not occur. In spite of the fact that the other students had not solved the exercise, they however questioned the model (exponential model) that student B suggested. The other students suggested that the data could also be modeled by the class of power functions. Student A: How did you know that it is an exponential model? Students B: You can see it on that [points at the figure (figure 22) in the exercise] Student C: Why not a power? Student B: No. That is how it is The teacher: Why not? Student B: It just is an exponential function Student D: But why? Student E: Can you not remember the rules for it? [Silence talk in the groups] Student E: It looked best The teacher: How could you find out if it is a power or exponential? [No students came with an answer ] The teacher: It is a good discussion you have started. It is actually what we will talk about in this module. How do we find out which model to use, when we get some numbers. [13] The students realized that they had a problem, since neither of them could justify their choice of model. This motivated the students to work with questions related to Q1 , as they were curious to found out the technique to determine the class of functions. I will not comment on the exploratory moment, as the students did not elaborate new techniques in the work with Q1 . The students developed their existing techniques and included the phenomena behind the data to solve the tasks. 10.2 analysis of the realized teaching sequence 10.2.2.2 The technical moment The technical moment related to Q1 turned out to be feasible, because the students mastered the instrumented techniques (make plots, regressions and calculate the coefficient of determination). Dealing with the technical work, it was surprising that the students evaluated the instrumented technique to make linear regression. For instance, one student discovered a limitation of the instrumented technique, when the students determined the linear model in exercise 11. The students worked in TI-Nspire (figure 28), when the following dialogue occurred. Student C: It does not make sense to me Student D: It does make sense Student C: So, I start at minus second. That does not make sense. But I can figure out how to do it Student D: Try to exchange x and y. Some time is x and Student C: No Student D: But what is it that make no sense? Student C: No no. If you zoom in. It takes the mean of it. It knows not better [When the students said it, the student referred to Nspire] Student C: Then the write determine b in the linear model. I just have to write zero because it looks like this [Silence] Student C: Yes yes, but I think about zero. I start at zero. If we think logically [Later in the discussion] Student B [ask student C]: Why do you understand that the b-values have to be negative? Student C: Logically you think it has to be zero. But you can say. You have to consider that Nspire do not know what the numbers must be [15] This illustrates that student C, during the process of determining the Figure 28: The linear model in TI-Nspire that the students discuss in the dialogue. Student C is confused about the intercept with the yaxis (0, -103.5), since it did not fit with the real life 111 112 a posteriori analysis linear model, wanted to justify the model but was confused when the phenomenon was taken into account, since the symbolic expression of the linear model did not fit the reality (the intersection should be 0). The student reflected about the technique and realized that TI-Nspire did not take the phenomenon into account. The example illustrates how the students made a connection between the model f 2 FL (Q3 ) and the evaluation of the model (Q4 ). 10.2.2.3 The technological-theoretical moment It turned out that the students were able to made qualitative arguments when choosing between a non-linear and linear class of functions, but had some trouble including the phenomenon to choose between the non-linear classes of functions (FE and FP ). In exercise 10, it turned out that the students only used the two techniques ⌧Qr2 and ⌧PS (see the a priori analysis). The students missed to use the properties of exponential and power function (technique 3 in the a priori analysis), hence they could not include the phenomenon of infant mortality in their arguments. The students’ arguments were like A1 and A2 in the a priori analysis. For instance, the students argued; From the data it is difficult to decide if it is a power or exponential function. But in this case I choose the exponential function f(x) = b · ax The observations and solutions indicate that the students were a bit confused that both models fit the data very well and wanted answers to which were best, however they realized, after discussing in the class, that more models can fit data. It is worth noting that the students were able to include the phenomenon in task 10d, which indicate that the students were able to relate to the phenomenon. Thus it is not possible to conclude, if the students did not considered the phenomenon in the first task (10a) or if they were unable to include the phenomenon in their arguments, because their knowledge about the properties of functions was insufficient. Dealing with the quality of the model in task 10d the students were able to discuss the predicted value and did rather well arguments for the restricted range of the model. For instance, one student argued: When we look at the model at our graph, then we can see that it decrease fast the years we have get. And we cannot expect that it will continue to decrease, so we have to consider the model against the reality. What realistically will happen. And it cannot continue to decrease all the time. Thus it be correct that the mortality rate has been 9.9 and not 2.4 [14] The students’ solutions to exercise 11 illustrate as well, that the students linked the data with the phenomenon and included the knowledge of run to justify their choice of model. For instance, one student argued; I will use the power function, since you can not run with the same 10.2 analysis of the realized teaching sequence speed in 42km. Even through the coefficient of determination are higher for the linear function [A.6]. Several of the arguments included the phenomenon and were similar to argument A5 in the a priori analysis. Based on the students’ solutions and the observations I infer that the students learn that the class of functions cannot be determined solely by the coefficients of determination. It was obvious that the students had no trouble to include the phenomenon in this case, since the two models look very different (very different properties). Few students concluded that none of the models could be used to model the phenomenon as it really occurs. One students explained: No, you did not move with the same speed. I will say that when running you start with faster speed and then you finished and then you finish with a spurt. So I would not actually say any of them [16]. This shows that the students realized the real world complication and that no model can describe real life perfect. The technical work and the technological-theoretical questioning result in that the students realized that no mathematical criteria of best class of functions can be made and to determine the best class of functions, the phenomenon and the purpose with the model has to be considered. In the students’ study process of Q1 , I observed that the students went forth and back between the four questions (Q1 -Q4 ) in the ERM and were able to establish an interrelation between the different praxeologies to answer Q1 . As should be clear from the analysis the students’ work with Q1 had the intended effect. The students used the evaluation of the models (Q4 ) and the phenomena to answer Q1 and links Q1 and Q4 . The students were able to determine the class of functions, but had trouble including the properties of the functions and the phenomenon to select between the non-linear models. The students rather well explain the choice between linear and non-linear models (Aim 3 of the design). Further the students were able to reflect about the range of the model (Aim 4 of the design). 113 11 DISCUSSION 11.1 redesign After analyzing the teaching, an interesting question is: What should be done different next time the design is used? In the following, suggestions to improvements of the design will be discussed. Next time the design will be used, it is important that more time is available. It was obvious that there was not enough time to establish the knowledge of sum and mean and further to solve the task tM . Another time, the students should be introduced to the sum notation and work with sums and mean before working with regression, such that the rationale of learning the techniques will be higher. In the design the praxeologies were only short visited. The praxeologies could be taught in relation to descriptive statistics. In addition, inclusion of more tasks like task 3j, where the tasks are solved by applying the rules of sums will hopefully make the technical work with task tM easier. In a redesign exercise 3 + 4 should be ommited. The knowledge of displacement turned out to be not to well established because the graphic representation failed (figure 24). Thus in a redesign, the graph window has to be fixed (figure 29). The window can be fixed by set the options of the coordinate system (the interval of the x and y variable). The exercise should be altered to include the options. For instance 5c should be: Make a plot of xk and yk - y. Adjust the coordinate system such that x 2 [-5, 20] and y 2 [-40; 280]. Explain what happens with the location of the data relative to the plot in 5a). Fixing the graph window should let the students visualize the displacement and be sufficiently to justify the connection between displacement and the symbolic expression of the line. In addition, it will be advantageous to use the knowledge of displacement to reconstruct the students praxeologies about regression. Several students know the technique to transform the independent variable when making regression (for instance from year to years after 2007) without knowing the technology. Thus, implementing the following exercise, the students should be able to justify the technique 115 116 discussion (a) Original data set (b) Vertical displacement (c) Centralized data set Figure 29: Plots of the displacement, when setting the graph window. The plots clearly show how the location of the data changes, while the mutual location are similar. of displacement. The exercise could be: The table shows the number of traffic casualty in the first half-year of the years 2007-2012 from the Roads Directorate. Year 2007 2008 2009 2010 2011 2012 Traffic casualty 195 190 161 110 107 82 Source: FDM It is assumed, that the development in the number of traffic casualty can be described by a linear model. a) Use the data of the table to determine a and b in the linear model y = ax + b, where y is the number of traffic casualty in the first half-year and x is the year b) Use the data of the table to determine a and b in the linear model y = ax + b, where y is the number of traffic casualty in the first half-year and x is years after 2007 c) Compare the models and explain the meaning of the parameter b in the two models. In the design, the students should work with task tM by themselves and this should not be changed in a redesign, since I believe the students are able to solve the subtasks, if the task and notation are throughly introduced (and the knowledge about sum and mean are established). Thus next time the task will be solved, the introduction should be altered. First, the teacher should explain the symbols and notations included in the tasks to ensure that the students are aware of the notation used. Second, the teacher has to problematize the task of minimizing the function S(a, b) and let the students discuss how to solve this problem. First after the students have discussed how to solve the problem, the technique should be presented. The intro- 11.2 the rationale for teaching more theoretically duction to the notation should help the students with the subtasks, especially 2a and 2b, where it turned out that the notation were the main obstacles. Furthermore the problematization of the task should help the students to connect the subtasks. As described in the analysis, I had not planned a moment of evaluation in the design. If a moment of evaluation should be included it could be done by recalling the students’ criteria from the first encounter and investigate to what extent these first incomplete techniques fit the general technique. Another possibility it to let the students work with concrete data, where the techniques had some limitations, for instance data set with outliers. The last change I would made is related to tasks of type TR , since it turned out the students had trouble choosing between the class of exponential and power functions. The new exercise should include data, for which the students have the conditions to include the phenomenon to make qualitative arguments for their choice. An example of such a phenomenon is a car’s braking distance. A car’s braking distance depends on the speed and can be modeled with a power model (f(x) = k · x2 ). Using a phenomenon like that, the students should be able to make qualitative arguments for their choice. Maybe the students know that double the speed quadruple the braking length, else they can found the information online. 11.2 the rationale for teaching more theoretically Investigation of the curricula showed that the students only have to learn to apply regression and for that reason it is interesting to consider the question: Why teaching more theoretically in regression? In the following, reasons to teach the technological-theoretical discourse of Q3 will be discussed. Consider the rationale to teach the technological-theoretical discourse of regression is closely related to what the purpose with mathematics in high school should be and what the curricula reasonable can and should provide. The development of CAS has in the recent years had an influence on the content of mathematics and the way mathematics is taught. A major issue that are emerged in relation with the development of CAS is the question about in which extent the students have to be presented for the theory behind the instrumented technique and what the reason to present the knowledge should be. In the teaching sequence it turned out that the students had rather well ideas of the best line and the question of "best line" seems to 117 118 discussion motivate the students to search for an answer. Based on the analysis I conclude that the technological discourse can be learned without major troubles and that the discourse can be reach at all level. Nevertheless, it can be discussed whether it is important for the students to learn the discourse, since the students can make the line without the knowledge. However, the knowledge is relevant when evaluate models and when the students treat problems in other disciplines. The a posteriori analysis also shows, how a student was able to justify the model and be aware of the limitation with the technique, because the students knew the discourse of the technique. Further it let the students reflect and consider the problem of linear regression and make the students able to justify the instrumented technique. The a posteriori analysis concerning the theoretical discourse highlights didactic restrictions coming from different levels. It turned out that the obstacles were not related to the subject-theme level of regression, but came from other subject-theme levels and higher level. Thus, teaching more theoretical in regression cannot be restricted to the subject-theme level of regression, but will require a reconstruction of other themes, for instance sums, means and transformation. For that reason it will not be simple to include the discourse in the teaching and it can be discussed whether the theory helps the students to reach the aims in the curricula. In addition, it can be difficult to motivate the work, since the students never will use the formula. Nonetheless, the work can be justified, since it let the students connect different techniques coming from different themes and make it possible to reconstruct praxeologies, for instance about optimization. Furthermore, students learn to make mathematical reasoning and the task illustrate how mathematics problems are solved by simplification. In short, rather well arguments for including both the technological and the theoretical discourse can be made, however teaching the technological discourse can be argued to be sufficient to avoid the black box effect of regression. 12 CONCLUSION The aim of the thesis was to examine the bodies of knowledge regarding regression and transform this knowledge to design a teaching sequence investigating the possibilities to work more theoretically with one-dimensional regression to avoid the black box effect. The analysis of the scholar knowledge showed that two technological discourses and three theoretical discourses exist about linear regression, but only the technological discourse about minimization and the theoretical discourse of elementary algebra are based on mathematics well known to students in high school. The other discourses require knowledge that goes beyond the mathematics presented in the curricula and consequently were not relevant in the design of the teaching sequence. Analyzing the knowledge to be taught showed that the students only is presented for a minimum of the technologicaltheoretical discourses and reading the curricula showed that the students only have to learn to apply regression. The analysis of the external didactical transposition resulted in four main questions which constitute the ERM. The ERM had a double function in the thesis; it was basis for the design of the teaching sequence and it enabled to make a priori and a posteriori analysis. In the teaching sequence two of the questions were taught: What do we mean by best linear model? What class of functions in Fi is suitable to model the data? The analysis of the teaching highlighted some interesting possibilities and obstacles in relation to teach more theoretically with regression. The teaching showed that it was possible to let the students consider the question of best linear model, and that the students had a sensible basis to reach the technological discourse. The technical work to reach the theoretical discourse turned out to be a challenge, where the didactic restrictions coming from levels beyond the subject-theme were the main obstacles. Especially the rules of sum and the algebraic notation was found to be obstacles in the students’ didactic process. The work with the question of best class of functions appeared to be based on praxeologies the students had. The students were able to made qualitative arguments when choosing between non-linear and linear classes of functions, but had some trouble including the phenomenon to choose between the non-linear classes of functions. 119 120 conclusion Discussing the rationale for teaching the technological-theoretical discourse of linear regression rather well arguments for including both the technological and the theoretical discourse can be made, however teaching the technological discourse can be argued to be sufficient to avoid the black box effect of regression. BIBLIOGRAPHY Arzarello, F., Bosch, M., Gascón, J., and Sabena, C. (2008). The ostensive dimension through the lenses of two didactic approaches. ZDM, 40(2):179–188. Barbé, J., Bosch, M., Espinoza, L., and Gascón, J. (2005). Didactic restrictions on the teacher’s practice: The case of limits of functions in spanish high schools. In Beyond the apparent banality of the mathematics classroom, pages 235–268. Springer. Bosch, M. and Gascón, J. (2006). Twenty-five years of the didactic transposition. ICMI Bulletin, 58:51–63. Bosch, M. and Gascón, J. (2014). Introduction to the anthropological theory of the didactic (atd). In Networking of Theories as a Research Practice in Mathematics Education, pages 67–83. Springer. Brydensholt, M. and Ebbesen, G. R. (2010). Lærebog i matematik: Bind 1. Systime, 1. edition. Brydensholt, M. and Ebbesen, G. R. (2011). Lærebog i matematik: Bind 2. Systime, 1. edition. Dictionary.com (2015). Regression. Retrieved 22/1 2015 at http:// dictionary.reference.com/browse/regression?s=t. Ditlevsen, S. and Sørensen, H. (2009). En introduktion til statistik, chapter 6: Lineær regression, pages 85–109. Københavns universitet: Institut for matematiske fag. Drijvers, P. and Gravemeijer, K. (2005). Computer algebra as an instrument: Examples of algebraic schemes. In The didactical challenge of symbolic calculators, pages 163–196. Springer. Garcia, F. J., Pérez, J. G., Higueras, L. R., and Casabó, M. B. (2006). Mathematical modelling as a tool for the connection of school mathematics. ZDM, 38(3):226–246. Grimmett, G. R. and Stirzaker, D. R. (2001). Probability and Random Processes, chapter 7.9: Prediction and conditional expectation, pages 343–346. Oxford: University press, 3. edition. Grøn, B. (2004). Evalueringsrapport: Matematik studentereksamen og hf. Grøn, B., Felsager, B., Bruun, B., and Lyndrup, O. (2013). Hvad er matematik? A. Lindhardt and Ringhof, 1. edition. 121 122 list of references Grøn, B., Felsager, B., Bruun, B., and Lyndrup, O. (2014). Hvad er matematik? C. Lindhardt and Ringhof, 2. edition. Gyöngyösi, E., Solovej, J. P., and Winsløw, C. (2011). Using cas based work to ease the transition from calculus to real analysis. Pytlak, M., Rowland, T., & Swoboda, E., Proceedings of CERME, 7:2002–2011. Kachapova, F. and Kachapov, I. (2010). Orthogonal projection in teaching regression and financial mathematics. Journal of Statistics Education, 18(1):1–18. Katz, V. J. (2009). A History of Mathematics: An Introduction, chapter 23: Probability and Statistics in the Nineteenth Century, pages 818–832. Addison-Wesley, 3rd edition. Key, E. S. (2005). A painless approach to least squares. College Mathematics Journal, pages 65–67. Khanacademy (2014). The least squares approximation for otherwise unsolvable equations (video). Retrieved 11/12 2014 at https://www.khanacademy.org/math/ linear-algebra/alternate_bases/orthogonal_projections/v/ linear-algebra-least-squares-approximation. Kro, A. (2003). Funksjoner af flere variable. Introduktion til Matematik. Københavns Universitet: Institut for matematiske fag. Lawson, A. E. (2009). Teaching inquiry science in middle and secondary schools, chapter 5-7, pages 81–129. Sage. Miller, S. J. (2006). The method of least squares. Mathematics Department Brown University Providence, RI, 2912:1–7. Nielsen, J. A. (2015). Regression med mindste kvadraters metode. LMFK, (2):28. Petersen, P. B. and Vagner, S. (2003). Studentereksamensopgaver i matematik 1806-2000. Matematik lærerforeningen. Rodríguez, E., Bosch, M., and Gascón, J. (2008). A networking method to compare theories: metacognition in problem solving reformulated within the anthropological theory of the didactic. ZDM, 40(2):287–301. Schomacker, G., Clausen, F., and Tolnø, J. (2010). Gyldendals Gymnasiematematik C Grundbog. Gyldendal Uddannelse, 2. edition. Siemsen, E. and Bollen, K. A. (2007). Least absolute deviation estimation in structural equation modeling. Sociological Methods & Research, 36(2):227–265. list of references STXBekendtgørelsen (1999). Bilag 23-matematik b, 1999. Retrieved 15/1 2015 at http://fc.silkeborg-gym.dk/fag/matematikweb/ fagbilag.htm. STXBekendtgørelsen (2013). Bek nr 776, bilag 36-matematik b 2013. Retrieved 15/1 2015 at https://www.retsinformation.dk/Forms/ R0710.aspx?id=152507#Bil36. Svendsen, J. (2009). Matematiklærerens forberedelse. observationer af samspillet mellem gymnasielærerens forberedelse og undervisning. Master’s thesis, Institut for Naturfagenes Didaktik. Københavns Universitet. Thomas, R. (1996). Proto-mathematics and/or real mathematics. For the learning of mathematics, pages 11–18. Undervisningsministeriet (24 maj 2013). Matematik B - Studentereksamen. Van Nes, F., Abma, T., Jonsson, H., and Deeg, D. (2010). Language differences in qualitative research: is meaning lost in translation? European journal of ageing, 7(4):313–316. Waterhouse, W. C. (1990). Gauss’s first argument for least squares. Archive for history of exact sciences, 41(1):41–52. Weisstein, E. W. (2015). Least squares fitting. MathWorld - A Wolfram Web Resource. Retrieved 23/1 2015 at http://mathworld.wolfram. com/LeastSquaresFitting.html. Winsløw, C. (2003). Semiotic and discursive variables in cas-based didactical engineering. Educational Studies in Mathematics, 52(3):271– 288. Winsløw, C. (2011). Anthropological theory of didactic phenomena: Some examples and principles of its use in the study of mathematics education. Un panorama de la TAD. An overview of ATD. CRM Documents, 10:533–551. Winsløw, C. (2013). Didaktiske elementer. En indføring i matematikkens og naturfagenes didaktik. Biofolia, 4. edition. Winsløw, C. (2015). Fodgængerversion af lineær regression. LMFK, (1):21. 123 Part III APPENDIX A APPENDIX a.1 the tasks bedste rette linje Opgave 1: Tabellen viser sammenhørende værdier af alder og længde for en population af spækhuggere. Alder 1 2 3 4 5 6 7 8 9 Cm 310 365 395 424 440 500 578 580 582 a) Lav et plot af punkterne. Kan gøres ved kommandoerne 3: Data, 9: Hurtiggraf. Klik på akserne for at tilføje de to variable. b) Indtegn den linje, du synes passer bedst til punkterne. Linjen tegnes ved hjælp af kommandoen 4: Undersøg data, 2: Tilføj flytbare linjer. Hvilken forskrift har linjen? c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at andre ville finde frem til præcis samme linje som dig, hvis de hørte din begrundelse. Opgave 2: På en skole har man undersøgt alder og højde for nogle børn. De sammenhørende værdier mellem alder og højde kan ses i tabellen. I kan hente tabellen på lectio. Alder Højde Alder Højde Alder Højde 6 108 9 150 12 148 6 123 10 132 13 143 7 123 10 143 13 169 7 138 10 150 14 148 8 128 11 144 14 167 8 139 11 156 15 148 9 122 12 141 15 177 127 128 list of references a) Lav et plot af punkterne. b) Indtegn den linje, du synes passer bedst til punkterne. Hvilken forskrift har linjen? c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at alle vil finde frem til præcis samme linje som dig, hvis de hører din begrundelse. Er dine begrundelser de samme som i forrige opgave? Hvorfor/hvorfor ikke? De næste spørgsmål skal besvares i grupperne. d) Sammenlign forskrifterne for den bedste linje i opg. 1 + 2. Har I fået den samme? Diskuter jeres overvejelser og begrundelser fra 1c) og 2c). e) Diskuter i gruppen hvad I forstår ved den bedste rette linje. Find kriterier for den bedste linje. Summer Man benytter en særlig notation til at opskrive summer komP pakt ved at bruge et specielt symbol, nemlig sumtegnet . Dette tegn er det græske bogstav sigma. For eksempel kan vi skrive summen af fire tal som 4 X xk = x1 + x2 + x3 + x4 k=1 Indekset k = 1 og 4 angiver at vi skal sætte 1, 2, 3, 4 ind på k’ plads. Mere generelt gælder at n X xk = x1 + x2 + x3 + . . . + xn k=1 hvor n er et naturligt tal. Tallet n betyder at vi skal summe op til xn . Vi får derfor n led. I det første eksempel var n = 4, så vi summede 4 led. Herunder ses eksempler på, hvordan man regner med summer: 3 X 4 = 4 + 4 + 4 = 12 k=1 5 X k=1 k = 1 + 2 + 3 + 4 + 5 = 15 list of references Opgave 3 Udregn summerne. Skriv udregningen til højre på papiret. Skriv resultatet i tabellen. P a) Udregn 4k=1 2 b) Udregn c) Udregn d) Udregn e) Udregn P4 k=1 3 P4 2 k=1 k P4 k=1 4k P4 k=1 5k f) Udregn 5 g) Udregn h) Udregn P4 k=1 2 2 P4 k=1 k P4 k=1 (2 + 5k) P4 2 k=1 (5k + k ) P4 k=1 3 P4 2 k=1 k P4 k=1 4k P4 2 k=1 5k 5 P4 k=1 k P4 k=1 (2 + 5k) P4 2 k=1 (5k + k ) i) Betragt summerne og dine resultater i tabellen. Redegør for sammenhængen mellem nogle af summerne og formuler disse mere generelt. j) Udregn når det oplyses at P100 k=1 100 X (2k2 + k) k=1 k2 = 338350 og Regneregler for summer Vi kan summe en konstant n X k=1 P100 k=1 k = 5050 (1) c = n·c Man kan sætte en konstant ude for summationstegnet n X k=1 c · xk = c n X k=1 xk (2) 129 130 list of references Man kan dele summer op n X (xk + yk ) = k=1 n X (xk - yk ) = k=1 n X k=1 n X k=1 xk + xk - n X k=1 n X yk (3) yk (4) k=1 Opgave 4 I skal nu arbejde videre med datasættet fra opgave 2 med alder og højde. a) Beregn gennemsnittet af alderen i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. b) Beregn gennemsnittet af højden i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. Symbolet x bruges til at beskrive gennemsnittet af x-værdier (x1 , x2 , . . . , xn ) i et datasæt. Tilsvarende bruges y til at beskrive gennemsnittet af y-værdier (y1 , y2 , . . . , yn ). Eksempelvis kan vi skrive gennemsnittet af 1, 2, 3, 4, 5 som 5 1+2+3+4+5 1X x= = k 5 5 k=1 c) Opstil en generel formel til at udregne gennemsnittet (x) af xværdierne (x1 , x2 , . . . , xn ) ved brug af sumtegn. d) Opstil en generel formel til at udregne gennemsnittet (y) af yværdierne (y1 , y2 , . . . , yn ) ved brug af sumtegn. Opgave 5 a) Plot datasættet fra opgave 2. Indtegn linjer, der viser gennemsnittet af x og y. Linjen, der viser gennemsnittet af x indtegnes ved 4: Undersøg data, 8: Plot værdi. Linjen, der viser gennemsnittet af y indtegnes ved 4: Undersøg data, 4: Plot funktion. Funktionen, der skal plottes, er y = y. b) Bestem den bedste rette linje for datasættet plottet i a). Skriv forskriften for linjen list of references c) Tilføj en kolonne til datasættet, hvor du udregner yk - y for hvert punkt i datasættet. Lav et plot af xk og yk - y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a). d) Bestem den bedste rette linje for datasættet plottet i c). Skriv forskriften for linjen P e) Udregn n k=1 (yk - y). f) Tilføj en ny kolonne til datasættet, hvor du udregner xk - x for hvert punkt i datasættet. Lav et plot af xk - x og yk - y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a) og c). g) Bestem den bedste rette linje for datasættet plottet i f). Skriv forskriften for linjen h) Sammenlign forskrifterne bestemt i b), d) og g). Gør rede for sammenhængen mellem forskrifterne P i) Udregn n k=1 (xk - x). j) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4c), at det altid gælder at n X (xk - x) = 0 k=1 k) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4d), at det altid gælder at n X (yk - y) = 0 k=1 Bestemme bedste rette linje for et centraliseret datasæt I skal bestemme den bedste rette linje y = ax + b for et datasæt (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ), bestående af n punkter, hvor n > 2. Vi antager at datasættet er centraliseret, dvs. at n n 1 X 1 X x= xk = 0 og y = yk = 0 n n k=1 k=1 Vi skal bestemme a og b, så kvadratsummen ◆2 n ✓ X yk - axk + b k=1 131 132 list of references bliver mindst mulig. Trin 1) Det første trin består i at omskrive summen, så at vi kan minimere denne i trin 2. 2 a) Gør rede for at yk - (axk + b) = y2k - 2axk yk - 2byk + a2 x2k + 2abxk + b2 P P 2 2 b) Vi har at n = n k=1 yk - (axk + b) k=1 yk - 2axk yk - 2byk + a2 x2k + 2abxk + b2 . Brug regnereglerne for summer (1-4) til at dele summen op, sætte a, b foran summationen og summe over b. c) Brug antagelsen om at gennemsnittet af x og y er 0 til at simplificere udtrykket fra b). Trin 2) I trin 1) fandt vi at ◆2 n ✓ n n n X X X X yk - axk + b = a2 x2k - 2a xk yk + yk + nb2 k=1 k=1 k=1 k=1 Vi skal nu finde de værdier for a og b, som minimere kvadratsummen. Summen n n n X X X 2 2 a xk - 2a xk yk + yk + nb2 k=1 k=1 k=1 kan vi skrive som a2 A - 2aB + C + nb2 a) Gør rede for hvad A, B, C svarer til i summen og forklar hvorfor A, B, C er konstanter. I udtrykket er A, B, C konstanter, som ikke afhænger af de to variable a, b. b) Udtrykket a2 A - 2aB + C + nb2 kan betragtes som en sum af to andengradspolynomier med hhv. a og b som variable. Skriv de to andengradspolynomier som udtrykket består af. c) Bestem a og b, så a2 A - 2aB + C + nb2 bliver mindst mulig. Brug fra b) at udtrykket kan betragtes som to andengradspolynomier, som hver skal minimeres. d) Opskriv værdierne for a og b, som du fandt i c). Disse værdier gør P 2 kvadratsummen n mindst mulig. k=1 yk - axk + b Opgave 6 Betragt datasættet fra opgave 2 med alder og højde. list of references a) Brug formlen for a og b til at bestemme den bedste rette linje. Benyt dine udregninger fra i går, hvor du udregnede yk - y, xk - x for alle punkter. b) Bestem forskriften for datasættet i TI-Nspire. Er forskriften ens? Opgave 7: Verdensrekorder Tabellen nedenfor viser tiderne for verdensrekorderne i maratonløb (målt i sekunder) for mænd i perioden 1998-2007. År 1998 1999 2002 2003 2007 Tid (sek) 7565 7542 7538 7495 7466 Udviklingen af verdensrekorderne kan beskrives ved en lineær model f(x) = ax + b, hvor x angiver antal år efter 1998 og f(x) verdensrekorden i sekunder. a) Bestem a og b i modellen b) Hvad kan vi forvente at verdensrekorden bliver i 2014 ifølge modellen? c) Verdensrekorden blev i 2014 slået, så den nu er på 7377. Hvordan passer det med b)? Benyt oplysningen til at kommentere på modellens rækkevidde. Opgave 8: Dykkersyge Hvis en dykker opholder sig et (længere) stykke tid på en vis dybde, skal han stige langsomt op til overfladen igen for at forhindre dykkersyge, dvs. at der opstår livsfarlige bobler af kuldioxyd i blodet. Der er en sammenhæng mellem dybde og det antal minutter, som dykkerne kan opholde sig på den pågældende dybde uden at få dykkersyge. Den kan ses i følgende tabel x (m) 10 12 14 16 18 20 22 25 ... 40 42 y (min) 219 147 98 72 56 45 37 29 ... 9 8 I en model antages det, at dybden og tiden kan beskrives med en potens funktion. a) Opstil en model, der beskriver sammenhængen mellem dybde og antal minutter. Bestem forskriften for modellen. b) Benyt modellen til at udregne hvor længde dykkeren kan opholde sig på en dybde på 30m. 133 134 list of references c) Benyt modellen til at udregne den dybde, dykkeren kan dykke, hvis han gerne vil dykke så dybt som muligt og opholde sig på dybden i 20min. Opgave 9: Vindmølleenergi Figuren herunder viser udviklingen på verdensplan i kapaciteten af vindmølleenergi. a) Opstil en model, der beskriver sammenhængen mellem årstal og kapaciteten af vindmølleenergi. Bestem forskriften for modellen. b) Benyt modellen til at beregne kapaciteten af vindmølleenergi i 2014 og sammenlign med den faktiske kapacitet af vindmølleenergi i 2014, som var 369.6 MW. Opgave 10: Spædbørnsdødelighed Spædbørnsdødeligheden angiver, hvor mange promille af de levendefødte, der dør inden for det første leveår. Tabellen nedenfor er hentet fra Danmarks Statistik og den viser udviklingen i spædbørnsdødeligheden i perioden 1933-1973. Årstal 1933 1938 1943 1948 1953 1958 1963 1968 1973 Spædbørnsdødelighed 71.4 60.0 48.4 37.8 27.4 23.0 19.6 15.4 11.5 a) Lav et plot af datasættet. Indfør passende variable og opstil en model, der beskriver sammenhængen mellem årstal og spædbørnsdødeligheden i perioden 1933-1973. Forklar dit valg af model. b) Bestem parameterne i modellen fra a). list of references c) Forudsig på baggrund af modellen hvor stor en promille af de levendefødte, der dør inden for det første leveår i 2008. d) Det oplyses at spædbørnsdødeligheden i 2008 var 9.9 promille. Hvordan passer det med modellen? Benyt oplysningen til at kommentere modellens rækkevidde. Opgave 11: Løb Tabellen viser sammenhørende værdier af løbedistance og den tid, der tager en kondiløber at løbe distancen hurtigst muligt. Løbedistance (km) 0.5 1 3 5 7 10 15 20 25 Løberens tid på distancen (sek) 98 215 779 1364 1989 2869 4341 5904 7413 På baggrund af disse data vil kondiløberen gerne udregner hvor lang tid det vil tage at løbe et maraton (42.195km). Nogle eksperter mener at kondiløberens tid som funktion af distancen kan beskrives ved en lineær model f(x) = ax + b, hvor x er distancen (målt i km) og f(x) er kondiløberens tid (målt i sekunder). a) Bestem a, b i den lineære model og brug denne model til at udregne hvor lang tid kondiløberen vil være om at løbe et maraton. Andre eksperter mener at kondiløberens tid som funktion af distancen skal beskrives ved en potens model f(x) = b · xa , hvor x er distancen (målt i km) og f(x) er kondiløberens tid (målt i sekunder). b) Bestem a, b i potens modellen og brug denne model til at udregne hvor lang tid kondiløberen vil være om at løbe et maraton. c) Er kondiløberens tid på et maraton den samme ved de to modeller? d) Hvilken model ville du bruge til at bestemme kondiløberens tid? Skriv begrundelserne ned. Diskuter dine begrundelser med gruppen Opgave 12: Trafikulykker Hvert år opgøres antallet af trafikulykker i Danmark. Antallet af trafikulykker i Danmark i årene 1979-1999 kan ses i tabellerne herunder. I 1995 blev antallet af trafikulykker ikke opgjort, derfor mangler der data for dette år. Datasættet ligger på lectio. 135 136 list of references År Antal ulykker År Antal ulykker År Antal ulykker 1979 9267 1986 8301 1993 8427 1980 8477 1987 7357 1994 8105 1981 8546 1988 7321 1995 ... 1982 8427 1989 7266 1996 8672 1983 8105 1990 9267 1997 8301 1984 8523 1991 8477 1998 7357 1985 8672 1992 8546 1999 7321 a) Bestem på baggrund af tabellen hvor mange trafikulykker der ca. har været i 1995. Skriv din model, dine overvejelser om valg af model samt dine udregninger. b) Forudsig på baggrund af tabellen antallet af trafikulykker i 2016. Skriv din model, dine overvejelser om valg af model samt dine udregninger. Opgave 13: Kaniner på en øde ø På en øde ø uden kaniner blev der sat 50 vilde kaniner ud 1. januar 2000. I de efterfølgende år talte man hvert år antallet af kaniner på øen d. 1. januar. Tabellen herunder viser hvordan antallet af vilde kaniner på øen vokser med tiden, målt i år efter 2000. Datasættet ligger på lectio. År Antal kaniner År Antal kaniner 1 100 8 17030 2 240 9 26010 3 450 10 30070 4 1000 11 32500 5 2000 12 33810 6 4100 13 33890 7 9050 14 33870 a) Hvilken sammenhæng er der mellem tiden og antal kaniner. b) Bestem ud fra data i tabellen hvor mange kaniner der har været på øen 1. juli 2006. Skriv din model, udregninger og overvejelser. c) Bestem ud fra data i tabellen hvor mange kaniner der vil være på øen 1. januar 2016 Skriv din model, udregninger og overvejelser. list of references a.2 the tasks with solutions 137 Bedste rette linje Opgave 1: Tabellen viser sammenhørende værdier af alder og længde for en population af spækhuggere. Alder 1 2 3 4 5 6 7 8 9 Cm 310 365 395 424 440 500 578 580 582 a) Lav et plot af punkterne. Kan gøres ved kommandoerne 3: Data, 9: Hurtiggraf. Klik på akserne for at tilføje de to variable. b) Indtegn den linje, du synes passer bedst til punkterne. Linjen ved hjælp af kommandoen 4: Undersøg data, 2: Tilføj flytbare linjer. Hvilken forskrift har linjen? Mange muligheder for svar. Min linje: f (x) = 37.1x + 281 c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at andre ville finde frem til præcis samme linje som dig, hvis de hørte din begrundelse. Forskellige begrundelser afhængigt af hvordan eleverne har lagt linjen. Beskrevet i opgave 2e) Opgave 2: På en skole har man undersøgt alder og højde for nogle børn. De sammenhørende værdier mellem alder og højde kan ses i tabellen. I kan hente tabellen på lectio. Alder 6 6 7 7 8 8 9 Højde 108 123 123 138 128 139 122 Alder 9 10 10 10 11 11 12 Højde 150 132 143 150 144 156 141 Alder 12 13 13 14 14 15 15 Højde 148 143 169 148 167 148 177 a) Lav et plot af punkterne. b) Indtegn den linje, du synes passer bedst til punkterne. Hvilken forskrift har linjen? Mange muligheder for svar. Min linje: f (x) = 4.76x + 92 c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at alle vil finde frem til præcis samme linje som dig, hvis de hører din begrundelse. Er dine begrundelser de samme som i forrige opgave? Hvorfor/hvorfor ikke? Forskellige begrundelser afhængigt af hvordan eleverne har lagt linjen. Beskrevet i opgave 2e) De næste spørgsmål skal besvares i grupperne. d) Sammenlign forskrifterne for den bedste linje i opg. 1 + 2. Har I fået den samme? Forklar hver af jeres overvejelser og begrundelser fra 1c) og 2c).ă Forskrifterne vil være forskelligt fra gruppe til gruppe. Fælles for grupperne vil være at de har fundet forskellige forskrifter, som de skal begrunde. e) Diskuter hvad I forstår ved den bedste rette linje. Find kriterier for den bedste linje. P • Minimere den horisontale afstand nk=1 (xk x) P • Minimere den vertikale afstand nk=1 (yk y) • Minimere den vinkelrette afstand • Minimere den vertikale afstanden numerisk Pn |yk Pn k=1 • Minimere kvadratet af de ovenstående afstande • Lige mange punkter på hver side af linjen y| k=1 (yk y)2 • Skal gå igennem den mindste og største værdi • Skal gå igennem gennemsnittet af værdierne • Skal gå igennem to punkter • Fastlægge skæringen (b-værdien) hvis vi kender denne Summer Man benytter en særlig notation til Pat opskrive summer kompakt ved at bruge et specielt symbol, nemlig sumtegnet . Dette tegn er det græske bogstav sigma. For eksempel kan vi skrive summen af fire tal som 4 X xk = x1 + x2 + x3 + x4 k=1 Indekset k = 1 og 4 angiver at vi skal sætte 1, 2, 3, 4 ind på k’ plads. Mere generelt gælder at n X xk = x1 + x2 + x3 + . . . + xn k=1 hvor n er et naturligt tal. Tallet n betyder at vi skal summe op til xn . Vi får derfor n led. I det første eksempel var n = 4, så vi summede 4 led. Herunder ses eksempler på, hvordan man regner med summer: 3 X 4 = 4 + 4 + 4 = 12 k=1 5 X k=1 k = 1 + 2 + 3 + 4 + 5 = 15 Opgave 3 Udregn summerne. Skriv udregningen til højre på papiret. Skriv resultatet i tabellen. a) Udregn b) Udregn 4 X 4 X 2 k=1 4 X k=1 4 X 3 k=1 c) Udregn d) Udregn e) Udregn 4 X k2 k=1 4 X g) Udregn h) Udregn 4 X 4k 2 k=1 4 X 5k 4 X k 8 k=1 2 P4 12 k=1 5k = 5 · 1 + 5 · 2 + 5 · 3 + 5 · 4 = 50 5 k=1 k=1 4 X (2 + 5k) k=1 4 X 4k 2 = 4 · 12 + 4 · 22 + 4 · 32 + 4 · 42 = 120 k=1 4 X k = 5(1 + 2 + 3 + 4) = 5 · 10 = 50 (2 + 5k) = (2 + 5) + (2 + 10) + (2 + 15) + (2 + 20) = 58 (5k + k 2 ) k=1 P4 k 2 = 12 + 22 + 32 + 42 = 30 k=1 4 X k=1 f ) Udregn 5 · 3 = 3 + 3 + 3 + 3 = 12 k=1 4 X k=1 4 X 2=2+2+2+2=8 k=1 4 X (5k + k 2 ) = (5 + 12 ) + (10 + 22 ) + (15 + 32 ) + (20 + 42 ) = 80 k=1 3 P4 30 k=1 k2 P4 k=1 120 4k 2 P4 50 k=1 5k P 5 4k=1 k 50 P4 58 k=1 (2 + 5k) P4 80 k=1 (5k + k2) i) Betragt summerne og dine resultater i tabellen. Redegør for sammenhængen mellem nogle af summerne. Diskuter i grupperne og forklar sammenhængene. 4 X a) k=1 4 X b) k=1 n X c) k=1 n X d) e) k=1 4 X f) k=1 n X g) k=1 4 X (a + e) h) k=1 4 X (c + e) k=1 n X 2=2·4 3=3·4 c=n·c 2 4k = 4 · 5k = 5 4 X k2 k=1 4 X k k=1 n X cxk = c xk k=1 (2 + 5k) = 2 4 X 2+ k=1 4 X (5k + k ) = (xk + yk ) = k=1 k=1 n X k=1 4 X 5k k=1 5k + xk + 4 X k=1 n X k2 yk k=1 j) Udregn når det oplyses at 100 X k=1 2 P100 k=1 100 X (2k 2 + k) k=1 k 2 = 338350 og (2k + k) = 100 X 2 2k + k=1 100 X =2 k=1 k2 + 100 X k=1 100 X k=1 P100 k=1 k = 5050 k k = 2 · 338350 + 5050 = 681750 Regnereglerne for summer er: Vi kan summe en konstant n X k=1 (1) c=n·c Man kan sætte en konstant ude for summationstegnet n n X X c · xk = c xk k=1 (2) k=1 Man kan dele summer op. n X k=1 n X k=1 (xk + yk ) = (xk yk ) = n X k=1 n X k=1 xk + xk n X k=1 n X yk (3) yk (4) k=1 Opgave 4 I skal nu arbejde videre med datasættet fra opgave 2 med alder og højde. a) Beregn gennemsnittet af alderen i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. P Bruger sum i TI-Nspire. 21 k=1 xk = 220 Kan også bruge =mean(alder). x = 220 = 10.4762 21 Dette betyder at børnene, som er blevet undersøgt, i gennemsnittet er 10.5 år. b) Beregn gennemsnittet af højden i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. P Bruger sum i TI-Nspire. 21 k=1 yk = 2991 Kan også bruge =mean(højde). x = 2991 = 142.429 21 Dette betyder at børnene, som er blevet undersøgt, i gennemsnittet er 142.4 cm høje Symbolet x bruges til at beskrive gennemsnittet af x-værdier (x1 , x2 , . . . , xn ) i et datasæt. Tilsvarende bruges y til at beskrive gennemsnittet af y-værdier (y1 , y2 , . . . , yn ). Eksempelvis kan vi skrive gennemsnittet af 1, 2, 3, 4, 5 som 5 1+2+3+4+5 1X x= = k 5 5 k=1 c) Opstil en generel formel til at udregne gennemsnittet (x) af x-værdierne (x1 , x2 , . . . , xn ) ved brug af sumtegn. n x1 + x2 + . . . + xn 1X x= = xk n n k=1 d) Opstil en generel formel til at udregne gennemsnittet (y) af y-værdierne (y1 , y2 , . . . , yn ) ved brug af sumtegn. n y1 + y2 + . . . + yn 1X y= = yk n n k=1 Opgave 5 a) Plot datasættet fra opgave 2. Indtegn linjer, der viser gennemsnittet af x og y. Linjen, der viser gennemsnittet af x indtegnes ved 4: Undersøg data, 8: Plot værdi. Linjen, der viser gennemsnittet af y indtegnes ved 4: Undersøg data, 4: Plot funktion. Funktionen, der skal plottes, er y = y. b) Bestem den bedste rette linje for datasættet i a). Skriv forskriften for linjen. Forskriften for linjen er: y = 4.755x + 92.61 c) Tilføj en kolonne til datasættet, hvor du udregner yk y for hvert punkt i datasættet. Lav et plot af xk og yk y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a). Kolonnen med yk y kaldes forskely. Den udregnes ved at skrive skrive = b1 - $ c $4 og trække kolonnen ned. På plottet kan vi se at punkterne er blevet forskudt vertikalt. Gennemsnittet af punkterne er nu 0 for y-værdierne. Plottet ligner det for før, men værdierne på y-aksen er ændret. d) Bestem den bedste rette linje for datasættet plottet i c). Skriv forskriften for linjen y = 4.755x 49.8147 Trukket y = 142.429 fra. P e) Udregn nk=1 (yk y). Udregnes ved sum(d1:d21). Giver i TI-Nspire 3E 11 . f) Tilføj en ny kolonne til datasættet, hvor du udregner xk x for hvert punkt i datasættet. Lav et plot af xk x og yk y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a) og c). Kolonnen med xk x kaldes forskelx. Den udregnes ved at skrive skrive = a1 - $ c $3 og trække kolonnen ned. På plottet kan vi se at punkterne er blevet forskudt horisontalt i forhold til c). Gennemsnittet af punkterne er nu 0 for både x- og y-værdierne. I forhold til plottet i a) er punkterne forskudt både horisontalt og vertikalt. g) Bestem den bedste rette linje for datasættet i f). Skriv forskriften for linjen. y = 4.755x h) Sammenlign forskrifterne bestemt i b), d) og g). Gør rede for sammenhængen mellem forskrifterne. a) y = 4.755x + 92.61 c) y = 4.755x 49.8147 g) y = 4.755x Hældningen (a) er den sammen for forskrifterne, men skæringen med y-aksen (b) ændrer sig ved forskydning. P i) Udregn nk=1 (xk x). Udregnes ved sum(e1:e21). Giver i TI-Nspire 1E 11 . j) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4c), at det altid gælder at n X (xk x) = 0 k=1 n X (xk x) = k=1 HUSK fra opg. 4c): x = 1 n n X n X xk k=1 k=1 = nx Pn k=1 x nx = 0 xn , nx = Pn k=1 xn k) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4d), at det altid gælder at n X (yk y) = 0 k=1 n X (yk k=1 HUSK fra opg. 4d): y = 1 n y) = n X n X yk k=1 = ny Pn k=1 yn , ny = y k=1 ny = 0 Pn k=1 yn Bestemme bedste rette linje for et centraliseret datasæt I skal bestemme den bedste rette linje y = ax + b for et datasæt (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ), bestående af n punkter, hvor n P 2. ViPantager at datasættet er centraliseret, dvs. at x = n1 nk=1 xk = 0 og y = n 1 k=1 yk = 0. n Vi skal bestemme a og b, så kvadratsummen n ✓ X yk axk + b k=1 ◆2 bliver mindst mulig. Trin 1) Det første trin består i at omskrive summen, så at vi kan minimere denne i trin 2. a) Gør rede for at yk Mulige teknikker: yk (axk + b) 2 (axk + b) = yk 2 = yk2 axk b 2byk + a2 x2k + 2abxk + b2 2axk yk 2 = yk axk b · yk axk b = yk2 yk axk yk b axk yk + a2 x2k + axk b = yk2 2axk yk 2byk + a2 x2k + 2abxk + b2 yk (axk + b) 2 byk + baxk + b2 = (yk )2 + (axk + b)2 2yk (axk + b) = yk2 + a2 x2k + b2 + 2axk b 2yk (axk + b) = yk2 + a2 x2k + b2 + 2abxk 2ayk xk 2byk P P 2 b) Vi har at nk=1 yk (axk + b) = nk=1 yk2 2axk yk 2byk + a2 x2k + 2abxk + b2 . Brug regnereglerne for summer (1-4) til at dele summen op, sætte a, b foran summationen og summe over b. n X yk (axk + b) 2 = k=1 = n X yk2 + k=1 = n X n X k=1 yk2 +a 2 k=1 = n X k=1 a2 x2k + n X k=1 yk2 + a2 n X k=1 x2k + n ✓ X k=1 n X 2 yk2 b + k=1 n X + a2 x2k n X 2 + b + 2abxk 2abxk k=1 2 b + 2ab k=1 x2k + nb2 + 2ab n X 2ayk xk k=1 n X xk 2a k=1 n X k=1 xk 2ayk xk n X 2byk k=1 n X yk xk 2b k=1 2a 2byk n X k=1 yk xk n X k=1 2b n X k=1 yk yk ◆ c) Brug antagelsen om at gennemsnittet af x og y er 0 til at simplificere udtrykket fra b). n X yk2 +a k=1 = n X k=1 2 n X x2k 2 + nb + 2ab k=1 yk2 +a 2 n X x2k + nb2 k=1 n X xk k=1 n X 2a 2a n X yk xk k=1 2b n X yk k=1 yk xk k=1 Trin 2) I trin 1) fandt vi at n ✓ X yk axk + b k=1 ◆2 = a2 n X x2k 2a k=1 n X k=1 xk yk + n X yk + nb2 k=1 Vi skal nu finde de værdier for a og b, som minimere kvadratsummen. Summen n n n X X X 2 2 a xk 2a xk yk + yk + nb2 k=1 kan vi skrive som k=1 a2 A k=1 2aB + C + nb2 a) Gøre rede for hvad A, B, C svarer til i summen og forklar hvorfor A, B, C er konstanter. A= n X x2k k=1 B= n X xk yk k=1 C= n X yk2 k=1 xk , yk er konstanter og derfor er summen af disse også konstanter. I udtrykket er A, B, C konstanter, som ikke afhænger af de to variable a, b. b) Udtrykket a2 A 2aB + C + nb2 kan betragtes som en sum af to andengradspolynomier med hhv. a og b som variable. Skriv de to andengradspolynomier som udtrykket består af. Summen kan skrives som en sum af to polynomier. Mulighed 1) a2 A 2aB + C og nb2 . Mulighed 2) a2 A 2aB og C + nb2 c) Bestem a og b, så a2 A 2aB + C + nb2 bliver mindst mulig. Brug fra b) at udtrykket kan betragtes som to andengradspolynomier, som hver skal minimeres. For at summen bliver mindst mulig skal hver af polynomierne være mindst mulige. Teknik 1: Toppunktformel a2 A 2aB + C har toppunkt når a = (2A2B) = B . A Da A > 0 er dette et minimum. nb2 er mindst mulig, når b = 0. Da n 2 er dette et minimum. Teknik 2: Differentiering: 2aA 2B = 0 , 2aA = 2B , a = B A B B Hvis B > 0, så er 12 B < < 2 . Dette giver at: A A A 2 12 B A 2B = B 2B = B < 0 og 2 · 2B A 2B = 4B 2B = 2B > 0. A A B Derfor er a = A er et minimum. Hvis B < 0, så er 2 B 0. A B Derfor er a = A er et minimum. 2nb = 0 , b = 0 Da n 2 er 2nb < 0 for b 2] 1; 0[ og 2nb > 0 for b 2]0; 1[, dvs. b = 0 er et minimum ✓ ◆2 Pn d) Opskriv værdierne for a og b som gør kvadratsummen k=1 yk axk + b mindst mulig. Vi har fundet at kvadratsummen er minimeret, når Pn yk xk B b = 0 og a = = Pk=1 n 2 A k=1 xk Opgave 6 Betragt datasættet fra opgave 2 med alder og højde. a) Brug dine formler for a og b til at bestemme forskriften for den bedste rette linje. Benyt dine udregninger fra i går, hvor du udregnede yk y, xk x for alle punkter. P Først udregnes nk=1 (yk y)(xk x). Dette gøres i kolonne f ved kommandoen = d1 · e1. Herefter udregnes summen ved = sum(f 1 : f 21). I kolonne g udregnes (xk x)2 og bagefter findes summen ved brug af sum. a = 785.714 = 4.755 165.238 b = 142.429 4.755 · 10.4762 = 92.6147 Forskriften for den rette linje er y = 4.755x + 92.61 b) Bestem forskriften for datasættet i TI-Nspire. Er forskriften ens? Der er to teknikker til at finde forskriften i TI-Nspire. 4) Statistisk, 1) Statistiske beregninger, lineær regression. 3) Data, 9) Hurtiggraf, 6) Regression, Vis lineær. Begge teknikker giver forskriften y = 4.755x + 92.61, som er den samme som blev bestemt i a). Opgave 7: Verdensrekorder Tabellen nedenfor viser tiderne for verdensrekorderne i maratonløb (målt i sekunder) for mænd i perioden 1998-2007. År 1998 1999 2002 2003 2007 Tid (sek) 7565 7542 7538 7495 7466 Udviklingen af verdensrekorderne kan beskrives ved en lineær model f (x) = ax + b, hvor x angiver antal år efter 1998 og f (x) verdensrekorden i sekunder. a) Bestem a og b i modellen Model: f (x) = 10.626x + 7561.58, hvor f (x) angiver sekunder og x år efter 1998. b) Hvad kan vi forvente at verdensrekorden bliver i 2014 ifølge modellen? f (2014 1998) = f (16) = 7391.56 c) Verdensrekorden blev i 2014 slået, så den nu er på 7377. Hvordan passer det med c)? Benyt oplysningen til at kommentere på modellens rækkevidde. Ifølge modellen skulle verdensrekorden i 2014 være på 7391.56sek. Dette stemmer fint overens med virkeligheden, da verdensrekorden her var på 7377sek, dvs. en forskel på 14.56sek. Dette viser at modellen godt kan bruge til at forudsige tiden i 2014. Opgave 8: Dykkersyge Hvis en dykker opholder sig et (længere) stykke tid på en vis dybde, skal han stige langsomt op til overfladen igen for at forhindre dykkersyge, dvs. at der opstår livsfarlige bobler af kuldioxyd i blodet. Der er en sammenhæng mellem dybde og det antal minutter, som dykkerne kan opholde sig på den pågældende dybde uden at få dykkersyge. Den kan ses i følgende tabel x (m) 10 12 14 16 18 20 22 25 ... 40 42 y (min) 219 147 98 72 56 45 37 29 ... 9 8 I en model antages det, at dybden og tiden kan beskrives med en potens funktion. a) Opstil en model, der beskriver sammenhængen mellem dybde og antal minutter. Bestem forskriften for modellen. Sammenhængen mellem dybde og antal minutter kan beskrives ved en potens model f (x) = bxa , hvor f (x) angiver antal minutter, dykkeren kan opholde sig på en dybde x. x angiver dybden i meter. Forskriften for modellen er: f (x) = 42529x 2.289 b) Benyt modellen til at udregne hvor længe dykkeren kan opholde sig på en dybde på 30m. Vi kender dybden x = 30m. f (30) = 17.7min c) Benyt modellen til at udregne den dybde, dykkeren kan dykke, hvis han gerne vil dykke så dybt som muligt og opholde sig på dybden i 20min. Vi kender tiden y = 20min. 1 20 2.289 f (x) = 20 , 42529x 2.289 = 20 , x = 42529 = 28.43m Eleverne kan også bruge solve(f (x) = 20, x) Opgave 9: Vindmølleenergi Figuren herunder viser udviklingen på verdensplan i kapaciteten af vindmølleenergi. a) Opstil en model, der beskriver sammenhængen mellem årstal og kapaciteten af vindmølleenergi. Bestem forskriften for modellen. På grafen er det svært at se om udviklingen af vindmølleenergi vokser eksponentielt eller potens, så vi prøver at lave begge modeller til at beskrive sammenhængen mellem årstal og kapaciteten af vindmølleenergien. Eksponentiel model: f (x) = bax , hvor f (x) er vindmølleenergi målt i MW og x er år efter 1996. g(x) = bax , hvor g(x) er vindmølleenergi målt i MW og x er årstal. f (x) = 6.476 · 1.278x med r2 = 0.998 g(x) = 1.589 · E 212 · 1.278x med r2 = 0.998 Potens model: f (x) = bxa , hvor f (x) er vindmølleenergi målt i MW og x er årstal. g(x) = bxa , hvor g(x) er vindmølleenergi målt i MW og x er år efter 1996 h(x) = bxa , hvor h(x) er vindmølleenergi målt i MW og x er år efter 1995 (for at undgå 0) f (x) = 0 · 491.409x med r2 = 0.998. Error (pga. ln(0) ) g(x) = 2, 759 · 1, 405x med r2 = 0.899. På baggrund af r2 -værdierne vælger vi at at beskrive sammenhængen med en eksponentiel model. Dette virker også realistisk i forhold til virkeligheden, pga. en smitteeffekt som påvirker folk til at købe vindmøller. Derfor vil størrelsen hvormed vindmølleenergien ændrer sig være proportional med x vindmølleenergien, b·ab·a·ln(a) = ln(a). x b) Benyt modellen til at beregne kapaciteten af vindmølleenergien i 2014 og sammenlign med den faktiske kapacitet af vindmølleenergi i 2014, som var 369.6M W . f (2014 1996) = f (18) = 535.4M W g(2014) = 535.4M W I følge modellen vil kapaciteten af vindmølleenergi i 2014 være 535.4M W . Dette stemmer ikke overens med den faktiske kapacitet, da denne er 369.6M W . Modellen overestimerer kapaciteten med 535.4 369.6M W = 165.8M W . Når vi anvender modellen til at forudsige vindmølleenergien forudsætter vi at vindmøllekapaciteten fortsætter med at stige med 27,8% pr. år, hvilket ikke er realistisk i længden, da kapaciteten af vindmølleenergi på et tidspunkt må stagnere. Modellen kan derfor ikke bruges til at forudsige vindmølleenergien i 2014. Opgave 10: Spædbørnsdødelighed Spædbørnsdødeligheden angiver, hvor mange promille af de levendefødte, der dør inden for det første leveår. Tabellen nedenfor er hentet fra Danmarks Statistik og den viser udviklingen i spædbørnsdødeligheden i perioden 1933-1973. Årstal Spædbørnsdødelighed 1933 71.4 1938 60.0 1943 48.4 1948 37.8 1953 27.4 1958 23.0 1963 19.6 1968 15.4 1973 11.5 a) Lav et plot af datasættet. Indfør passende variable og opstil en model, der beskriver sammenhængen mellem årstal og spædbørnsdødeligheden i perioden 1933-1973. Forklar dit valg af model. Hvis vi plotter datasættet ser vi at datasættet aftager, men det er svært at afgøre om det skal beskrives med eksponentiel eller potens funktion. Eksponentiel: På plottet af datasættet kan vi se at modellen passer fint på datasættet. Eksponentiel model har r2 = 0.996. Model: f (x) = bax , hvor f (x) angiver dødeligheden i promille og x angiver år efter 1933. Potens: På plottet af datasættet kan vi se at modellen passer fint på datasættet. Potens model med r2 = 0.996. Model: f (x) = bxa , hvor f (x) angiver dødeligheden i promille og x angiver årstal. På baggrund af graferne og forklaringsgraden kan vi ikke afgøre hvilken model, der bedst beskriver sammenhængen. Hvis vi forholder os til virkeligheden vælges eksponentiel modellen, da midlerne som bruges til at forbedre spædbørnsdødeligheden må formodes at afhænge af spædbørnsdødeligheden. Derfor jo mindre spædbørnsdødeligheden er, desto færre midler bruges på at formindske denne. Derfor vil størrelsen, hvormed spædbørnsdødeligheden ændres formodes at være proportional med spædbørnsdødeligheden. b·ax ·ln(a) = ln(a). b·ax b) Bestem parameterne i modellen fra a). Eksponentiel model f (x) = 73.58 · 0.955x c) Forudsig på baggrund af modellen hvor stor en promille af de levendefødte, der dør inden for det første leveår i 2008. Eksponentiel: f(2008-1933)=f(75) = 2.402. I følge modellerne vil 2.402 promille af de levendefødte dø inden for det første leveår i 2008 d) Det oplyses at spædbørnsdødeligheden i 2008 var 9.9 promille. Hvordan passer det med modellen? Benyt oplysningen til at kommentere modellens rækkevidde. Hvis vi bruger modellen til at forudsige spædbørndødeligeheden i 2008 underestimeres antallet af spædbørn der dør. I virkeligheden var der 9.9 promille, mens vi ved modellen beregnede 2.4, dvs. en forskel på 7.5. Når vi anvender modellen til at forudsige spædbørndødeligheden i 2008 antages vi at den vil falde hvert år med 4,5%. Dette er ikke realistisk at den vil fortsætte med det så mange år efter de faktiske tal. Modellen kan derfor bruges til at beskrive spædbørnsdødeligheden i 1933-1973 og få år efter, men ikke til at lave ekstrapolation så mange år. Opgave 11: Løb Tabellen viser sammenhørende værdier af løbedistance og den tid, der tager en kondiløber at løbe distancen hurtigst muligt. Løbedistance (km) Løberens tid på distancen (sek) 0.5 98 1 215 3 779 5 1364 7 1989 10 2869 15 4341 20 5904 25 7413 På baggrund af disse data vil kondiløberen gerne udregner hvor lang tid det vil tage at løbe et maraton (42.195km). Nogle eksperter mener at kondiløberens tid som funktion af distancen kan beskrives ved en lineær model f (x) = ax + b, hvor x er distancen (målt i km) og f (x) er kondiløberens tid (målt i sekunder). a) Bestem a, b i den lineære model og brug denne model til at udregne hvor lang tid kondiløberen vil være om at løbe et maraton. Modellen er givet ved: f (x) = 299.47x 103.54, hvor f (x) angiver tiden i sek og x angiver distancen i km. f (42.195) = 12532.5sek Andre eksperter mener at kondiløberens tid som funktion af distancen skal beskrives ved en potens model f (x) = b · xa , hvor x er distancen (målt i km) og f (x) er kondiløberens tid (målt i sekunder). b) Bestem a, b i potens modellen og brug denne model til at udregne hvor lang tid kondiløberen vil være om at løbe et maraton. Modellen er givet ved: f (x) = 219.91x1.11 , hvor f (x) angiver tiden i sek og x angiver distancen i km. f (42.195) = 13879.9sek c) Er kondiløberens tid på et maraton den samme ved de to modeller?. Hvis nej, hvor stor er forskellen. Lineær model: f (42.195) = 12532.5sek Potens model: f (42.195) = 13879.9sek Forskel: 13879.9 12532.5sek = 1347.4sek ⇡ 22min d) Hvilken model ville du bruge til at bestemme kondiløberens tid? Skriv begrundelserne ned. Diskuter dine begrundelser med gruppen. Lineær model: r2 = 0.9998 Potens model: r2 = 0.9993 Viden om løb: Tiden angiver hvor lang tid personen er om at løbe en given distance hurtigst muligt. Vi ved at desto længere man løber, desto langsommere løber man pr. km. Ved en lineær model antager vi at den maksimale hastighed er den samme uanset distancen, hvilket ikke er realistisk. Den lineære model giver at tiden øges med 299.47sek hver gang personen løber 1km mere. De faktiske tider pr. km for de forskellige distancer er: Løbedistance (km) Løberens hastighed (sek/km) 0.5 196 1 215 3 259.7 5 272.8 7 284.1 10 286.9 15 289.4 20 295.2 25 296.5 I tabellen kan vi se at hastigheden ikke er konstant og at det faktisk kun er ved de længere distancer (15-25) at sek/km stemmer nogenlunde overens med modellen. Ved en potens model antages det at hastigheden ikke er konstant og at den maksimale hastighed falder desto længere man løber. Ved at bruge potens modellen ses det at løberen vil løbe med 13879.9 = 328.9sek/km, dvs. omkring 42.195 30sek langsommere pr. km end ved 25km på et marathon. På grafen kan vi også se at tiden bliver overestimeret en smule. Da den lineære model ikke er realistisk i forhold til løb vil jeg bruge potensmodellen til at forudsige løberens tid på marathon. Opgave 12: Tilskadekommende i trafikken Hvert år opgøres antallet af tilskadekommende i trafikken i Danmark. Antallet af tilskadekommende i Danmark i årene 1979-1999 kan ses i tabellerne herunder. I 1995 blev antallet af tilskadekommende i trafikken ikke opgjort, derfor mangler der data for dette år. År 1979 1980 1981 1982 1983 1984 1985 Antal ulykker 9267 8477 8546 8427 8105 8523 8672 År 1986 1987 1988 1989 1990 1991 1992 Antal ulykker 8301 7357 7321 7266 6396 6231 6031 År 1993 1994 1995 1996 1997 1998 1999 Antal ulykker 5735 5661 ... 5339 4424 4071 4217 a) Bestem på baggrund af tabellen hvor mange tilskadekommende i trafikken der ca. har været i 1995. Skriv din model, dine overvejelser om valg af model samt dine udregninger. På scatterplottet kan vi se at antallet af tilskadekommende ser ud til at aftage lineært, så jeg har lavet lineær regression. Ved lineær regression fås modellen f (x) = 256.6x + 9407.8, hvor f (x) er antal tilskadekommende og x antal år efter 1979. Ved lineær regression fås r2 = 0.95. Antallet af tilskadekommende i trafikken i 1995 er f (1995 1979) = f (16) = 5301.5, hvilket stemmer overens med tabellen. b) Forudsig på baggrund af tabellen antallet af tilskadekommende i trafikken i 2016. Skriv din model, dine overvejelser om valg af model samt dine udregninger. Antallet af tilskadekommende i trafikken, hvis den lineære udvikling fortsætter er: f (2016 1979) = f (37) = 87.9 Dette tal er urealistisk, da det er negativt. Det er heller ikke realistisk at antallet af tilskadekommende i trafikken vil fortsætte med at aftage lineært efter 1999, da vi så på et tidspunkt får 0 eller et negativt antal tilskadekommende. Derfor kan den lineære modellen ikke bruges til ekstrapolation. Vores antal af tilskadekommende i 1995 er meget realistisk hvis vi ser på antallet i 1994 og 1996. Modellen kan derfor anvendes til at beskrive årstallene mellem 1979-1999 og i en kort periode efter. For at udregne antallet i 2016 vil vi derfor modellere data med en anden model. På plottet ser vi at data aftager med en lineær tendens, men vi forventer ikke at de fortsætter med at aftage lineært, da antallet på et tidspunkt må stagnere. Både en eksponentiel og potens model vil stagnere, så antallet af tilskadekommende vil aftage år for år, men aldrig blive negativt. Ved eksponentiel regression fås: f (x) = 9886.6 · 0.961x hvor f (x) er antal tilskadekommende og x år efter 1979. Forklaringsgraden for modellen er 0.924. Ved denne model er antallet af tilskadekommende i 2016 f (37) = 2262. Ved potens regression fås: g(x) = 2.0096 · E 265 · x 79.2696 hvor g(x) er antal tilskadekommende og x er årstallet. Forklaringsgraden for modellen er 0.923. Ved denne model er antallet af tilskadekommende i 2016 g(2016) = 2278. Afhængigt af hvilken model vi vælger vil forskellen kun være på 16. Da forklaringsgraden er ens for begge modeller og de begge stagnere (ved at det er svært at undgå tilskadekommende i trafikulykker) er der ingen kvalitative argumenter som kan bruges til at afgøre hvilken model vi skal bruge. Da modellerne forudsiger antallet til at være henholdsvis 2262 og 2278 kan vi forudsige at antallet af tilskadekommende i 2016 vil være omkring 2270. Den lineære model beskriver data fra 1979-1999 godt (højest r2 -værdi) og kan derfor bruges til interpolation, mens en eksponentiel eller potens model er bedre egnet til extrapolation uden for modellens rækkevidde. Opgave 13: Kaniner på en øde ø På en øde ø blev der sat 50 vilde kaniner ud 1. januar 2000. I de efterfølgende år talte man hvert år antallet af kaniner på øen d. 1. januar. Tabellen herunder viser hvordan antallet af vilde kaniner på øen vokser med tiden, målt i år efter 2000. År 1 2 3 4 5 6 7 Antal kaniner 100 240 450 1000 2000 4100 9050 År 8 9 10 11 12 13 14 Antal kaniner 17030 26010 30070 32500 33810 33890 33870 a) Hvilken sammenhæng er der mellem tiden og antal kaniner. På scatterplottet kan vi se at antallet af kaniner vokser eksponentielt indtil 8-9år, hvorefter at udvikling mindskes og efter ca. 12 år er antallet af kaniner stabilt på ca. 33800-33900. Kan beskrives ved logistisk vækst (hørt dette fra biologi?) 34089.8 Logistisk regression (d=0): f (x) = 1+3509.79·e 1.02543x . b) Bestem ud fra data i tabellen hvor mange kaniner der har været på øen 1. juli 2006. Skriv din model, udregninger og overvejelser. Kaninerne blev sat ud 1. januar 2000, så 1. juli 2006 svarer til x = 6.5. Ved den logistiske model (d=0) får vi at antallet er f (6.5) = 6228.89. Da den første del af udviklingen kan beskrives ved en eksponentiel funktion, laver jeg eksponentiel regression på data fra de første 8 år og bruger denne model til at bestemme antallet af kaniner 1. juli 2006. Modellen, der kan bruges til at beskrive udviklingen de første 8 år er givet ved: f (x) = 51.6 · 2.08x med r2 = 0.999. Dette viser at udviklingen de første 8 år kan beskrives ved en eksponentiel model. Antallet af kaniner d. 1 juli 2006 er da f (6.5) = 5985.99 c) Bestem ud fra data i tabellen hvor mange kaniner der vil være på øen 1. januar 2016 Skriv din model, udregninger og overvejelser. Ifølge tabellen og grafen kan vi se at antallet af kaniner er stabilt efter 12år på ca. 33800-33900, så antallet af kaniner efter 16 år (2016) vil være i dette interval. Vi kan beskrive de sidste år (år 12-14) med en lineær model som har hældning 0. Udregning ved logistisk model: f (16) = 34080.8. Dette stemmer overens med hvad vi forventede ud fra data. 162 list of references a.3 tables of the didactic process Description of the mathematical praxeologies and the didactic process in module 1-3. Module 1 Episode Problem Task Introduction to best line Q3 : What do we mean by best line? TRL Q3 : What do we mean by best line? Exercise 1+2 Summary of exercise 1+2. Definition of best line Like above Exercise 3 Calculate sums Summary of exercise 3 Exercise 4 + 5ai Calculate sums Summary of exercise 4 + 5a-i Exercise 5j (5k) Calculate mean/sums. Displacement of data TRL Like above Prove Pn 0 Summary k=1 (xk x) = Technique of task Technology Theory of specific activity Didactic moment of MO 1) with best line Draw the best line by eye. Calculate sums by hand Calculate sums. Draw dataset I Technique ⌧RL Like above Pn x = n1 k=1 xk Rules for sums Explain their choice of the line. Specify what best means Definition of best line. ✓P M : Minimize n k=1 (yk (axk + b))2 The rules sums Changing data do change slope of the for the not the line Definition of mean. Change of data do not change the slope Subtract x, y centralize the dataset D Definition of best line Rules of sums Data can always be centralized 1) with best line 2) Elaborate techniques 3) Technological discourse 4) Technical work with techniques to the technical moment 4) Rules of sums 4) Technical work with techniques to the technical moment 4) Technique to displace data 4) Technique to centralize data Module 2 Episode Summary module 1 Problem Task Technique of Formula of best line for centralized data Summary of the proof Minimize Pn k=1 (yk (axk + b))2 Part 1) Rewrite the sum Part 2) Minimize the sum of two polynomials Like above Formula of best line for a data set Minimize Pn k=1 (yk (axk + b))2 Exercise 6 TRL Introduction to best exponential/power model Q2 : What do we mean by best exp/power model? Q2 : How can we determine the models? Technology Theory Didactic moment Definition and results from module 1 4) Techniques to technical work 4) Develop technique Theorem of best line for centralized data Theorem of best line for data set 5) Institutionalization Realize that the result of the two techniques are identical 6) Evaluation of the technique and the instrumented technique 1) with best exp/power model 5) of definition and techniques min(s(a,b)) = min( (a)) + min( (b)) Multiplication Quadratic identity Rules of sums Differentiation Formula of vertex Like above 1) Displacement of data 2) Theorem of centralized data A ⌧RL : Use the formula I ⌧RL : TI-Nspire Transformation / linearization Linear regression Definition of best exp/power model by linearization 5) Institutionalization Module 3 Episode Problem Task Technique Technology Theory Didactic moment Summary of exercise 7-9 TRL , TR , TRP , TY M , TXM , TQP ⌧ I : TI-Nspire I I ⌧RL , ⌧RP , solve 1) with Q1 , 1) reencounter with old techniques Exercise 10 +11 Related to Q1 , Q4 . 10) TR , TY M , TQP 11) TRL , TRP , TY M Nspire. ExamI I ples: ⌧RL , ⌧RP , Scatterplot r2 Calculate speed Summary of exercise 10 + 11 Like above Like above Realize that models can describe data well in a restricted range, but not useful for predictions Restricted range for predictions. The phenomenon are relevant for choice of Fi . Like above Exercise 12 + 13 Related to Q1 , Q4 . 12) TR , TY D , 13) TR , TY D The same as in exercise 10+11. Regression of restricted data set Logistics regression Summary of exercise 12 + 13 Like above Like above Summary module 3 Q1 and Q4 of As in exercise 10 + 11. The model with highest r2 are not necessarily the best. New class of functions Choice of Fi depends on the purpose. Not solely use r2 to determine Fi . Investigate the phenomenon. The previous technological and theoretical elements 3) specify what best mean. 4) Develop techniques to choose Fi 3) State precise what best mean. 4) Technique to choice best. 5) Institutionalization of best. 3) Explain best Fi , 4) Develop techniques 5) Institutionalization of Q1 and Q4 5) Institutionalization of Q1 and Q4 166 list of references a.4 the planned didactic process Lektion 1 Tid Aktivitet 5min 3min Introduktion til forløb. Introduktion til bedste linje og regression. 13min Opgave 1 + 2 12min Repræsentation /arbejdsform Vise datasæt. S: Hvilken linje er bedst? Motivation til forløb og timen. Grafisk via plots. Tegne bedste linje. Individuel/gruppe Lærer Fortælle om emne, Lytte, stille spørgsmål dataindsamling Introducerer problem Lytte, stille spørgsmål. om hvad vi mener med bedst? Observere, lytte til diskussioner, identificerer gode argumenter. Opklare spørgsmål Opsamling opg. 1 Grafisk illustration af 1) Italesat kriterier for +2. kriterier. Brug af TI- bedste linje. Input fra Introduktion til Nspire til at vise linjer. elever. MKM + summer. Algebraisk og grafisk 2) Definition af repræsentation af bedste linje. MKM. MKM. 3) Forklare summen af kvadreret fejl. Forklare sumtegn. 15min Opgave 3 5min 6min PAUSE Opsamling på opg. 3. Institutionaliserin g af regneregler. 20min Opgave 4 + 5a-i Elev Tegne bedste linje. Overveje og diskutere kriterier. Definere bedste linje. 1) Eleverne gennemgår kriterier og forslag til bedste linje. 2) Genkalde definition af bedste linje. 3) Lytte, stille spørgsmål. Numerisk. Udregne Observere. Afklare Udregne summer. summer. Finde spørgsmål. Lytte til Diskutere sammenhænge/nye diskussioner. sammenhænge/nye teknikker. Algebraisk - Identificerer grupper, teknikker i grupper. skrive regleregler for der finder Hvis en gruppe finder summer. sammenhængene. regneregler skrives Individuel/gruppe. denne på tavlen. Algebraisk. Valideres via numeriske eksempler. Regneregler skrives op. Regneregler for summer institutionaliseres. Regneregler opskrives via input fra eleverne. Aktiv deltage med formulering af regneregler/teknikker fundet i opg. 3. Forklare hvorfor regnereglerne gælder. Udregning af gns. Overvære eleverne for Arbejde med summer Generalisering af at få fornemmelse for i TI-Nspire. Opstille udtryk. Udregning og forståelse og generel formel. grafisk repræsentation problemer. Forståelse for ændring af ændret datasæt. Identificere grupper af datasæt - se grafisk. Gruppearbejde. med gode Diskussion af besvarelser/forklaring forflytning. er. Lektion 1 Opsamling 4 + 5a- Fælles gennemgang af i. Introduktion til 4 + 5a-i. Grafisk centraliseret illustration af datasæt. gennemsnit + forflytning. Nyt datasæt. 8min Gennemgang af 5j (5k) 5min 3min 95min Gennemgå formel for Enkelte elever: gennemsnit. Sikre Gennemgå løsning af forklaring af flytning 5a-f. De andre tjekker er tilstrækkelig. udregninger, stiller Supplere med spørgsmål. Forklare spørgsmål. Pointe: hvad der sker. Flytte datasæt, så det bliver centraliseret. Algebraisk: Bevis for at Gennemgå hvorfor Deltage aktivt i summen af sum er 0. Stille beviset ved at bidrage centraliseret data er 0. spørgsmål til eleverne - med ideer til inddrage regneregler udledning. for summer og formel for gns. Huske at give eleverne tid til huske regneregler + gns. Opsamling. Gennemgå vigtige Pointere vigtige Definition af pointer, som er samlet pointer. Motivation bedste model på et slide. for næste time. (arbejde videre med i morgen). Bruge summer, centraliseret data. Lytte, stille spørgsmål, tage noter. Lektion 2 Repræsentation /arbejdsform Tid Aktivitet 5min Opsamling på Vigtige pointer skrives vigtige pointer fra på tavlen. lektion 1, som skal bruges i beviset. Introduktion til bevis. 5min Introduktion til bevis af bedste rette linje 43min Udlede formlen Algebraisk. Udlede for bedste rette beviset via trin. linje for et Gruppearbejde. centraliseret datasæt. 1) Omskrive sum 2) Minimere sum 5min 15min PAUSE Gennemgang af Algebraisk. Trinnene bevis for et og mellemregningerne centraliseret er skrevet og vises datasæt. trinvis. Kan ske over to omgange (først trin 1, derefter trin 2), hvis det er nødvendigt at opbryde forrige aktivitet. Bevis for ikkeGrafisk, algebraisk. centraliseret data. Kobling ml. bevis og opg. 5. 7min Grafisk. Algebraisk Lærer Elev Spørger eleverne til vigtige pointer og skriver disse på tavlen. Sikre at følgende genkaldes: 1)Bedste linje (MKM) 2) Centraliseret data 3) Regneregler for summer Introducere opg. Om formel for a,b. Skrive problem op. Grafisk vise hvordan datasæt kan rykkes. Starte med at bestemme for centraliseret data. Hjælpe eleverne med hints ved at refererer til allerede kendt viden. Validere resultaterne i trinnene. Observere om der er behov for gennemgang efter trin 1. Eleverne genkalder og gengiver pointer fra lektion 1. Spørge eleverne om forklaringer til de beviset (skrevet på forhånd). Validere og forklare trinnene/udregninger ne. Vigtigt at pointere de to trin. Forklare at vi har en formel til at beregne bedste linje. S: Hvad hvis data ikke er centraliseret? Eleverne får tid til at tænke over muligheder? Bruge formel fra før, da vi kan centralisere data. Vise forskydning grafisk. I alm. gælder vise formel. Forklare udregningerne og udledningen af formlerne. Lytte, stille opklarende spørgsmål. Lytte, stille spørgsmål. Arbejde i grupper på egen hånd (i det omfang det er muligt). Udlede formlen via trin. Opdage formlen for parameterne a og b for et centraliseret datasæt. Besvare spørgsmål. Tid til at tænke over løsning. Bruge viden fra lektion 1. Deltage i det omfang, de kan. Lektion 2 8min Opgave 6 7min Introduktion til Grafisk via eksempel i bedste TI-Nspire. Vise eksponentiel/pote transformationen. ns model. Linearisering. 95min Udregning af bedste rette linje via formel. Individuel/par Lade eleverne arbejde på egen hånd. Kan bidrage med teknikker (kommandoer) til at udregne sum, produkt og kvadrat. Arbejde med formlen udledt i forrige aktivitet. Opdage at formlen og CAS giver det samme. Validere selv deres udregning. Vise hvordan data kan transformeres, så de bliver lineære. Genkalde viden om eks/pot funktioner og logaritmer. Finde den bedste lineære model af transformeret data (ændre datasæt) Lytte, stille spørgsmål. Opdage/introduceres for sammenhæng mellem eksponentiel, potens og lineær regression. Lektion 3 Tid Aktivitet Repræsentation /arbejdsform 12min Opsamling på TI-Nspire opg. 7 - 9 (lektier) 3min Introducere problemet med valg af model. 27min Opgave 10 + 11 Numerisk / grafisk. Arbejde i TI-Nspire. Grupper. 10min Opsamling på opg. 10 + 11 TI-Nspire 5min 25min PAUSE Opgave 12 + 13 8min Opsamling på opg. 12 + 13 Lærer Elev Udvælge 1-2elever, 1-2 elever gennemgår som gennemgår opg. 8-opgaverne. 9. Validere Andre: Stille besvarelserne. Stille opklarende spørgsmål, supplere spørgsmål, tjekke og præcisere svar. resultater. Læreren gennemgår at Lytte, stille spørgsmål model typen sjældent er givet. Arbejde med at bestemme typen af model. Besvare opklarende spørgsmål. Observere elevernes arbejde. Udvælge elever med gode overvejelser/ besvarelser, som gennemgår efterfølgende. Validere besvarelserne. Supplere og præcisere svar. Bidrage med andre muligheder for at afgøre model. Arbejde med valg af model. Diskutere modeller og overvejelser om valg af model. 1-2 elever gennemgår. De andre bidrager med kommentarer, forslag og deres overvejelser samt stiller spørgsmål. Besvare opklarende Arbejde med valg af spørgsmål. model. Diskutere Hjælpe/give hints med modeller og opg. 13. Observere overvejelser om valg elevernes arbejde. af model. Udvælge elever med gode overvejelser/ besvarelser. TI-Nspire Validere besvarelserne. Supplere og præcisere svar. Pointere at der aldrig findes en rigtig model. Kommer an på formål. 1-2 elever gennemgår. De andre bidrager med kommentarer, forslag og deres overvejelser samt stiller spørgsmål. Lektion 3 5min 95min Opsamling på valg Væsentlige teknikker til Inddrage eleverne. af model at afgøre model skrives Understrege pointer. på tavlen. Kommer an på formål med modellen. Aktiv deltagelse. Bidrage med teknikker til at vælge model. Noterer vigtige pointer. Plan for undervisning Inden undervisningsforløbet bestemmes grupper, som eleverne arbejder i. Eleverne vil både komme til at arbejde individuelt og i grupper igennem forløbet. Diskussioner og overvejelser vil altid foregå i grupperne. Eleverne vil til hvert model få udlevet opgaver til dagens model, hvor jeg har indsat tekstbokse, som de kan skrive deres besvarelser i (se opgaverne). Efter opgaverne vil der være en tom side, hvor eleverne kan skrive noter osv. i. Tekstboksene og noterne gør det muligt for mig at dokumentere og analysere elevernes arbejde efterfølgende. Da det er svært at lave en tidsplan over forløbet har læreren mulighed for at ændre undervejs. I de enkelte moduler vil jeg skrive mulige ændringer, såsom at undlade opgaver, lave opsamling før, selv gennemgå opgaver og lignende justeringer. Modul 1: Torsdag d. 9/4 8.00-9.35 Formål/indhold • Kriterier for bedste rette linje, som giver anledning til definition af mindste kvadraters metode (MKM) • Introduktion til summer og arbejde med summer. Regneregler for summer • Centralisering af datasæt ved at forskyde dette Introduktion til forløb Jeg introducerer kort forløbet. Jeg fortæller om emnet; regression og hvad de skal arbejde med i de tre moduler. Jeg fortæller om dataindsamling (optagelser af deres samtaler ved diktafoner og billeder af opgaveark, observationer). Introduktion til bedste linje og regression Læreren fortæller at vi skal repetere regression og stiller spørgsmålet om hvad vi mener med bedste linje. Læreren fortæller at vi de næste to dage (modul 1 + 2) skal arbejde med matematikken omkring regression, så vi kan forstå hvordan TI-Nspire udregner den bedste rette linje. Læreren gør det klart at TI-Nspire selvfølgelig finder det bedste rette linje ud fra nogle kriterier og at der ligger matematik bag. Det er vigtigt at læreren pointerer at vi kan forklare udregningerne med matematik og derfor også hvad TI-Nspire gør. Derefter fortæller læreren grupperne til eleverne og opgave 1-2 udleveres. Eleverne arbejde individuelt med opgave 1 + 2a-c). Herefter arbejder eleverne i grupperne (2d-e) Opsamling på opgave 1 + 2 Når læreren fornemmer at eleverne har arbejdet med opg. 1-2 og fået diskuteret og overvejet hvad de forstår ved bedste rette linje starter læreren en opsamling på klassen. Ved opsamlingen har læreren datasættet for opg. 2 plottet i TI-Nspire, så læreren grafisk kan illustrere nogle af kriterierne for bedste rette linje. Læreren udvælger forskellige grupper og spørge hvilke kriterier de har. De forskellige kriterier diskuteres. Kriterier som vertikale, horisontale og vinkelrette afstande kan illustreres på tavlen. Hvis eleverne har få kriterier kan læreren fremsætte nogle kriterier, som eleverne kan diskutere (se mulige kriterier herunder). Forhåbentlig vil nogle eleverne fremsætte kriteriet om at kvadrere de vertikale afstande, idet eleverne tidligere har arbejdet med lineær regression og blevet kort introduceret til MKM. Dette kriterie kan illustreres via TI-Nspire. Mulige kriterier og "begrundelser" for hvorfor de ikke bruges: P • Minimere den horisontale afstand nk=1 (xk x) - Ved dette minimeres fejl i x-værdierne. Ofte måler vi en fejl i y-værdien (eks: kender årstal, alder, osv.). P • Minimere den vertikale afstand nk=1 (yk y) - Måler fejl i y-værdierne. • Minimere den vinkelrette afstand - god ide (både fejl i x og y). - Rigtig svær at beregne. • Minimere den vertikale afstanden numerisk - Svær at beregne. Kan give flere svar. Pn k=1 • Minimere kvadratet af de ovenstående afstande |yk Pn y| k=1 (yk y)2 - Undgår at positive og negative værdier udligner hinanden. Simpel løsning, som er let at regne. • Lige mange punkter på hver side af linjen - Eksempel med 1 outliner • Skal gå igennem den mindste og største værdi - Eksempel med 1 outliner. Linjen ligge helt skævt • Skal gå igennem gennemsnittet af værdierne - Gør den også med MKM. • Skal gå igennem to punkter - Vise hvor skævt det bliver ved at tage to punkter fra datasæt opg. 2. Vigtigt at vi anvender alle punkter. • Fastlægge skæringen (b-værdien) hvis vi kender denne - Der kan altid være en vis usikkerhed. Hvis eleverne ikke selv fremsætterPkriteriet om MKM genkalder læreren dette kriterie hos eleverne og præsenterer nk=1 e2k , hvor e er den vertikale afstand. Læreren illustrerer kvadraterne for datasættet i opg. 2 og forklarer hvad sumtegnet betyder. Herefter skrives definitionen op: Datasæt , yn ). Den bedste rette linje y = ax + b findes ved at minP (x1 , y1 ), . . . , (xnP imere nk=1 (yk y)2 = nk=1 (yk (axk + b))2 . Valgt kriterie kan vi forklare og retfærdiggøre linjen ved brug af matematik. Læreren fortæller at eleverne nu skal arbejde med summer, idet vi finder den bedste rette linje ved at summe fejlene. Eleverne arbejder herefter med opg. 3 omkring summer. Hvis elever opdager regneregler for summer skrives disse på tavlen. Imens eleverne arbejder med opgave 3 skriver læreren tabellen med summer på tavlen. Opsamling på opgave 3 De sidste spørgsmål i + j) vil være krævende for eleverne og derfor kan denne opsamling startes før i + j) er regnet, hvis læreren vurderer at eleverne ikke selv når frem til regnereglerne for summer. Læreren spørger eleverne om resultaterne til 3a-h og de skrives ind i tabellen. Hvis nogle elever har fundet sammenhænge står disse på tavlen og eleverne forklare disse til resten af klassen. Hvis ikke, kan læreren spørge til sammenhæng mellem nogle af summerne (a-b, c-d, e-f, g-(a+e), h-(e+c). Denne opsamling kan være mere eller mindre lærerstyret afhængig af hvordan eleverne klarer opgaverne og tiden. Hvis eleverne ikke når opg. 3j) tages denne i samspil med eleverne på tavlen efter at regnereglerne er gennemgået. Det er vigtigt at eleverne bliver introduceret for regnereglerne, da de skal bruge disse i beviset (modul 2). Regneregler for summer Vi kan summe en konstant n X c=n·c k=1 Man kan sætte en konstant ude for summationstegnet n X k=1 c · xk = c n X xk k=1 Man kan dele summer op. n X (xk + yk ) = k=1 n X k=1 (xk n X xk + k=1 yk ) = n X k=1 n X yk k=1 xk n X yk k=1 Regnereglerne skal skrives op på tavlen. Eleverne vil få regnereglerne på papir, så det er ikke vigtigt at disse noteres. Læreren udleverer opgave 4 + 5, som eleverne herefter arbejder med. Opsamling på opgave 4 + 5a-i Opgave 4 gennemgås ved at læreren spørger eleverne, som forklare hvordan gennemsnittet regnes (opg. a+b) og herefter Pn fortæller generel formel for gennemsnit. Læreren understreget at hvis x = 0 er k=1 xk = 0. Opgave 5 gennemgås ved at læreren har udvalgt 2 elever, som har gode besvarelser/ forklaringer. Eleverne viser via TI-Nspire hvordan gennemsnittet flyttes i de forskellige situationer. Læreren supplerer med forklaringer og sikrer at eleverne indser at ændring af datasættet medfører til en forflytning af datasættet og at hældningen på den bedste rette linje er uændret. Læreren gentager og illustrerer hvis nødvendigt at datasættet flyttes vertikalt med y og horisontalt med x. Læreren introducerer at vi faktisk ændrer datasættet fra (x1 , y1 ), . . . , (xn , yn ) til (x1 x, y1 y), . . . , (xn x, yn n) ved forskydningen. Det nye datasæt kaldes et centraliseret datasæt, da gennemsnittet af dette vil være i (0, 0). Gennemgang af opgave 5j (samme med 5k) P Læreren beviser at nk=1 xk x = 0. Dette gøres ved at inddrage eleverne aktivt i udledningen. Læreren skriver summen op og spørger eleverne hvordan vi kan udlede dette. Det er vigtigt at læreren giver eleverne lidt tid til at tænke. Forhåbentlig vil eleverne genkalde regnereglen for at summe en konstant og formel for gennemsnit. Hvis ikke må læreren gennemgå udledningen. Der forklares hvordan 5h tilsvarende kan bevises. Læreren understreger at denne sum altid vil være 0 og at vi derfor altid kan centralisere et datasæt ved at trække gennemsnittet fra. Hvis tiden er knap gennemgås 5k ikke. Opsamling på modul 1 Læreren gennemgår de vigtigste pointer fra dagens model, som er skrevet på et slide. • Definition af bedste linear model: Givet datasæt (x1 , y1 ), . . . , (xn , yn ). Den bedste rette linje y = ax + b findes ved at minimere n X 2 (yk y) = k=1 n X (axk + b))2 (yk k=1 • Regneregler for summer Vi kan summe en konstant n X k=1 c=n·c Man kan sætte en konstant ude for summationstegnet n X k=1 c · xk = c n X xk k=1 Man kan dele summer op. n X (xk + yk ) = k=1 n X k=1 (xk n X xk + k=1 yk ) = n X k=1 n X yk k=1 xk n X yk k=1 • Et datasæt (x1 , y1 ), . . . , (xn , yn ) med x = 0 og y = 0 kaldes et centraliseret datasæt. Et datasæt kan centraliseres ved at forskyde datasættet vertikalt og horisontalt med x og y. Modul 2: Fredag d. 10/4 8.00-9.35 Formål/indhold • Formlen for bedste rette linje for et centraliseret datasæt • Formlen for bedste rette linje for et ikke-centraliseret datasæt • Introduktion til bedste eksponentiel og potens model. I dette modul er det vigtigt at eleverne har tid til at arbejde med formlen for bedste rette linje. Jeg har lavet en enkelt opgave, hvor eleverne skal bruge formlen til at beregne bedste rette linje, men da fokus er på formlen og beviset skal denne kun udleveres og laves, hvis der er tid til det. Læreren må afgøre hvorvidt der er tid til denne. Læreren skal prioritere opgave 6 højere end introduktion til eksponentiel og potens modeller, da denne kan udskydes til modul 3. Opsamling på vigtige pointer fra modul 1 Modul 2 indledes med en opsamling af de vigtigste pointer fra gårdagens undervisning. Læreren spørger eleverne til vigtige pointer fra modul 1 og disse skrives på tavlen. Læreren skal sikre at pointerne fra opsamling på modul (fra modul 1) kommer på tavlen, da eleverne skal bruge regnereglerne til at udlede formlen for bedste rette linje. Introduktion til bevis af bedste rette linje Læreren introducerer efterfølgende opgaven om at udlede formlen for bedste rette linje for et datasæt. Læreren pointere hvorfor vi finder denne og læreren kan motivere eleverne ved at trække en parallel til formlen for to punkter og forklare at vi nu skal finde en formel for a og b, når vi har mere end to punkter. Læreren understreger at TI-Nspire bruger denne formel til at bestemme den bedste rette linje (lineær regression), intet mystisk. Læreren illustrerer grafisk hvordan vi kan flytte datasættet, så det bliver centraliseret og fortæller at dette gør det lettere at finde a og b. Læreren introducerer at eleverne nu skal finde forskriften for a, b, når datasættet er centraliseret. Læreren giver eleverne opgaven med trin 1) for forskrift for ret linje. Eleverne skal i grupperne arbejder med trin 1 i opgaven. Når eleverne er færdige med trin 1) udleveres trin 2), som eleverne herefter arbejder med. Læreren skal være opmærksom på at redegørelsen af kvadratet kan være krævende for eleverne, da de normalt kun udregner kvadratet af 2leddet størrelser. Læreren kan hjælpe eleverne ved enten at huske dem på hvordan man ganger parenteser eller at bruge kvadratsætningen på yk og (ax + b). Læreren har mulighed for at reagere undervejs og skal observere om der er behov for at gennemgå trin 1), hvor summen omskrives inden eleverne arbejder videre med trin 2), hvor summen minimeres. Hvis læreren vurderer at eleverne har brug for at få valideret trin 1) inden der arbejdes videre med trin 2) kan hun dette. Udledningen af formlen behøves ikke at blive gennemgået samlet, men kan deles op i de to trin. Gennemgang af bevis for centraliseret data Udledningen af formlen for a og b gennemgås enten samlet eller i to omgange, afhængig af hvad læreren vurderer. Udledningen foregår ved at læreren spørger eleverne om forklaringer til beviset. Beviset er skrevet på forhånd og de enkelte skridt vises løbende, mens eleverne forklarer udregningerne. Læreren validerer og trækker paralleller til gårdagens og tidligere kendt viden. Læreren understreger at vi nu har fundet en formel til at beregne den bedste linje, for et centraliseret datasæt. Læreren skriver på tavlen: Givet et datasæt (x1 , y1 ), . . . , (xn , yn ) hvor x = 0 og y = 0 er den bedste rette linje givet ved Pn xk yk y = ax + b hvor a = Pk=1 og b = 0 n 2 k=1 xk Udledning af formel for ikke-centraliseret data Efter gennemgangen af beviset for centraliseret data går læreren tilbage til det generelle datasæt, hvor data ikke er centraliseret. Læreren viser igen datasættet fra introduktionen og spørger eleverne hvordan vi kan bruge formlen for centraliseret datasæt til at bestemme a, b for et generelt datasæt. Læreren giver eleverne 1-2min til at tænke over dette. Forhåbentlig vil nogle eleverne komme med forslag til at vi kan rykke datasættet, så det bliver centraliseret og læreren kan tage udgangspunkt i dette ved at spørge ind til hvordan datasættet så ændres fra (xk , yk ) til (xk x, yk y). Herefter skal læren illustrere via skitse at hældningen på datasættet ikke ændres og derved forklare at vi kan bruge formlen fra før med de nye værdier for datasættet. Derefter viser læreren hvordan vi kan bestemme b. Forklaing: Betragt datasættet (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ) med n punkter, hvor n 2. Vi ved at for et centraliseret datasæt er forskriften for y = ax + b givet ved Pn xk yk a = Pk=1 og b = 0 n 2 k=1 xk For et vilkårligt datasæt (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ), som har tyngdepunkt i (x, y) kan vi centralisere datasættet, så det får tyngdepunkt i (0, 0). Vi centraliserer datasættet ved at forskyde det både vertikalt og horisontalt. Det centraliseret datasæt bliver da (x1 x, y1 y), (x2 x, y2 y), . . . , (xn x, yn y). Forskydningen af det nye datasæt ændrer ikke på hældningen for den bedste rette linje, så hældningen bliver: Pn (x x)(yk y) Pn k a = k=1 x)2 k=1 (xk Vi ved at b = 0 for det centraliseret datasæt. Vi har derfor at y y = a(x x) , y = a(x x) + y Dette giver at b = y ax. Det er vigtigt at læreren trækker paralleller til modul 1 og illustrere grafisk hvordan et datasæt kan centraliseret, for at opnå forståelse for at vi nu har udledt formlen for a og b i det generelle tilfælde. Læreren forklarer at vi nu har fundet en formel for a og b, når vi har mere end 2 punkter og skriver afslutningsvis på tavlen: For et datasæt (x1 , y1 ), . . . , (xn , yn ) er den bedste rette linje y = ax + b givet ved Pn (x x)(yk y) Pn k a = k=1 og b = y ax x)2 k=1 (xk Læreren fortæller at vi nu har forklaret matematisk hvordan TI-Nspire udregner den bedste rette linje og vi faktisk også selv kan udregne den. Læreren vurderer herefter afhængigt af tiden om opgave 6 skal regnes eller droppes. Introduktion til bedste eksponentiel/ potens model Timen afsluttes med at læreren forklarer at der også eksisterer tilsvarende formler for eksponentiel og potens regression, men at vi ikke skal bestemme disse. Læreren viser, via eksempel, hvordan data der tilnærmelsesvis kan beskrives ved en eksponentiel/potens model kan transformeres til at blive lineær. Læreren viser hvordan data transformeres i TI-Nspire og laver plot af data. Læreren inddrager eleverne ved at spørge om de kan forklare/huske, hvorfor transformeringen af data giver en lineær sammenhæng. Herefter forklarer læreren at modellen kan findes ved lineær regression på det transformerede datasæt. Læreren skal pointere at ved at ændre datasættet kan vi opnå en lineær sammenhæng, som vi kan lave lineær regression på. Denne introduktion kan rykkes til modul 3, hvis der ikke er tid til det i slutningen af timen. Læreren giver eleverne opgave 7-9 for som lektie. Modul 3: Mandag d. 13/4 9.50-11.25 Formål/indhold • Valg af model (formål) • Matematik kan ikke udelukkende bruges til at vælge model. Inddrage viden omkring datasæt. Ingen modeller er rigtige, men nogle er brugbare Opsamling på opgave 7-9 Eleverne forventes at have lavet opg. 7-9 hjemme. Læreren udvælger to elever, som hhv. gennemgår opg. 8 og 9. Læreren supplere med besvarelser og præciserer svar. Herefter introducerer læreren problemet med bestemmelse af model. Læreren udleverer opgave 10-11 og eleverne arbejder med disse. Opsamling på opgave 10 + 11 Opsamling på opgave 10+11 starter, når læreren fornemmer at eleverne er igennem opgaverne, hvis opgaverne tager kortere tid end afsat. Læreren udvælger to grupper, som gennemgår opg. 10 og 11. Læreren validerer elevernes besvarelser og præciserer svarene. Læreren skal understrege at selvom en model passer godt på data i et interval kan vi ikke være sikre på at modellen kan bruges til at lave fremskrivninger (10d) og der derfor også skal være faglig viden. Læreren skal sikre at eleverne får diskuteret og begrundet valg af model i opg. 11. Mulige begrundelser til at vælge model i opg. 11: • Data følger tilnærmelsesvis en lineær tendens • Begge modeller passer godt på data, da vi har r2 = 0.999 for begge modeller. • Beregning af løberens hastighed sek/km. Dette viser at løberens hastighed ikke er konstant, hvormed en lineær model ikke kan bruges. • Viden om løb: Når distancen øges mindskes hastigheden. Derfor kan den lineær model ikke bruges, da denne forudsætter at hastigheden er uafhængig af distancen. Vi har at tiden øges med 299.47sek pr. km. • Potens model. Ved en potens model vil hastigheden ikke være konstant og vi har at løberens hastighed øges med antallet af kilometer, som vi må forvente. • Ved den lineære model får vi at løberen vil løbe med 297sek/km på maraton, dvs. at han kan løbe et maraton med samme hastighed som 25km. Ved potens model vil løberen løbe med 328.9sek/km på maraton, dvs. løbe over 30sek langsommere hver km end ved 25km. Hvis eleverne ikke selv inddrager r2 -værdien skal læreren inddrage denne. Eleverne arbejder herefter videre med opgave 12+13. Opsamling på opgave 12 + 13 Læreren kan påbegynde opsamling på opgave 12 + 13, når hun fornemmer at eleverne har arbejdet med opgave 13 og er gået i stå med denne. Det forventes at eleverne går i stå med opgave 13, da denne ikke ligner tidligere opgaver og at eleverne ikke kender til model som kan beskrive denne form for udvikling. Læreren udvælger to elever som gennemgår opgave 12. Opsamlingen foregår ligesom ved opgave 10+11. Ud fra lærerens observation af arbejdet med opgave 13 vælger læreren om eleverne skal gennemgå denne eller om hun selv gennemgår den ved at inddrage eleverne. Det er vigtigt at læreren får pointeret at der aldrig findes en rigtig model til at beskrive et datasæt, men at nogle modeller selvfølgelig er mere brugbare end andre. Det er vigtigt at læreren understreger at formålet med modellen er vigtigt, når man vælger model ved at inddrage opgave 10 + 12. Opsamling på valg af model Læreren samler op på de væsentlige teknikker og redskaber til at vælge model, når denne ikke kendes ved at inddrage eleverne. De forskellige teknikker, overvejelser og muligheder for at afgøre valg af model skrives på tavlen. • Hvad er formålet med modellen? • Hvad ved vi om data? (viden om data) • Plot (logaritme plot) • Forklaringsgraden r2 • Ekstrapolation / intrapolation Læreren understreger endnu en gang at der ikke findes noget kriterie for at vælge hvilke type af model der bedst egner sig til at beskrive et datasæt, men når vi har bestemt hvilken type af model så kan vi begrunde med matematikken hvilken der er bedst. list of references a.5 the realized didactic process 183 Modul 1 11 elever (4 fraværende). Der dannes tre grupper. Opsamling på opg. 4 +5a-i, gennemgang af 5j-k + opsamling på modul 1 udskydes Tid 8min 20min 12min 12min 10min 8min 7min 18min Aktivitet Introduktion til bedste rette linje Opgave 1 + 2 Opsamling opg. 1 + 2. Introduktion til bedste rette linje Opgave 3 Opsamling opg. 3 Pause Læreren gennemgår teknik til 4a+b + opg. 5 Opgave 4 + 5 Opgaver der ikke nås/droppes 3i+j 3j til modul 2. Modul 2 14 elever (1 fraværende - 4 nye elever). De nye elever fordeles i de eksisterende grupper. Der går 5min før timen starter. Gennemgang af 5j-k blev droppet. Gennemgang af bevis samt introduktion til Tid 11min 5min 9min 20min 7min 15min 22min Aktivitet Opgave 5 Opsamling på modul 1 Opsamling opg. 5. Introduktion til bevis Bevis. Trin 1 Pause Opsamling trin 1 Trin 2 Opgaver der ikke nås/droppes 5g-5k Dropper 5j-k b+c Kun 1 gruppe når til c bedste eksponentiel/potens model udskydes til modul 3. Opgave 6 droppes. Modul 3 14 elever. 3 grupper. Opgave 8 blev ikke gennemgået, da kun to havde lavet lektier. Tid 20min 6min 8min 24min 7min 18min 6min Aktivitet Gennemgang af bevis (trin 2 for centraliseret data) samt den generelle tilfælde Introduktion til bedste eksponentiel/potens model Opsamling på opg. 9 Opgave 10 Pause Opgave 11 Opsamling opg. 11 + forløb Opgaver der ikke nås/droppes Opgave 12 + 13 blev droppet, da aktiviteter fra modul 2 blev medtaget. Modul 1 11 elever (4 fraværende). Der dannes tre grupper. Opsamling på opg. 4 +5a-i, gennemgang af 5j-k + opsamling på modul 1 udskydes Tid 8min 20min 12min 12min 10min 8min 7min 18min Aktivitet Introduktion til bedste rette linje Opgave 1 + 2 Opsamling opg. 1 + 2. Introduktion til bedste rette linje Opgave 3 Opsamling opg. 3 Pause Læreren gennemgår teknik til 4a+b + opg. 5 Opgave 4 + 5 Opgaver der ikke nås/droppes 3i+j 3j til modul 2. Modul 2 14 elever (1 fraværende - 4 nye elever). De nye elever fordeles i de eksisterende grupper. Der går 5min før timen starter. Gennemgang af 5j-k blev droppet. Gennemgang af bevis samt introduktion til Tid 11min 5min 9min 20min 7min 15min 22min Aktivitet Opgave 5 Opsamling på modul 1 Opsamling opg. 5. Introduktion til bevis Bevis. Trin 1 Pause Opsamling trin 1 Trin 2 Opgaver der ikke nås/droppes 5g-5k Dropper 5j-k b+c Kun 1 gruppe når til c bedste eksponentiel/potens model udskydes til modul 3. Opgave 6 droppes. Modul 3 14 elever. 3 grupper. Opgave 8 blev ikke gennemgået, da kun to havde lavet lektier. Tid 20min 6min 8min 24min 7min 18min 6min Aktivitet Gennemgang af bevis (trin 2 for centraliseret data) samt den generelle tilfælde Introduktion til bedste eksponentiel/potens model Opsamling på opg. 9 Opgave 10 Pause Opgave 11 Opsamling opg. 11 + forløb Opgaver der ikke nås/droppes Opgave 12 + 13 blev droppet, da aktiviteter fra modul 2 blev medtaget. 188 list of references a.6 the students’ solutions Bedste rette linje Opgave 1: a) Lav et plot af punkterne. Kan gøres ved kommandoerne 3: Data, 9: Hurtiggraf. Klik på akserne for at tilføje de to variable. b) Indtegn den linje, du synes passer bedst til punkterne. Linjen ved hjælp af kommandoen 4: Undersøg data, 2: Tilføj flytbare linjer. Hvilken forskrift har linjen? Mange svarmuligheder: f (x) = 34x + 276 m(x) = 36, 6x + 274 f (x) = 37.9x + 275 y = 38, 5x + 273 f (x) = 37x + 275 f (x) = 37.1x + 281 f (x) = 33x + 301 f (x) = 37x + 282 34.9x + 278 f (x) = 38.05x + 270 c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at andre ville finde frem til præcis samme linje som dig, hvis de hørte din begrundelse. Begrundelser: • Har prøvet at sætte den nogenlunde i midten af punkterne samt at den skulle starte i 0 • Går gennem grafens start punkt (venstre hjørne) for at kunne få udviklingen fra år 0 hvor der allerede er en længde. • Grunden til at jeg placerede linjer der er fordi jeg har rettet den til at den starter i (0,0). • Jeg valgte denne linje, da jeg synes den passer bedst til punkterne. Jeg fandt hældningen (a) først, så den lå på 38,5. • At den rette linje skulle være så tæt på flest punkter som muligt. • Jeg valgte denne linje, da det var der, hvor flest punkter blev ramt. Desuden satte jeg "starten" af linjen ved skæringen af y- og x-aksen. 5 punkter bliver her ramt. Vi havde glemt at den ikke skal være i (0,0). • Linjen er sat efter at skabe det tætteste gennemsnit af alle punkterne. Hvilket vil sige man prøver at berører alle punkter i en vis grad. • Jeg tænkte at linjen skulle ramme så mange punkter som muligt, hvilket betød at linjen ikke rammer punkterne helt, men snitter dem. • Jeg valgte linje da det var den linje jeg synes beskrev sammenhængen mellem punkterne bedst og desuden er den placeret i midten af punkterne. Opgave 2: a) Lav et plot af punkterne. b) Indtegn den linje, du synes passer bedst til punkterne. Hvilken forskrift har linjen? Mange svarmuligheder: m(x) = 8.69x + 50 f (x) = 9x + 49 f (x) = 8.42x + 51 8.63x + 48 f (x) = 9.53x + 46 f (x) = 9x + 48 Egen gæt: f (x) = 7.43x + 65. f (x) = 8.85x + 51 f (x) = 3.63x + 102 linjen har jeg placeret i midten af alle punkterne. c) Begrund dit valg af linje: hvorfor valgte du netop den linje i b)? Prøv at formulere din begrundelse på en måde, så du kan være sikker på, at alle vil finde frem til præcis samme linje som dig, hvis de hører din begrundelse. Er dine begrundelser de samme som i forrige opgave? Hvorfor/hvorfor ikke? Begrundelser: • Har prøvet at lave en linje som passer bedst muligt. Lagt den nogenlunde i midten, samt den skærer i 0. • 3 underliggende punkter blev ramt, og derfor valgte jeg denne. Intet punkt ligger nemlig langt væk. Starten blev sat ved skæringen af y- og x-aksen. • Nogenlunde samme begrundelse som før, men da der er mange punkter i øst og vest her laves (sættes) linjen mere udfra hvad jeg mener er gennemsnittet af punkterne. • Linjen har jeg placeret i midten af alle punkterne. • Igen har jeg prøvet og lægge linjen, så den lægger tættest på alle punkterne. De næste spørgsmål skal besvares i grupperne. d) Sammenlign forskrifterne for den bedste linje i opg. 1 + 2. Har I fået den samme? Forklar hver af jeres overvejelser og begrundelser fra 1c) og 2c).ă • Ikke de samme, men nogen der minder om hinanden. Vi har samme fremgangsmåde • I opgave 1 der har vi meget ens ligninger. Samme i opgave 2. • Næ- kun en havde ret e) Diskuter hvad I forstår ved den bedste rette linje. Find kriterier for den bedste linje. • Den der er i midten. Den der er tættest på flest mulige punkter. • Den linje der er i midten og rammer flest mulige punkter. • Tættest på flest mulige punkter (nogenlunde i midten) Noter: N-spire finder regression ved at finde den mindste sum af kvadraterne (axk + b))2 = mindst mulige kvadrat sum. Man skal minimere. P (yk Summer Opgave 3 Udregn summerne. Skriv udregningen til højre på papiret. Skriv resultatet i tabellen. a) Udregn 4 X 2=2+2+2+2=8 k=1 b) Udregn 4 X 3 = 3 + 3 + 3 + 3 = 12 k=1 c) Udregn d) Udregn e) Udregn 4 X k 2 = 12 + 22 + 32 + 42 = 30 k=1 4 X k=1 4 X 4k 2 = 4 · 12 + 4 · 22 + 4 · 32 + 4 · 42 = 120 5k = 5 · 1 + 5 · 2 + 5 · 3 + 5 · 4 = 50 k=1 f ) Udregn 5 · g) Udregn h) Udregn 4 X 4 X (2 + 5k) = (2 + 5 · 1) + (2 + 5 · 2) + (2 + 5 · 3) + (2 + 5 · 4) = 58 k=1 4 X k=1 k = 5(1 + 2 + 3 + 4) = 50 k=1 (5k + k 2 ) = (5 + 12 ) + (10 + 22 ) + (15 + 32 ) + (20 + 42 ) = 80 i) Betragt summerne og dine resultater i tabellen. Redegør for sammenhængen mellem nogle af summerne. Diskuter i grupperne og forklar sammenhængene. j) Udregn når det oplyses at P100 k=1 100 X (2k 2 + k) k=1 k 2 = 338350 og P100 k=1 k = 5050 Opgave 4 I skal nu arbejde videre med datasættet fra opgave 2 med alder og højde. a) Beregn gennemsnittet af alderen i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. 10.47 ⇠ 10.5år. At gennemsnits alderen er 10.5 år. Tallet er den gennemsnitlige alder af de 21 personer b) Beregn gennemsnittet af højden i datasættet fra opgave 2. Gør rede for hvad dette tal betyder. 142.43. At gennemsnits højden er 142.43. Den gennemsnitlige højde afhængig af alderen af de 21 personer. c) Opstil en generel formel til at udregne gennemsnittet (x) af x-værdierne (x1 , x2 , . . . , xn ) ved brug af sumtegn. Pn x k x = Pk=1 n n xP = k=1 xk /n n 1 x n Pk=n k n 1 k=1 n Pxnk 1 x = n k=1 xk d) Opstil en generel formel til at udregne gennemsnittet (y) af y-værdierne (y1 , y2 , . . . , yn ) ved brug af sumtegn. Pn y k y = Pk=1 n n y= Pn k=1 yk /n 1 y n Pk=n k n 1 k=1 n Pyk y = n1 nk=1 yk Opgave 5 a) Plot datasættet fra opgave 2. Indtegn linjer, der viser gennemsnittet af x og y. Linjen, der viser gennemsnittet af x indtegnes ved 4: Undersøg data, 8: Plot værdi. Linjen, der viser gennemsnittet af y indtegnes ved 4: Undersøg data, 4: Plot funktion. Funktionen, der skal plottes, er y = y. b) Bestem den bedste rette linje for datasættet i a). Skriv forskriften for linjen. f (x) = 4.755x + 92.6 f (x) = 4.75504x + 92.6138 f (x) = 4, .6x + 99 y = 4.20x + 100 f (x) = 4.15x + 99 c) Tilføj en kolonne til datasættet, hvor du udregner yk y for hvert punkt i datasættet. Lav et plot af xk og yk y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a). • Det er rykket længere ned på koordinatsystemet. • Centraliseret datasæt. • Når man trækker yk y går datasættet ned i koordinatsystemet. Når man trækker xk x går datasættet mod venstre. • Punkterne ligger på samme måde. De er bare forskudt. • De ligger på samme måde, men er blevet forskudt. • Prikkerne har ikke ændret udseende, men koordinatsættet er gået i minus • Man kan se på plottet at det starter i negativ ligeledes som på grafen • De ligger stadig ens i forhold til hinanden, men de er rykket i grafen. d) Bestem den bedste rette linje for datasættet plottet i c). Skriv forskriften for linjen f (x) = 4.755x 49.81 f (x) = 5.4x 0.83 y = 5.5x + 0.9 f (x) = 3.72x 0.1 P e) Udregn nk=1 (yk y). =0 Udregnet til 3.1E 11 ⇡ 0 e 11 3P ⇡0 n y) = 3E 11 = 0 k=1 (yk f) Tilføj en ny kolonne til datasættet, hvor du udregner xk x for hvert punkt i datasættet. Lav et plot af xk x og yk y. Forklar hvad der er sket med placeringen af datasættet i forhold til plottet i a) og c). Placeringen er centraliseret (er placeret omkring 0) g) Bestem den bedste rette linje for datasættet i f). Skriv forskriften for linjen. h) Sammenlign forskrifterne bestemt i b), d) og g). Gør rede for sammenhængen mellem forskrifterne. P i) Udregn nk=1 (xk x). j) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4c), at det altid gælder at n X (xk x) = 0 k=1 k) Vis, ved brug af regnereglerne for summer og formlen for gennemsnittet (fra opg. 4d), at det altid gælder at n X (yk y) = 0 k=1 Bestemme bedste rette linje for et centraliseret datasæt I skal bestemme den bedste rette linje y = ax + b for et datasæt (x1 , y1 ), (x2 , y2 ), . . . , (xn , yn ), bestående af n punkter, hvor n P 2. ViPantager at datasættet er centraliseret, dvs. at x = n1 nk=1 xk = 0 og y = n 1 k=1 yk = 0. n Vi skal bestemme a og b, så kvadratsummen n ✓ X yk axk + b k=1 ◆2 bliver mindst mulig. Trin 1) Det første trin består I at omskrive summen, så at vi kan minimere denne i trin 2. a) Gør rede for at yk Mulige teknikker: yk (axk + b) 2 (axk + b) = yk 2 = yk2 2axk yk (axk + b) · yk 2byk + a2 x2k + 2abxk + b2 (axk + b) = yk axk b · yk axk b = yk2 yk axk yk b axk yk + a2 x2k + axk b = yk2 2axk yk 2byk + a2 x2k + 2abxk + b2 byk + baxk + b2 yk (axk + b) · yk (axk + b) regne direkte på denne 2 2 yk + (axk + b) 2yk (axk + b) 2 2 2 2 = yk + a xk + b + 2axk 2yk axk + 2yk b (yk (yk yk 2 (axk + b)) · (yk (axk + b)) axk b) · (yk axk b) axkyk byk axyyk 2 + axk 2 + axkb Herunder ses de forskellige besvarelser: ykb + axkb + b2 P P 2 b) Vi har at nk=1 yk (axk + b) = nk=1 yk2 2axk yk 2byk + a2 x2k + 2abxk + b2 . Brug regnereglerne for summer (1-4) til at dele summen op, sætte a, b foran summationen og summe over b. c) Brug antagelsen om at gennemsnittet af x og y er 0 til at simplificere udtrykket fra b). Trin 2) a) Gøre rede for hvad A, B, C svarer til i summen og forklar hvorfor A, B, C er konstanter. • Viser med streger hvad der er A, B, C. A, B, C er konstanter, da det er de givne tal. • Har skrevet A, B, C under de tre summer. De er konstanter fordi de ikke afhænger af a og b og fordi punkterne er fastlagt. P P P • A = nk=1 x2k , C = nk=1 yk , B = nk=1 xk yk . for at gøre det lettere at regne med. • De er konstanter fordi de ikke afhænger af a og b. Punkterne er fastlagt, derfor er det også en konstant. • De er konstanter, fordi de ikke afhænger af de to variabler a,b og fordi x og y er fastlagt. • Vi har fået opgivet nogle værdier, så derfor er de konstanter. • Konstanterne A, B, C svarer til de værdier som bliver opgivet som datasæt. • x og y er variable, mens A, B, C er konstanter, fordi disse er givet i et datasæt. • Fordi punkterne er fastlagt. b) Udtrykket a2 A 2aB + C + nb2 kan betragtes som en sum af to andengradspolynomier med hhv. a og b som variable. Skriv de to andengradspolynomier som udtrykket består af. • a2 A • 1) a2 A 2aB. nb2 + C. 2aB + C. 2) nb2 • Vi kan variere a og b. Ene polynomier: a2 A 2aB +C. Andet polynomier: nb2 . • Vi kan variere a og b. a2 A 2 • Aa + 2Ba + C = 0 • Aa2 2Ba = 0 2aB + C. 2 nb + C = 0 nb2 + C = 0 • a2 A 2aB omskrevet til ) Aa2 + 2Ba + 0 = 0 og nb2 + C omskrevet til ) nb2 + c = 0. c) Bestem a og b, så a2 A 2aB + C + nb2 bliver mindst mulig. Brug fra b) at udtrykket kan betragtes som to andengradspolynomier, som hver skal minimeres. b=0 fordi det skal være mindst muligt. b=0. Toppunktformlen: 2ab ) 2B :B 2A A b=0 Toppunktsformlen 2B =B 2A PA xkyk a= xk2 b=0 Toppunktsformlen: 2B =B 2A PA a = Pxxk 2yk b 2a b 2a k d) Opskriv værdierne for a og b som gør kvadratsummen mindst mulig. Pn k=1 ✓ yk axk + b ◆2 Opgave 7: Verdensrekorder a) Bestem a og b i modellen b) Hvad kan vi forvente at verdensrekorden bliver i 2014 ifølge modellen? c) Verdensrekorden blev i 2014 slået, så den nu er på 7377. Hvordan passer det med c)? Benyt oplysningen til at kommentere på modellens rækkevidde. Opgave 8: Dykkersyge a) Opstil en model, der beskriver sammenhængen mellem dybde og antal minutter. Bestem forskriften for modellen. b) Benyt modellen til at udregne hvor længe dykkeren kan opholde sig på en dybde på 30m. c) Benyt modellen til at udregne den dybde, dykkeren kan dykke, hvis han gerne vil dykke så dybt som muligt og opholde sig på dybden i 20min. Opgave 9: Vindmølleenergi a) Opstil en model, der beskriver sammenhængen mellem årstal og kapaciteten af vindmølleenergi. Bestem forskriften for modellen. b) Benyt modellen til at beregne kapaciteten af vindmølleenergi i 2014 og sammenlign med den faktiske kapacitet af vindmølleenergi i 2014, som var 369.6M W . Opgave 10: Spædbørnsdødelighed Besvarelser: a) Model: bax . Den eksponentielle funktion har den højeste r2 -værdi. b) Ved regression a = 0.96, b = 1.48. c) ba2008 d) Det passer ikke. a) Eksponentiel regression passer bedst, da r2 -værdien er tættest på 1. b) a = 0.955 og b = 1.484 · e40 c) f (x) = 1.484 · e40 · 0.9552008 d) Modellen afviger en smule og er derfor ikke helt troværdig a) Forklaringsgraden er næsten lige så god for dem begge. Derfor har jeg bare valgt den eksponentielle funktion f (x) = b · ax . b) f (x) = 1.484 · 0.955x c) Promillen for børnedødelighed i 2008 er på 2.402. d) Dårligt - modellen vil ikke nødvendigvis være realistisk. Det kan ikke fortsætte med at falde så meget og det er ikke præcist at bruge det til at forudsige noget. a) Ud fra datasættet er det svært at bestemme om det er en potens eller eksponentiel funktion. Men i dette tilfælde vælges der eksponentiel funktion f (x) = b · ax . b) Tager det ud fra år efter 1933. f (x) = 73.58 · (0.96)x c) 73.58 · (0.96)7 5 = 3.44 a) f (x) = 1.48368·1040 ·(0.955397)x . Vi tjekkede forklaringsgraden. De passede begge godt, men den eksponentielle passede best. Jo tættere på 1, jo bedre passer den. b) Se a) c) x sættes til 2008 og det beregnes at være 2.40192 d) Det passer dårligt. Modellen er et udtryk for en sammenhæng der ikke nødvendigvis er. Opgave 11: Løb Besvarelser: a) a = 299.5 og b = 103.5 b) a = 1.1, b = 219.9. 13879.9 a) f (x) = 299.467x 103.53.f (x) 12532.5sek.208.875min = 3.48timer. b) 4.55timer. c) Nej. = 299.467 · 42.195 103.54 = a) 299.467x +-103.544 Et maraton tager 12474.1sek b) f (x) = 219 · x1.1076 . c) Potens funktion - 13809sek. Lineær funktion - 12474.1sek. d) Jeg ville bruge potens funktion, da der i den lineære skal vedkommende løbe med en konstant fart. Hvilket nok ikke er så realistisk. a) a = 299.467, b = 103.544. f (x) = 299.467 · 45.195 + ( 103.544) = 12532.5sek tager et maraton. b) a = 219.913, b = 1.1076. f (x) = 219.913x1.1076 . Maraton = 13880sek. c) Nej! 13880 12532.5 = 1347.5sek forskel på de to modeller. d) Potensfunktionen, da er mere virkelighedstro, da der er svingende hastighed. a) f (x) = 299.467x 103.544. Tid om et maraton: 12532.5sek. b) f (x) = 219.913 · x1,1076 . Sekunder for et maraton: 13879.9 c) Nej det er den ikke. d) Vi ville bruge potensfunktionen, da man ikke kan holde den samme fart i 42km. Selvom at forklaringen passer bedst for den lineære funktion. 202 list of references a.7 the transcriptions in danish 1. Gruppe 1 diskuterer valg af linje i opg. 1 C: Hvilken forskrift har I fået for jeres linje? (intet svar) D: Den er faktisk meget god. Men jeg ved ikke hvorfor. Vi har prøvet at tage hensyn til at der er sådan nogenlunde samme mængde på begge sider. Altså med sådan afstand. B: Altså det er ikke et punkt er det. A: Jo. D: Men du kan se at der er to som er rimelig langt væk fra.. Fra vores linje. Og der er også to her der ligger rimelig meget på. A: Det faktisk svært at begrunde. 2. Gruppe 2 diskuterer valg af linje i opg. 2 B: Men hvad er jeres begrundelser for at I har valgt den? C: At den skulle ramme så mange punkter som muligt. A: Ramme, altså ramme dem. C: Ja, sådan her. A: Så det er lige meget den skulle bare ramme. B: Hvad med den der sidder i midten? D: Der må være nogle der er uden for. C: Man kunne måske godt have lagt den sådan lidt mere hernede. Det ville nok være mere rigtigt. A: Altså jeg har skrevet det der med at der skal være nogenlunde lige mange punkter med samme side. Altså med samme afstand til linjen. C: Men også bare. Hvis man ligger den her. Så er de jo bare virkelig langt væk fra denne her. Fra det her punkt. 3. Gennemgang af opgave 2 på tavlen. Eleverne kommer med kriterier C: At afstanden fra punkterne til linjerne skal være nogenlunde lige stor på begge sider L: Så tænker du noget med at den her afstand ind til linjerne skal C: Ja, altså alle punkterne tilsammen skal have nogenlunde samme afstand. C: Altså alle punkterne på den ene side, deres samlede afstand skal være nogenlunde den samme som den anden side. L: Ja, så den samlede afstand. C: Ja, skal være lige stor på begge sider. L: Ja. 4. Kommentarer i gruppe 1, da læreren viser den bedste rette linje C: Arg B: Det kan du godt glemme alt om. A: Fuck. C: Så ved man at man kan ikke bruge øjemål. Det er (laver en list of references skuffende lyd) B: Øjemål. Dårligt. 5. Gruppe 1 arbejder med opg. 3 A: Jeg forstår ikke helt. Hvad er k. K lig 1. Indekset. C: Det er det det starter fra. A: Nej det er det ikke. C: Eller det stiger med en. A: Nej. Prøv at kigge på eksemplerne B: Er det så ikke bare at man skal sætte 2 op fire gange. A: Jo det vil jeg også skyde. D: Så det er 2 + 2 +2 +2 6. Gruppe 1 diskuterer opg. 4c B: Vi skal lave. Det 5 tal der står over sumtegnet om til n. Og så skal vi sige. A: Nej B: Jo A: Det der står under er n. B: ja, det kommer også til at hedde n, men det der står over kommer også til at hedder n. C: Men er det ikke bare sumtegn så ovenover n og så det der dytelyt. Og så k. Og så dividere man 7. Gennemgang af del 1 i beviset. L: Okay. Vi har fået delt det op på hvert sit sumtegn. Vi har brugt en regneregel. Nu ser det sådan her ud. Hvad tænker I så. Hvilke regneregler kan vi så bruge nu? [Forklare de to regneregler] B: Kan vi ikke bruge et’eren? L: Jo, hvor vil du bruge et’eren henne. B: Er det ikke der hvor der står k L: k? C: Nej. A: Altså jeg vil bruge, ja nr. 1 som han siger. Men der hvor vi har en konstant. Og det er for eksempel der hvor der står 2 foran et bogstav. L: Ja, det er rigtigt. 2 er en konstant. Er der andre bogstaver, jeg ved godt det er bogstaver, men er der andre bogstaver der er konstanter heroppe? Hvis vi lige prøver at tænke. Hvad er vores variable og hvad er vores konstanter? A: Dem der står i anden. Så det er a i anden. 8. Gruppe 2 arbejder med forflytning af datasættet A: Der er sq da ikke sket noget. [ Kalder på læren ] A: I c. Så skulle man øhh lave de der udregninger vi har lavet og 203 204 list of references indsætte dem med et plot. Men på vores giver det det samme som altså de samme prikmæssigt. L: Når I har trukket gennemsnittet fra. A: Ja, altså sådan her. Vi tegner dem. L: Så dem I plotter nu A: Prikkerne ligger jo ens nu. Er det meningen? L: Ja B: Men prøv at se tallene her. L: Men de er ikke helt ens. Hvad er forskellen? B: Den går i minus A: Ja ja, y og x værdien’s værdier. L: Ja. Så det at du siger det er de samme. Hvad mener du med det? A: jeg mener bare at prikkerne ligger samme sted, men L: Ja B: Bare på forskellige, de ligger i forhold til C: Hvis du satte det op i et andet koordinatsystem så ville de ikke ligge de samme steder. 9. Starten på opsamlingen af opg. 5 L: Opgave 5. Hvad går den ud på? A: Med at skrive en masse tal ind i Nspire. L: Ja. Det har I gjort. Men hvad går opg. 5 ud på. Hvad har I fundet ud af? A: Udregnet gennemsnit og forskel. L: Så gennemsnittet af hvad A: Gennemsnittet af alder og højde L: Ja, så vi har. Gennemsnittet af alderen og gennemsnittet af højden. Hvad skulle vi bruge det til? A: Til at indsætte nogle plots. I.. L: Ja ? og øhh? A: Vi er ikke nået så langt endnu. L: Har du noget at supplere med elev B? Eller hvad ville du at sige? B: Nej, øm. Nææ. Jo, så har vi bestemt. Den lineære funktion for den. L: Ja, for hvad. B: For plottet, for punkterne. L: Ja, så B: Ja, vi har lavet lineær regression. L: Alder og højde. B: Ja. L: Ja, så hvis vi har her. Alderen er vores x’er, højden er vores y’er. Så ligger vores datasæt et eller andet sted heroppe (viser på tavlen). Det har I alle fundet Klassen:Ja L: Hvad bruger vi det her gennemsnit til? Hvad har I brugt list of references gennemsnittet til. A: Udregnet forskellen. L: Så forskel mellem x og gennemsnittet. Det samme med y. Og nu skriver jeg det med lille k hernede. Taget alle x?erne og fratrukket gennemsnittet. L: De her to forskelle. Hvad har I brugt dem til? D: Til at lave et nyt plot/graf. L: Ja, så har vi lavet et nyt plot. Så har vi trukket gennemsnittet fra i x’erne og y’erne. Hvordan ser det nye plot ud? E: Magen til. Bare i minus. L: I minus? F: Noget af det. L: Noget af det er i minus? F: Halvdelen er ca. i minus L: (Læreren viser på tavlen hvordan datasættet er flyttet) L: Hvordan ser datasættet ud? Hvad er det centraliseret om? Elev mumler. 0 G: 0. L: Ja, 0.0 komma hvad? Klassen. 0 L: Ja, så vi har forskudt vores datasæt fra at være heroppe til at komme herned (viser på tavlen). 10. Gruppe 2 diskuterer bevis opg. 2b B: Jeg forstår ikke engang spørgsmålet. Som en sum af to andengrads polynomier. E: Kan de der a,b og c ikke være det samme som x. Altså ax i anden + bx + c. Bum L: Ja. Hvad kunne så være x. E: Det kunne være a, b og c. E: Store A, B og C. L: Store A,B, C var hvad? B + E: Det var konstanter. L: Ja, så hvad er det der skal variere her? E: Det skal lille a, b og c. B: Variere a og b. L: Hvilket kunne være et polynomium? B: Skal man se bort fra konstanterne L: Må man se bort fra konstanterne når man laver et polynomium C + E: Nej. L: Så hvad kunne det være polynomium være? Tænk på a’erne som det I plejer at have som x’ere. 11. Gruppe 1 diskuterer bevis 2b B: Jeg kan ikke forstå hvorfor det er to andengradspolynomier. Jeg kan godt se ideen med at det er et, men jeg kan ikke se hvor- 205 206 list of references dan det skal blive til to. B: Men jeg kan ikke få det til at passe. B: vi forstår ikke hvordan det er to andengradspolynomier? C: For der mangler et C? L: Hvis I nu tænker på a og b, som de variable. Det er dem vi normalt kender som x’er. Hvordan kan I så dele det der udtryk op, så a’erne for sig og b’erne for sig. Og nu snakker jeg om de små bogstaver. Kan I dele det op på en måde så a’erne er for sig og b’erne er for sig. C: Dele det op. L: I får at vide at der skal være to andengradspolynomier B: Ja L: så I skal have to udtryk. Hvad kunne det ene udtryk hedde? Noget med a’erne. De led hvor små a’er i. B: a i anden + nb i anden L: a’erne og b’erne må ikke være sammen. For det skal være to andengradspolynomier og de har dem som variable. Så I skal tænke på at de er x’er nu. Som normale x?er i et andengradspolynomium. E: Så bare sig. B: jamen så er der jo 3. E: De små a’er er x’er. dvs. stort A f.eks. lille a i anden ? 2aB. E: + C. Nej det må jeg ikke. L: Jo. + C kan du også. For den kan høre med til begge for den afhænger ikke af noget. E: Okay. [Læren viser på tavlen ] L: Hvad er der tilbage? C: nb i anden. E: Ja. B: Det kan jo godt passe. B: Store C må du ikke bruge to gange. Så kun med til det ene polynomium. L: Kan I se det. I kor: Ja! Got it. E: Jeg kan godt forstå det. Det er bare forvirrende at man skal bytte om på tallene. Men det er til at forstå. 12. Kommentarer efter beviset er gennemgået A: Så det er ikke sådan her vi skal gøre når vi skal beregne det. Det er bare et bevis! L: Heldigvis gør Nspire det her for jer. Flere: ja tak Men det er det her, der ligger til grunds for det. B (i gruppen til A): Så nu ved du hvad Nspire laver. L: Så hver gang I trykker regression i Nspire, lineær regression i Nspire, så regner den a værdien på den her måde og b værdien list of references på den her måde. A: Det er et bevis! L: Lige præcis. 13. 2 elever gennemgår opg. 9. Kun den ene (B) har lavet opgaven hjemme. A: Hvorfra ved I at det er en eksponentiel model? B: Det kan man se på den der (peger på figuren) C: Hvorfor ikke en potens? B: Nej. Sådan er det bare. L: Hvorfor ikke? Det er gode spørgsmål B: Det er bare en eksponentiel funktion D: Jamen hvorfor? E: Kan I ikke huske reglerne for det? [Snak i grupperne] E: Det så pænest ud. L: Hvordan ville I finde ud af om det var en potens eller eksponentiel? [larm] L: Det er nogle rigtige diskussioner I har gang i. Det er faktisk det vi skal snakke om i den her time. Hvordan finder vi ud af hvilken model vi skal bruge når vi har nogle tal. 14. Gennemgang af opg. 10 L: Hvordan har I grebet opg. 10 an? A: Til at starte med kiggede vi lidt på den der forklaringsgrad. Så vi lavede regression på eksponentiel og potens funktion. Ens forklaringsgrad. Bruge begge modeller. Valg eksponentiel, da den passe lige lidt bedre. [ Opgave b + c bliver gennemgået ] L: Hvad siger I til d’eren? B: Dårligt C: Når vi kigger på modellen på vores hurtiggraf så kan vi se at den falder sådan rimelig meget de der årstal der er blevet oplyst. Og vi kan ikke forvente at den bliver ved med at falde, så vi bliver nødt til at sætte modellen op mod virkeligheden. Realistisk kommer til at ske. Og det kan jo ikke blive ved med at gå ned hele tiden. Derfor kan det godt passe at dødeligheden har været 9.9 og ikke 2.4. 15. Gruppe 3 arbejder med opg. 11a C: Har I lavet den lineær funktion? D: Jeg er ikke kommet ind endnu. C: Det giver ikke mening for mig. D: Det gør det da. C: Altså jeg starter på minus det der sekunder. Det giver ikke mening. Men jeg kan godt finde ud af at gøre det D: Prøv at bytte om på x og y. Så tid er x og 207 208 list of references C: Nej. D: Hvad er det der ikke giver mening C: Nej nej. Hvis du zoomer ind. Den tager jo gennemsnittet af det. Den ved jo ikke bedre. C: Så står der bestem b i den lineære model. Skal jeg ikke bare skrive nul fordi sådan ser den ud C: Ja ja, men jeg tænker bare sådan nul. Jeg starter på 0. Altså hvis vi tænker logisk. L: Det rigtigt. ( Forklare) [Diskutere videre] B: Til elev C - Hvorfor forstod du at b-værdien skulle være minus (lineære model) C: Logisk tænker man at den skal være 0. Men man kan sige. Man skal tænke på at Nspire ved jo ikke hvad tallet skal være. B: Nej C: Så du skal bare skrive det Nspire siger. 16. Gruppe 3 diskuterer opg. 11d L: Hvilken model vil I vælge? A: Jeg tænker umiddelbart.. Umiddelbart potens E: Lineær L: Hvorfor? E: Fordi tallene ser mest realistiske ud L: Ser det mest realistisk ud. Hvad med ud i fremtiden? Bevæger man sig med.. C: Nej, man bevæger sig ikke med samme hastighed.. Man altså. Jeg vil sige at når man løber starter man med hurtigere fart og så slutter man af og så slutter man af med en slutspurt. Så jeg vil faktisk ikke sige nogen af dem. L: nej, okay. Så det kan godt være at ingen af modeller er specielt gode til maraton. E: Det rigtigt nok. B: Men ifølge den der r værdi så passer de jo godt. colophon This document was typeset using the typographical look-and-feel classicthesis developed by André Miede. The style was inspired by Robert Bringhurst’s seminal book on typography “The Elements of Typographic Style”. classicthesis is available for both LATEX and LYX: http://code.google.com/p/classicthesis/ Happy users of classicthesis usually send a real postcard to the author, a collection of postcards received so far is featured here: http://postcards.miede.de/ Final Version as of June 27, 2015 (classicthesis myversion).