752
Views
3
CrossRef citations to date
0
Altmetric
Practice Paper

‘A Picture Is Worth 10,000 Words’: A Module to Test the ‘Visualization Hypothesis’ in Quantitative Methods Teaching

, &

Abstract

Inadequate quantitative methods (QM) training provision for undergraduate social science students in the United Kingdom is a well-known problem. This paper reports on the design, implementation and assessment of an induction module created to test the hypothesis that visualization helps students learn key statistical concepts. The induction module is a twelve-week compulsory unit taught to first year social science students at a UK university, which they complete prior to a more traditional statistical, workshop-based QM module. A component of the induction module focuses on the use of visualization through Geographic Information Systems (GIS), to teach the process of hypothesis generation to students while they also are introduced to the basics of QM research design and univariate and bivariate forms of data analysis. Self-reflexive evaluation indicates that visualization could assist students with more advanced QM statistical skills.

Background

Statistically numerate and critically informed social science graduates are vitally important to the British economy (CitationBritish Academy 2012). Yet QM provision varies in the United Kingdom (CitationLinden 2012, MacInnes 2009, Parker et al. 2010). Inadequate numerical skills are noted as a problem across the UK (CitationLynch et al. 2007, McVie et al. 2008). Curriculum initiatives have been sought by British funding councils and learned societies (CitationNuffield Foundation, ESR, HEFCE, 2012) to ameliorate the quantitative skills deficit (CitationChamberlain et al. 2014).

Graphical methods are a powerful means of conveying statistical messages to less numerically literate audiences, as effectively exemplified by Professor. Hans CitationRosling (2014) in productions such as The Joy of Statistics (CitationBBC 2013). Visualization has been proposed as a mechanism to teach QM to social science students more effectively (CitationMacInnes 2012). This recommendation follows CitationParker et al. (2010, p12) who assert, based on a review by CitationGarfield & Ben-Zvi (2007), that “exploring visual aspects of data helps students to learn basic statistical concepts”. Visual approaches are also more widely advocated, for instance for younger students (CitationBalka et al. 2007, CitationLewis 2010, CitationSobanski 2002). Moreover, the belief that a visual approach conveys an idea more effectively than words is succinctly expressed in the saying “A picture is worth 10,000 words” (CitationLarkin & Simon 1987, p65).

The work reported in this paper is part of a research project (see CitationChamberlain et al. 2014) to provide an evidential base to assess the use of visual methods in QM teaching to social science undergraduates, who are typically anxious and report low levels of statistical ability (CitationUnrau & Beck 2004, CitationZeiffler et al. 2008). Utilizing developments in visualization software is commonly argued to improve performance at all mathematical levels (e.g. CitationEliëns & Ruttkay 2009, CitationKim & Kim 2003), and may lead to innovative QM teaching (CitationMacInnes 2012). However, the authors of this paper chose to integrate visualization through using mapping technology (CitationMacEachren et al. 2004, CitationSmith et al. 2013), because mapping technology is seen as being increasingly important for employers and social science research in a range of areas, such as poverty, health, and crime (CitationBoba 2005, CitationLloyd 2001, CitationRiner et al. 2004).

The visual component of the QM induction module was structured around an experiential learning model (CitationKolb 1984), problem-based learning (PBL) (CitationSolem 2001) and continuous feedback and reflection (CitationFink 2003) with a focus on hypothesis generation and repeated interaction with key statistical concepts. Experiential learning or ‘learning by doing’ is a cyclical process based on the idea that experiences gained through activities (for instance, real-life experiences or simulated in a learning context) facilitate learning (CitationFry et al. 1999). Despite the critiques of Kolb’s experiential learning model (see for instance CitationKayes 2002), it remains one of the most influential learning models (CitationHealey & Jenkins 2000). As will be explained in the next section, it is composed of four distinctive learning stages all of which are important for attaining effective learning. While the Kolb’s model describes the process through which the GIS workshops were organised, PBL – defined as “the ability to conduct independent inquiry into real-life issues” (CitationMacVaugh 2009, p1) – presented real problems to students with the aim to elicit their interest and drive their learning (for a useful list of web resources for PBL see CitationHEA 2005).

The module used Geographic Information Systems (GIS) to support this approach through the use of aspects of the process of visualization that include the ‘exploration’ of data (CitationMacEachren et al. 2004, CitationSmith et al. 2013); the aims of this being to (i) support students throughout their course, (ii) introduce the key statistical concept of hypothesis generation, and (iii) alleviate student anxiety of QM.

The GIS component

The GIS component is part of a twelve-week induction module for first year students enrolled either in a BSc (Hons) Sociology programme or BSc (Hons) Criminology and Social Policy programme. The module is delivered in semester one, preparing students for a second part more traditional QM statistics module taking place in semester two (see CitationChamberlain et al. 2014). The module consists of eleven one-hour long weekly lectures and ten weekly workshops lasting two hours, with the GIS component delivered in its last five workshops (). The initial five weeks were designed to introduce students gradually to key concepts (research design, sampling, levels of measurement) that are prerequisites for the hypothesis generation work that the GIS element aims to illuminate. The lectures introduced the theory, while in order to lessen students’ fear of QM the workshops put the theory into practice in an environment which facilitated interaction with the tutors. Critically, numbers and equations were avoided in teaching the concepts with the intention of preventing students erecting barriers towards QM.

Table 1 Induction module structure.

GIS-based visualization is thought to motivate students in their learning, with stimulation tools being used to explore patterns and relationships between visualized variables (CitationForbes 2012, CitationOrford et al. 1998). These aspects of visualization have been described in a framework consisting of four functions; from ‘exploration’ and ‘analysis’ of data to the ‘synthesis’ and ‘presentation’ of information (CitationMacEachren et al. 2004, (CitationSmith et al. 2013). The GIS component was designed in this framework with the intention of pushing students beyond the presentation of static images into the exploration area of visualization ‘space’ (CitationMacEachren et al. 2004).

The problem-based learning approach used during the GIS workshops aimed to stimulate and drive student learning by setting a ‘question’ (CitationSolem 2001). This provided a subject-relevant focus for each workshop, directly developing students’ ability to investigate social data. The questions () reflect real issues in criminology, social policy and sociology, which students explored through active learning (CitationHorton 2001) following the experiential learning cycle (CitationKolb 1984). The structure of each workshop and its relationship to the stages of the cycle are shown in and described below. In another level of cyclicity, although they used a progression in the statistics employed () the four main workshops deliberately contained similar and repeated aspects. The intended purpose of the repetition was to reduce student anxiety through accustoming them to similar scenarios where hypotheses needed to be formulated. A fifth workshop focused on how to produce clean, well laid out maps which students would then include in their project work.

Table 2 GIS course structure, problem-based questions and activities.

Figure 1 Workshop structure and the learning cycle (adapted from CitationKolb 1984).

Each workshop started with a presentation by the tutor, followed by a guided exercise and an assignment. The tutor’s presentation introduced the ‘problem’ using relevant data from the Crime Survey for England and Wales, the Index of Multiple Deprivation Crime Dimension and 2011 Census data: ‘concrete experience’ (CE) in the Kolb cycle. After this, students experienced the set ‘problem’ hands-on through a guided exercise. Each guided exercise had a geographical focus on the students’ locality (i.e. Leicestershire) and took them through the process of hypothesis generation which underpins all inferential statistics. These exercises were punctuated by ‘reflective observation’ (RO), as this enriches active learning (CitationFink 2003, (CitationWall 2013), and interaction with tutors. Throughout each exercise students had to explicitly reflect on the data visualized, the patterns identified, and specifically reflect on the problem through their personal knowledge or academic theories. In the course of these reflection steps students were led to summarize their thinking about the patterns identified through the development of null and research hypotheses: ‘abstract conceptualization’ (AC). Students were thus prepared for the main AC stage of the cycle, which is the presentation of findings in the form of maps and a short report after undertaking ‘active experimentation’ (AE) in doing the assignment. The assignment required students to choose their own variables and go through the visualization ‘functions’ of exploration and analysis of geographical patterns, followed by synthesis and presentation of findings in the form of maps and a short report. To provide further support, feedback was given to each student on a draft of their assignment (), after which students would revise their work. This is conducive to deep learning (CitationFry et al. 1999) – that is learning aimed at understanding rather than reproducing material for the purpose of passing an exam (CitationCase 2008).

So, while the ‘GIS problem-based learning cycle’ of the induction module was developed to test the assertion that ‘exploring visual aspects of data helps students learn basic statistical concepts’ by considering hypothesis generation, the GIS component also implements other principles of effective teaching of statistics (CitationGarfield & Ben-Zvi 2007), by providing an environment where students can learn by (i) constructing knowledge out of their coursework, (ii) being actively involved in learning activities, and (iii) receiving timely feedback.

The GIS component in practice

The GIS component was developed with the use of the mapping software ArcMap10.1, part of ArcGIS desktop (ESRI). As the main purpose of the GIS component was to help students familiarize themselves with the process of hypothesis generation and lessen their anxiety towards QM, the use of ArcMap was carefully structured in order to minimize technical problems which may hamper learning. This was achieved by detailed step-by-step instructions within the exercises and by strongly advising students to use ArcMap during the workshop session where technical help could be provided in person by the tutor.

The first two GIS workshops illustrate well the methods and principles used, so the rest of this section will examine them in more detail. The first GIS workshop was on learning how to generate a research hypothesis about a single variable (crime). Students were guided in the visual investigation of the geographical patterns with the question: ‘Is crime higher in urban areas than rural areas?’ To investigate this, students were asked to inspect colour-coded maps (e.g. ) which displayed crime data for Leicestershire; to identify spatial patterns such as whether crime is high or low in different areas and to consider how the use of colour affects readers’ visual impression of the underlying data (CitationMitchell 1999). Specifically they were asked if high crime levels are displayed in urban areas compared to rural areas. Then, students converted such questions, for example:

  • Crime levels are greater in urban areas than in rural areas in Leicestershire (Research Hypothesis; H1).

  • There is no difference in crime levels between urban areas and rural areas in Leicestershire (Null Hypothesis; H0).

Figure 2 Map of crime deprivation in Leicestershire. shows the map of crime deprivation in Leicestershire as quantified by the crime score variable and used to colour Middle Layer Super Output Area (MSOA) units. District boundaries are bold, whilst MSOA boundaries are fainter. Districts and more significant urban areas are named, with the town and city names underlined in white.

The assignment replicated this procedure for another variable of each student’s choice.

The statistical aim of the second workshop was to generate research hypotheses about the relationship between the two variables and interpret them. The question asked for the guided exercise was: ‘Is there an association between crime and fear of crime?’ Students generated hypotheses and expectations of any link based on their visual interpretation of the two variables, initially on a scatter plot (). They formulated testable hypotheses:

  • H1: There is an association between crime score and fear of crime.

  • H0: There is no association, and any apparent link results from random variation in the selected sample.

Figure 3 Scatter plot of crime score and fear of crime variables at the Middle Layer Super Output Area level. In the scatter plot in there are 120 data (circles), and a trend line has been fitted by ordinary least squares (OLS); note that OLS is arguably not appropriate where error exists in both x and y variables, but it is commonly used and adequate where the relationship is strong.

Then they investigated these variables on maps to include the spatial information which a scatter plot cannot (), and refined their hypotheses and expectations accordingly, for example:

  • H1: Levels of fear of crime are greater in urban areas where crime is high than in rural areas in Leicestershire where crime is low.

Figure 4 Map of Leicestershire showing the distribution of crime. The map of Leicestershire showing the distribution of crime in is illustrated by Middle Layer Super Output Area in shades of grey, against the fear of crime variable which is displayed by the density of the red dots (1 dot = 0.03 units that is 3% of people in fear of crime in an area. Dots are scattered randomly within each area and do not indicate actual location of people.)

The assignment replicated the guided exercise; students selected either the crime index or the fear of crime variable and explored its association with another social variable of their choice from the 2011 Census.

Student assessment

The induction module has a weighting of ten credits, and represents 100 hours of directed and self-guided work. It is summatively assessed through a portfolio, with the four GIS components comprising 50% of the total (). Such a portfolio of achievement (CitationHounsell et al. 2007) is well suited to evaluating students’ progress in understanding key concepts (CitationThomson et al. 2008). It was also intended to support the students’ learning.

Table 3 Portfolio elements and weights.

Firstly, one of the intended functions of the induction module was to relieve students’ anxiety by introducing continuous formative assessment, which emphasises how much they have learned (CitationRust 2002). Providing feedback to students throughout the GIS workshops was included to promote deep learning as students could understand and progress from their ‘mistakes’.

Secondly, a key factor in the module design was the focus on getting students to positively engage with QM. The GIS weekly assignments were intended to encourage students to take ownership of their learning, a sort of ‘homework’ (as one student put it), supplemented by the dedicated time in each workshop to work on the assignment while tutors were present.

To deter disengagement, an approach was adopted where students were asked to present an acceptable first draft of their GIS assignment for ‘signing off’ by a tutor before the end of the subsequent week’s practical session. A ‘good draft’ included sections stating the hypotheses and a number of relevant maps, and verbal feedback was given individually. This one-to-one exchange between tutor and student also gave students the opportunity to ask questions.

Final, compiled summative portfolios included at least the formatively assessed materials from the workshops, but students were free to move beyond this minimum requirement. For the GIS component of the module, assignment templates were provided (see Appendix) with the intent to support students with the development of a ‘good draft’. A ‘reflection’ document reporting student thoughts on their learning was also required to be in the portfolio. Though the document had no influence on the final module mark, it was designed to enhance student engagement, but also it provided pedagogical evidence of the impact of the course. Although the portfolios were not anonymous, students appeared to feel free to provide frank comments as evidenced by some critical but constructive comments on the use of GIS mapping. This augmented the information gleaned by the tutors as they used the formative feedback provided to monitor progress towards the intended learning outcomes. Students were also able to provide continuous feedback through weekly interactions with tutors. As in the cohort prior to this one (CitationChamberlain et al. 2014), an identical survey at the beginning and end of the module demonstrated that students are anxious about QM.

Self-reflexive evaluation

Critical reflection involves carefully considering how a learning activity went (CitationSchon 1983), with the aim of improving students’ learning. Reflections that considered what went well, what worked less well and suggested potential changes to the GIS course, were gathered by the tutors during its delivery. The high level of student–tutor interaction and the ‘reflection’ document within the submitted portfolios also meant that tutors could reflect (reflection in action) through the eyes of the students.

A reflective cycle (CitationGibbs 1988) provides relevant questions in order to structure reflection around Kolb’s experiential learning cycle. Specifically, a focus of educators is on whether students can relate ‘theory to practice’ (CitationBrigden & Purcell 2004, p21). Here, the varied evidence is assessed in order to evaluate how well students related ‘theory to practice’ throughout the GIS workshops, and whether the induction module achieved its aims.

Four main themes summarize the reflections of the students and instructors and provide a provisional answer as to whether (i) students felt supported, (ii) students related the theory of hypothesis generation to the GIS workshop practice, and (iii) visualization can lessen students’ anxiety towards QM.

Theme 1: The learning cycle

Students typically found the repetitive, cyclical approach to hypothesis generation useful. For instance, many found the process of hypothesis generation for two variables easier because they had already done similar work for one variable. Feedback from the tutors was used to develop their learning. In particular an important moment of the learning cycle was the review of the GIS draft and the feedback received from the tutor, which would allow students to produce an improved final version of their GIS assignment that reflected their increasing understanding. Indeed, some students habituated the process of hypothesis development. Unfortunately, this also resulted in instances of uncritical repetition of the same formula whatever the topic, missing the vital point that research hypotheses should be tailored to a specific question. As an example, workshop GIS 3 required students to develop hypotheses about whether the majority of people in Leicestershire live in areas of high crime, but some students completely ignored the question and focused on a related but tangential question about the link between crime and population density covered in GIS 2. Based on student–tutor interaction for workshop GIS 3 and feedback provided to students, it was noted that some students found it difficult to ‘estimate’ the majority of people in Leicestershire and consequently opted for an easier (in their view) albeit tangential answer to the assignment question. Tutors related this difficulty to the fact that some students were unwilling (rather than unable) to ‘estimate’ i.e., to engage with ‘numbers’ (this issue will be explained in Theme 3: Student anxiety towards numeracy). There is evidence that the cycle approach to learning meant that students felt supported as intended by one of the aims set, though there is mixed evidence regarding the ability to relate the theory about hypothesis generation to the GIS practice.

Theme 2: GIS technologies and visualization

The focus of the induction module was to learn with GIS about hypothesis generation rather than learning about GIS, and the intention was to avoid the technology hampering the learning process rather than facilitating it. This is non-trivial to achieve as GIS packages are complex and require technical ability. No technical problems were experienced, so the presence of tutors and detailed instruction sets appear to have worked. Students who thought the software was difficult to use or experienced technical problems were those who had not attended the workshop and were trying to catch up in their own time.

While a minority of students thought that ‘looking at maps’ was not relevant for their sociology course, during the workshops it was generally observed that students engaged with the technology in the sense that it was clear to them that the focus of their work was the hypothesis generation process, as highlighted in several student reflection documents, rather than GIS. The aim to introduce the key statistical concept of hypothesis generation through GIS seems to have been fulfilled, i.e. the mapping tools utilized throughout the workshops were a means towards a learning outcome (understanding the hypothesis generation process) rather than an end in itself.

Theme 3: Student anxiety towards numeracy

The GIS course seems to have eased student anxiety toward numeracy, as one student observed: “I do not feel confident learning about [numbers] but I am becoming increasingly confident through the module”. Throughout the GIS workshops, the tutors also observed raised student confidence in engaging with other QM concepts (for instance sampling and levels of measurement) that had been dealt with in earlier non-visual workshops and a willingness to revise those assignments because of their increased understanding. However, observations in the workshops also showed that some students prefer not to engage with numbers (as reported in Theme 1: The learning cycle), even with the use of visualizations such as dot density maps (). A minority of students found it difficult to link the number of dots in an area to the number representing the proportion of people in fear of crime. Others explained the dots as the actual location of people in fear of crime, even if the presentation and the guided exercise made it clear that dots did not indicate location of people. The misreading of dot density maps by some students might be the result of a mismatch between these students’ learning preferences, (who might be, for instance, ‘verbalizers’ rather than ‘visualizers’) and the visual (GIS) learning and teaching approach used (on the issue of learning styles see CitationFelder & Spurlin 2005, for instance). However, a match between student preferred learning styles and learning method used (i.e. using visual methods only with ‘visualizers’) do not necessarily result in better learning outcomes as cognitive ability (in particular ‘spatial visualization’) seems to have a more prominent role in learning outcomes than learning styles only (CitationKollöffel 2012). In fact, during the course of the GIS workshops it was noted that for some students increased easiness with visualizations was matched by an increased easiness about the hypothesis generation process. However, as this evidence is anecdotal more research on the issues of visual learning styles and teaching methods is warranted.

If student understanding of visual aids (e.g. maps or graphs) is fully formed they can assist learning (CitationBooth & Koedinger 2012), but in contrast CitationGarfield & Ben-Zvi (2007) recognize that students fail, if not carefully guided, to use graphs as reasoning tools, and instead employ them as illustrations. For this reason, as explained above, several methods were used to ensure that GIS facilitated rather than hampered learning: i) tutor-led introduction to the task ii) detailed step-by-step instructions, and iii) the continual presence of tutors and teaching assistants who proactively engaged the students.

Theme 4: The wider context

As a corollary of the significant interaction and feedback, the course was time and resource intensive for both students and instructors. A positive aspect of this was that students appreciated repetition, interaction with tutors and fellow students, and spending time in the classroom as this gave them time to assimilate their understanding of key concepts. Feedback was particularly appreciated. For instance, during each GIS workshop the original plan was to provide each student with 2–3 minutes of verbal feedback. In practice 5–10 minutes was required in order to provide detailed suggestions. A further strain was placed on resources by student requests for feedback outside workshop time, although this could be dealt with by (i) using online learning environments to handle more queries in a forum rather than individual basis (ii) expectation management and (iii) self-support in groups.

It seems that on balance the GIS component achieved the aim of introducing students to the hypothesis generation process in a non-technical manner, though the full assessment of the induction module impact will only be clear after the conclusion of the more traditional QM module at the end of semester two (June 2014). A full assessment of impact also will be conducted following the completion of the research project which is currently collecting and analysing data to explore student anxiety issues before and after the implementation of the induction module (CitationChamberlain et al. 2014).

Conclusions

Based on student feedback and tutors’ self-reflective evaluation, the following conclusions about the GIS component of the induction module can be drawn.

Visualization using GIS appears to have the capacity to lower the barrier to some QM concepts for some students. Specifically, it can introduce key concepts (e.g. hypothesis generation) in QM without the use of numbers. The extent to which shielding students from their anxiety towards numeracy is an effective learning approach which helps them when they are faced with a traditional QM module, however, is open to debate. A more robust answer will be provided once the full impact of this approach is assessed in June 2014.

In line with previous work, students respond positively to repetitive learning when facing difficult concepts and to a high level of interactivity, and may just have found the work less daunting as additional curriculum time was targeted at it. These effects, however, are difficult to disentangle from anecdotal evidence, which will require analysis of questionnaire data collected before and after the implementation of the induction module.

Given the initial positive results of a visual learning and teaching approach, the authors of this paper would also consider integrating statistical visualization software (CitationGarfield & Ben-Zvi 2007, MacInnes 2012) together with the GIS component in order to develop a fully visual module. However, it is important to reiterate the findings of previous work that the additional effort may not be repaid unless QM teaching is integrated fully into the curriculum (CitationChamberlain et al. 2014).

Notwithstanding the relatively small sample (85 students) of the study, it is clear that students respond positively to feedback, interaction with tutors, formative assessment and the ‘GIS problem-based learning cycle’, and while the high level of interactivity in the course made it time and resource intensive, in the future solutions can be sought through the use of online learning environments and ‘self-help’ forums.

All the materials of the GIS component including tutor presentations and guided exercises together with the rest of the induction module will be available for use across the social sciences in July 2014 via the ESRC’s Quantitative Methods Initiative website: www.quantitativemethods.ac.uk.

References

  • Anselin, L. (1995) Local indicators of spatial association – LISA. Geographical Analysis 27 (2), 93–115.
  • Balka, D., Bana, J., Hoover, C., Marshall, L. and Swan, P. (2007) Visual math dictionary. p130. Didax.
  • BBC (2013) The joy of statistics. Available at http://www.bbc.co.uk/programmes/b00wgq0l (accessed 31 January 2014).
  • Boba, R. (2005) Crime analysis and crime mapping. Thousand Oaks, CA: Sage.
  • Booth, J.L. and Koedinger, K.R. (2012) Are diagrams always helpful tools? Developmental and individual differences in the effect of presentation format on student problem solving. British Journal of Educational Psychology 82 (3).
  • Brigden, D.N. and Purcell, N. (2004) Becoming a reflective practitioner, Special Focus, The newsletter of LTSN, 01.6, Available at http://www.medev.ac.uk/static/uploads/resources/01_newsletter/LTSN016_lo_res.pdf (accessed 31 January 2014).
  • British Academy (2012) A position statement – society counts, quantitative skills in the social sciences and humanities. Available at http://www.britac.ac.uk/policy/Society_Counts.cfm (accessed 31 January 2014).
  • Case, J. (2008) Education theories on learning: an informal guide for the engineering education scholar. Higher Education Academy Engineering Subject Centre. Available at http://exchange.ac.uk/downloads/scholarart/education-theories.pdf (accessed 6 May 2014).
  • Chamberlain, J.M., Hillier, J. and Signoretta, P. (2014). Counting better? An examination of the impact of quantitative method teaching on undergraduate sociology and criminology students’ statistical anxiety and confidence to complete statistical tasks. Active Learning in Higher Education 15 (3).
  • Dorling, D. (1996) Area cartograms: their use and creation. CATMOG, 59. Institute of British Geographers.
  • Eliëns, A.P.W. and Ruttkay, Z.M. (2009) Math game(s) – an alternative (approach) to teaching math? In Proceedings of 10th Int. Conference on Intelligent Games and Simulation, GAME-ON 09, 2009, pp68–72. Dusseldorf, Germany: EUROSIS.
  • ESRI (2014) Available at www.esri.com (accessed 31 January 2014).
  • Felder, R.M. and Spurlin, J. (2005) Applications, reliability and validity of the index of learning styles. International Journal of Engineering Education 21 (1), 103–112.
  • Fink, L.D. (2003) Creating significant learning experiences: an integrated approach to designing college courses. San Francisco, CA: Jossey-Bass.
  • Forbes, S.D. (2012) Data visualisation: motivational and teaching tool in official statistics. Technology Innovations in Statistics Education 6 (1), 1–19.
  • Fry, H., Kettering, S. and Marshall, S. (1999) Teaching and learning in higher education: enhancing academic practice. Kogan Page.
  • Garfield, J. and Ben-Zvi, D. (2007) How students learn statistics revisited: a current review of research on teaching and learning statistics. International Statistical Review 75 (3).
  • Gibbs, G. (1988) Learning by doing: a guide to teaching and learning methods. Further Education Unit. Oxford: Oxford Brookes University.
  • HEA (2005) Web resources for problem-based Learning. Where are the leading exemplars? The Higher Education Academy. Available at http://www.heacademy.ac.uk/assets/ps/documents/primers/ps0090_web_resources_for_problem_based_learning_mar_2005.pdf (accessed 06 May 2014).
  • Healey, M. and Jenkins, A. (2000) Kolb’s experiential learning theory and its application in geography in higher education. The Journal of Geography 99 (5), 185–195.
  • Horton, B. (2001) ‘I hear and I forget, I see and I remember, I do and I understand’ – putting learning models into practice. Planet 2, 12–14.
  • Hounsell, D., Falchikov, N., Hounsell, J., Klampfleitner, M., Huxham, M., Thomson, K. and Blair, S. (2007) Innovative assessment across the disciplines. An analytical review of the literature. Final Report. HEA.
  • Kayes, D.C. (2002) Experiential learning and its critics: preserving the role of experience in management learning and education. Academy of Management Learning & Education. 1 (2), 137–149.
  • Kim, H.S. and Kim, Y.M. (2003) Visualization for math-education using Mathematica. International Mathematica Symposium.
  • Kolb, D. (1984) Experiential learning: experience as the source of learning and development. New Jersey: Prentice-Hall.
  • Kollöffel, B. (2012) Exploring the relation between visualizer–verbalizer cognitive styles and performance with visual or verbal learning material. Computers and Education 58 (2).
  • Larkin, J.H. and Simon, H.A. (1987) Why a diagram is (sometimes) worth ten thousand words. Cognitive Science, 11 (1), 65–100.
  • Lewis, B. (2010) Help your kids with math: a visual problem solver for kids and parents. DK Publishing.
  • Linden, A.M. (2012) Teaching quantitative literacy through a regression analysis of exam performance. Teaching Sociology 40 (1), 50–59.
  • Lloyd, W.J. (2001) Integrating GIS into the undergraduate learning environment. The Journal of Geography 100 (5), 158.
  • Lynch, R., Maio, G., Moore, G., Moore, L., Orford, S., Robinson, A., Taylor, C. and Whitfield, K. (2007) ESRC/HEFCW scoping study into quantitative methods capacity building in Wales. Final report to the ESRC and HEFCW. Swindon: ESRC.
  • MacEachren, A.M., Gahegan, M., Pike, W., Brewer, I., Cai, G., Lengerich, E. and Hardistry, F. (2004) Geovisualization for knowledge construction and decision support. Computer Graphics and Applications 24 (1), 13–17.
  • MacInnes, J. (2009) Proposals to support and improve the teaching of quantitative research methods at undergraduate level in the UK. Available at http://www.esrc.ac.uk/_images/undergraduate_quantitative_research_methods_tcm8-2722.pdf (accessed 31 January 2014).
  • MacInnes, J. (2012) Quantitative methods teaching in UK higher education: the state of the field and how it might be improved. Available at http://www.heacademy.ac.uk/assets/documents/events/SS_assets/Blog/MacInnes_fullpaper1.pdf (accessed 31 January 2014).
  • MacVaugh, J. (2009) Problem based learning. In: Handbook of Sustainability Literacy. Available at http://arts.brighton.ac.uk/__data/assets/pdf_file/0017/6254/Problem-Based-Learning.pdf (accessed 06 May 2014).
  • McVie, S., Coxon, A., Hawkins, P., Palmer, J. and Rice, R. (2008) ESRC/SFC scoping study into quantitative methods capacity building in Scotland. Final report. Edinburgh: Scottish Funding Council.
  • Mitchell, A. (1999) The ESRI guide to GIS analysis, Volume 1: Geographic patterns and relationships. ESRI Redlands: California.
  • Nuffield Foundation, ESR, HEFCE (2012) Programme Background – Promoting a step-change in the quantitative skills of social science undergraduates.
  • Orford, S., Dorling, D., Harris, R. (1998) Review of visualization in the social sciences: a state of the art survey and report. Report for the Advisory Group on Computer Graphics, Final Report 27th July 1998.
  • Parker, J., Dobson, A., Scott, S., Wyman, M. and Sjöstedt Landén, A. (2010) International bench-marking review of best practice in the provision of undergraduate teaching in quantitative methods in the social sciences. Available at http://www.esrc.ac.uk/_images/International_benchmarking_undergraduate_quantitative_methods_tcm8-2725.pdf (accessed 31 January 2014).
  • Riner, M.E., Cunningham, C. and Johnson, A. (2004) Public health education and practice using geographic information system technology. Public Health Nursing 21 (1), 57–65.
  • Rosling, H. (2014) Available at http://www.gapminder.org/ (accessed 31 January 2014).
  • Rust, C. (2002) Purposes and principles of assessment. Learning and Teaching Briefing Papers Series. Oxford Centre for Staff and Learning Development.
  • Schon, D. (1983) The reflective practitioner. Temple Smith: London.
  • Smith, M.S., Hillier, J.K., Otto, J-C. and Geilhausen, M. (2013) Geovisualization. In Treatise on Geomorphology, Volume 3 ( ed. J.F. Schroder and M.P. Bishop), pp299–325. San Diego: Academic Press.
  • Sobanski, J. (2002) Visual math: see how math makes sense. NY: Learning Express.
  • Solem, M. (2001) Using geographic information systems and the Internet to support problem-based learning. Planet, Special Edition Two, 22–24. Available at http://www.gees.ac.uk/planet/p4/ms.pdf (accessed 31 January 2014).
  • Thomson, R., Jefferson, H. and Brigden, D. (2008) Student experiences of portfolios. Available at HEA http://www.medev.ac.uk/static/uploads/resources/01_newsletter/0116_lo_res.pdf (accessed 31 January 2014).
  • Unrau, Y.A. and Beck, A.R. (2004) Increasing research self-efficacy among students in professional academic programs. Innovative Higher Education 28 (3), 187–204.
  • Wall, M.L. (2013) Using balanced learning course design to reduce resistance to diversity curricula. ELiSS 5 (2).
  • Zeiffler, A., Garfield, J., delMas, B. and Reading, C. (2008) A framework to support research on informal inferential reasoning. Statistics Education Research Journal 7 (2), 40–58. Available at http://iase-web.org/documents/SERJ/SERJ7(2)_Zieffler.pdf (accessed 31 January 2014).

Appendix: Template for GIS 1 Assignment

Title e.g. An investigation into the distribution of crime in Leicestershire

[Short and concise]

Aims

[Briefly explain what you plan to do. Bullet points are acceptable.]

Background

[This should explain the context of your proposed analysis. This can be based upon your general knowledge or popular press, but you should support this by referring to the primary literature on crime (i.e. books or journal articles). Note that this is not a literature review. For instance, refer to literature that can support your hypotheses regarding the distribution of crime in urban and rural areas.]

Graphics Methods

[Describe what data you used for your analysis and why. Explain how you plan to display the data and justify why – be specific about which visualization option you selected and why etc …]

Visual Analysis

[What are the results of your analysis? Explain your findings in the context of the literature referred to in the background i.e. set up the justification for the null and research hypotheses.

Hypotheses

[Full statement of the hypotheses and precisely what testing these (either visually or numerically) will allow you to determine.]

Conclusion

[Brief summary statement of the conclusion of your analysis i.e. research hypothesis, your best estimate of you confidence in it, and likelihood of being able to reject the null hypothesis.]

References

[Correctly formatted reference list. If in doubt, please refer to your department’s referencing guidelines.]

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.