On Tuesday, my good friend Dr Michelle Dalrymple won this year’s Prime Minister’s Science Teacher award. It was so great to be able to fly down to Christchurch with Maxine Pfannkuch to watch the live streaming of the award ceremony with Michelle, her family and her colleagues at Cashmere High School. Michelle was the first mathematics and statistics teachers to win the prize, and it couldn’t have gone to a more deserving teacher!

You can read more about the awesomeness of Michelle in the links below:

- https://www.pmscienceprizes.org.nz/2019-prime-ministers-science-teacher-winner/
- https://www.stuff.co.nz/national/education/121954122/maths-teacher-takes-out-pms-top-science-prize-with-help-from-her-dogs

In her acceptance speech, Michelle thanked me for being her “statistics hero”. Well, turns out she’s also mine and here’s just one example of why!

After a year or so after I moved from teaching high school statistics to teaching a very large introductory statistics course, I had conversation with Michelle where I complained about how much I missed doing the kinds of hands-on interactive activities that are so important for teaching statistics. I told her what I was being told by others at the university level: that you just can’t do those kinds of things with large lectures, there’s too many students, it won’t work, things could go wrong, not all the students will want to do this, etc.

Michelle listened to me first and then suggested that I try doing something small initially. She told me about one of her activities – comparing how long it takes to eat M&Ms using a plastic fork vs chopsticks – and suggested doing this with just 10 of my 500 students. She explained that I could ask for volunteers, bring them down to the front of the lecture theatre, record the data live, and then use this within the same lecture. I tried this activity out and it worked brilliantly – just imagine a whole lecture theatre of students cheering on students eating M&M’s!

In her pragmatic way, Michelle helped me remember that there’s always a way to do what you know is best for teaching and learning. Her encouragement and attitude to “make it happen” inspired the first of many interactive activities I have since developed to use in my teaching of intro stats. It’s natural to focus on the limitations that a teaching environment or system presents, especially for very large introductory statistics classes of over 300 students. But what Michelle helped me re-affirm in terms of my teaching approach for “large scale teaching” is that it can be more helpful and rewarding to think of the opportunities that working with such a large group of students offers.

Which is one of the reasons why we (Rhys Jones, Emma Lehrke and I) have set up a new sub blog that focuses specifically on teaching large introductory statistics courses. It’s called “Go big or go home!“. In this blog we will share our experiences with trying to build more interactivity and engagement within our very large lecture-based classes. I know that many people reading this blog are statistics teachers based at the school level, so I haven’t assumed you will want to receive emails about new posts for this sub blog. Check out the Go big or go home! blog if you’re interested in reading more and subscribing to this new blog.

]]>Actually, it’s not a new tool exactly, more a re-working of the existing modelling tool I’ve already shared on this blog, but with a new name and web location – the *probability distribution explorer*!

I developed the *probability distribution explorer* as part of my Masters research into teaching probability distribution modelling. The proposed teaching framework and the tool were developed in response to use of data for distribution modelling for AS91586, in particular the need for students to demonstrate use of methods related to the *distribution of true probabilities* versus *distribution of model estimates of probabilities* versus *distribution of experimental estimates of probabilities*.

The tool was developed primarily to support comparisons of the “distribution of experimental estimates of probabilities” and “distribution of model estimates of probabilities”. When reviewing research literature, I found limited examples of how to teach this comparison using an informal approach i.e. not using a Chi-square goodness-of-fit test. Consequently, I also found a lack of statistically sound criteria to enable drawing of conclusions in such resources as textbooks, workbooks and assessment exemplars.

This led to my research, which involved a small group of New Zealand high school statistics teachers. Focusing on the Poisson distribution, the criteria used by ten Grade 12 teachers for informally testing the fit of a probability distribution model was investigated. I found that criteria currently used by the teachers were unreliable as they could not correctly assess model fit, in particular, sample size was not taken into account.

After exploring the goodness-of-fit using my visual inference tool, teachers reported a deeper understanding of model fit. In particular, that the tool had allowed them to take into account sample size when testing the fit of the probability distribution model through the visualisation of expected distributional shape variation. I’ve re-developed the tool this year to support NZQA as they explore opportunities for assessment within a digital environment. A team of teachers are developing prototype assessment activities for AS91586 and these will be trialled with students in schools later in the year.

The video below gives a general introduction to the tool, using data on how many times I say “um” when I’m teaching. The video itself provides another source of data because, um … well, you’ll see if you watch!

More videos, teaching notes and related resources can be found here: stat.auckland.ac.nz/~fergusson/prob_dist_explorer/teachers/

]]>Just a quick post to let you know that the mathstatic.co.nz site is hopefully only temporarily down, and I am working with my hosting company to get it back online ASAP. This affects the random redirect tool, the BYOP sampler tool and the experiment lab page, which will not be available until this gets sorted. I’ll update this post soon with a progress update!

**UPDATE ONE**

It seems the issue is that some overseas dodgy folk have been using the random redirect tool for fraudalent things like phishing scams. So, I’m going to restrict the URLs that can be used – which means analysis time to identify which sites/URL patterns to accept e.g. Google forms, survey monkey etc.

**UPDATE TWO**

mathstatic.co.nz is back up and running! It probably was a couple of hours ago, but I have been rewriting the code that processes the random redirect requests. Below are the main changes to the random redirect tool to better prevent issues in the future.

*Due to abuse of this tool by dodgy folk, **only links with domains on the approved list will now be accepted! **Please **complete this form** to request a domain to be added to the approved list, but donāt expect any new additions to happen any time soon (this is a free tool remember and was created for simple classroom-based randomised experiments with Google forms).*

*Any random redirect URLs created using this tool can be disabled at any time. If this has happened to you and you are a legitimate teacher, educator or researcher, then send me an email and I might be able to help you.*

**UPDATE THREE**

After emailing me this morning to say everything was sorted with mathstatic.co.nz, my webhosting company then decided to set my site to “maintenance” mode this afternoon and remove some crucial code used to redirect the URLs to the right locations on my website I’m trying to get things rest back to what they were now.

**UPDATE FOUR**

Well, I had been meaning to retire the old mathstatic.co.nz website anyway! I’m not sure when mathstatic will be online again, so:

- You can now set up random redirects here: https://allocate.monster (this will be the place to do this from now on!)
- You can access the BYOP sampler here: https://statistics-is-awesome.org/BYOP (this will be the place to do this from now on!)
- At some point soon I’ll move the experiment lab to https://learning.statistics-is-awesome.org but not today, so for now it is unavailable.

I think that’s everything. If there is something else not working, then please let me know!

]]>Today I demonstrated some in-class interactive activities that I had developed for my super large intro statistics lectures at a teaching and learning symposium. I’ve shared a summary of the activities and the data below.

If you haven’t already, check out learning.statistics-is-awesome.org/different_strokes/, where you can sample some cat (and other) drawings and learn more about how people draw in the Google game Quick, Draw!

I also get students to draw things in class and use their drawings as data. Below are all the drawings of cats made from the demonstration today, and also from the awesome teachers who helped me out last night. If you click/touch and hold a drawing you will be able to drag it around. How many different ways can you sort the drawings into groups?

]]>It was awesome to be back in Wellington, as not only did I complete a BMus/BSc double degree at Victoria University, I actually taught music at Hutt Valley High School (the venue for the conference) while I was training to become a high school teacher (in maths/stats and music). I didn’t talk much in my keynote about the relationship between music and data analysis, but I did describe my thoughts a few years ago (see below):

All music has some sort of structure sitting behind it, but the beauty of music is in the variation. When you learn music, you learn about key ideas and structures, but then you get to hear how these same key ideas and structures can be used to produce so many different-sounding works of art. This is how I think we need to help students learn statistics ā minimal structure, optimal transfer, maximal experience. Imagine how boring it would be if students learning music only ever listened to Bach.

https://www.stat.auckland.ac.nz/en/about/news-and-events-5/news/news-2017/2017/08/the-art-of-teaching-statistics-to-teenagers.html

Due to some unforeseen factors, I ended up ZOOMing my slides from one laptop at the front of the hall to another laptop in the back room which was connected to the data projector. Since I was using ZOOM, I decided to record my talk. However, the recording is not super awesome due to not really thinking about the audio side of things (ironically). If you want to try watching the video, I’ve embedded it below:

You can also view the slides here: bit.ly/followthedataNZAMT. I’m not sure they make a whole lot of sense by themselves, so here’s a quick summary of some of what I talked about:

- Currently, we pretty much choose data to match the type of analysis we want to teach, and then “back fit” the investigative problem to this analysis. This is not totally a bad thing, we do it in the hope that when students are out there in the real world, they think about all the analytical methods they’ve learned and choose the one that makes sense for the thing they don’t know and the data they have to learn from. But, there’s a whole lot of data out there that we don’t currently teach students about how to learn from, which comes from the computational world our students live in. If we “follow the data” that students are interacting with, what “new” ways of thinking will our students need to make sense of this data?
- Album covers are a form of data, but how do we take something we can see visually and turn this into “data”. For the album covers I used from one week of 1975 and one week of 2019, we can see that the album covers from 1975 are not as bright and vibrant as those from 2019, similarly we can see that people’s faces feature more in the 1975 album covers. We could use the image data for each album cover, extract some overall measure of colour and use this to compare 1975 and 2019. But what measure should we use? What is luminosity, saturation, hue, etc.? How could we overfit a model to predict the year of an album cover by creating lots of super specific rules? What pre-trained models can we use for detecting faces? How are they developed? How well do they work? What’s this thing called a “confusion matrix”?
- An intended theme across my talk was to compare what humans can do (and to start with this), with what we could try to get computers to do, and also to emphasise how important human thinking is. I showed a video of Joy Buolamwini talking about her Gender Shades project and algorithmic bias: https://www.youtube.com/watch?v=TWWsW1w-BVo and tried to emphasise that we can’t teach about fun things we can do with machine learning etc. without talking about bias, data ethics, data ownership, data privacy and data responsibility. In her video, Joy uses faces of members of parliament – did she need permission to use these people’s faces for her research project since they were already public on websites? What if our students start using photos of our faces for their data projects?
- I played the song that was number one the week I was born (tragedy!) as a way to highlight the calendar feature of the nztop40 website – as long as you were born after 1975, you can look up your song too. Getting students to notice the URL and how it changes as you navigate a web page is a useful skill – in this case, if you navigate to different chart weeks, you can notice that the “chart id” number changes. We could “hack” the URL to get the chart data for different weeks of the years available. If the website terms and conditions allow us, we could also use “web scraping” to automate the collection of chart data from across a number of weeks. We could also set up a “scheduler” to copy the chart data as it appears each week. But then we need to think about what each row in our super data set represents and what visualisations might make sense to communicate trends, features, patterns etc. I gave an example of a visualisation of all the singles that reached number one during 2018, and we discussed things I had decided to do (e.g. reversing the y axis scale) and how the visualisation could be improved [data visualisation could be a whole talk in itself!!!]
- There are common ways we analyse music – things like key signature, time signature, tempo (speed), genre/style, instrumentation etc. – but I used one that I thought would not be too hard to teach during the talk: whether a song is in the major or minor key. However, listening to music first was really just a fun “gateway” to learn more about how the Spotify API provides “audio features” about songs in its database, in particular supervised machine learning. According to Spotify, the Ed Sheeran song
*Beautiful people*is in the minor key, but me and guitar chords published online think that it’s in the major key. What’s the lesson here? We can’t just take data that comes from a model as being the truth. - I also wanted to talk more about how songs make us feel, to extend thinking about the modality of the song (major = happy, minor = sad), to the lyrics used in the song as well. How can we take a set of lyrics for a song and analyse these in terms of overall sentiment – positive or negative? There’s lots of approaches, but a common one is to treat each word independently (“bag of words”) and to use a pre-existing
*lexicon*. The slides show the different ways I introduce this type of analysis, but the important point is how common it is to transfer a model trained within one data context (for the*bing*lexicon, customer reviews online) and use it for a different data context (in this case, music lyrics). There might just be some issues with doing this though! - Overall, what I tried to do in this talk was not to showcase computer programming (coding) and mathematics, since often we make these things the “star attraction” in talks about data science education. The talk I gave was totally “powered by code” but do we need to start with code in our teaching? When I teach statistics, I don’t start with pulling out my calculator! We start with the data context. I wanted to give real examples of ways that I have engaged and supported all students to participate in learning data science: by focusing on what humans think, feel and see in the modern world first, then bringing in (new) ways of thinking statistically and computationally, and then teaching the new skills/knowledge needed to support this thinking.
- We have an opportunity to introduce data science in a real and meaningful way at the school level, and we HAVE to do this in a way that allows ALL students to participate – not just those in enrichment/extension classes, coding clubs, and schools with access to flash technology and gadgets. While my focus is the senior levels (Years 11 to 13), the modern world of data gives so many opportunities for integrating statistical and computational thinking to learn from data across all levels. We need teachers who are confident with exploring and learning from modern data, and we need new pedagogical approaches that build on the effective ones crafted for statistics education. We need to introduce computational thinking and computer programming/coding (which are not the same things!) in ways that support and enrich statistical thinking.

If you are a NZ-based teacher, and you are interested in learning more about teaching data science, then please use the “sign-up” form at undercoverdata.science (the “password” is *datascience4everyone*). I’ll be sending out some emails soon, probably starting with learning more about APIs (for an API in action, check out learning.statistics-is-awesome.org/popularity_contest/ ).

I’m pretty excited about the talks and workshops I’m doing over the next month or so! Below are the summaries or abstracts for each talk/workshop and when I get a chance I’ll write up some of the ideas presented in separate posts.

** Keynote: Searching for meaningful sampling in apple orchards, YouTube videos, and many other places!** (AMA, Auckland, September 14, 2019)

In this talk, I shared some of my ideas and adventures with developing more meaningful learning tasks for sampling. Using the āApple orchardā exemplar task, I presented some ideas for ārenovatingā existing tasks and then introduced some new opportunities for teaching sample-to-population inference in the context of modern data and associated technologies. I shared a simple online version of the apple orchard and also talked about how my binge watching of DIY YouTube videos led to my personal (and meaningful) reason to sample and compare YouTube videos.

** Workshop: Expanding your toolkit for teaching statistics** (AMA, September 14, Auckland, 2019)

In this workshop, we explored some tools and apps that I’ve developed to support studentās statistical understanding. Examples were: an interactive dot plot for building understanding of mean and standard deviation, a modelling tool for building understanding of distributional variation, tools for carrying out experiments online and some new tools for collecting data through sampling.

The slides for both the keynote and workshop are embedded below:

*Talk***:**** Introducing high school statistics teachers to code-driven tools for statistical modelling** (VUW/NZCER, Wellington, September 30, Auckland, 2019)

**Abstract: **The advent of data science has led to statistics education researchers re-thinking and expanding their ideas about tools for teaching and learning statistical modelling. Algorithmic methods for statistical inference, such as the randomisation test, are typically taught within NZ high school classrooms using GUI-driven tools such as VIT. A teaching experiment was conducted over three five-hour workshops with six high school statistics teachers, using new tasks designed to blend the use of both GUI-driven and code-driven tools for learning statistical modelling. Our findings from this exploratory study indicate that teachers began to enrich and expand their ideas about statistical modelling through the complementary experiences of using both GUI-driven and code-driven tools.

** Keynote: Follow the data** (NZAMT, Wellington, October 3, 2019)

**Abstract: **Data science is transforming the statistics curriculum. The amount, availability, diversity and complexity of data that are now available in our modern world requires us to broaden our definitions and understandings of what data is, how we can get data, how data can be structured and what it means to teach students how to learn from data. In particular, students will need to integrate statistical and computational thinking and to develop a broader awareness of, and practical skills with, digital technologies. In this talk I will demonstrate how we can* follow the data* to develop new learning tasks for data science that are inclusive, engaging, effective, and build on existing statistics pedagogy.

**Workshop: ****Just hit like! Data science for everyone, including cats (and maybe dogs)** (NZAMT, Wellington, October 2, 2019)

**Abstract: **Data science is all about integrating statistical and computational thinking with data. In this hands-on workshop we will explore a collection of learning tasks I have designed to introduce students to the exciting world of image data, measures of popularity on the web, machine learning, algorithms, and APIs. Weāll explore questions such as āAre photos of cats or dogs more popular on the web?ā, āWhat makes a good black and white photo?ā, āHow can we sort photos into a particular order?ā, āHow can I make a cat selfie?ā and many more. Weāll use familiar statistics tools and approaches, such as data cards, collaborative group tasks and sampling activities, and also try out some new computational tools for learning from data. Statistical concepts covered include features of data distributions, informal inference, exploratory data analysis and predictive modelling. Weāll also discuss how each task can also be extended or adapted to focus on specific aspects and levels of the statistics curriculum. Please bring along a laptop to the workshop.

I’m also presenting a workshop at NZAMT with Christine Franklin on what makes a good statistical task. I’ve been assisting Maxine Pfannkuch and members of the NZSA education committee to set up a new teaching journal, which we will be launching at the workshop!!

]]>Since I already had a tool that creates data cards from the *Quick, Draw! *data set, I’ve created a prototype for the kind of tool that would support this approach using the same data set.

I’ve written about the Quick, Draw! data set already:

- http://teaching.statistics-is-awesome.org/cat-and-whisker-plots-sampling-from-the-quick-draw-dataset/
- http://teaching.statistics-is-awesome.org/you-say-data-i-say-data-cards/
- http://teaching.statistics-is-awesome.org/the-power-of-pixels-modelling-with-images/

For this new tool, called different strokes, users sort drawings into two or more groups based on something visible in the drawing itself. Since you have the drag the drawings around to manually “classify” them, the larger the sample you take, the longer it will take you.

There’s also the novelty and creativity of being able to create your own rules for classifying drawings. I’ll use cats for the example below, but from a teaching and assessment perspective there are SO many drawings of so many things and so many variables with so many opportunities to compare and contrast what can be learned about how people draw in the *Quick, Draw!*

Here’s a precis of the kinds of questions I might ask myself to explore the general question **What can we learn from the data about how people draw cats in the Quick, Draw! game?**

- Are drawings of cats more likely to be heads only or the whole body? [I can take a sample of cat drawings, and then sort the drawings into heads vs bodies. From here, I could bootstrap a confidence interval for the population proportion].
- Is how someone draws a cat linked to the game time? [I can use the same data as above, but compare game times by the two groups I’ve created – head vs bodies. I could bootstrap a confidence interval for the difference of two population means/medians]
- Is there a relationship between the number of strokes and the pause time for cat drawings? [And what do these two variables actually measure – I’ll need some contextual knowledge!]
- Do people draw dogs similarly to cats in the
*Quick, Draw!*game? [I could grab new samples of cat and dog drawings, sort all drawings into “heads” or “bodies”, and then bootstrap a confidence interval for the difference of two population proportions]

**Check out the tool and explore for yourself here: ** http://learning.statistics-is-awesome.org/different_strokes/

A little demo of the tool in action!

Here’s a scenario. You buy a jumbo bag of marshmallows that contains a mix of pink and white colours. Of the 120 in the bag, 51 are pink, which makes you unhappy because you prefer the taste of pink marshmallows.

Time to write a letter of complaint to the company manufacturing the marshmallows?

The thing we work so hard to get our statistics students to believe is that there’s this crazy little thing called chance, and it’s something we’d like them to consider for situations where random sampling (or something like that) is involved.

For example, let’s assume the manufacturing process overall puts equal proportions of pink and white marshmallows in each jumbo bag. This is not a perfect process, there will be variation, so we wouldn’t expect exactly half pink and half white for any one jumbo bag. But how much variation could we expect? We could get students to flip coins, with each flip representing a marshmallow, and heads representing white and tails representing pink. We then can collate the results for 120 marshmallows/flips – maybe the first time we get 55 pink – and discuss the need to do this process again to build up a collection of results. Often we move to a computer-based tool to get more results, faster. Then we compare what we observed – 51 pink – to what we have simulated.

I use these kind of activities with my students, but I wanted something more so I made a very simple app earlier this year. You can find it here: learning.statistics-is-awesome.org/threethings/. You can only do three things with it (in terms of user interactions) but in terms of learning, you can do way more than three things. Have a play!

In particular, you can show that models other than 50% (for the proportion of pink marshmallows) can also generate data (simulated proportions) consistent with the observed proportion. So, not being able to reject the model used for the test (50% pink) doesn’t mean the 50% model is the **one true thing**. There are others. Like I told my class – just because my husband and I are compatible (and I didn’t reject him), doesn’t mean I couldn’t find another husband similarly compatible.

*Note: **The app is in terms of percentages, because that aligns to our approach with NZ high school students when using and interpreting survey/poll results. However, I first use counts for any introductory activities before moving to percentages, as demonstrated with this marshmallow example. The app rounds percentages to the closest 1% to keep the focus on key concepts rather than focusing on (misleading) notions of precision. I didn’t design it to be a tool for conducting formal tests or constructing confidence intervals, more to support the reasoning that goes with those approaches.*

If you’ve been keeping track of my various talks & workshops over the last year or so, you will have noticed that I’ve become a little obsessed with analysing images (see power of pixelsĀ and/orĀ read more here).Ā As part of my PhD research, I’ve been using images to broaden students’ awareness of what is data, and data science, and it’s been so much fun!

If you’re in the Auckland area next week, you could come along to a workshop I’m running for R-Ladies and have some fun yourself using the statistical programming language R to explore images.Ā The details for the workshop and how to sign up are here:Ā https://www.meetup.com/rladies-auckland/events/255112995/

This is not a teaching-focused workshop, it’s more about learning fun and cool things you can do with images, like making GIFs like the one below….

…. and other cool things, like classifying photos as cats or dogs, or finding the most similar drawing of a duck!

It will be at an introductory level,Ā and you don’t need to be a “lady” to come along, just supportive of gender diversity in the R community (or more broadly, data science)! **If you’ve never used R before, don’t worry – just bring yourself along with a laptop and we’ll look after you **

]]>This long weekend (in Auckland anyway!), I spent some time updating the Quick! Draw! sampling tool (read more about it hereĀ Cat and whisker plots: sampling from the Quick, Draw! dataset). You may need to clear your browser cache/data to see the most recent version of the sampling tool.

One of the motivations for doing so was a visit to my favourite kind of store – a stationery store – where I saw (and bought!) this lovely gadget:

It’s a circle punch with a 2″/5 cm diameter. When I saw it, my first thought was “oh cool I can make dot-shaped data cards”, like a normal person right?

Using data cards to make physical plots is not a new idea – see censusatschool.org.nz/resource/growing-scatterplots/Ā by Pip Arnold for one example:

But I haven’t seen dot-shaped ones yet, so this led me to re-develop the Quick! Draw! sampling tool to be able to create some

I was also motivated to work some more on the tool after the fantastic Wendy Gibbs asked me at the NZAMT (New Zealand Association of Mathematics Teachers) writing camp if I could include variables related to the times involved with each drawing. I suspect she has read this super cool post by Jim VallandinghamĀ (while you’re at his site, check out some of his other cool posts and visualisations) which came out after I first released the sampling tool and compares strokes and drawing/pause times for different words/concepts – including cats and dogs!

So, with Quick! Draw! sampling tool you can now get the following variables for each drawing in the sample:

The drawing and pause times are in seconds. The drawing time captures the time taken for each stroke from beginning to end and the pause time captures all the time between strokes. If you add these two times together, you will get the total time the person spent drawing the word/concept before either the 20 seconds was up, or Google tried to identify the word/concept. Below the word/concept drawn is whether the drawing was correctly recognised (true) or not (false).

I also added three ways to use the data cards once they have been generated using the sampling tool (scroll down to below the data cards). You can now:

- download a PDF version of the data cards, with circles the same size as the circle punch shown above (2″/5cm)
- download the CSV file for the sample data
- show the sample data as a HTML table (which makes it easy to copy and paste into a Google sheet for example)

In terms of options (2) and (3) above, I had resisted making the data this accessible in the previous version of the sampling tool. One of the reasons for this is because I wanted the drawings themselves to be considered as data, and as human would be involved in developed this variable, there was a need to work with just a sample of all the millions of drawings. I still feel this way, so I encourage you to get students to develop at least one new variable for their sample data that is based on a feature of the drawing For example, whether the drawing of a cat is the face only, or includes the body too.

There are other cool things possible to expand the variables provided. Students could create a new variable by adding** drawing_time** and **pause_time**Ā together. They could also create a variable which compares the **number_strokes** to the **drawing_****time **e.g. average time per stroke.Ā Students could also use the **day_sketched** variable to classify sketches as weekday or weekend drawings. Students should soon find theĀ **hemisphere** is not that useful for comparisons, so could explore another country-related classification likeĀ continent.Ā More advanced manipulations could involve working with the time stamps, which are given for all drawings using UTC time. This has consequences for the variableĀ **day_sketched** as many countries (and places within countries) will be behind or ahead of the UTC time.

**If you’ve made it this far in the post…. why not play with a little R **

I wonder which common household pet

Quick!drawers tend to use the most strokes to draw? Cats, dogs, or fish?

Have a go at modifying the R code below, using the iNZightPlots package by Tom Elliott and my [very-much-in-its-initial-stages-of-development] iNZightR package, to see what we can learn from the data If you’re feeling extra adventurous, why not try modifying the code to explore the relationship between number of strokes and drawing time!

]]>