NSS 2017: One data set, many dataviz approaches.

I often say that one dashboard cannot answer all the questions we may have for our data and that the way we visualise our data depends strongly on the question(s) we pursue to answer. On 24th October 2017, the latest meeting of the Midlands Tableau User Group (#MidlandsTUG) community took place in Leicester and, as part of it, attendees had the opportunity to present their take on visualising a public dataset – the 2017 National Student Survey (NSS) results. As a result, a few the attendees presented their work, thus showcasing the variety of angles an analyst can take when trying to gain insight into their data.

Further in this post you will have the opportunity to see some of these examples, presented in no particular order, but first, please remember that the analysts whose work is referenced in this post work in a wide range of sectors. Some may be more familiar to the data than others, but they may not necessarily be an expert on it, and, therefore, their visualisations should not be interpreted as in-depth analyses!

The main purpose of the ‘data hackathon’ exercise during the event was to encourage creativity and a range of approaches when analysing a common data set, so please keep that in mind when exploring their visualisations.

About the data:

The 2017 NSS results were published in July and the data is available on the hefce website. The survey itself is ‘aimed at mainly final-year undergraduates, it gathers opinions from students about their experience of heir courses, asking them to provide honest feedback on what has been like to study on their course at their institution’ (http://www.thestudentsurvey.com/about.php).

Approach #1: Elena Hristozova

Elena Hristozova 1.png

Link to interactive viz: https://public.tableau.com/profile/elena.hristozova#!/

Let me introduce you to my approach. I had worked with this data before which allowed me straightaway to bring together some context in my analysis by providing the themes in which the 27 questions are grouped. My main question for the data was: are there any trends that can be seen in the questions’ results across the subjects and the teaching institutions? In particular, I wanted to make it easier for the user to see whether any groups of questions tend to score lower than others.

I used a heatmap to show the results for all questions and subject or institutions, allowing users to make a choice. Based on no scientific evidence, I chose 75% agree score as my middle point and used colour to encode the ‘less than or higher than 75%’ results.

Approach # 2: Ali Motion

Ali Motion.png

Link to interactive viz: https://public.tableau.com/profile/ali.motion#!/

Ali’s visualisation is a great example of a benchmarking type of dashboard where one can easily see how a selected institution has performed for each of the subjects taught at it, and how the results compare to those of other providers for a selected question. Ali has used a jitter plot where the results for each of the providers are plotted on a single axis, but the marks in the charts are given ‘some space to breathe’ and to show groupings a bit better.

Colour has been used to highlight the selected institution, which also changes between green and orange to indicate of the said institution’s score was above or below the benchmark, or the sector average. Furthermore, a user is also supported in the interpretation of the results through a very clear to read and understand set of tooltips that appear on hover.

Approach # 3: David Clutterbuck:

David Clutterbuck.png

Link to interactive viz: https://public.tableau.com/profile/david.clutterbuck#!/

David’s approach was very straightforward: show the scores for the Top 10 and Bottom 10 providers, show movement from previous year and breakdown of satisfaction per question. He used both colour and shapes very effectively to show the insights and he even went a step further obtaining the previous year’s data to demonstrate the change over time in the overall score.

David has also added a couple of very small touches that make his visualisation easy to explore: by clicking on one of the providers in the table, the provider’s logo appears underneath and the provider’s results become highlighted in the detailed question breakdown.

Approach #4: James Linnett

James Linnett.png

Link to interactive viz: https://public.tableau.com/profile/james.linnett#!/

‘The way I approach it was for the end user (who may or may not have any analytical experience) to easily understand what the dashboard is portraying. By that I mean easily being able to compare their institution and/or subject to the national average.

The traffic light colours indicate how the specific institution, question group or individual question compare against the said average.’

Approach #5: Rob Radburn

Rob Radburn.png

Link to interactive viz: https://public.tableau.com/profile/robradburn#!/

Rob’s approach is very interesting in the sense that it has incorporated a way to visualise both, the mean score of students answers, which is a value from 1 to 5 where 1 = ‘Strongly Disagree’ and 5 = ‘Strongly Agree’ , and the level of agreement (or consensus), which is a measure he has calculated that has a value between 0 and 1, where 0 = disagreement and 1 = agreement.

Rob’s visualisation is very clear and he has provided the readers with a very simple explanation of how to read the chart: ‘The dot shows the average score for each question. The length of the line from the dot measures how in agreement students were in answering the question. This uses Tastle and Wierman’s measure of dispersal. The shorter the line, there is more agreement in the answers for the question. The longer the line, less agreement.  The questions are then ordered from more to less agreement.’

Approach #6: Neil Richards

Neil Richards.png

Link to interactive viz: https://public.tableau.com/profile/neil.richards#!/

‘For my visualisation I was aware that there were some great examples in Big Book of Dashboards chapter 3, so essentially I just wanted to recreate a simple version of the jitter plot of the start of the chapter, with one for each question. I didn’t actually look at the book until just now writing these bits, so I didn’t realise how well I’d remembered the look of the chart!

I added the overall average line after feedback on twitter, and fixed the chosen provider to always show in the middle (with all other x positions just placed at random).’

Approach #7: David Hoskins

David Hoskins.png

Link to interactive viz: https://public.tableau.com/profile/hoskerdu#!/

‘I decided to take a more personal approach to the dataset and compare responses for three institutions where my step daughter Sophie is thinking of studying Social Work next year: Bournemouth, Leeds Trinity and Nottingham Trent.

I knew I had the perfect photo (taken at my partner’s graduation), and soon decided on a grid structure for my dashboard, with a column for each institution showing KPIs for overall comparison and the questions grouped into two categories to keep the layout uncluttered.

Narrowing the focus also allowed me to display the proportion of responses for each question as diverging stacked bars (using instructions at https://t.co/LyB0ikBSmB ) and show the detail behind the aggregated metrics.’

Approach #8: Jennie Holland

Jennie Holland.png

Link to interactive viz: https://public.tableau.com/profile/jennie.holland#!/

‘For my analysis I used the summary data at country level. The first sheet   in the workbook shows survey results for the question groups at country level. As the data wasn’t too big I wanted to be able to show this all on one sheet, with highlights on country and UK comparison to show the variation at a glance.

The second page looks at the questions asked within each question group. To help with consistency between the two sheets, I grouped the questions into the same categories used in the first sheet, and allowed the user to select the group of questions they are interested in. I was quite keen on using one scale for all question groupings that were selected in the filter, to enable the viewer to see the distribution of scores across the question groups.’

Approach # 9: Neil Davidson

Neil Davidson.png

Link to interactive viz: https://public.tableau.com/profile/neild#!/

Neil has not had the chance to provide a quick summary of his approach but straightaway it is easy to see that he was familiar with the data and the HE sector. In his visualisation he has demonstrated the correlation between an institution’s overall score for each of the questions and the said institutions’ rank in the Complete University Guide (CUG) league table.

Not all providers who have results for the NSS appear in the league table and so only higher education establishments are visualised. Though Neil’s work is still a work in progress (as he has admitted it himself), it still demonstrates that the strongest correlation of the CUG ranks appears to be with Q15: The course is well organised and running smoothly (the question with line of best fit closest to 45 degrees). Of course, there is no argument against the fact that a league table is not based merely on the results of the NSS, but the bottom line is that visualising a set of small multiples to show correlation between two variables can work well.

Approach #10: Elena Hristozova & Dave Kirk

Elena Hristozova 2.png

Link to interactive viz: https://public.tableau.com/profile/elena.hristozova#!/

The last approach to demonstrate is somewhat a joint approach between myself and Dave. Dave had the idea to create a visualisation that shows how many of the questions scored below the score for Q27 – Overall Satisfaction. Due to time constraint, he was unable to complete his visualisation but after a couple of exchanged messages I realised that this was an ideal question to answer using the sunburst chart I was currently learning how to make.

The end result is the visualisation below which compares the scores for all 26 questions in the survey to the overall satisfaction score for each of the country regions in the UK.  The analysis is presented through encoding colour – negative difference presented in red and positive difference shown in blue. The intensity of the colour is an indicator of how big or small the difference is.

To sum up…

I hope you have enjoyed exploring the different ways in which the Tableau Community in the Midlands approached the 2017 NSS results. Some analysts drew inspiration from books and the literature, others took a very narrow focus on the data, and the rest just practised visualising survey data and they all had a different set of questions driving their analyses and visualisations!

Thank you,

Elena | #VisualisingHE

P.S. If there were any more submissions to the #MidlandsTUG data challenge that have been missed from this post, do get in touch!

Advertisement

3 thoughts on “NSS 2017: One data set, many dataviz approaches.

  1. What a wonderful blog post – well done to everyone, it’s excellent seeing an exercise like this succeed. I really like that the brief was to be more functional than wee see in MakeoverMonday. Great work.

    Like

  2. […] we at #VizHE decided to pick up from where we left off last year (NSS 2017: One data set, many dataviz approaches.) and pull out a few interesting nuggets for […]

    Like

  3. […] one standout post for me, of which I wasn’t personally involved, but admire a lot, it is the NSS 2017 – one dataset many dataviz approaches post, where Dave and Elena opened up the concept of #VisualisingHE to the #MidlandsTUG,  the […]

    Like

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s