UX Camp Brighton 2018: Jobs-To-Be-Done-(R) #uxcb18

Why am I here (as in you, at this blog, and reading this now)?

I hope you are here because you would like an introduction to using RStudio for data visualisation. I also hope that you are here because you are interested in Jobs-To-Be-Done (abbrev~ JTBD; Ulwick, 2016). This blog provides an example of how to apply one of the approaches that sits under the mantle of JTBD, an Opportunity Landscape (Ulwick, 2016; VIII. Target Hidden Growth Opportunities section).

What are we going to do?

We are going to create a plot, or Opportunity Landscape (SEE also Ulwick, 2017), based on data from a quick opinion survey that was distributed via social media. Survey respondents were asked to rate particular tasks that they regularly complete on a smart watch according to importance and satisfaction on 5-part Likert scales.

At this point I should add a caveat, for those familiar with JTBD. These questions were not based on full job statements or stories but were abbreviated task statements. This is largely because based on survey testing, I found that the time it took to read and consider longer Job stories, based on Situation->Motivation->Expected Outcome (Klement, 2016) required too much mental effort for a <15min survey. I have found Job stories to be indispensable at the interview stage, when derived from transcripts as a means to uncover motivations. In the context of this post I would say that my methodology is a bit of a mash-up of JTBD and Gerry McGovern’s top tasks (2010; 2017).

Why are creating a plot?

I believe that the true purpose of any plot is to start conversations. Data visualisations have the power to both subvert truth and to challenge commonly held assumptions, I aim for the latter. Always.

Why are we using RStudio with ggplot2?

The beauty of using RStudio with ggplot2 is…well beauty. Most people can make standard Excel visualisations without too much effort, but even with a lot of tweaking they still have a heady whiff. In the past I used Adobe Illustrator, but speaking of my own failings it was too easy to leave the land of accuracy and enter the land of surface-treatment.

Here are my top reasons why you should consider using RStudio with ggplot2:

  • You know how to use Microsoft Excel, but would like to create better visualisations…
  • You have lots of data and you would like to segment it and look at a smaller portions without getting caught in a scrolling cycle-of-doom in a humongous spreadsheet…
  • You want something inexpensive – the only cost here is time, but it is well worth it…
  • Once you have built it, it can be used over and over and over…and tweaked quickly for whatever the need…
  • Bonus: RStudio and gpplot2 can be used to prototype data visualisations for interface design and depending on your audience a realistic looking representation/visualisation could be very important.

If any of the statements above have piqued your interest then keep reading.

Step 1: Collect the Data, Show me the Data

In true Blue Peter style, here is one I made earlier…

Image of an Opportunity Landscape plot

The plot above is based on Webcredible’s fantastic blog post on the topic of JTBD (Josephy, 2016).

 

I should add that your plots can be much more extravagant than this, but essentially this is the basic type of plot that we are going to create to get started. The descriptive lines above are superfluous*. They are useful to keep commented-out in your code though as they can be used to create quick descriptive visuals for stakeholders when needed.

What we can read from the visualisation above is that I could have asked about a lot more jobs; it is a bit sparse, but 12 jobs is what this particular survey could accommodate. Also the majority of the jobs fall into the ‘overserved’ domain. This means satisfaction has been rated higher than importance and according to Ulwick (2016) this is prime territory for disruption. So although initially, when I looked at this visualisation I was like “oh blast”, as I was looking for some untapped potential related to smart watch jobs and there is none. Actually it is more that customers are overserved for jobs that aren’t perceived as that important and therefore the landscape is primed for some new competitor to come in and do the job in a cheaper and perhaps less good, but cheaper way, which though market adoption and iteration of ‘said’ cheap version may eventually lead to something better (Ulwick, 2016). Of course another dimension would be to ask questions about jobs that have not yet been considered for smart watch, but these were out of scope for this particular survey. In a way this plot perfectly sums up smart watches as they are still finding their footing, they are still superfluous to the smartphone which is almost ubiquitous.

The point I am getting to here is that this plot asks questions of the data, that we might not normally ask if we were looking at each question in isolation. It is only when we bring these jobs together that we reach a new level of meaning.

*Edward Tufte (2001) would likely frown to see these lines – “erase non-data-ink” (p. 105).

Collect the data

You will need a survey question something like this…
Screenshot of survey question
…Or split it into two versions of the same question depending on the allowances of your survey tool.

“DANGER, WILL ROBINSON!!!!”

Q: Hey didn’t you say earlier that you need a 5 part Likert scale? But you only have four options in the ‘satisfaction’ scale above?

The answer is I had to take a difficult decision about the experience of users on smartphones using my survey. One more column and it was over the precipice into the land of non-function. As I am only interested in ‘Quite Important’ and ‘Very Important’ and ‘Satisfied’ and ‘Highly Satisfied’ in this context I sacrificed the final column.

So the results for the first question looked like this…
Image of results

For the Satisfaction element, I created an IFS Function in my final spreadsheet, to make 4 equaled 5, 3 equaled 4, 2 equaled 3 and 1 equaled 2. This put everything on the same scale.

Download the data

Any good survey tool should allow you to download the results to CSV format (Comma Separated Values) and any survey tool worth its salt should also provide a SPSS or data analysis format which codes each response with a number. So in my case I wanted to the most positive response to have the highest rank value. I am going to use Excel to sort the data, as that is/was my comfort zone and I think that it is good to start from a position of confidence and trust in the original figures prior to visualising them. Personally speaking after only a few months of working with RStudio, I now feel happy to complete the entire workflow in RStudio.

Tip: Split up your Importance (IMP) and Satisfaction (SAT) onto separate tabs in your Excel spreadsheet

Believe me this will make your life SO much easier, when it comes to doing your calculations.

Image of sort data slide

Step 2: Do the Maths

Having separated your data, create a new tab in your spreadsheet and create the rows defined in the column labelled A below. Then use the data on your Importance tab to populate column B. Start with one question-worth of data, we’ll get to the rest of your questions soon.

  • So for row 3 “total importance”: count all the entries in the importance column for that particular job from the first data row to the last – indicated by Ax [COUNTA].
  • For row 4 we want to count how many of these entries had a rating of 4 (Quite Important) and 5 (Very Important). So we use a COUNTIF formula to look at values greater than 3. As we aren’t dealing with decimals (yet), “>3” is accurate enough for this purpose.
  • For row 5 we need to find out what percentage of overall responses are Quite and Very Important. So we divide row 4 by row 3 and then times by 10.

1

A

B

2

Job I am a job In Excel speak

3

total importance How many entries are there in the column? =COUNTA(‘IMP data’!A2:Ax)

4

imp >3 How many entries are there in the column >3? =COUNTIF(‘IMP data’!A2:Ax, “>3”)

5

imp % Importance/total importance  X 10 =B4/B3*10

Create identical rows for satisfaction (after in my case, adjusting it upwards first, due to having a satisfaction scale of 4 parts), so that you end up with this…

The Opportunity Algorithm

Image of the opportunity algorithm

So you will need a final row, row 10 to calculate the Opportunity Score. This is outlined above, but to explain the contents of the image. You take the outcome of importance, which is your percentage, so 9.6 in this case and take away the outcome of satisfaction, again a percentage. So that would leave you with -0.1, then add back on the outcome percentage which is is 9.6 and you end up with the total opportunity score of 9.5.

outcome importance +
(outcome importance – outcome satisfaction)
= Opportunity Score

9.6 + (9.6 – 9.7) = 9.5

This is not the most exciting opportunity score, the most exciting ones are more than 10, or better yet more than 12.

Oh and here is why I said that you should start with the formulae for one question-worth of data and to split Importance and Satisfaction onto separate sheets…Once you have created one column for Importance and Satisfaction data you can simply drag the formulae across all of your questions.

Create R Food

Having done all that you need to get your data into columns for RStudio, specifically we want four columns to start off with: Job; Importance; Satisfaction; Opportunity; and Score.

So that it looks like this…

Image showing the columns in the spreadsheet

Tip to get your data from rows to these columns use the transpose(array) function in Excel, here it is in action…

Step 3: Get Sta(R)ted

Step 4: OoooH aRrrrr

Launch RStudio and have a go!

Image of the RStudio UI

Image of the RStudio UI

 

So now that you have done all of the preparation you can actually start to create plots. Before you do anything else you need to load your library. In this case we are using ggplot2 which is part of Hadley Wickham’s (and others) Tidyverse – packages for R with standardised concepts and grammar. This is a good place to start as I have found that some of the R packages are quite idiosyncratic and it can take a bit of time to wrap your head around the internal logic, whereas the Tidyverse stuff all made a lot of sense. This could be because I originally learned to code in the JavaScript iteration of Processing.

RStudio comes with a lot of packages already, so…

library(gpplot2)

If you needed to install it first and say you wanted the whole Tidyverse it would intuitively be…

install.packages(tidyverse)

Next you need to define your data and as you will recall we set that us as a .csv file on the desktop so lets call the data ‘jtdb’…

library(gpplot2)

jtbd <- read.csv("~/Desktop/JTBD_R/jtbd.csv", header = T)

#This is a comment in RMarkup, so lets also find out about the data itself by using the summary function

summary(jtbd)

Result in the console:

Screenshot of results in console

Well…that is jolly helpful information!

Now for the plot part:

#the plot mapping - telling where your data to map on the axis
ggplot(jtbd, aes(x=Importance, y=Satisfaction))+

# the "+" allows you to add another function, each function goes on its own layer, 
#so the order is important for example geom_point defines your data as points on the plot
  geom_point(color="blue", alpha=0.5, size=4)+

# To map aesthetics in a more intentional way - for instance adding labels, 
# we can use another function geom label which requires two variables 
#(in this case Importance and Satisfaction) 
# and tell it which column to get the labels from, in this case, the "Job" column
geom_label(aes(label = jtbd$Job), nudge_y = 0.08, nudge_x = 0.06, hjust="right", alpha=0.8, size=4)

This gives a starter plot like this…
Image of starter plot

Just to prove that it is easy to get started I gave myself a little challenge…

After that it is really all about the aesthetics with this particular plot. Specifically I added a data.frame (a container for data) for each aesthetic element such as text, lines, shaded polygons and areas. These were then called as shown below…

#Labels for the areas of the plot
labelover <- data.frame(
  x = 1, 
  y = 9.5, 
  label="OVERSERVED"
)

#Potential for disruption line
segment4 <- data.frame(
  x = 0,
  y = 2,
  xend = 7,
  yend = 10
)
#further down in the code where the plot is created - I call these#

ggplot(jtbd, aes(x=Importance, y=Satisfaction)) +

#Calling the text
geom_text(data = labelover, aes(x = x, y = y, label = label), size=5) +

#Calling the 'potential for disruption line' and adding some annotation to it
geom_segment(data = segment4, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + 
annotate("text", x = 5.1, y = 9, label = "Potential for disruption")

Step 5: Aesthetics

Well that provides an overview of the basics the rest will be based on your own aesthetic decisions. So I have provided the full code work-up with annotation below and recommend that you have a go and explore. Also if any of you are R and ggplot2 ‘pros’ I am open to feedback as I am sure that it could be improved.

To summarise, from my perspective using RStudio has provided the following benefits:

  • A fast way to create reproducible visualisations of JTBD data
  • An effective way to create prototype data visualisations for a range of design contexts
  • An inexpensive option when compared to other data packages. Although I can get a student discount on software at the moment, I was thinking about the long term and developing a skill with an open source package
  • It has helped me to think differently about data and to iterate visualisations in the same way that I would with other design projects
  • It is immensely satisfying

Further Reading:

RStudio online learning materials: https://www.rstudio.com/online-learning/#DataScience

RMarkdown gallery for ideas: https://rmarkdown.rstudio.com/gallery.html

RStudio Cheatsheets (they taught me pretty much everything that I couldn’t figure out from the built-in docs): https://www.rstudio.com/resources/cheatsheets/

Check out the Tidyverse: https://www.tidyverse.org/

References:

Josephy, J. (2016) ‘Jobs to be done – a better way to innovate’, Webcredible blog, 19 May. Available at: https://www.webcredible.com/blog/may16-jobs-be-done-better-way-innovate/ (Accessed: 23 March 2018).

Klement, A. (2016) ‘Designing features using job stories’, in Intercom on Jobs-to-be-Done’. pp. 28-80. Available at: https://www.intercom.com/books/jobs-to-be-done (Downloaded: 24 March 2017).

McGovern, G. (2010) The stranger’s long neck: How to deliver what your customers really want online. London: A & C Black.

McGovern, G. (2017) ‘Measuring customer effort with Top Tasks’ [PowerPoint presentation]. Available at: https://www.slideshare.net/uxbri/measuring-customer-effort-with-top-tasks-gerry-mcgovern (Accessed: 23 February 2017).

Tufte, E. R. (2001) The visual display of quantitative information. 2nd edn. USA, CT: Graphics Press. Reprint, 2013.

Ulwick, A. W. (2016) Jobs to be Done: Theory to Practice. Available at: http://amzn.eu/7BpXFp0 (Downloaded: 3 March 2017).

Ulwick, A. W. (2017) ‘The path to growth: The opportunity algorithm’, The marketing journal, 9 July. Available at: http://www.marketingjournal.org/the-path-to-growth-the-opportunity-algorithm-anthony-ulwick/ (Accessed: 23 March 2018).

 


Creative Commons Licence
Code for JTBD data visualisation by Fiona MacNeill is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Based on a work at http://www.marketingjournal.org/the-path-to-growth-the-opportunity-algorithm-anthony-ulwick/.

#Code for JTBD data visualisation (2018) by Fiona MacNeill is licensed under a Creative 
#Commons Attribution-ShareAlike 4.0 International License: 
#http://creativecommons.org/licenses/by-sa/4.0/
#Based on a work at 
#http://www.marketingjournal.org/the-path-to-growth-the-opportunity-algorithm-anthony-ulwick/.

#Loading the libraries needed for the visualisation
library(extrafont)
library(ggplot2)
font_import("Trebuchet MS")

#This data is provided for illustration only and is simulated - please use your own data
jtbd <- structure(list(id = structure(c("Look at time", "Check emails", "Review notifications", "Review activity", "Review heart rate", "Check Social Media", "Log exercise", "Text messaging", "Use voice control", "Pay by Touch", "Unlock by proximity", "Using apps"), .Label
                                      = c("Look at time", "Check emails", "Review notifications", "Review activity", "Review heart rate", "Check Social Media", "Log exercise", "Text messaging", "Use voice control", "Pay by Touch", "Unlock by proximity", "Using apps")),
                       Importance = c(9, 4, 8, 7, 5, 4, 7, 3, 1, 2, 2, 5),
                       Satisfaction = c(9.5, 6, 9, 9, 8, 6, 9, 6, 4, 4, 5, 7),
                       Oportunity.Score = c(8.5, 2, 7, 5, 2, 2, 5, 0, -2, 0, -1, 3)), .Names = c("Jobs", "Importance", "Satisfaction", "Opportunity Score"), class = "data.frame")
  

#Labels for the areas of the plot
labelover <- data.frame(
  x = 1, 
  y = 9.5, 
  label="OVERSERVED"
)

labeljust <- data.frame(
  x = 3, 
  y = 0.5, 
  label="JUST RIGHT"
)

labelunder <- data.frame(
  x = 9, 
  y = 0.5, 
  label="UNDERSERVED"
)

#Line segments for the midsection of the plot

#Top line
segment1 <- data.frame(
  x = 0, 
  y = 0, 
  xend = 10, 
  yend = 10
)

#Bottom line
segment2 <- data.frame(
  x = 5, 
  y = 0, 
  xend = 10, 
  yend = 10
)

#Ripe for disruption line
segment3 <- data.frame(
  x = 0,
  y = 4,
  xend = 5,
  yend = 10
)

#Potential for disruption line
segment4 <- data.frame(
  x = 0,
  y = 2,
  xend = 7,
  yend = 10
)

#Limited Opportunity
segment5 <- data.frame(
  x = 5,
  y = 0,
  xend = 10,
  yend = 10
)

#Solid Opportunity
segment6 <- data.frame(
  x = 6,
  y = 0,
  xend = 10,
  yend = 8
)

#High Opportunity
segment7 <- data.frame(
  x = 7.2,
  y = 0,
  xend = 10,
  yend = 5
)

#top triangle
tri.top <- data.frame(
  x = c(0, 10, 0), 
  y = c(10, 10, 0)
)

#middle triangle
tri.mid <- data.frame(
  x = c(5, 10, 0), 
  y = c(0, 10, 0)
)

#bottom triangle
tri.btm <- data.frame(
  x = c(10, 10, 5), 
  y = c(10, 0, 0)
)

#ulwick triangle
tri.ulw <- data.frame(
  x = c(0, 10), 
  y = c(2, 6)
)

#Definining global x and y variables
x <- jtbd$Importance
y <- jtbd$Satisfaction

#The plot itself 
ggplot(jtbd, aes(x=Importance, y=Satisfaction)) + 

#A grey area like you would find in Tony Ulwick's Opportunity Landscape visualisation 
geom_area(data = tri.ulw, aes(x = x, y = y), alpha=0.2, fill ="purple") + 

#A trend line to show the data trend which may or may not be useful 
#geom_smooth(method = "lm", se = FALSE) + 

#The triangle for OVERSERVED 
geom_polygon(data = tri.top, aes(x = x, y = y), fill ="grey90", alpha="0.6") + 

#The triangle for JUST RIGHT 
geom_polygon(data = tri.mid, aes(x = x, y = y), fill ="grey70", alpha="0.5") + 

#The triangle for UNDERSERVED 
geom_polygon(data = tri.btm, aes(x = x, y = y), fill ="grey65", alpha="0.7") + 

#If you wanted to highlight that sweet spot, these areas could be turned on 

#geom_segment(data = segment1, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + 
#geom_segment(data = segment2, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + 
geom_segment(data = segment3, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + annotate("text", x = 2, y = 8, label = "Ripe for disruption") + 
geom_segment(data = segment4, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + annotate("text", x = 5.1, y = 9, label = "Potential for disruption") + 
geom_segment(data = segment5, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + annotate("text", x = 9.3, y = 8, label = "Limited Opportunity", size = 3) + 
geom_segment(data = segment6, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + annotate("text", x = 9, y = 6, label = "Solid Opportunity (Opp >10)", size = 3) +
geom_segment(data = segment7, aes(x = x, y = y, xend = xend, yend = yend), alpha=0.2, linetype=2) + annotate("text", x = 8.8, y = 4, label = "High Opportunity (Opp >12)", size = 3) +
  annotate("text", x = 9, y = 1.5, label = "Extreme Opportunity (Opp >15)", size = 3) +
  
#Our very helpful labels
geom_text(data = labelover, aes(x = x, y = y, label = label), size=5) +
geom_text(data = labeljust, aes(x = x, y = y, label = label), size=5) +
geom_text(data = labelunder, aes(x = x, y = y, label = label), size=5) +
  
#Our job labels
#geom_label(aes(label = jtbd$Job), nudge_y = 0.24, nudge_x = 0.04, hjust="right", alpha=0.8, size=4) +
  
#The points for our jobs
geom_point(color="blue", alpha=0.5, size=4) +
  
#Set the scales the way that I want them - aka not as decimals!
scale_x_continuous(limits=c(0, 10), breaks = c(0, 2, 4, 6, 8, 10), labels = scales::comma_format(digits = 1)) +
scale_y_continuous(limits=c(0, 10), breaks = c(0, 2, 4, 6, 8, 10), labels = scales::comma_format(digits = 1)) +
  
#Hide the legend as it is not helpful this time
theme(legend.position = "none") +
  
#Labelling the plot
  labs(
    title ="DEMO: Tasks on a smart watch", 
    subtitle = "a subtitle") +
  theme(panel.background = element_rect(fill = "white", colour = "black"), legend.position = "none")



Print Friendly, PDF & Email
, , ,

Leave a Reply

Your email address will not be published. Required fields are marked *

Skip to toolbar