Empirical Project 8 Working in R

Download the code

To download the code chunks used in this project, right-click on the download link and select ‘Save Link As…’. You’ll need to save the code download to your working directory, and open it in RStudio.

Don’t forget to also download the data into your working directory by following the steps in this project.

Getting started in R

For this project you will need the following packages:

If you need to install any of these packages, run the following code:

install.packages(c("tidyverse", "readxl", "knitr"))

You can import the libraries now, or when they are used in the R walk-throughs below.

library(tidyverse)
library(readxl)
library(knitr)

Part 8.1 Cleaning and summarizing the data

Learning objectives for this part

  • detect and correct entries in a dataset
  • recode variables to make them easier to analyse
  • calculate percentiles for subsets of the data.

So far we have been working with data that was formatted correctly. However, sometimes the datasets will be messier and we will need to clean the data before starting any analysis. Data cleaning involves checking that all variables are formatted correctly, all observations are entered correctly (e.g. no typos), and missing values are coded in a way that the software you are using can understand. Otherwise, the software will either not be able to analyse your data, or only analyse the observations it recognizes, which could lead to incorrect results and conclusions.

In the data we will use, the European Values Study (EVS) data has been converted to an Excel spreadsheet from another program, so there are some entries that were not converted correctly and some variables that need to be recoded (for example, replacing words with numbers, or replacing one number with another).

Download the EVS data and documentation:

Although we will be performing our analysis in R, the data has been provided as an Excel spreadsheet, so look at the spreadsheet in Excel first to understand its structure.

  1. The Excel spreadsheet contains multiple worksheets, one for each wave, that need to be joined together to create a single ‘tibble’ (like a spreadsheet for R). The variable names currently do not tell us what the variable is, so we need to provide labels and short descriptions to avoid confusion.

R walk-through 8.1 Importing the data into R

As we are importing an Excel file, we use the read_excel function from the readxl package. The file is called Project-8-datafile.xlsx and contains four worksheets that contain the data, named ‘Wave 1’ through to ‘Wave 4’. We will load the worksheets one by one and add them to the previous worksheets using the rbind function, which attaches objects with the same number of rows to each other. The final output is called lifesat_data.

The variable names provided in the spreadsheet are not very specific (a combination of letters and numbers that don’t tell us what the variable measures), we can assign a label (labels) and a short description (shortDescription, taken from the Variable Correspondence file) to each variable for reference later. The attr function does this for us (note that the order in each list should be the same as the order of the variable names).

attr(lifesat_data, "labels")  
c("EVS-wave", "Country/region",
  "Respondent number",
  "Health", "Life satisfaction",
  "Work Q1", "Work Q2", "Work Q3",
  "Work Q4", "Work Q5",
  "Sex", "Age", "Marital status",
  "Number of children",
  "Education", "Employment",
  "Monthly household income")

attr(lifesat_data, "shortDescription")  
c("EVS-wave",
  "Country/region",
  "Original respondent number",
  "State of health (subjective)",
  "Satisfaction with your life",
  "To develop talents you need to have a job",
  "Humiliating to receive money w/o working for it",
  "People who don't work become lazy",
  "Work is a duty towards society",
  "Work comes first even if it means less spare time",
  "Sex",
  "Age",
  "Marital status",
  "How many living children do you have",
  "Educational level (ISCED-code one digit)",
  "Employment status",
  "Monthly household income (× 1,000s PPP euros)")

Throughout this project we will refer to the variables using their original names, but you can use the attr function again to look up the attributes of each variable (the following code uses the variable X028 to show how you can do this):

attr(lifesat_data, "labels")[
  attr(lifesat_data, "names") == "X028"]
## [1] "Employment"
attr(lifesat_data, "shortDescription")[
  attr(lifesat_data, "names") == "X028"]
## [1] "Employment status"

In general, data dictionaries and variable correspondences are useful because they contain important information about what each variable represents and how it is measured, which usually cannot be summarized in a short variable name.

  1. Now we will take a closer look at how some of the variables were measured.

Before doing any data analysis, it is important to check that all variables are coded in a way that the software can recognize. This process involves checking how missing values are coded (usually these need to be coded in a particular way for each software), and that numerical variables (numbers) only contain numbers and not text (in order to calculate summary statistics).

  1. We will now check and clean the dataset so it is ready to use.

R walk-through 8.2 Cleaning data and splitting variables

Inspect the data and recode missing values

R stores variables as different types depending on the kind of information the variable represents. For categorical data, where, as the name suggests, data is divided into a number of groups, such as country or occupation, the variables can be stored as text (chr). Numerical data (numbers that do not represent categories) should be stored as numbers (num), but sometimes R imports numerical values as text, which can cause problems later when we use these variables in calculations. The str function shows what type each variable is being stored as.

str(lifesat_data)
## Classes 'tbl_df', 'tbl' and 'data.frame':    164997 obs. of  17 variables:
##  $ S002EVS: chr  "1981-1984" "1981-1984" "1981-1984" "1981-1984" ...
##  $ S003   : chr  "Belgium" "Belgium" "Belgium" "Belgium" ...
##  $ S006   : num  1001 1002 1003 1004 1005 ...
##  $ A009   : chr  "Fair" "Very good" "Poor" "Very good" ...
##  $ A170   : chr  "9" "9" "3" "9" ...
##  $ C036   : chr  ".a" ".a" ".a" ".a" ...
##  $ C037   : chr  ".a" ".a" ".a" ".a" ...
##  $ C038   : chr  ".a" ".a" ".a" ".a" ...
##  $ C039   : chr  ".a" ".a" ".a" ".a" ...
##  $ C041   : chr  ".a" ".a" ".a" ".a" ...
##  $ X001   : chr  "Male" "Male" "Male" "Female" ...
##  $ X003   : num  53 30 61 60 60 19 38 39 44 76 ...
##  $ X007   : chr  "Single/Never married" "Married" "Separated" "Married" ...
##  $ X011_01: chr  ".a" ".a" ".a" ".a" ...
##  $ X025A  : chr  ".a" ".a" ".a" ".a" ...
##  $ X028   : chr  "Full time" "Full time" "Unemployed" "Housewife" ...
##  $ X047D  : chr  ".a" ".a" ".a" ".a" ...
##  - attr(*, "labels")= chr  "EVS-wave" "Country/region" "Respondent number" "Health" ...
##  - attr(*, "shortDescription")= chr  "EVS-wave" "Country/region" "Original respondent number" "State of health (subjective)" ...

You can see that some variables have the value .a, which is how missing values were coded in the original dataset. We start by recoding all missing values as NA (which is how R recognizes and usually codes missing values).

lifesat_data[lifesat_data == ".a"]  NA

Recode the life satisfaction variable

To recode the life satisfaction variable (A170), we can use logical indexing (asking R to check if a certain condition is satisfied or not) to select all observations where the variable A170 is equal to either ‘Dissatisfied’ or ‘Satisfied’ and assign the value 1 or 10 respectively. This variable was imported as a chr (text) instead of num variable because some of the entries included text (the .a for missing values). After changing the text into numerical values, we use the as.numeric function to convert the variable into a numerical type.

lifesat_data$A170[lifesat_data$A170 == 
  "Dissatisfied"]  1
lifesat_data$A170[lifesat_data$A170 == 
  "Satisfied"]  10
lifesat_data$A170  as.numeric(lifesat_data$A170)

Recode the variable for number of children

We repeat this process (using logical indexing and then converting to numerical variable) for the variable indicating the number of children (X011_01).

lifesat_data$X011_01[lifesat_data$X011_01 == 
  "No children"]  0
lifesat_data$X011_01  
  as.numeric(lifesat_data$X011_01)

Replace text with numbers for multiple variables

When we have to recode multiple variables with the same mapping of text to numerical value (i.e. a word/phrase corresponds to the same number for all variables), we can use the tidyverse library and the piping (%>%) operator to run the same sequence of steps in one go for all variables, instead of repeating the process used above for each variable. The punctuation %>% links the commands together. (At this stage, don’t worry if you can’t understand exactly what this code does, as long as you can adapt it for similar situations. For a more detailed introduction to piping, see the University of Manchester’s Econometric Computing Learning Resource.)

lifesat_data  lifesat_data %>%
  mutate_at(c("C036", "C037", "C038", "C039", "C041"),
    funs(recode(., 
      "Strongly disagree" = "1", 
      "Disagree" = "2", 
      "Neither agree nor disagree" = "3",
      "Agree" = "4",
      "Strongly agree" = "5"))) %>%
  mutate_at(c("A009"), funs(recode(., 
    "Very poor" = "1",
    "Poor" = "2",
    "Fair" = "3",
    "Good" = "4",
    "Very good" = "5"))) %>%
  mutate_at(c("A009", "C036", "C037",
    "C038", "C039", "C041", "X047D"), funs(as.numeric))

Split a variable containing numbers and text

To split the education variable (X025A) into two new columns, we use the separate function, which creates two new variables called Education_1 and Education_2 containing the numeric value and the text description respectively. Then we use the mutate_at function to convert Education_1 into a numeric variable.

lifesat_data  lifesat_data %>% 
  separate(X025A, c("Education_1", "Education_2"),
    " : ", remove = FALSE) %>%
  mutate_at("Education_1", funs(as.numeric))

Although the paper we are following only considered individuals aged 25–80 who were not students, we will retain the other observations in our analysis. However, we need to remove any observations that have missing values for A170, X028, or any of the other variables except X047D. Removing missing data ensures that any summary statistics or analysis we do is done on the same set of data (without having to always account for the fact that some values are missing), and is fine as long as the data are missing at random (i.e. there is no particular reason why certain observations are missing and not others).

  1. In your dataset, remove all rows in all waves that have missing data for A170. Do the same for:

    • X003, X028, X007 and X001 in all waves
    • A009 in Waves 1, 2, and 4 only
    • C036, C037, C038, C039, C041 and X047D in Waves 3 and 4 only
    • X011_01 and X025A, in Wave 4.

    R walk-through 8.3 gives guidance on how to do this.

R walk-through 8.3 Dropping specific observations

As not all questions were asked in all waves, we have to be careful when dropping observations with missing values for certain questions, to avoid accidentally dropping an entire wave of data. For example, information on self-reported health (A009) was not recorded in Wave 3, and questions on work attitudes (C036 to C041) and information on household income are only asked in Waves 3 and 4. Furthermore, information on the number of children (X011_01) and education (X025A) are only collected in the final wave.

include  names(lifesat_data) %in% c(
  "X003", "A170", "X028", "X007", "X001")
lifesat_data  lifesat_data[
  complete.cases(lifesat_data[, include]), ]

Next we will look at variables that were only present in some waves. For each variable/group of variables, we have to only look at the particular wave(s) in which the question was asked, then keep the observations with complete information on those variables. As before, we make lists of variables that only feature in Waves 1, 2, and 4 (A009—stored in include_wave_1_2_4), Waves 3 and 4 (C036 to C041, X047D—stored in include_wave_3_4), and Wave 4 only (X011_01 and X025A—stored in include_wave_4). Again we use the complete.cases function, but combine it with the logical OR operator | to include all observations for waves that did not ask that question, along with the complete cases for that question in the other waves. For example, the first command keeps an observations if it is in Wave 1 (1981-1984) or Wave 2 (1990-1993), or is in Waves 3 or 4 with complete information.

# A009 is not in Wave 3.
include_wave_1_2_4  names(lifesat_data) %in% c("A009")

include_wave_3_4  names(lifesat_data) %in% 
# Work attitudes and income are in Waves 3 and 4.
c("C036", "C037", "C038", "C039", "C041", "X047D")

# Number of children and education are in Wave 4.
include_wave_4  names(lifesat_data) %in% 
c("X011_01", "X025A")

lifesat_data  lifesat_data %>%
  filter(S002EVS == "1981-1984" | 
    S002EVS == "1990-1993" | 
    complete.cases(lifesat_data[, include_wave_3_4])) 

lifesat_data  lifesat_data %>%
  filter(S002EVS != "2008-2010" |  
    complete.cases(lifesat_data[, include_wave_4]))

lifesat_data  lifesat_data %>%
  filter(S002EVS == "1999-2001" |  
    complete.cases(lifesat_data[, include_wave_1_2_4]))
  1. We will now create two variables, work ethic and relative income, which we will use in our comparison of life satisfaction.

R walk-through 8.4 Calculating averages and percentiles

Calculate average work ethic score

We use the rowMeans function to calculate the average work ethic score for each observation (workethic) based on the five survey questions related to working attitudes (C036 to C041).

lifesat_data$work_ethic  rowMeans(lifesat_data[, 
  c("C036", "C037", "C038", "C039", "C041")])

Calculate income percentiles for each individual

R provides a handy function (ecdf) to obtain an individual’s relative income as a percentile. As we want a separate income distribution for each wave, we use the group_by function. Finally, we tidy up the results by converting each value into a percentage and rounding to a single decimal place.

The ecdf function will not work if there is missing data. Since we don’t have income data for Waves 1 and 2, we will have to split the data when working out the percentiles. First we store all observations from Waves 3 and 4 separately in a temporary dataset (df.new), only keep observations that have income values (!is.na(X047D)), calculate the percentile values with the ecdf function, and save the values in a variable called percentile (which will be ‘NA’ for Waves 1 and 2). Then we recombine this data with the original observations from Waves 1 and 2.

# Create the percentile variable for Waves 3 and 4
df.new  lifesat_data %>%
  # Select Waves 3 and 4
  subset(S002EVS %in% c("1999-2001", "2008-2010")) %>%
  # Drop observations with missing income data
  filter(!is.na(X047D)) %>%
  # Calculate the percentiles per wave
  group_by(S002EVS) %>%
  # Calculate relative income as a percentile
  mutate(., percentile = ecdf(X047D)(X047D)) %>%
  # Express the variable as a percentage
  mutate(., percentile = round(percentile * 100, 1))

# Recombine Waves 3 and 4 with the Waves 1 and 2
lifesat_data  lifesat_data %>%
  # Select Waves 1 and 2
  subset(S002EVS %in% c("1981-1984", "1990-1993")) %>%
  # Create an empty variable so old and new dataframes are 
  # same size
  mutate(., percentile = NA) %>%
  # Recombine the data
  bind_rows(., df.new)
  1. Now we have all the variables we need in the format we need. We will make some tables to summarize the data. Using the data for Wave 4:
Male Female
Mean Standard deviation Mean Standard deviation
Life satisfaction
Self-reported health
Work ethic
Age
Education
Number of children

Summary statistics by gender, European Values Study.

Figure 8.1 Summary statistics by gender, European Values Study.

 

R walk-through 8.5 Calculating summary statistics

Create a table showing employment status, by country

One of the most useful R functions to obtain summary information is summarize, which produces many different summary statistics for a set of data. First we select the data for Wave 4 and group it by country (S003) and employment type (X028). Then we use the summarize function to obtain the number of observations for each combination of country and employment type. With the data effectively grouped by country only (a single observation contains the number of observations for each level (category) of X028), we can use the sum function to obtain a total count (number of observations) per country and calculate relative frequencies (saved as freq).

lifesat_data %>%
  # Select Wave 4 only
  subset(S002EVS == "2008-2010") %>%
  # Group by country and employment
  group_by(S003, X028) %>%
  # Count the number of observations by group
  summarize(n = n()) %>%
  # Calculate the relative frequency (percentages)
  mutate(freq = round(n / sum(n), 4) * 100) %>%
  # Drop the count variable
  select(-n) %>%
  # Reshape the data into a neat table
  spread(X028, freq) %>%
  kable(., format = "markdown")

Calculate summary statistics by gender

We can also obtain summary statistics on a number of variables at the same time using the summarize_at function. To obtain the mean value for each of the required variables, grouped by the gender variable (X001), we can use the piping operator again:

# Obtain mean value
lifesat_data %>%
  # We only require Wave 4.
  subset(S002EVS == "2008-2010") %>%
  # Group by gender
  group_by(X001) %>%
  summarize_at(c("A009", "A170", "work_ethic", "X003",
    "Education_1", "X011_01"), mean)
## # A tibble: 2 x 7
##   X001    A009  A170 work_ethic  X003 Education_1 X011_01
##                       
## 1 Female  3.60  6.93       3.64  47.3        3.05    1.69
## 2 Male    3.77  7.03       3.72  46.9        3.14    1.55

We can repeat the same steps to obtain the standard deviation, again grouped by gender.

# Obtain standard deviation
lifesat_data %>%
  subset(S002EVS == "2008-2010") %>%
  group_by(X001) %>%
  summarize_at(c("A009", "A170", "work_ethic", "X003",
    "Education_1", "X011_01"), sd)
## # A tibble: 2 x 7
##   X001    A009  A170 work_ethic  X003 Education_1 X011_01
##                       
## 1 Female 0.969  2.32      0.765  17.5        1.40    1.39
## 2 Male   0.929  2.28      0.757  17.4        1.31    1.43

We can combine these two operations into a single code block and ask summarize_at to perform multiple functions on multiple variables at the same time.

# Obtain mean and standard deviations
lifesat_data %>%
  subset(S002EVS == "2008-2010") %>%
  group_by(X001) %>%
  summarise_at(c("A009", "A170", "work_ethic", "X003",
    "Education_1", "X011_01"), funs(mean, sd))
## # A tibble: 2 x 13
##   X001   A009_mean A170_mean work_ethic_mean X003_mean Education_1_mean
##                                          
## 1 Female      3.60      6.93            3.64      47.3             3.05
## 2 Male        3.77      7.03            3.72      46.9             3.14
## # ... with 7 more variables: X011_01_mean , A009_sd ,
## #   A170_sd , work_ethic_sd , X003_sd ,
## #   Education_1_sd , X011_01_sd 

In this example, R automatically appends the function name (in this case, _mean or _sd) to the end of the variables. If you need these variables in subsequent calculations, you need to reformat the data and variable names.

Part 8.2 Visualizing the data

Note

You will need to have done Questions 1–5 in Part 8.1 before doing this part.

Learning objectives for this part

  • use column charts, line charts, and scatterplots to visualize data
  • calculate and interpret correlation coefficients.

We will now create some summary charts of the self-reported measures (work ethic and life satisfaction), starting with column charts to show the distributions of values. Along with employment status, these are the main variables of interest, so it is important to look at them carefully before doing further data analysis.

The distribution of work ethic and life satisfaction may vary across countries but may also change over time within a country, especially since the surveys are conducted around once a decade. To compare distributions for a particular country over time, we have to use the same horizontal axis, so we will first need to make a separate frequency table for each distribution of interest (survey wave). Also, since the number of people surveyed in each wave may differ, we will use percentages instead of frequencies as the vertical axis variable.

  1. Use the data from Wave 3 and Wave 4 only, for three countries of your choice:

R walk-through 8.6 Calculating frequencies and percentages

First we need to create a frequency table of the work_ethic variable for each wave. This variable only takes values from 1.0 to 5.0 in increments of 0.2 (since it is an average of five whole numbers), so we can group by each value and count the number of observations in each group using the summarize function. Once we have counted the number of observations that have each value (separately for each wave), we compute the percentages by dividing these numbers by the total number of observations for that wave. For example, if there are 50 observations between 1 and 1.2, and 1,000 observations in that wave, the percentage would be 5%.

waves   c("1999-2001", "2008-2010")
df.new  lifesat_data %>% 
  # Select Waves 3 and 4
  subset(., S002EVS %in% waves) %>%
  # We use Germany in this example.
  subset(., S003 == "Germany") %>%
  # Group by each wave and the discrete values of 
  # work_ethic
  group_by(S002EVS, work_ethic) %>%
  summarize (freq = n()) %>%
  mutate(percentage = (freq / sum(freq))*100)

The frequencies and percentages are saved in a new dataframe called df.new. If you want to look at it, type view(df.new) into the console window. Now that we have the percentages and frequency data, we use ggplot to plot a column chart. To overlay the column charts for both waves and make sure that the plot for each wave is visible, we use the alpha option in the geom_bar function to set the transparency level (try changing the transparency to see how it affects your chart’s appearance).

p  ggplot(df.new, aes(work_ethic, percentage, 
  fill = S002EVS, color = S002EVS)) +    
  # Set the colour and fill of each wave 
  geom_bar(stat = "identity", position = "identity", 
    alpha = 0.2) + 
  theme_bw() +
  ggtitle(paste("Distribution of work ethic for Germany"))

print(p)

Distribution of work ethic scores for Germany.

Figure 8.2 Distribution of work ethic scores for Germany.

  1. We will use line charts to make a similar comparison for life satisfaction (A170) over time. Using data for countries that are present in Waves 1 to 4:

R walk-through 8.7 Plotting multiple lines on a chart

Calculate average life satisfaction, by wave and country

Before we can plot the line charts, we have to calculate the average life satisfaction for each country in each wave.

In R walk-through 8.5 we produced summary tables, grouped by country and employment status. We will copy this process, but now we only require mean values. Countries that do not report the life satisfaction variable for all waves will have an average life satisfaction of ‘NA’. Since each country is represented by a row in the summary table, we use the rowwise and na.omit functions to drop any countries that do not have a value for the average life satisfaction for all four waves.

df.new  lifesat_data %>%
  group_by(S002EVS, S003) %>%
  # Calculate average life satisfaction per wave and country
  summarize(avg_lifesat = round(mean(A170), 2)) %>%
  # Reshape the data to one row per country
  spread(S002EVS, avg_lifesat) %>%
  # Set subsequent operations to be performed on each row
  rowwise() %>%
  # Drop any rows with missing observations
  na.omit() %>%
  print()
## # A tibble: 11 x 5
##    S003             `1981-1984` `1990-1993` `1999-2001` `2008-2010`
##                                           
##  1 Belgium                 7.37        7.6         7.42        7.63
##  2 Denmark                 8.21        8.17        8.31        8.41
##  3 France                  6.71        6.77        6.98        7.05
##  4 Germany                 7.22        7.03        7.43        6.77
##  5 Iceland                 8.05        8.01        8.08        8.07
##  6 Ireland                 7.82        7.88        8.21        7.82
##  7 Italy                   6.65        7.3         7.18        7.4 
##  8 Netherlands             7.75        7.77        7.83        7.99
##  9 Northern Ireland        7.66        7.88        8.07        7.82
## 10 Spain                   6.6         7.15        6.97        7.29
## 11 Sweden                  8.03        7.99        7.62        7.68

Create a line chart for average life satisfaction

To draw the lines for all countries on a single plot, we need to use the gather function to transform the data into the correct format to use in ggplot.

df.new %>%
  gather(Wave, mean, `1981-1984`:`2008-2010`) %>%
  # Group the data by country 
  ggplot(., aes(x = Wave, y = mean, color = S003)) + 
    geom_line(aes(group = S003, linetype = S003), 
      size = 1) +
    theme_bw()

Line chart of average life satisfaction across countries and survey waves.

Figure 8.3 Line chart of average life satisfaction across countries and survey waves.

correlation coefficient
A numerical measure, ranging between 1 and −1, of how closely associated two variables are—whether they tend to rise and fall together, or move in opposite directions. A positive coefficient indicates that when one variable takes a high (low) value, the other tends to be high (low) too, and a negative coefficient indicates that when one variable is high the other is likely to be low. A value of 1 or −1 indicates that knowing the value of one of the variables would allow you to perfectly predict the value of the other. A value of 0 indicates that knowing one of the variables provides no information about the value of the other.

After describing patterns in our main variables over time, we will use correlation coefficients and scatter plots to look at the relationship between these variables and the other variables in our dataset.

  1. Using the Wave 4 data:
Variable Life satisfaction Work ethic
Age
Education
Full-time employment
Gender
Self-reported health
Income
Number of children
Relative income
Life satisfaction 1
Work ethic 1

Correlation between life satisfaction, work ethic and other variables, Wave 4.

Figure 8.4 Correlation between life satisfaction, work ethic and other variables, Wave 4.

R walk-through 8.8 Creating dummy variables and calculating correlation coefficients

To obtain the correlation coefficients between variables, we have to make sure that all variables are numeric. However, the data on sex and employment status are coded using text, so we need to create two new variables (gender and employment). (We choose to create a new variable rather than overwrite the original variable so that even if we make a mistake, the raw data is still preserved). We can use the ifelse function to make the value of the variable conditional on whether a logical statement (e.g. X001 == "Male") is satisfied or not. As shown below, we can nest ifelse statements to create more complex conditions, which is useful if the variable contains more than two values. We used two ifelse statements for the unemployment variable (X028) so that the new variable will be 1 for full-time employed, 0 for unemployed, and NA if neither condition is satisfied.

Although we had previously removed observations with missing values (‘NA’), we have reintroduced ‘NA’ values in the previous step when creating the employment variable. This would cause problems when obtaining the correlation coefficients using the cor function, so we include the option use = "pairwise.complete.obs" to exclude observations containing ‘NA’ from the calculations.

lifesat_data %>%
  # We only require Wave 4.
  subset(S002EVS == "2008-2010") %>%
  # Create a gender variable
  mutate(gender = as.numeric(
    ifelse(X001 == "Male", 0, 1))) %>%
  # Provide nested condition for employment
  mutate(employment = as.numeric(
    ifelse(X028 == "Full time", 1,
      ifelse(X028 == "Unemployed", 0, NA)))) %>%
  select(X003, Education_1, employment,
    gender, A009, X047D, X011_01,
    percentile, A170, work_ethic) %>%
  # Store output in matrix       
  cor(., use = "pairwise.complete.obs") -> M

# Only interested in two columns, so use matrix indexing
M[, c("A170", "work_ethic")]
##                    A170  work_ethic
## X003        -0.08220504  0.13333312
## Education_1  0.09363900 -0.14535643
## employment   0.18387286 -0.02710835
## gender      -0.02105535 -0.04781557
## A009         0.37560287 -0.07320865
## X047D        0.23542497 -0.15226999
## X011_01     -0.01729372  0.08925452
## percentile   0.29592968 -0.18679869
## A170         1.00000000 -0.03352464
## work_ethic  -0.03352464  1.00000000

Next we will look at the relationship between employment status and life satisfaction, and investigate the paper’s hypothesis that this relationship varies with a country’s average work ethic.

  1. Using the data from Wave 4, carry out the following:

R walk-through 8.9 Calculating group means

Calculate average life satisfaction and differences in average life satisfaction

We can achieve the tasks in Question 4(a) and (b) in one go using an approach similar to that used in R walk-through 8.5, although now we are interested in calculating the average life satisfaction by country and employment type. Once we have tabulated these means, we can compute the difference in the average values. We will create two new variables: D1 for the difference between the average life satisfaction for full-time employed and unemployed, and D2 for the difference in average life satisfaction for full-time employed and retired individuals.

# Set the employment types that we want to report
employment_list = c("Full time", "Retired", "Unemployed")
df.employment  lifesat_data %>%
  # Select Wave 4
  subset(S002EVS == "2008-2010") %>%
  # Select only observations with these employment types
  subset(X028 %in% employment_list) %>%
  # Group by country and then employment type
  group_by(S003, X028) %>%
  # Calculate the mean by country/employment type group
  summarize(mean = mean(A170)) %>%
  # Reshape to one row per country
  spread(X028, mean) %>%
  # Create the difference in means
  mutate(D1 = `Full time` - Unemployed, 
    D2 = `Full time` - Retired)

kable(df.employment, format = "markdown")

Make a scatterplot, sorted according to work ethic

In order to plot the differences ordered by the average work ethic, we first need to get all data from Wave 4 (using subset), summarize the work_ethic variable by country (group_by then summarize), and store the results in a temporary dataframe (df.work.ethic).

df.work_ethic  lifesat_data %>%
  subset(S002EVS == "2008-2010") %>%    
  select(S003, work_ethic) %>%
  group_by(S003) %>%
  summarize(mean_work = mean(work_ethic))

We can now combine the average work_ethic data with the table containing the difference in means (using inner_join to match the data correctly by country) and make a scatterplot using ggplot. This process can be repeated for the difference in means between full-time employed and retired individuals by changing y = D1 to y = D2 in the ggplot function.

df.employment %>%
  # Combine with the average work ethic data
  inner_join(., df.work_ethic, by = "S003") %>%
  ggplot(., aes(y = D1, x = mean_work)) +
    geom_point(stat = "identity") + 
    xlab("Work ethic") + ylab("Difference") +
    ggtitle("Difference in wellbeing between the 
      full-time employed and the unemployed") +
    theme_bw() + 
    # Rotate the country names
    theme(axis.text.x = element_text(
      angle = 90, hjust = 1), 
      plot.title = element_text(hjust = 0.5))

Difference in life satisfaction (wellbeing) between the full-time employed and the unemployed vs average work ethic.

Figure 8.5 Difference in life satisfaction (wellbeing) between the full-time employed and the unemployed vs average work ethic.

To calculate correlation coefficients, use the cor function. You can see that the correlation between average work ethic and difference in life satisfaction is quite weak for employed vs unemployed, but moderate and positive for employed vs retired.

# Full-time vs unemployed
cor(df.employment$D1, df.work_ethic$mean_work)
## [1] -0.1575654
# Full-time vs retired
cor(df.employment$D2, df.work_ethic$mean_work)
## [1] 0.4842609

So far we have described the data using tables and charts, but have not made any statements about whether what we observe is consistent with an assumption that work ethic has no bearing on differences in life satisfaction between different subgroups. In the next part, we will assess the relationship between employment status and life satisfaction and assess whether the observed data lead us to reject the above assumption.

Part 8.3 Confidence intervals for difference in the mean

Note

You will need to have done Questions 1–5 in Part 8.1 before doing this part. Part 8.2 is not necessary but is helpful to get an idea of what the data looks like.

Learning objectives for this part

  • calculate and interpret confidence intervals for the difference in means between two groups.

The aim of this project was to look at the empirical relationship between employment and life satisfaction.

When we calculate differences between groups, we collect evidence which may or may not support a hypothesis that life satisfaction is identical between different subgroups. Economists often call this testing for statistical significance. In Part 6.2 of Empirical Project 6, we constructed 95% confidence intervals for means, and used a rule of thumb to assess our hypothesis that the two groups considered were identical (at the population level) in the variable of interest. Now we will learn how to construct confidence intervals for the difference in two means, which allows us to make such a judgment on the basis of a single confidence interval.

Remember that the width of a confidence interval depends on the standard deviation and number of observations. (Read Part 6.2 of Empirical Project 6 to understand why.) When making a confidence interval for a sample mean (such as the mean life satisfaction of the unemployed), we use the sample standard deviation and number of observations in that sample (unemployed people) to obtain the standard error of the sample mean.

When we look at the difference in means (such as life satisfaction of employed minus unemployed), we are using data from two groups (the unemployed and the employed) to make one confidence interval. To calculate the standard error for the difference in means, we use the standard errors (SE) of each group:

This formula requires the two groups of data to be independent, meaning that the data from one group is not related, paired, or matched with data from the other group. This assumption is reasonable for the life satisfaction data we are using. However, if the two groups of data are not independent, for example if the same people generated both groups of data (as in the public goods experiments of Empirical Project 2), then we cannot use this formula.

Once we have the new standard error and number of observations, we can calculate the width of the confidence interval (distance from the mean to one end of the interval) as before. The width of the confidence interval tells us how precisely the difference in means was estimated, and this precision tells us how compatible the sample data is with the hypothesis that there is no difference between the population means. If the value 0 falls outside the 95% confidence interval, this implies that the data is unlikely to have been sampled from a population with equal population means. For example, if the estimated difference is positive, and the 95% confidence interval is not wide enough to include 0, we can be reasonably confident that the true difference is positive too. In other words, it tells us that the p-value for the difference we have found is less than 0.05, so we would be very unlikely to find such a big difference if the population means were in fact identical.

In Figure 8.6, for Great Britain, we can be reasonably confident that the true difference in average life satisfaction between the full-time employed and the unemployed is positive. However, for Spain, we do not have strong evidence of a real difference in mean life satisfaction.

95% confidence intervals for the average difference in life satisfaction (full-time employed minus unemployed) in Great Britain and Spain.

Figure 8.6 95% confidence intervals for the average difference in life satisfaction (full-time employed minus unemployed) in Great Britain and Spain.

  1. We will apply this method to make confidence intervals for differences in life satisfaction. Choose three countries: one with an average work ethic in the top third of scores, one in the middle third, and one in the lower third. (See R walk-through 6.3 for help on calculating confidence intervals and adding them to a chart.)

R walk-through 8.10 Calculating confidence intervals and adding error bars

We will use Turkey, Spain, and Great Britain as example countries in the top, middle, and bottom third of work ethic scores respectively.

In the tasks in Questions 1(a) and (b) we will obtain the means, standard errors, and 95% confidence intervals step-by-step, then for Question 1(c) we show how to use a shortcut to obtain confidence intervals from a single function.

Calculate confidence intervals manually

We obtained the difference in means in R walk-through 8.9 (D1 and D2), so now we can calculate the standard error of the means for each country of interest.

# List chosen countries
country_list  c("Turkey", "Spain", "Great Britain")
df.employment.se  lifesat_data %>%
  # Select Wave 4
  subset(S002EVS == "2008-2010") %>% 
  # Select the employment types we are interested in
  subset(X028 %in% employment_list) %>%
  # Select countries
  subset(S003 %in% country_list) %>%
  # Group by country and employment type
  group_by(S003, X028) %>%
  # Calculate the standard error of each group mean
  summarize(se = sd(A170) / sqrt(n())) %>%
  spread(X028, se) %>%
  # Calculate the SE of difference
  mutate(D1_SE = sqrt(`Full time`^2 + Unemployed^2), 
    D2_SE = sqrt(`Full time`^2 + Retired^2)) 

We can now combine the standard errors with the difference in means, and compute the 95% confidence interval width.

df.employment  df.employment %>%
  # Select chosen countries
  subset(S003 %in% country_list) %>%
  # We only need the differences.
  select(-`Full time`, -Retired, -Unemployed) %>%
  # Join the means with the respective SEs
  inner_join(., df.employment.se, by = "S003") %>%
  select(-`Full time`, -Retired, -Unemployed) %>%
  # Compute confidence interval width for both differences
  mutate(CI_1 = 1.96 * D1_SE, CI_2 = 1.96 * D2_SE) %>%
  print()
## # A tibble: 3 x 7
## # Groups:   S003 [3]
##   S003             D1     D2 D1_SE D2_SE  CI_1  CI_2
##                  
## 1 Great Britain 1.51  -0.398 0.267 0.154 0.524 0.302
## 2 Spain         0.145  0.127 0.265 0.168 0.520 0.330
## 3 Turkey        0.737 -0.107 0.229 0.231 0.449 0.452

We now have a table containing the difference in means, the standard error of the difference in means, and the confidence intervals for each of the two differences. (Recall that D1 is the difference between the average life satisfaction for full-time employed and unemployed, and D2 is the difference in average life satisfaction for full-time employed and retired individuals.)

Calculate confidence intervals using the t.test function

We could obtain the confidence intervals directly by using the t.test function. First we need to prepare the data in two groups. In the following example we go through the difference in average life satisfaction for full-time employed and unemployed individuals in Turkey, but the process can be repeated for the difference between full-time employed and retired individuals by changing the code appropriately (also for your other two chosen countries).

We start by selecting the data for full-time and unemployed people in Turkey, and storing it in two separate temporary matrices (arrays of rows and columns) called turkey_full and turkey_unemployed respectively, which is the format needed for the t.test function.

turkey_full  lifesat_data %>%
  # Select Wave 4 only
  subset(S002EVS == "2008-2010") %>%
  # Select country
  subset(S003 == "Turkey") %>%
  # Select employment type
  subset(X028 == "Full time") %>%
  # We only need the life satisfaction data.
  select(A170) %>%
  # We have to set it as a matrix for t.test.
  as.matrix()

# Repeat for second group
turkey_unemployed  lifesat_data %>%
  subset(S002EVS == "2008-2010") %>%    
  subset(S003 == "Turkey") %>%
  subset(X028 == "Unemployed") %>%
  select(A170) %>%
  as.matrix()

We can now use the t.test function on the two newly created matrices. We also set the confidence level to 95%. The t.test function produces quite a bit of output, but we are only interested in the confidence interval, which we can obtain by using $conf.int.

turkey_ci  t.test(turkey_full, turkey_unemployed, 
  conf.level = 0.95)$conf.int
turkey_ci
## [1] 0.2873459 1.1875704
## attr(,"conf.level")
## [1] 0.95

We can then calculate the difference in means by finding the midpoint of the interval ((turkey_ci[2] + turkey_ci[1])/2 is 0.7374582), which should be the same as the figures obtained in Question 1(b) (df.employment[3, 2] is 0.7374582).

Add error bars to the column charts

We can now use these confidence intervals (and widths) to add error bars to our column charts. To do so, we use the geom_errorbar option, and specify the lower and upper levels of the confidence interval for the ymin and ymax options respectively. In this case it is easier to use the results from Questions 1(a) and (b), as we already have the values for the difference in means and the CI width stored as variables.

ggplot(df.employment, aes(x = S003, y = D1)) +
  geom_bar(stat = "identity") + 
  geom_errorbar(aes(ymin = D1 - CI_1, ymax = D1 + CI_1), 
    width = .2) +
  ylab("Difference in means") + 
  xlab("Country") + 
  theme_bw() +
  ggtitle("Difference in wellbeing 
    (full-time and unemployed)")

Difference in life satisfaction (wellbeing) between full-time employed and unemployed.

Figure 8.7 Difference in life satisfaction (wellbeing) between full-time employed and unemployed.

Again, this can be repeated for the difference in life satisfaction between full-time employed and retired. Remember to change y = D1 to y = D2, and change the upper and lower limits for the error bars.

  1. The method we used to compare life satisfaction relied on making comparisons between people with different employment statuses, but a person’s employment status is not entirely random. We cannot therefore make causal statements such as ‘being unemployed causes life satisfaction to decrease’. Describe how we could use the following methods to assess better the effect of being unemployed on life satisfaction, and make some statements about causality: