Exploring data normalization and analysis in large TMT experimental designs

Part 4 - Statistical testing with similar samples

Phil Wilmarth, OHSU PSR Core, February 2018

In Part 1, several normalization methods were used on a developing mouse lens TMT study that spanned three 6-plex TMT labeling experiments. It was clear that an IRS-like procedure is critical to combining data from multiple TMT experiments because the different TMT experiments act like different batches. We also saw that increases in expression of several highly abundant lens proteins during the time course created a compositional bias in the samples that could be corrected by procedures like TMM.

In Part 2, we compared two rather different conditions: early development (E15 plus E18) to later development (P6 plus P9) because we could have some reasonable expectation of the differences due to what is known about the lens. We looked at using ratios of samples to a standard as an alternative to scaling experiments based on the standard values in Part 3.

In this installment, we will look at comparing P0 to P3 with and without IRS normalization. These samples should be relatively similar, so we will use TMM normalization. We will stick with an exact test in edgeR so that we are not changing too many things at once.

Robinson, M.D., McCarthy, D.J. and Smyth, G.K., 2010. edgeR: a Bioconductor package for differential expression analysis of digital gene expression data. Bioinformatics, 26(1), pp.139-140.

Robinson, M.D. and Oshlack, A., 2010. A scaling normalization method for differential expression analysis of RNA-seq data. Genome biology, 11(3), p.R25.

Data from:

Khan, S.Y., Ali, M., Kabir, F., Renuse, S., Na, C.H., Talbot, C.C., Hackett, S.F. and Riazuddin, S.A., 2018. Proteome Profiling of Developing Murine Lens Through Mass Spectrometry. Investigative ophthalmology & visual science, 59(1), pp.100-107.


Let's start fresh and reload the data

In [17]:
# Analysis of IOVS mouse lens data (Supplemental Table S01):
# Khan, Shahid Y., et al. "Proteome Profiling of Developing Murine Lens Through Mass Spectrometry." 
# Investigative Ophthalmology & Visual Science 59.1 (2018): 100-107.

# load libraries
library(tidyverse) # modern R packages for big data analysis
library(limma) # edgeR will load this if we do not
library(edgeR)

# read the Supplemental 01 file (saved as a CSV export from XLSX file)
data_start <- read_csv("iovs-58-13-55_s01.csv")

# filter out proteins not seen in all three runs
data_no_na <- na.omit(data_start)

# fix the column headers
col_headers <- colnames(data_no_na)
col_headers <- str_replace(col_headers, " {2,3}", " ")
col_headers <- str_replace(col_headers, "Reporter ion intensities ", "")
colnames(data_no_na) <- col_headers

# save the annotation columns (gene symbol and protein accession) for later and remove from data frame
annotate_df <- data_no_na[1:2]
data_raw <- as.data.frame(data_no_na[3:20])
row.names(data_raw) <- annotate_df$`Protein Accession No.`
Parsed with column specification:
cols(
  .default = col_double(),
  `Gene Symbol (NCBI)` = col_character(),
  `Protein Accession No.` = col_character()
)
See spec(...) for full column specifications.

We need to do SL and IRS normalizations

In [18]:
# separate the TMT data by experiment
exp1_raw <- data_raw[c(1:6)]
exp2_raw <- data_raw[c(7:12)]
exp3_raw <- data_raw[c(13:18)]

# figure out the global scaling value
target <- mean(c(colSums(exp1_raw), colSums(exp2_raw), colSums(exp3_raw)))

# do the sample loading normalization before the IRS normalization
# there is a different correction factor for each column
# seems like a loop could be used here somehow...
norm_facs <- target / colSums(exp1_raw)
exp1_sl <- sweep(exp1_raw, 2, norm_facs, FUN = "*")
norm_facs <- target / colSums(exp2_raw)
exp2_sl <- sweep(exp2_raw, 2, norm_facs, FUN = "*")
norm_facs <- target / colSums(exp3_raw)
exp3_sl <- sweep(exp3_raw, 2, norm_facs, FUN = "*")

# make a pre-IRS data frame after sample loading norms
data_sl <- cbind(exp1_sl, exp2_sl, exp3_sl)

# make working frame with row sums from each frame
irs <- tibble(rowSums(exp1_sl), rowSums(exp2_sl), rowSums(exp3_sl))
colnames(irs) <- c("sum1", "sum2", "sum3")

# get the geometric average intensity for each protein
irs$average <- apply(irs, 1, function(x) exp(mean(log(x))))

# compute the scaling factor vectors
irs$fac1 <- irs$average / irs$sum1
irs$fac2 <- irs$average / irs$sum2
irs$fac3 <- irs$average / irs$sum3

# make new data frame with normalized data
data_irs <- exp1_sl * irs$fac1
data_irs <- cbind(data_irs, exp2_sl * irs$fac2)
data_irs <- cbind(data_irs, exp3_sl * irs$fac3)

Exact pair-wise testing in edgeR

Setting up the experiment design and loading the data

We start with a data frame where the rows are the different protein expression levels and the columns are the biological samples. We need to get the data into an edgeR DGEList object. This data container holds the count data and the sample mapping information. We have to tell edgeR which samples belong to which groups. When we perform the statistical testing, we will have to specify which two groups are being compared. We will do an analysis of the SL data without IRS first. EdgeR will work better if we load all 18 samples in the DGEList object and estimate dispersions on the full dataset. We will limit the pair-wise comparison to the P0 versus P3 samples below.

In [19]:
# set up the sample mapping
group <- rep(c("E15", "E18", "P0", "P3", "P6", "P9"), 3)

# make group into factors and set the order
group <- factor(group, levels = c("E15", "E18", "P0", "P3", "P6", "P9"))

# create a DGEList object with our data
y_sl <- DGEList(counts = data_sl, group = group)

# we need to run TMM norm and estimate the dispersion
y_sl <- calcNormFactors(y_sl)
y_sl <- estimateDisp(y_sl)
y_sl$samples
plotBCV(y_sl, main = "Biological variation SL only (no IRS)")

# we need to transform the SL data with TMM factors for later plotting
sl_tmm <- calcNormFactors(data_sl)
data_sl_tmm <- sweep(data_sl, 2, sl_tmm, FUN = "/") # this is data after SL and TMM on original scale
Design matrix not provided. Switch to the classic mode.
grouplib.sizenorm.factors
E15_Set1E15 98264753371.7553031
E18_Set1E18 98264753371.4989245
P0_Set1P0 98264753371.3713938
P3_Set1P3 98264753371.1606568
P6_Set1P6 98264753370.9922070
P9_Set1P9 98264753370.8278740
E15_Set2E15 98264753371.5307406
E18_Set2E18 98264753371.3668135
P0_Set2P0 98264753371.1875263
P3_Set2P3 98264753370.8306536
P6_Set2P6 98264753370.7672969
P9_Set2P9 98264753370.7383877
E15_Set3E15 98264753371.2234557
E18_Set3E18 98264753370.9421124
P0_Set3P0 98264753370.9034106
P3_Set3P3 98264753370.7316805
P6_Set3P6 98264753370.6116590
P9_Set3P9 98264753370.5334659

Do the Exact test between early and late and see how many candidates (FDR < 0.05)

We have loaded the data, performed TMM normalizations, and estimated dispersion factors for the statistical modeling. Now we will perform a modified Fisher's exact test between the P0 and P3 groups.

In [20]:
# the exact test object has columns like fold-change, CPM, and p-values
et_sl <- exactTest(y_sl, pair = c("P0", "P3"))
summary(decideTestsDGE(et_sl)) # this counts up, down, and unchanged genes (here it is proteins)
       P0+P3
Down       0
NotSig  3155
Up         0

The high variance without doing IRS eliminates all DE candidates

We might expect that P0 and P3 are not too different, but there probably should be some changes in expression.

In [21]:
# the topTags function adds the BH FDR values to an exactTest data frame. Make sure we do not change the row order!
tt_sl <- topTags(et_sl, n = Inf, sort.by = "none")
tt_sl <- tt_sl$table # tt_sl is a list. We just need the data frame table

# add the default value as a new column
tt_sl$candidate <- "no"
tt_sl[which(tt_sl$FDR <= 0.10 & tt_sl$FDR > 0.05), dim(tt_sl)[2]] <- "low"
tt_sl[which(tt_sl$FDR <= 0.05 & tt_sl$FDR > 0.01), dim(tt_sl)[2]] <- "med"
tt_sl[which(tt_sl$FDR <= 0.01), dim(tt_sl)[2]] <- "high"
tt_sl$candidate <- factor(tt_sl$candidate, levels = c("high", "med", "low", "no"))

# what does tt_sl look like?
head(tt_sl)
logFClogCPMPValueFDRcandidate
P246220.7688977 17.18071 0.036007050.9999957 no
Q9JJU90.5134693 15.60409 0.294993170.9999957 no
Q9WVJ50.3971478 16.04543 0.383892850.9999957 no
P043450.7573050 15.33568 0.053042460.9999957 no
P626962.6182098 13.56508 0.016036790.9999957 no
Q615970.6582113 14.13001 0.354238760.9999957 no

Is the statistical test behaving correctly?

Non-differentially-expressed proteins should produce a uniformly distributed (flat) background of p-values (at all values from 0.0 to 1.0). Any true differential candidates will be a second distribution peaked at very small p-values. This is a good example of what the p-value distribution should not look like! The variances are so large compared to the means, that the distribution is skewed.

In [22]:
# what does the test p-value distribution look like?
ggplot(tt_sl, aes(PValue)) + 
  geom_histogram(bins = 100, fill = "white", color = "black") + 
  geom_hline(yintercept = mean(hist(tt_sl$PValue, breaks = 100, plot = FALSE)$counts[26:100])) +
  ggtitle("P0 vs P3 (SLNorm only) p-value distribution")
In [23]:
# for plotting results, we will use the average intensities for the 3 P0 and the 3 P3 samples
P0 <- c(3, 9, 15)
P3 <- c(4, 10, 16)
de_sl <- data.frame(rowMeans(data_sl_tmm[P0]), rowMeans(data_sl_tmm[P3]), tt_sl$candidate)
colnames(de_sl) <- c("P0", "P3", "candidate")
volcano_sl <- data.frame(log2(rowMeans(data_sl[P0])/rowMeans(data_sl[P3])), log10(tt_sl$FDR)*(-1), tt_sl$candidate)
colnames(volcano_sl) <- c("FoldChange", "FDR", "candidate")

Make plots and color code by candidate status

We will do three types of plots for visualizing the DE candidates: mean-difference plots (common in genomics), scatter plots (my favorite), and a volcano plot. For the MA plots and scatter plots, we will also separate the plots by candidate status.

The solid diagonal lines are the unity line and the dotted lies are plus/minus 2-fold. Since we have no candidates, the non-candidates are orange points, and the separate plots are the same.

In [24]:
# start with MA plot
library(scales)
temp <- data.frame(log2((de_sl$P0 + de_sl$P3)/2), log2(de_sl$P3/de_sl$P0), de_sl$candidate)
colnames(temp) <- c("Ave", "FC", "candidate")
ggplot(temp, aes(x = Ave, y = FC)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_continuous("FC (P3 / P0)") +
  scale_x_continuous("Ave_intensity") +
  ggtitle("Without IRS P0 vs P3 (MA plot)") + 
  geom_hline(yintercept = 0.0, color = "black") + # one-to-one line
  geom_hline(yintercept = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_hline(yintercept = -1.0, color = "black", linetype = "dotted") # 2-fold down

# make separate MA plots
ggplot(temp, aes(x = Ave, y = FC)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_continuous("FC (P3 / P0)") +
  scale_x_continuous("Ave_intensity") +
  geom_hline(yintercept = 0.0, color = "black") + # one-to-one line
  geom_hline(yintercept = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_hline(yintercept = -1.0, color = "black", linetype = "dotted") + # 2-fold down
  facet_wrap(~ candidate) +
  ggtitle("Without IRS, separated by candidate (MA plots)")

# make the combined candidate corelation plot
ggplot(de_sl, aes(x = P0, y = P3)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_log10() +
  scale_x_log10() +
  ggtitle("Without IRS P0 vs P3") + 
  geom_abline(intercept = 0.0, slope = 1.0, color = "black") + # one-to-one line
  geom_abline(intercept = 0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_abline(intercept = -0.301, slope = 1.0, color = "black", linetype = "dotted") # 2-fold down

# make separate corelation plots
ggplot(de_sl, aes(x = P0, y = P3)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_log10() +
  scale_x_log10() +
  geom_abline(intercept = 0.0, slope = 1.0, color = "black") + # one-to-one line
  geom_abline(intercept = 0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_abline(intercept = -0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold down
  facet_wrap(~ candidate) +
  ggtitle("Without IRS, separated by candidate")

# make a volcano plot
ggplot(volcano_sl, aes(x = FoldChange, y = FDR)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  xlab("Fold-Change (Log2)") +
  ylab("-Log10 FDR") +
  ggtitle("Without IRS Volcano Plot")

Now let's see what effect IRS has

Same steps, using the IRS data instead of the SL data

In [25]:
# create a DGEList object with the IRS data
y_irs <- DGEList(counts = data_irs, group = group)

# we need to normalize and estimate the dispersion terms (global and local)
y_irs <- calcNormFactors(y_irs)
y_irs <- estimateDisp(y_irs)
y_irs$samples
plotBCV(y_irs, main = "Biological variation with IRS")

# we need to transform the IRS data with TMM factors for later plotting
irs_tmm <- calcNormFactors(data_sl)
data_irs_tmm <- sweep(data_irs, 2, irs_tmm, FUN = "/") # this is data after SL and TMM on original scale
Design matrix not provided. Switch to the classic mode.
grouplib.sizenorm.factors
E15_Set1E15 92706647741.5114732
E18_Set1E18 93637612941.2419268
P0_Set1P0 94516224361.1403959
P3_Set1P3 95318373940.9451601
P6_Set1P6 95378376570.8136767
P9_Set1P9 95669963260.6791176
E15_Set2E15 92530604681.5830751
E18_Set2E18 91358307211.4090527
P0_Set2P0 95970661161.1219356
P3_Set2P3 95287573060.7748380
P6_Set2P6 96409790740.7140989
P9_Set2P9 95670261960.6842127
E15_Set3E15 96687114031.5692792
E18_Set3E18 94903358011.1766499
P0_Set3P0 94250867061.1308403
P3_Set3P3 93896577870.8922979
P6_Set3P6 93537577710.7594531
P9_Set3P9 93951704130.6671584

Biological variation is greatly reduced after IRS and we have some DE candidates

P-value distribution now looks more typical

In [26]:
# the exact test object has columns like fold-change, CPM, and p-values
et_irs <- exactTest(y_irs, pair = c("P0", "P3"))
summary(decideTestsDGE(et_irs)) # this counts up, down, and unchanged genes
       P0+P3
Down      38
NotSig  3066
Up        51
In [27]:
# the topTags function adds the BH FDR values to an exactTest data frame. Make sure we do not change the row order!
tt_irs <- topTags(et_irs, n = Inf, sort.by = "none")
tt_irs <- tt_irs$table # tt_sl is a list. We just need the data frame table

# add the default value as a new column
tt_irs$candidate <- "no"
tt_irs[which(tt_irs$FDR <= 0.10 & tt_irs$FDR > 0.05), dim(tt_irs)[2]] <- "low"
tt_irs[which(tt_irs$FDR <= 0.05 & tt_irs$FDR > 0.01), dim(tt_irs)[2]] <- "med"
tt_irs[which(tt_irs$FDR <= 0.01), dim(tt_irs)[2]] <- "high"
tt_irs$candidate <- factor(tt_irs$candidate, levels = c("high", "med", "low", "no"))

# what does tt_sl look like?
head(tt_irs)

# what does the test p-value distribution look like?
ggplot(tt_irs, aes(PValue)) + 
  geom_histogram(bins = 100, fill = "white", color = "black") + 
  geom_hline(yintercept = mean(hist(tt_irs$PValue, breaks = 100, plot = FALSE)$counts[26:100])) +
  ggtitle("P0 vs P3 (after IRS) p-value distribution")
logFClogCPMPValueFDRcandidate
P246220.7871157 17.21966 1.700120e-050.003575919 high
Q9JJU90.5603518 15.60201 1.140208e-030.044355404 med
Q9WVJ50.4211099 16.05431 3.718613e-020.260743098 no
P043450.8195890 15.37622 7.927053e-050.009262908 high
P626962.7169591 13.62826 4.201875e-030.093608202 low
Q615970.7694345 14.04192 1.945807e-030.059029043 low

DE candidate plots after IRS

The solid lines are the unity line and the dotted lies are plus/minus 2-fold. Proteins with FDR values less than 0.01 are "high" significance candidates (orange), proteins with FDR between 0.05 and 0.01 are "medium" (sea green), proteins with FDR between 0.10 and 0.05 are "low" (teal), and proteins with FDR greater than 0.10 are non-candidates ("no", in purple).

In [28]:
# for plotting results, we will use the average intensities for the P0 and the P3 samples
de_irs <- data.frame(rowMeans(data_irs_tmm[P0]), rowMeans(data_irs_tmm[P3]), tt_irs$candidate)
colnames(de_irs) <- c("P0", "P3", "candidate")
volcano_irs <- data.frame(-1*tt_irs$logFC, -1*log10(tt_irs$FDR), tt_irs$candidate)
colnames(volcano_irs) <- c("FoldChange", "FDR", "candidate")

# start with MA plot
temp <- data.frame(log2((de_irs$P0 + de_irs$P3)/2), log2(de_irs$P3/de_irs$P0), de_irs$candidate)
colnames(temp) <- c("Ave", "FC", "candidate")
ggplot(temp, aes(x = Ave, y = FC)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_continuous("FC (P3 / P0)") +
  scale_x_continuous("Ave_intensity") +
  ggtitle("After IRS P0 vs P3 (MA plot)") + 
  geom_hline(yintercept = 0.0, color = "black") + # one-to-one line
  geom_hline(yintercept = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_hline(yintercept = -1.0, color = "black", linetype = "dotted") # 2-fold down

# make separate MA plots
ggplot(temp, aes(x = Ave, y = FC)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_continuous("FC (P3 / P0)") +
  scale_x_continuous("Ave_intensity") +
  geom_hline(yintercept = 0.0, color = "black") + # one-to-one line
  geom_hline(yintercept = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_hline(yintercept = -1.0, color = "black", linetype = "dotted") + # 2-fold down
  facet_wrap(~ candidate) +
  ggtitle("After IRS, separated by candidate (MA plots)")

# make the combined candidate corelation plot
ggplot(de_irs, aes(x = P0, y = P3)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_log10() +
  scale_x_log10() +
  ggtitle("After IRS P0 vs P3") + 
  geom_abline(intercept = 0.0, slope = 1.0, color = "black") + # one-to-one line
  geom_abline(intercept = 0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_abline(intercept = -0.301, slope = 1.0, color = "black", linetype = "dotted") # 2-fold down

# make separate corelation plots
ggplot(de_irs, aes(x = P0, y = P3)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  scale_y_log10() +
  scale_x_log10() +
  geom_abline(intercept = 0.0, slope = 1.0, color = "black") + # one-to-one line
  geom_abline(intercept = 0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold up
  geom_abline(intercept = -0.301, slope = 1.0, color = "black", linetype = "dotted") + # 2-fold down
  facet_wrap(~ candidate) +
  ggtitle("IRS, separated by candidate")

# make a volcano plot
ggplot(volcano_irs, aes(x = FoldChange, y = FDR)) +
  geom_point(aes(color = candidate, shape = candidate)) +
  xlab("Fold-Change (Log2)") +
  ylab("-Log10 FDR") +
  ggtitle("After IRS Volcano Plot")

The statistical testing after IRS seems more reasonable

Even though the time delta is just 3 days, we should still have more expression of major lens proteins at P3 compared to P0. We do indeed see some up-regulated proteins starting to emerge.

Let's look at the scatter plots of the sample averages for P0 and P3, before and after IRS

We will not highlight candidates and we will add marginal histograms.

In [29]:
library(ggExtra)
# add marginal distrubution histograms to basic correlation plot (good starting point)
ggplot()
corr_plot <- ggplot(de_sl, aes(x = log10(P0), y = log10(P3))) +
  geom_point() + ggtitle("Before IRS")
ggMarginal(corr_plot, type = "histogram")

ggplot()
corr_plot <- ggplot(de_irs, aes(x = log10(P0), y = log10(P3))) +
  geom_point() + ggtitle("After IRS")
ggMarginal(corr_plot, type = "histogram")

Average intensities are not very informative about underlying variance

As we saw in Part 2, the average values across TMT experiments do not depend too much on whether the IRS method is done or not. The variance, a key component in the statistical testing, is dramatically different with and without IRS.

Random sampling of MS2 scans between LC runs negates the precision of TMT labeling

Reference standards are truly needed in TMT experiments. Either IRS normalization or ratios to a common reference are required to remove the random sampling effect. Despite some data summary visualizations suggesting that standard normalization methods are comparable to the IRS method, they completely fail to prepare the data for statistical testing. The high variance from random MS2 sampling simply kills the statistical testing.

In [30]:
# log the R session for records
sessionInfo()
R version 3.5.0 (2018-04-23)
Platform: x86_64-apple-darwin15.6.0 (64-bit)
Running under: macOS High Sierra 10.13.4

Matrix products: default
BLAS: /Library/Frameworks/R.framework/Versions/3.5/Resources/lib/libRblas.0.dylib
LAPACK: /Library/Frameworks/R.framework/Versions/3.5/Resources/lib/libRlapack.dylib

locale:
[1] en_US.UTF-8/en_US.UTF-8/en_US.UTF-8/C/en_US.UTF-8/en_US.UTF-8

attached base packages:
[1] stats     graphics  grDevices utils     datasets  methods   base     

other attached packages:
 [1] ggExtra_0.8     scales_0.5.0    edgeR_3.22.2    limma_3.36.1   
 [5] forcats_0.3.0   stringr_1.3.1   dplyr_0.7.5     purrr_0.2.4    
 [9] readr_1.1.1     tidyr_0.8.1     tibble_1.4.2    ggplot2_2.2.1  
[13] tidyverse_1.2.1

loaded via a namespace (and not attached):
 [1] Rcpp_0.12.17         locfit_1.5-9.1       lubridate_1.7.4     
 [4] lattice_0.20-35      assertthat_0.2.0     digest_0.6.15       
 [7] psych_1.8.4          mime_0.5             IRdisplay_0.5.0     
[10] R6_2.2.2             cellranger_1.1.0     plyr_1.8.4          
[13] repr_0.15.0          evaluate_0.10.1      httr_1.3.1          
[16] pillar_1.2.2         rlang_0.2.0          lazyeval_0.2.1      
[19] uuid_0.1-2           readxl_1.1.0         rstudioapi_0.7      
[22] miniUI_0.1.1.1       labeling_0.3         splines_3.5.0       
[25] foreign_0.8-70       munsell_0.4.3        shiny_1.1.0         
[28] broom_0.4.4          compiler_3.5.0       httpuv_1.4.3        
[31] modelr_0.1.2         pkgconfig_2.0.1      base64enc_0.1-3     
[34] mnormt_1.5-5         htmltools_0.3.6      tidyselect_0.2.4    
[37] crayon_1.3.4         later_0.7.2          grid_3.5.0          
[40] xtable_1.8-2         nlme_3.1-137         jsonlite_1.5        
[43] gtable_0.2.0         magrittr_1.5         cli_1.0.0           
[46] stringi_1.2.2        reshape2_1.4.3       promises_1.0.1      
[49] bindrcpp_0.2.2       xml2_1.2.0           IRkernel_0.8.12.9000
[52] tools_3.5.0          glue_1.2.0           hms_0.4.2           
[55] parallel_3.5.0       colorspace_1.3-2     rvest_0.3.2         
[58] pbdZMQ_0.3-3         bindr_0.1.1          haven_1.1.1