1 Background

Researchers are studying the impact of protein sources and protein levels in the diet on the weight of rats. They feed the rats with diets of beef, cereal and pork and use a low and high protein level for each diet type. The researchers can include 60 rats in the experiment. Prior to the experiment, the rats were divided in 10 homogeneous groups of 6 rats based on characteristics such as initial weight, appetite, etc.

Within each group a rat is randomly assigned to a diet. The rats are fed during a month and the weight gain in grams is recorded for each rat.

2 Experimental design

  • There are three explanatory variables in the experiment: the factor diet type with three levels (beef, cereal and pork), factor protein level with levels (low and high) and a group blocking factor with 10 levels.

  • There are 6 treatments: beef-high, cereal-high, pork-high, beef-low, cereal-low, pork-low protein.

  • The rats are the experimental units (the unit to which a treatment is applied): in this design, there is a randomisation restriction: Within a block, a rat is randomly assigned to a diet.

  • The rats are the observational units (the unit on which the response is measured): The weight is weighted for each rat.

  • The weight gain is the response variable.

  • The experiment is a randomized complete block (RCB) design

Load libraries

library(tidyverse)

3 Data import

diet <- read.table("https://raw.githubusercontent.com/statOmics/PSLS21/data/dietRats.txt",
                   header=TRUE)
head(diet)

4 Tidy data

diet <- diet %>%
  mutate(block = as.factor(block),
         protSource = as.factor(protSource),
         protLevel = as.factor(protLevel)) %>%
  mutate(protLevel = fct_relevel(protLevel, "l"))

5 Data exploration

  • Boxplot of the weight gain against protein source, protein level with coloring according to block
diet %>% 
  ggplot(aes(x=protLevel, y=weightGain)) + 
    scale_fill_brewer(palette="RdGy") +
    theme_bw() +
    geom_boxplot(outlier.shape=NA) + 
    geom_jitter(aes(color=block)) +
    ggtitle("Weight gain dependence on protein source") +
    ylab("Weigth gain (g)") + 
    #stat_summary(fun = mean, geom="point", shape=5, size=3, color="black", fill="black") +
    facet_wrap(.~protSource)

  • Lineplot of the weight gain against protein source, protein level with coloring and grouping according to block
diet %>% 
  ggplot(aes(x=protSource:protLevel, y=weightGain)) + 
    scale_fill_brewer(palette="RdGy") +
    theme_bw() +
    geom_line(aes(group=block, color=block)) +
    geom_point(aes(color=block))+
    ggtitle("Weight gain dependence on protein source") +
    ylab("Weigth gain (g)") 

    #stat_summary(fun = mean, geom="point", shape=5, size=3, color="black", fill="black")
  • An increase in the weight of the rats seems to depend on the protein source received in the diet.

  • The increase in the weight of the rats seems to depend on the level of protein received in the diet

  • There also seems to be an interaction effect between the protein level and the protein source on the gain in weight of the rats. For the beef and the pork diets the effect of high protein levels in the data seems to be much stronger than in the cereal diet.

  • There is also a strong effect of the block. Blocking implies a randomisation restriction, hence, we will have to include the block effect anyway.

6 Multivariate linear regression analysis

6.1 Assumptions

List assumptions:

  1. The observations are independent
  2. Linearity between the response and predictor variable
  3. The residuals of the model must be normally distributed
  4. Homoscedasticity of the data

The first assumption is met if we correct for block in the model because the rats were randomized to the treatment within block. The other three assumptions can be assessed by fitting the linear model and calling the plot() function as follows.

lm1 <- lm(weightGain ~ block + protSource*protLevel, data=diet)
plot(lm1)

All assumptions are met for this dataset.

6.2 Hypothesis testing

We here fit a linear model with a blocking factor for block and main and interaction effects for protein source and protein level.

summary(lm1)
## 
## Call:
## lm(formula = weightGain ~ block + protSource * protLevel, data = diet)
## 
## Residuals:
##     Min      1Q  Median      3Q     Max 
## -21.017  -5.183  -1.017   4.092  18.983 
## 
## Coefficients:
##                        Estimate Std. Error t value Pr(>|t|)    
## (Intercept)             88.3500     4.1728  21.173  < 2e-16 ***
## block2                  -9.1667     4.8183  -1.902 0.063520 .  
## block3                  -4.8333     4.8183  -1.003 0.321166    
## block4                 -20.3333     4.8183  -4.220 0.000117 ***
## block5                  -9.3333     4.8183  -1.937 0.059031 .  
## block6                   0.6667     4.8183   0.138 0.890572    
## block7                 -31.3333     4.8183  -6.503 5.58e-08 ***
## block8                 -10.1667     4.8183  -2.110 0.040450 *  
## block9                 -19.8333     4.8183  -4.116 0.000162 ***
## block10                 -0.1667     4.8183  -0.035 0.972559    
## protSourcec              1.1000     3.7322   0.295 0.769556    
## protSourcep              1.3000     3.7322   0.348 0.729226    
## protLevelh              20.0000     3.7322   5.359 2.76e-06 ***
## protSourcec:protLevelh -18.2000     5.2782  -3.448 0.001236 ** 
## protSourcep:protLevelh  -1.7000     5.2782  -0.322 0.748883    
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Residual standard error: 8.346 on 45 degrees of freedom
## Multiple R-squared:  0.7712, Adjusted R-squared:    0.7 
## F-statistic: 10.83 on 14 and 45 DF,  p-value: 3.726e-10

6.3 Interpretation of the regression parameters

library(ExploreModelMatrix)
ExploreModelMatrix::VisualizeDesign(diet, ~ block + protSource * protLevel)$plotlist
## $`block = 1`

## 
## $`block = 10`

## 
## $`block = 2`

## 
## $`block = 3`

## 
## $`block = 4`

## 
## $`block = 5`

## 
## $`block = 6`

## 
## $`block = 7`

## 
## $`block = 8`

## 
## $`block = 9`

There are 3 levels for protein source, 2 levels for protein level and 10 levels for the blocking variable. We will have one reference level for each respective variable: beef, low, block 1. So we need 2, 1 and 9 dummy variables to introduce the factors protein source, protein level and block in the linear model, respectively.

Hence, we can write down the linear model as follows:

\(y_i=\beta_0+\beta_cx_{i,c} +\beta_px_{i,p}+\beta_hx_{i,h}+\beta_{ch}x_{i,c}x_{i,h}+\beta_{ph}x_{i,p}x_{i,h}+\beta_{b2}x_{i,b2}+\ldots+\beta_{b10}x_{i,b10}+\epsilon_{i}\)

with: \(y_i\) the observed weight gain for rat i, \(x_{i,h}\) a dummy variable which is 1 if rat i receives a high protein diet and is 0 otherwise,
\(x_{i,c}\) a dummy variable which is 1 if rat i receives a cereal diet and is 0 otherwise,
\(x_{i,p}\) a dummy variable which is 1 if rat i receives a pork diet and is 0 otherwise and \(x_{i,bk}\) is a dummy variable which is 1 if rat i belongs to block \(bk\) and is 0 otherwise, with \(k \in 2,\ldots, 10\), and \(\epsilon_i\) an error term which is normally distributed with mean 0 and variance \(\sigma^2\), i.e. \(\epsilon_i \sim N(0,\sigma^2)\).

  • Rats that are assigned to block \(k\) and receive a beef based low protein diet have a covariate pattern \(x_{i,h}=0\), \(x_{i,c}=0\), \(x_{i,p}=0\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). Their mean weight gain is thus equal to \(\mu_{l,b,bk}=\beta_0+\beta_{bk}\)

  • Rats that are assigned to block \(k\) and receive a beef based high protein diet have a covariate pattern \(x_{i,h}=1\), \(x_{i,c}=0\), \(x_{i,p}=0\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). Their mean weight gain is thus equal to \(\mu_{h,b,bk}=\beta_0+\beta_h+\beta_{bk}\)

  • Rats that are assigned to block \(k\) and receive a cereal based low protein diet have a covariate pattern \(x_{i,h}=0\), \(x_{i,c}=1\), \(x_{i,p}=0\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). Their mean weight gain is thus equal to \(\mu_{h,c,bk}=\beta_0+\beta_c+\beta_{bk}\)

  • Rats that are assigned to block \(k\) and receive a cereal based heigh protein diet have a covariate pattern \(x_{i,h}=1\), \(x_{i,c}=1\), \(x_{i,p}=0\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). Their mean weight gain is thus equal to \(\mu_{h,c,bk}=\beta_0+\beta_h+\beta_c+\beta_{ch}+\beta_{bk}\)

  • Rats that are assigned to block \(k\) and receive a pork based low protein diet have a covariate pattern \(x_{i,h}=0\), \(x_{i,c}=0\), \(x_{i,p}=1\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). Their mean weight gain is thus equal to \(\mu_{l,p,bk}=\beta_0+\beta_p+\beta_{bk}\)

  • Rats that are assigned to block \(k\) and receive a pork based heigh protein diet have a covariate pattern \(x_{i,h}=1\), \(x_{i,c}=0\), \(x_{i,p}=1\), \(x_{i,bm}=0\) with \(m\neq k\) and \(x_{i,bk}=1\). There mean weight gain is thus equal to \(\mu_{h,p,bk}=\beta_0+\beta_h+\beta_p+\beta_{ph}+\beta_{bk}\)

We can now relate this to the output of the lm function:

  • The intercept \(\beta_0\) is thus the average weight increase in the low beef diet for rats in block 1.

  • The parameter \(\beta_c\): the average difference in weight gain between cereal-low and beef-low diet is 1.1g.

  • The parameter \(\beta_p\): the average difference in weight gain between pork-low and beef-low diet is 1.3g.

  • The parameter \(\beta_h\): the average difference in weight gain between beef-high and beef-Low diet is 20g.

  • The parameter \(\beta_{ch}\) is the difference in the average weight gain difference due to the high protein level as compared to the low protein level for cereal diets as compared to the weight gain difference that occurs due to the protein level in the reference class (here beef diet). Here this is negative, i.e. -18.2g, thus the weight gain for the cereal protein source increases on average less between high and low protein diets than in beef based diets.

  • The parameter \(\beta_{ph}\) is the difference in the average weight gain difference due to the high protein level as compared to the low protein level for pork diets as compared to the weight gain difference that occurs due to the protein level in the reference class (here beef diet). Here this is negative, i.e. -1.7g, thus the weight gain for the proc protein source increases on average less between high and low protein diets than in beef based diets.

6.4 Testing the overall (combined) effect of diet

Because there are multiple factors with different levels in the model, we can first assess the effect of the diet (protein Level, protein source and the interaction) by using anova. With this test we will assess the null hypothesis that the average weight gain in each treatment is equal: i.e. \(H_0: \mu_{b,l}=\mu_{b,h}=\mu_{c,h}=\mu_{c,l}=\mu_{p,h}=\mu_{p,h}\) versus the alternative hyptohesis \(H_1:\) that at least two treatment means are different.

lm0 <- lm(weightGain ~ block, data=diet)
anova(lm0, lm1)

We can conclude that there is an extremely significant effect of the diet type (protein source and/or protein level and/or protein source-protein level interaction) on the weight gain of rats (p << 0.001).

6.5 Assessing the interaction effect between protein source and protein level

library(car)
Anova(lm1, type="III")

There is a very significant interaction between the protein source and the protein level (p = 2e-03). This indicates that the average weight increase due to the protein level differs according to the protein source. Hence, we cannot assess the effect of the protein source and/or protein level independently because there effects of the protein source vary according to the protein level.

6.6 Assessing specific contrasts

Imagine that we are interested in assessing if there is an effect of

  1. protein source in the low protein diets
    • \(\mu_{c,l}-\mu_{b,l} = \beta_c\)
    • \(\mu_{p,l}-\mu_{b,l} = \beta_p\)
    • \(\mu_{c,l}-\mu_{p,l}= \beta_c-\beta_p\))
  2. protein source in high protein diets
    • \(\mu_{c,h}-\mu_{b,h}=\beta_c+\beta_{ch}\)
    • \(\mu_{p,h}-\mu_{b,h}=\beta_p+\beta_{ph}\)
    • \(\mu_{p,h}-\mu_{c,h}=(\beta_c+\beta_{ch})-(\beta_p+\beta_{ph})\)
  3. protein level for
    • beef diets (\(\mu_{b,h}-\mu_{b,l}=\beta_h\)),
    • cereal diets (\(\mu_{c,h}-\mu_{c,l}=\beta_h+\beta_{ch}\)) and
    • pork diets (\(\mu_{p,h}-\mu_{p,l}=\beta_h+\beta_{ph}\)).
  4. If the effect of the protein level differs between
    • beef and cereal \((\mu_{c,h}-\mu_{c,l}) - (\mu_{b,h}-\mu_{b,l})=\beta_{ch}\)
    • beef and pork \((\mu_{p,h}-\mu_{p,l}) - (\mu_{c,h}-\mu_{c,l})=\beta_{ph}\) and
    • cereal and pork diets \((\mu_{c,h}-\mu_{c,l}) - (\mu_{p,h}-\mu_{p,l})=\beta_{ch}-\beta_{ph}\).

These effects of interest are so-called contrasts, i.e. linear combinations of the parameters.

We can define the contrasts and assess the significance of the contrasts with the code below. The contrasts are given as input in the form of symbolic descriptions to the linfct argument of the glht function.

library(multcomp)
set.seed(75468) # to get reproducible results (small effect if removed)
lm1MultComp <- glht(
    model = lm1,
    linfct = c("protSourcec = 0",
               "protSourcep = 0",
               "protSourcec- protSourcep = 0",
               "protSourcec + protSourcec:protLevelh = 0",
               "protSourcep + protSourcep:protLevelh = 0",
               "(protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) = 0",
               "protLevelh = 0",
               "protLevelh +  protSourcec:protLevelh = 0",
               "protLevelh +  protSourcep:protLevelh = 0",
               "protSourcec:protLevelh = 0",
               "protSourcep:protLevelh = 0",
               "protSourcec:protLevelh - protSourcep:protLevelh = 0")
)
summary(lm1MultComp)
## 
##   Simultaneous Tests for General Linear Hypotheses
## 
## Fit: lm(formula = weightGain ~ block + protSource * protLevel, data = diet)
## 
## Linear Hypotheses:
##                                                                                      Estimate
## protSourcec == 0                                                                        1.100
## protSourcep == 0                                                                        1.300
## protSourcec - protSourcep == 0                                                         -0.200
## protSourcec + protSourcec:protLevelh == 0                                             -17.100
## protSourcep + protSourcep:protLevelh == 0                                              -0.400
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0  -16.700
## protLevelh == 0                                                                        20.000
## protLevelh + protSourcec:protLevelh == 0                                                1.800
## protLevelh + protSourcep:protLevelh == 0                                               18.300
## protSourcec:protLevelh == 0                                                           -18.200
## protSourcep:protLevelh == 0                                                            -1.700
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                  -16.500
##                                                                                      Std. Error
## protSourcec == 0                                                                          3.732
## protSourcep == 0                                                                          3.732
## protSourcec - protSourcep == 0                                                            3.732
## protSourcec + protSourcec:protLevelh == 0                                                 3.732
## protSourcep + protSourcep:protLevelh == 0                                                 3.732
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0      3.732
## protLevelh == 0                                                                           3.732
## protLevelh + protSourcec:protLevelh == 0                                                  3.732
## protLevelh + protSourcep:protLevelh == 0                                                  3.732
## protSourcec:protLevelh == 0                                                               5.278
## protSourcep:protLevelh == 0                                                               5.278
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                      5.278
##                                                                                      t value
## protSourcec == 0                                                                       0.295
## protSourcep == 0                                                                       0.348
## protSourcec - protSourcep == 0                                                        -0.054
## protSourcec + protSourcec:protLevelh == 0                                             -4.582
## protSourcep + protSourcep:protLevelh == 0                                             -0.107
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0  -4.475
## protLevelh == 0                                                                        5.359
## protLevelh + protSourcec:protLevelh == 0                                               0.482
## protLevelh + protSourcep:protLevelh == 0                                               4.903
## protSourcec:protLevelh == 0                                                           -3.448
## protSourcep:protLevelh == 0                                                           -0.322
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                  -3.126
##                                                                                      Pr(>|t|)
## protSourcec == 0                                                                       0.9995
## protSourcep == 0                                                                       0.9989
## protSourcec - protSourcep == 0                                                         1.0000
## protSourcec + protSourcec:protLevelh == 0                                              <0.001
## protSourcep + protSourcep:protLevelh == 0                                              1.0000
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0   <0.001
## protLevelh == 0                                                                        <0.001
## protLevelh + protSourcec:protLevelh == 0                                               0.9949
## protLevelh + protSourcep:protLevelh == 0                                               <0.001
## protSourcec:protLevelh == 0                                                            0.0118
## protSourcep:protLevelh == 0                                                            0.9993
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                   0.0265
##                                                                                         
## protSourcec == 0                                                                        
## protSourcep == 0                                                                        
## protSourcec - protSourcep == 0                                                          
## protSourcec + protSourcec:protLevelh == 0                                            ***
## protSourcep + protSourcep:protLevelh == 0                                               
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0 ***
## protLevelh == 0                                                                      ***
## protLevelh + protSourcec:protLevelh == 0                                                
## protLevelh + protSourcep:protLevelh == 0                                             ***
## protSourcec:protLevelh == 0                                                          *  
## protSourcep:protLevelh == 0                                                             
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                 *  
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## (Adjusted p values reported -- single-step method)
confint(lm1MultComp)
## 
##   Simultaneous Confidence Intervals
## 
## Fit: lm(formula = weightGain ~ block + protSource * protLevel, data = diet)
## 
## Quantile = 2.8765
## 95% family-wise confidence level
##  
## 
## Linear Hypotheses:
##                                                                                      Estimate
## protSourcec == 0                                                                       1.1000
## protSourcep == 0                                                                       1.3000
## protSourcec - protSourcep == 0                                                        -0.2000
## protSourcec + protSourcec:protLevelh == 0                                            -17.1000
## protSourcep + protSourcep:protLevelh == 0                                             -0.4000
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0 -16.7000
## protLevelh == 0                                                                       20.0000
## protLevelh + protSourcec:protLevelh == 0                                               1.8000
## protLevelh + protSourcep:protLevelh == 0                                              18.3000
## protSourcec:protLevelh == 0                                                          -18.2000
## protSourcep:protLevelh == 0                                                           -1.7000
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                 -16.5000
##                                                                                      lwr     
## protSourcec == 0                                                                      -9.6358
## protSourcep == 0                                                                      -9.4358
## protSourcec - protSourcep == 0                                                       -10.9358
## protSourcec + protSourcec:protLevelh == 0                                            -27.8358
## protSourcep + protSourcep:protLevelh == 0                                            -11.1358
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0 -27.4358
## protLevelh == 0                                                                        9.2642
## protLevelh + protSourcec:protLevelh == 0                                              -8.9358
## protLevelh + protSourcep:protLevelh == 0                                               7.5642
## protSourcec:protLevelh == 0                                                          -33.3827
## protSourcep:protLevelh == 0                                                          -16.8827
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                 -31.6827
##                                                                                      upr     
## protSourcec == 0                                                                      11.8358
## protSourcep == 0                                                                      12.0358
## protSourcec - protSourcep == 0                                                        10.5358
## protSourcec + protSourcec:protLevelh == 0                                             -6.3642
## protSourcep + protSourcep:protLevelh == 0                                             10.3358
## (protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) == 0  -5.9642
## protLevelh == 0                                                                       30.7358
## protLevelh + protSourcec:protLevelh == 0                                              12.5358
## protLevelh + protSourcep:protLevelh == 0                                              29.0358
## protSourcec:protLevelh == 0                                                           -3.0173
## protSourcep:protLevelh == 0                                                           13.4827
## protSourcec:protLevelh - protSourcep:protLevelh == 0                                  -1.3173

Note that the p-values and the confidence intervals are automatically corrected for multiple testing.

7 Conclusion

  • There is an extremely significant effect of the type of protein diet on the weight gain of rats (p<<1e-6). The interaction between protein type and protein source is also very significant (p=0.0019).

  • The average weight gain does not vary significantly according to protein source in the diets with low protein levels (all p>0.99).

  • The weight gain in the cereal diet at high protein concentration is on average 17.1g and 16.7g g lower than in the high protein beef and pork diet, respectively (95% CI [6.4, 27.8] and [6, 27.4]) and the differences are extremely significant (p<0.001). The average weight gains after meat based diets at high protein levels, however, do not differ significantly (p>0.99).

  • We also discovered an extremely significant difference in weight gain according to the protein level for beef and pork based diets (p<0.001).
    The weight gain on average increases with 20g and 18.3g in the high protein level as compared to the low protein beef and pork diet, respectively (95%CI [9.3, 30.7] and [7.6, 29]). The protein level effect is not significant for the cereal diet (p>0.99).

  • Finally significant interactions between protein level and protein source were found (p<0.05), i.e. the increase in weight gain due to protein level in cereal based diets was 18.2g and 16.5g lower than that of beef and pork based diets, respectively (95% CI [3g,33.4g] and [1.3g,31.7g]). The average difference in increase in weight gain due to protein level among the meat based diets was not significant (p>0.99).

All reported p-values and confidence intervals were corrected for multiple testing

---
title: "Exercise 8.x: Blocking on the rat diet dataset - solution"   
author: "Lieven Clement and Jeroen Gilis"
date: "statOmics, Ghent University (https://statomics.github.io)"  
output:
    html_document:
      code_download: true    
      theme: cosmo
      toc: true
      toc_float: true
      highlight: tango
      number_sections: true
    pdf_document:
      toc: true
      number_sections: true
      latex_engine: xelatex
---

# Background 

Researchers are studying the impact of protein sources and protein levels in 
the diet on the weight of rats. They feed the rats with diets of beef, cereal 
and pork and use a low and high protein level for each diet type. 
The researchers can include 60 rats in the experiment. Prior to the experiment,
the rats were divided in 10 homogeneous groups of 6 rats based on 
characteristics such as initial weight, appetite, etc.

Within each group a rat is randomly assigned to a diet. The rats are fed during 
a month and the weight gain in grams is recorded for each rat. 

# Experimental design 

- There are three explanatory variables in the experiment: the factor diet type
with three levels (beef, cereal and pork), factor protein level with levels 
(low and high) and a group blocking factor with 10 levels.

- There are 6 treatments: beef-high, cereal-high, pork-high, beef-low, 
cereal-low, pork-low protein. 

- The rats are the experimental units (the unit to which a treatment is applied): in this design, there is a randomisation restriction: Within a block, a rat is randomly assigned to a diet.

- The rats are the observational units (the unit on which the response is measured): The weight is weighted for each rat.

- The weight gain is the response variable. 

- The experiment is a randomized complete block (RCB) design

Load libraries

```{r, message=FALSE, warning=FALSE}
library(tidyverse)
```

# Data import

```{r}
diet <- read.table("https://raw.githubusercontent.com/statOmics/PSLS21/data/dietRats.txt",
                   header=TRUE)
head(diet)
```

# Tidy data

```{r}
diet <- diet %>%
  mutate(block = as.factor(block),
         protSource = as.factor(protSource),
         protLevel = as.factor(protLevel)) %>%
  mutate(protLevel = fct_relevel(protLevel, "l"))
```

# Data exploration

- Boxplot of the weight gain against protein source, protein level with coloring according to block

```{r}
diet %>% 
  ggplot(aes(x=protLevel, y=weightGain)) + 
    scale_fill_brewer(palette="RdGy") +
    theme_bw() +
    geom_boxplot(outlier.shape=NA) + 
    geom_jitter(aes(color=block)) +
    ggtitle("Weight gain dependence on protein source") +
    ylab("Weigth gain (g)") + 
    #stat_summary(fun = mean, geom="point", shape=5, size=3, color="black", fill="black") +
    facet_wrap(.~protSource)
```

- Lineplot of the weight gain against protein source, protein level with coloring and grouping according to block

```{r}
diet %>% 
  ggplot(aes(x=protSource:protLevel, y=weightGain)) + 
    scale_fill_brewer(palette="RdGy") +
    theme_bw() +
    geom_line(aes(group=block, color=block)) +
    geom_point(aes(color=block))+
    ggtitle("Weight gain dependence on protein source") +
    ylab("Weigth gain (g)") 
    #stat_summary(fun = mean, geom="point", shape=5, size=3, color="black", fill="black")
```

- An increase in the weight of the rats seems to depend on the protein source
received in the diet.

- The increase in the weight of the rats seems to depend on the level of protein
received in the diet

- There also seems to be an interaction effect between the protein level and the
protein source on the gain in weight of the rats. For the beef and the pork 
diets the effect of high protein levels in the data seems to be much stronger
than in the cereal diet.

- There is also a strong effect of the block. Blocking implies a randomisation 
restriction, hence, we will have to include the block effect anyway. 

# Multivariate linear regression analysis

## Assumptions 

List assumptions:

1. The observations are independent 
2. Linearity between the response and predictor variable
3. The residuals of the model must be normally distributed
4. Homoscedasticity of the data

The first assumption is met if we correct for block in the model because the rats were randomized to the treatment within block. The other three assumptions can be assessed by
fitting the linear model and calling the `plot()` function as follows.

```{r}
lm1 <- lm(weightGain ~ block + protSource*protLevel, data=diet)
plot(lm1)
```

All assumptions are met for this dataset.

## Hypothesis testing

We here fit a linear model with a blocking factor for `block` and main and 
interaction effects for protein source and protein level.

```{r}
summary(lm1)
```

## Interpretation of the regression parameters

```{r}
library(ExploreModelMatrix)
ExploreModelMatrix::VisualizeDesign(diet, ~ block + protSource * protLevel)$plotlist
```

There are 3 levels for protein source, 2 levels for protein level and 10 levels
for the blocking variable. We will have one reference level for each respective 
variable: beef, low, block 1. So we need 2, 1 and 9 dummy variables to introduce 
the factors protein source, protein level and block in the linear model, 
respectively.

Hence, we can write down the linear model as follows:

$y_i=\beta_0+\beta_cx_{i,c} +\beta_px_{i,p}+\beta_hx_{i,h}+\beta_{ch}x_{i,c}x_{i,h}+\beta_{ph}x_{i,p}x_{i,h}+\beta_{b2}x_{i,b2}+\ldots+\beta_{b10}x_{i,b10}+\epsilon_{i}$

with:
$y_i$ the observed weight gain for rat i, 
$x_{i,h}$ a dummy variable which is 1 if rat i receives a high protein diet 
and is 0 otherwise,  
$x_{i,c}$ a dummy variable which is 1 if rat i receives a cereal diet 
and is 0 otherwise,  
$x_{i,p}$ a dummy variable which is 1 if rat i receives a pork diet 
and is 0 otherwise and 
$x_{i,bk}$ is a dummy variable which is 1 if rat i belongs to block $bk$ 
and is 0 otherwise, with $k \in 2,\ldots, 10$, 
and $\epsilon_i$ an error term which is normally distributed with mean 0 and 
variance $\sigma^2$, i.e. $\epsilon_i \sim N(0,\sigma^2)$.

- Rats that are assigned to block $k$ and receive a beef based low protein diet have
	a covariate pattern $x_{i,h}=0$, $x_{i,c}=0$, $x_{i,p}=0$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	Their mean weight gain is thus equal to $\mu_{l,b,bk}=\beta_0+\beta_{bk}$

- Rats that are assigned to block $k$ and receive a beef based high protein diet have
	a covariate pattern $x_{i,h}=1$, $x_{i,c}=0$, $x_{i,p}=0$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	Their mean weight gain is thus equal to $\mu_{h,b,bk}=\beta_0+\beta_h+\beta_{bk}$

- Rats that are assigned to block $k$ and receive a cereal based low protein diet have
	a covariate pattern $x_{i,h}=0$, $x_{i,c}=1$, $x_{i,p}=0$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	Their mean weight gain is thus equal to $\mu_{h,c,bk}=\beta_0+\beta_c+\beta_{bk}$

- Rats that are assigned to block $k$ and receive a cereal based heigh protein diet have
	a covariate pattern $x_{i,h}=1$, $x_{i,c}=1$, $x_{i,p}=0$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	Their mean weight gain is thus equal to $\mu_{h,c,bk}=\beta_0+\beta_h+\beta_c+\beta_{ch}+\beta_{bk}$

- Rats that are assigned to block $k$ and receive a pork based low protein diet have
	a covariate pattern $x_{i,h}=0$,  $x_{i,c}=0$, $x_{i,p}=1$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	Their mean weight gain is thus equal to $\mu_{l,p,bk}=\beta_0+\beta_p+\beta_{bk}$

- Rats that are assigned to block $k$ and receive a pork based heigh protein diet have
	a covariate pattern $x_{i,h}=1$, $x_{i,c}=0$, $x_{i,p}=1$, $x_{i,bm}=0$ with $m\neq k$ and $x_{i,bk}=1$. 
	There mean weight gain is thus equal to $\mu_{h,p,bk}=\beta_0+\beta_h+\beta_p+\beta_{ph}+\beta_{bk}$
	
We can now relate this to the output of the `lm` function:

- The intercept $\beta_0$ is thus the average weight increase in the low beef 
diet for rats in block 1.

- The parameter $\beta_c$: the average difference in weight gain between cereal-low and 
beef-low diet is `r lm1$coef["protSourcec"]`g.

- The parameter $\beta_p$: the average difference in weight gain between pork-low and 
beef-low diet is `r lm1$coef["protSourcep"]`g.

- The parameter $\beta_h$: the average difference in weight gain between beef-high and 
beef-Low diet is `r lm1$coef["protLevelh"]`g.

- The parameter $\beta_{ch}$ is the difference in the average weight gain difference 
due to the high protein level as compared to the low protein level for cereal diets 
as compared to the weight gain difference that occurs due to the protein level
in the reference class (here beef diet). Here this is negative, 
i.e. `r lm1$coef["protSourcec:protLevelh"]`g, thus the weight gain 
for the cereal protein source increases on average less between high and low protein diets 
than in beef based diets.

- The parameter $\beta_{ph}$ is the difference in the average weight gain difference 
due to the high protein level as compared to the low protein level for pork diets 
as compared to the weight gain difference that occurs due to the protein level
in the reference class (here beef diet). Here this is negative, 
i.e. `r lm1$coef["protSourcep:protLevelh"]`g, thus the weight gain 
for the proc protein source increases on average less between high and low protein diets 
than in beef based diets.

## Testing the overall (combined) effect of diet

Because there are multiple factors with different levels in the model, 
we can first assess the effect of the diet (protein Level, protein source 
and the interaction) by using anova. With this test we will assess the null
hypothesis that the average weight gain in each treatment is equal: i.e. $H_0: \mu_{b,l}=\mu_{b,h}=\mu_{c,h}=\mu_{c,l}=\mu_{p,h}=\mu_{p,h}$ versus 
the alternative hyptohesis $H_1:$ that at least two treatment means are 
different. 

```{r}
lm0 <- lm(weightGain ~ block, data=diet)
anova(lm0, lm1)
```

We can conclude that there is an extremely significant effect of the diet type 
(protein source and/or protein level and/or protein source-protein level 
interaction) on the weight gain of rats (p << 0.001).

## Assessing the interaction effect between protein source and protein level

```{r, message=FALSE, warning=FALSE}
library(car)
Anova(lm1, type="III")
```

There is a very significant interaction between the protein source and 
the protein level (p = `r format(Anova(lm1, type=3)$"Pr(>F)",digits=1)[5]`). This indicates that the average weight increase due to 
the protein level differs according to the protein source. 
Hence, we cannot assess the effect of the protein source and/or protein level 
independently because there effects of the protein source vary according to 
the protein level.

## Assessing specific contrasts

Imagine that we are interested in assessing if there is an effect of
 
1. protein source in the low protein diets 
	- $\mu_{c,l}-\mu_{b,l} = \beta_c$
	- $\mu_{p,l}-\mu_{b,l} = \beta_p$
	- $\mu_{c,l}-\mu_{p,l}= \beta_c-\beta_p$)


2. protein source in high protein diets 
	- $\mu_{c,h}-\mu_{b,h}=\beta_c+\beta_{ch}$
	- $\mu_{p,h}-\mu_{b,h}=\beta_p+\beta_{ph}$
	- $\mu_{p,h}-\mu_{c,h}=(\beta_c+\beta_{ch})-(\beta_p+\beta_{ph})$


3. protein level for
	- beef diets ($\mu_{b,h}-\mu_{b,l}=\beta_h$), 
	- cereal diets  ($\mu_{c,h}-\mu_{c,l}=\beta_h+\beta_{ch}$) and 
	- pork diets  ($\mu_{p,h}-\mu_{p,l}=\beta_h+\beta_{ph}$).


4. If the effect of the protein level differs between
	- beef and cereal  $(\mu_{c,h}-\mu_{c,l}) - (\mu_{b,h}-\mu_{b,l})=\beta_{ch}$
	- beef and pork $(\mu_{p,h}-\mu_{p,l}) - (\mu_{c,h}-\mu_{c,l})=\beta_{ph}$ and
	- cereal and pork diets $(\mu_{c,h}-\mu_{c,l}) - (\mu_{p,h}-\mu_{p,l})=\beta_{ch}-\beta_{ph}$.

These effects of interest are so-called 
**contrasts, i.e. linear combinations of the parameters**.

We can define the contrasts and assess the significance of the contrasts with 
the code below. The contrasts are given as input in the form of symbolic
descriptions to the `linfct` argument of the `glht` function.

```{r,message=FALSE, warning=FALSE}
library(multcomp)
set.seed(75468) # to get reproducible results (small effect if removed)
lm1MultComp <- glht(
    model = lm1,
    linfct = c("protSourcec = 0",
               "protSourcep = 0",
               "protSourcec- protSourcep = 0",
               "protSourcec + protSourcec:protLevelh = 0",
               "protSourcep + protSourcep:protLevelh = 0",
               "(protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh) = 0",
               "protLevelh = 0",
               "protLevelh +  protSourcec:protLevelh = 0",
               "protLevelh +  protSourcep:protLevelh = 0",
               "protSourcec:protLevelh = 0",
               "protSourcep:protLevelh = 0",
               "protSourcec:protLevelh - protSourcep:protLevelh = 0")
)
```

```{r}
summary(lm1MultComp)
```

```{r}
confint(lm1MultComp)
```

Note that the p-values and the confidence intervals are automatically corrected 
for multiple testing.

# Conclusion

- There is an extremely significant effect of the type of protein diet on the 
weight gain of rats  (p<<1e-6). The interaction between protein type and protein 
source is also very significant (p=`r round(Anova(lm1,type=3)$Pr[5],4)`). 

- The average weight gain does not vary significantly according to protein 
source in the diets with low protein levels (all p>0.99). 

- The weight gain in the cereal diet at high protein concentration is on 
average `r confint(lm1MultComp)$confint["protSourcec + protSourcec:protLevelh",1] %>% abs %>% round(.,1)`g and `r confint(lm1MultComp)$confint["(protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh)",1] %>% abs %>% round(.,1)`g g lower than in the high protein beef and pork diet, 
respectively (95% CI [`r confint(lm1MultComp)$confint["protSourcec + protSourcec:protLevelh",-1] %>% abs %>% round(.,1) %>% sort`] and [`r confint(lm1MultComp)$confint["(protSourcec + protSourcec:protLevelh) - (protSourcep + protSourcep:protLevelh)",-1] %>% abs %>% round(.,1) %>% sort`]) and the differences are 
extremely significant (p<0.001). The average weight gains after meat based 
diets at high protein levels, however, do not differ significantly 
(p>0.99).

- We also discovered an extremely significant difference in weight gain 
according to the protein level for beef and pork based diets (p<0.001).  
The weight gain on average increases with `r round(confint(lm1MultComp)$confint["protLevelh",1],1)`g and `r round(confint(lm1MultComp)$confint["protLevelh + protSourcep:protLevelh",1],1)`g in the high protein 
level as compared to the low protein beef and pork diet, respectively 
(95%CI [`r round(confint(lm1MultComp)$confint["protLevelh",-1],1)`] and [`r round(confint(lm1MultComp)$confint["protLevelh + protSourcep:protLevelh",-1],1)`]). The protein level effect is not significant
for the cereal diet (p>0.99). 

- Finally significant interactions between protein level and protein source were
found (p<0.05), i.e. the increase in weight gain due to protein level in cereal
based diets was 18.2g and 16.5g lower than that of beef and pork based diets,
respectively (95% CI [3g,33.4g] and [1.3g,31.7g]). The average difference in 
increase in weight gain due to protein level among the meat based diets was not 
significant (p>0.99).

All reported p-values and confidence intervals were corrected for multiple 
testing
