Clicker Q

to go with Practicing Statistics by Kuiper & Sklar. Math 150 - Methods in Biostatistics.


In terms of the prerequisite for Math 150, Methods in Biostatistics, you should know at least a little bit (hopefully a lotta bit) about the following topics.

  1. Hypothesis test, confidence interval, sample mean, central limit theorem, standard deviation, standard error of a statistics, p-value, t-test, chi-square test.1

    1. Never heard of it
    2. Heard of it, but don’t know anything about it
    3. Know a little about it (or did once)
    4. Know something about it
    5. Confident about it

In terms of the prerequisite for Math 150, Methods in Biostatisitcs, you do not need to know the following topics

  1. Interaction, simple linear regression, multiple linear regression, logistic regression, survival analysis, R.2

    1. Never heard of it
    2. Heard of it, but don’t know anything about it
    3. Know a little about it (or did once)
    4. Know something about it
    5. Confident about it

  1. The Central Limit Theorem (CLT) says:3
    1. The sample average (statistic) converges to the true average (parameter)
    2. The sample average (statistic) converges to some point
    3. The distribution of the sample average (statistic) converges to a normal distribution
    4. The distribution of the sample average (statistic) converges to some distribution
    5. I have no idea what the CLT says

  1. The p-value is the probability:4
    1. that the null hypothesis is true given the observed data.
    2. of data as or more extreme than the observed data given that the null hypothesis is true.

  1. Why do we use a t distribution (instead of a z / normal distribution) in the t-test?5
    1. the technical conditions don’t hold
    2. the means are quite variable
    3. we like the letter t
    4. we have two samples
    5. we don’t know the true standard deviation parameter

  1. What happens if a t-test is used but isn’t appropriate (technical conditions don’t hold)?6
    1. the p-value isn’t actually the probability of our data or more extreme if H0 is true.
    2. the software won’t give a p-value as output
    3. the rejection region needs to be calculated in the opposite direction
    4. the world blows up

  1. We use linear regression to run a test of means (\(x_i = 0\) for controls, group 1; \(x_i = 1\) for cases, group 2) What is: \(\sum_i x_i\)?7
    1. \(n\)
    2. \(n_1\)
    3. \(n_2\)
    4. \(n_1 \cdot \overline{y}_1\)
    5. \(n_2 \cdot \overline{y}_2\)

  1. We use linear regression to run a test of means (\(x_i = 0\) for controls, group 1; \(x_i = 1\) for cases, group 2) What is: \(\sum_i x_iy_i\)?8
    1. \(n\)
    2. \(n_1\)
    3. \(n_2\)
    4. \(n_1 \cdot \overline{y}_1\)
    5. \(n_2 \cdot \overline{y}_2\)

  1. The regression technical conditions include:9
    1. The Y variable is normally distributed
    2. The X variable is normally distributed
    3. The residuals are normally distributed
    4. The slope coefficient is normally distributed
    5. The intercept coefficient is normally distributed

  1. We need the technical conditions to hold in order to calculate \(b_0\) and \(b_1.\)10

    1. TRUE
    2. FALSE
    3. It depends

  1. Why do we check technical conditions?11
    1. so that the inference is valid
    2. so that the estimates are valid
    3. so that the p-value is more likely to be small
    4. so that the confidence level is right
    5. for fun

  1. When writing the regression equation, why is there a hat ( ^) on the response variable?12
    1. because the prediction is an estimate
    2. because the prediction is an average
    3. because the prediction may be due to extrapolation
    4. a & b
    5. all of the above

  1. With a strong correlation and very small p-value, what can we conclude about happiness and life expectancy?13
    1. happiness causes longer lives
    2. longer lives cause happiness
    3. happiness and longer life are correlated
    4. happiness and longer life are perfectly predictive
    5. happiness and longer life are unrelated

  1. If there is no relationship in the population (true correlation = 0), then r = 0.14
    1. TRUE
    2. FALSE

  1. If there is no relationship in the population (true slope \(\beta_1 = 0\)), then \(b_1=0\).15
    1. TRUE
    2. FALSE

  1. Smaller variability around the regression line (\(\sigma\)):16
    1. increases the variability of \(b_1\).
    2. decreases the variability of \(b_1\).
    3. doesn’t necessarily change the variability of \(b_1\).

  1. Smaller variability in the explanatory variable (SD(X) = \(s_X\)):17
    1. increases the variability of \(b_1\).
    2. decreases the variability of \(b_1\).
    3. doesn’t necessarily change the variability of \(b_1\).

  1. A smaller sample size (\(n\)):18
    1. increases the variability of \(b_1\).
    2. decreases the variability of \(b_1\).
    3. doesn’t necessarily change the variability of \(b_1\).

  1. We transform our variables…19
    1. … to find the highest \(r^2\) value.
    2. … when the X variable is not normally distributed.
    3. … to make the model easier to interpret.
    4. … so that the technical conditions are met.

  1. In the Botox and Pain Relief example, the p-value is calculated. What does “probability” refer to?20
    1. random allocation
    2. random sample

p-value = probability of the observed data or more extreme given the null hypothesis is true.


  1. “Observed data or more extreme” is:21
    1. fewer than 9
    2. 9 or fewer
    3. 9 or more
    4. more than 9

  1. What is the mean value of the null sampling distribution for the number of Botox therapy who showed pain reduction?22
    1. 0
    2. 9
    3. 5.3
    4. 11
    5. 15

  1. What conclusion would you draw from the Back Pain and Botox study?23
    1. Not enough evidence to conclude that Botox is more effective than the placebo.
    2. Strong evidence that Botox is equally as effective as the placebo.
    3. Strong evidence that Botox is more effective than the placebo.

  1. If we consider those in the study with back pain to be representative of all people with back pain, what would you conclude about the percentage of people who will have reduced back pain if they use Botox?24
    1. Substantially greater than 50%
    2. Substantially less than 50%
    3. Close to 50%

  1. Material check-in
    1. So far, so good
    2. Concepts are good, R is confusing
    3. R is good, concepts are confusing
    4. Everything is confusing

  1. People check-in
    1. So far, so good
    2. I can go to office hours / mentor sessions, but I didn’t happen to this week.
    3. I can’t make the scheduled office hours / mentor sessions
    4. I’m looking for someone to study with

See Canvas front page for anonymous survey / feedback for the class. Also, if you are looking for people to work with, you could contact me directly (non-anonymously!) so that I can connect you to people.


  1. Sample 1,000,000 people who are over 6’ tall and 1,000,000 people who are under 6’ tall. Record if the person is in the NBA. What is measurable?25
    1. P(NBA if 6’ tall)
    2. P(6’ tall if in the NBA)
    3. both
    4. neither

  1. Sample 100 people who are in the NBA and 100 people who are not in the NBA. Record if the person is over 6’ tall. What is measurable?26
    1. P(NBA if 6’ tall)
    2. P(6’ tall if in the NBA)
    3. both
    4. neither

  1. Sample 10,000,000 people. Record their height and whether or not they are in the NBA. What is measurable?27
    1. P(NBA if 6’ tall)
    2. P(6’ tall if in the NBA)
    3. both
    4. neither

From the NYT, March 21, 2023, https://www.nytimes.com/2023/03/21/sports/basketball/tall-basketball-march-madness.html

The average W.N.B.A. player, at a shade taller than 6 feet, towers over the average American woman (5 feet 3.5 inches). American men who are between 6 feet and 6-2 — significantly taller than the 5-9 average — have about a five in a million chance of making the N.B.A., according to “The Sports Gene,” a 2013 book by David Epstein about the science of athletic performance. But if you hit the genetic lottery and happen to be 7 feet tall, your chances of landing in the N.B.A. are roughly one in six. (There are 38 players on active rosters who are 7 feet or taller, according to N.B.A. Advanced Stats; the average height of an N.B.A. player is 6 feet 6.5 inches.)

https://davidepstein.com/david-epstein-the-sports-gene/


  1. Calcium channel blockers have recently been reported to be associated with increased mortality. Cardiac patients who recently died of their heart disease were compared to control cardiac patients with similar disease who survive. Assume such a study had found that 40% of the recent cardiac deaths were taking calcium channel blockers at the time of death, as compared to 25% of the controls.28
    1. Case-control
    2. Cohort
    3. Cross-classification

  1. It is well known that the use of urinary catheters conveys a substantial risk of urinary tract infection (UTI). A group of physicians believe that, in an intensive care setting, use of one particular type of urinary catheter is more likely to encourage infection than use of other types. They therefore review medical records over a recent period for all uses of urinary catheters in an ICU. They find that 200 new UTIs occurred during 1000 ICU patient-days of catheterization with the suspect type of catheter, as compared to 100 new UTIs during 5000 ICU-patient days of catheterization with all other types. Noting the increased frequency of new UTIs when the suspect catheter type is used, they regard their hypothesis as confirmed. To reduce nosocomial UTIs, they recommend discontinuing use of that type of catheter in the ICU.29
    1. Case-control
    2. Cohort
    3. Cross-classification

  1. When we select individuals based on the explanatory variable, we cannot accurately measure30
    1. the proportion of people in the population in each explanatory category
    2. the proportion of people in the population in each response group
    3. anything about the population
    4. confounding variables

  1. Relative Risk is31
    1. the difference of two proportions
    2. the ratio of two proportions
    3. the log of the ratio of two proportions
    4. the log of the difference of two proportions

  1. The odds ratio is “invariant to which variable is explanatory and which is response” means:32
    1. we always put the bigger odds in the numerator
    2. we must collect data so that we can estimate the response in the population
    3. which variable is called the explanatory changes the value of the OR
    4. which variable is called the explanatory does not change the value of the OR

  1. In finding a CI for RR = p1/p2, why is it okay to exponentiate the end points of the interval for ln(p1/p2)?33
    1. Because if ln(p1/p2) is in the original interval, p1/p2 will be in the exponentiated interval.
    2. Because taking the natural log of the RR makes the distribution approximately normal.
    3. Because the natural log compresses values that are bigger than 1 and spreads values that are smaller than 1.
    4. Because we can get exact p-values using Fisher’s Exact Test.

  1. In order to find a CI for the true OR, our steps are:34
  1. find \(\widehat{\ln(\mbox{OR})}\)
  2. add \(\pm \ z^* \sqrt{\frac{1}{n_1 \hat{p}_1 (1-\hat{p}_1)} + \frac{1}{n_2 \hat{p}_2 (1-\hat{p}_2)}}\)
  3. take exp of the endpoints
  1. because the sampling distribution of \(\widehat{\mbox{OR}}\) is normal
  2. because OR is typically greater than 1
  3. because the \(\ln\) transformation makes the sampling distribution almost normal
  4. because OR is invariant to the choice of explanatory or response variable

  1. I know where to find: the solutions to the warm-ups, the clicker questions (with solutions), and the HW solutions35
    1. TRUE
    2. FALSE

  1. At the value \(x = -\beta_0 / \beta_1\), the probability of success is:36
    1. 0
    2. 0.5
    3. 1
    4. depends on \(\beta_0\)
    5. depends on \(\beta_1\)

  1. The logistic model gives probability of failure:37
    1. \(\frac{e^{\beta_0+ \beta_1 x}}{1+ e^{\beta_0+ \beta_1 x}}\)
    2. \(\frac{1}{1+ e^{\beta_0+ \beta_1 x}}\)
    3. \(e^{\beta_0+ \beta_1 x}\)
    4. \(e^{-(\beta_0+ \beta_1 x)}\)
    5. \(\beta_0+ \beta_1 x\)

  1. The logistic model gives odds of success:38
    1. \(\frac{e^{\beta_0+ \beta_1 x}}{1+ e^{\beta_0+ \beta_1 x}}\)
    2. \(\frac{1}{1+ e^{\beta_0+ \beta_1 x}}\)
    3. \(e^{\beta_0+ \beta_1 x}\)
    4. \(e^{-(\beta_0+ \beta_1 x)}\)
    5. \(\beta_0+ \beta_1 x\)

  1. The logistic model gives odds of failure:39
    1. \(\frac{e^{\beta_0+ \beta_1 x}}{1+ e^{\beta_0+ \beta_1 x}}\)
    2. \(\frac{1}{1+ e^{\beta_0+ \beta_1 x}}\)
    3. \(e^{\beta_0+ \beta_1 x}\)
    4. \(e^{-(\beta_0+ \beta_1 x)}\)
    5. \(\beta_0+ \beta_1 x\)

  1. With a logistic regression model, the relative risk of success (for a one unit increase in X) is:40
    1. \(- \beta_0/\beta_1\)
    2. \(\beta_0+ \beta_1 x\)
    3. \(e^{\beta_0+ \beta_1 x}\)
    4. a non-linear function of X (which depends on X )

  1. If we want the relative risk of survival (for a one unit increase in X) to be independent of X, we should use which link:41
    1. linear
    2. logistic
    3. complementary log-log
    4. log-linear

  1. You take a sample of size 4 from a binary population and get: FSFF. (failure, success, failure, failure) What is your guess for p = P(success)?42
    1. 0.05
    2. 0.15
    3. 0.25
    4. 0.5
    5. 0.75

  1. In a logistic regression model, the variability is given by43
    1. Normal Y given X
    2. Binomial Y given X
    3. Bernoulli Y given X
    4. Poisson Y given X

  1. When trying to find estimates for \(\beta_0\) and \(\beta_1\), we maximize the likelihood. \[\prod_{i=1}^n \bigg(\frac{e^{\beta_0+ \beta_1 x_i}}{1+ e^{\beta_0+ \beta_1 x_i}}\bigg)^{y_i}\bigg(\frac{1}{1+ e^{\beta_0+ \beta_1 x_i}}\bigg)^{1 - y_i}\] Take the derivative with respect to which variable(s):44
    1. X
    2. Y
    3. \(\beta_0\)
    4. \(\beta_1\)
    5. \(\beta_0\) and \(\beta_1\)

  1. Maximum likelihood estimation seeks to:45
    1. Find the data which are most likely under the model.
    2. Find the parameters which are most likely under the model.
    3. Find the parameters which make the data most likely under the model.
    4. Find the data which make the parameters most likely under the model.

  1. We use maximum likelihood estimation because:46
    1. It gives an principled approach for estimating the parameters.
    2. The estimates are asymptotically normally distributed.
    3. The estimates are always easy to compute.
    4. All of the above.
    5. Some of the above.

  1. We know that for a given data set (with MLEs of \(b_0\),\(b_1\)):47
    1. \(L(b_0,b_1)< L(b_0,\beta_1=0)\) always
    2. \(L(b_0,b_1)> L(b_0,\beta_1=0)\) always
    3. \(L(b_0,b_1) \leq L(b_0,\beta_1=0)\) always
    4. \(L(b_0,b_1) \geq L(b_0,\beta_1=0)\) always

  1. How many parameters did we estimate in the HERS warm-up with the additive model?48
    1. 1
    2. 3
    3. 4
    4. 2757
    5. 2761

  1. How many parameters did we estimate in the HERS warm-up with the interaction model?49
    1. 3
    2. 4
    3. 6
    4. 7
    5. 12

  1. What are the df for the LRT addressing whether interaction is needed in the HERS warm-up?50
    1. 2
    2. 3
    3. 2760
    4. 2754
    5. 2757

  1. (Bird nest example) How many parameters do we estimate when considering Length as a categorical variable? (the only variable)51
    1. 0
    2. 1
    3. 2
    4. 33
    5. 34

  1. (Bird nest example) How many df for the LRT addressing whether Length (as a categorical variable) belongs in the model?52
    1. 0
    2. 1
    3. 2
    4. 33
    5. 34

  1. (Bird nest example) How many df for the LRT addressing whether Incubate and Color belong in the model (given Length is determined to be in the model)?53
    1. 0
    2. 1
    3. 2
    4. 3
    5. 4

  1. An interaction term in a multiple logistic regression model may be used when:54
    1. the model fit is poor.
    2. there is a quadratic relationship between the response and explanatory variables.
    3. neither one of two explanatory variables contribute significantly to the regression model.
    4. the relationship between X1 and P(success) changes for differing values of X2.

  1. The interpretations of the main effects (on their own) make sense only when the interaction component is not significant.55
    1. TRUE
    2. FALSE

  1. If the interaction is significant but the main effects aren’t:56
    1. report on the significance of the main effects
    2. remove the main effects from the model
    3. avoid talking about main effects on their own
    4. test whether the main effects are significant without interaction in the model

  1. With two variables of interest, what should you test first?57
    1. Variable 1.
    2. Variable 2.
    3. The interaction between variables 1 and 2.
    4. None of the above.

  1. Consider variable 1 is continuous and variable 2 has 4 levels. How many degrees of freedom are associated with the drop in deviance test (LRT) of their overall interaction?58
    1. 1
    2. 2
    3. 3
    4. 4
    5. 5

  1. When selecting variables, it is important that59
    1. The model predicts training data well
    2. The model predicts test data well
    3. The coefficients on the variables are all significant
    4. The relationships between the variables make sense

  1. To get a sense of the true accuracy of the model, the test data should be assessed (for accuracy)60
    1. on the first model only.
    2. on the last model only.
    3. on every model in the process.

  1. If I am using all features of my dataset and I achieve 100% accuracy on my training set, but ~70% on testing set, what should I look out for?61
    1. Underfitting
    2. Nothing, the model is perfect
    3. Overfitting

  1. If I am picking and choosing between features of my dataset and I achieve 30% accuracy on my training set, and ~30% on testing set, what should I look out for?62
    1. Underfitting
    2. Nothing, the model is perfect
    3. Overfitting

  1. Cross validating will guarantee that the model does not overfit.63
    1. TRUE
    2. FALSE

  1. Suppose we want to compute 10-Fold Cross-Validation error on 200 training examples. We need to compute a model error rate N1 times, and the Cross-Validation error is the average of the errors. To compute each error, we need to train a model with data of size N2, and test the model on the data of size N3. What are the numbers for N1, N2, N3?64
    1. N1 = 1, N2 = 180, N3 = 20
    2. N1 = 10, N2 = 180, N3 = 20
    3. N1 = 10, N2 = 200, N3 = 20
    4. N1 = 10, N2 = 200, N3 = 200
    5. N1 = 20, N2 = 180, N3 = 20

  1. You are reviewing papers for Fancy Conference, and you see submissions with the following claims. Which ones would you consider accepting?65
    1. My method achieves a training error lower than all previous methods!
    2. My method achieves a test error lower than all previous methods! (Footnote: When variables are chosen so as to min test error.)
    3. My method achieves a test error lower than all previous methods! (Footnote: When variables are chosen so as to min CV error.)
    4. My method achieves a CV error lower than all previous methods! (Footnote: When variables are chosen so as to min CV error.)

  1. Which model is better (according to ROC)?66
    1. pink because it goes closer to (1,1)
    2. pink because it is closer to y=x
    3. blue because it is farther from y=x
    4. blue because it is steeper
    5. neither


  1. In ROC curve, the x-axis measures67
    1. True Pos Rate which we want high
    2. False Pos Rate which we want low
    3. True Neg Rate which we want high
    4. False Neg Rate which we want low

  1. Quiz on 11 topics (you know nothing). Your friends know topics:
    A: {1, 2, 3, 4, 5, 6, 7}
    B: {8, 9, 10}
    C: {1, 2, 3, 4, 8, 10}
    D: {5, 6, 7, 9, 11}
    Who should you choose to help you answer the questions?68
    1. A
    2. B
    3. C
    4. D
    5. can’t tell

  1. Who do you want to choose next?69
    A: {1, 2, 3, 4, 5, 6, 7}
    B: {8, 9, 10}
    C: {1, 2, 3, 4, 8, 10}
    D: {5, 6, 7, 9, 11}
    1. A
    2. B
    3. C
    4. D
    5. can’t tell

  1. If you can pick two people, who do you pick?70
    A: {1, 2, 3, 4, 5, 6, 7}
    B: {8, 9, 10}
    C: {1, 2, 3, 4, 8, 10}
    D: {5, 6, 7, 9, 11}
    1. A, B
    2. A, C
    3. A, D
    4. C, B
    5. C, D

  1. The variables in the k-variable model identified by forward selection are a subset of the variables in the (k+1)-variable model identified by forward selection.71
    1. TRUE (always TRUE)
    2. FALSE (not always TRUE)

  1. The variables in the k-variable model identified by backward selection are a subset of the variables in the (k+1)-variable model identified by backward selection.72
    1. TRUE (always TRUE)
    2. FALSE (not always TRUE)

  1. The variables in the k-variable model identified by backward selection are a subset of the variables in the (k+1)-variable model identified by forward selection.73
    1. TRUE (always TRUE)
    2. FALSE (not always TRUE)

  1. The variables in the k-variable model identified by forward selection are a subset of the variables in the (k+1)-variable model identified by backward selection.74
    1. TRUE (always TRUE)
    2. FALSE (not always TRUE)

  1. The variables in the k-variable model identified by best-subsets selection are a subset of the variables in the (k+1)-variable model identified by best-subsets selection.75
    1. TRUE (always TRUE)
    2. FALSE (not always TRUE)

  1. In a drop-in-deviance test (LRT), the reduced model corresponds to the null hypothesis being true.76
    1. TRUE
    2. FALSE

  1. In a drop-in-deviance test (LRT), the full model corresponds to the alternative hypothesis being true.77
    1. TRUE
    2. FALSE

  1. With model building:78
    1. There are many ways to find a good model.
    2. There is always one right answer.
    3. There is no end to the fun.
    4. Can we take a pure math class yet?

  1. When probability of being able to buy a candy bar is modeled as a function of the number of coins, the coefficient on number of coins is:79
    1. positive
    2. negative
    3. zero
    4. no intuition exists for being able to answer this question

  1. When probability of being able to buy a candy bar is modeled as a function of the number of low coins, the coefficient on number of low coins is:80
    1. positive
    2. negative
    3. zero
    4. no intuition exists for being able to answer this question

  1. When probability of being able to buy a candy bar is modeled as a function of the number of coins and number of low coins, the coefficient on number of coins is:81
    1. positive
    2. negative
    3. zero
    4. no intuition exists for being able to answer this question

  1. When probability of being able to buy a candy bar is modeled as a function of the number of coins and number of low coins, the coefficient on number of low coins is:82
    1. positive
    2. negative
    3. zero
    4. no intuition exists for being able to answer this question

  1. If we consider the censored times to be event times, the empirical survival curve will (on average)83
    1. underestimate the parameter
    2. overestimate the parameter
    3. sometimes under and sometimes overestimate the parameter

  1. \(n_i - d_i = n_{i+1}\) when:84
    1. there are no deaths at time \(t_i\)
    2. there is no censoring at time \(t_i\)
    3. there are no deaths at time \(t_{i+1}\)
    4. there is no censoring at time \(t_{i+1}\)
    5. there is no censoring at time \(t_{i-1}\)

  1. \(\frac{(n_i - d_i)}{n_i} = 1\) when:85
    1. there are no deaths at time \(t_i\)
    2. there is no censoring at time \(t_i\)
    3. there are no deaths at time \(t_{i+1}\)
    4. there is no censoring at time \(t_{i+1}\)
    5. there is no censoring at time \(t_{i-1}\)

  1. Prop survive > 50 days, treated (turquoise line)86
    1. ~0.65
    2. ~0.35
    3. ~0.45
    4. we only know it’s bigger than red
    5. we only know it’s smaller than red



  1. Kaplan Meier curves (Log-Rank p-value),87
    1. blue is clearly better
    2. red is clearly better
    3. can’t tell because they cross
    4. can’t tell because the p-value is big
    5. can’t tell because the p-value is small



  1. In the log-rank test, why is it okay to consider only one cell of the 2x2 table at time \(t_j\)?88
    1. Because the row totals are fixed.
    2. Because the column totals are fixed.
    3. Because the row and column totals are fixed.
    4. Because the total number of observations is fixed.

  1. What does it mean for the log rank test to be more powerful than the Wilcoxon test?89
    1. log rank is more likely to reject \(H_0\) when \(H_0\) is true.
    2. log rank is more likely to reject \(H_0\) when \(H_0\) is false.
    3. log rank is less likely to reject \(H_0\) when \(H_0\) is true.
    4. log rank is less likely to reject \(H_0\) when \(H_0\) is false.

  1. The hazard at time \(t\) represents:90
    1. the probability of the event
    2. the instantaneous rate of the event
    3. the relative risk of the event
    4. the odds ratio of the event

  1. The last entry in the table for the h(t) column is NA because:91
    1. the last observation was a death
    2. the last observation was censored
    3. the time interval is too big
    4. the time interval is too small

Table 9.6 [@KuiperSklar]

  1. Censored observations are\(\ldots\)?92
    1. More important than non-censored ones in survival analysis
    2. Are assumed to be normally distributed over time
    3. Are assumed to have the same survival chances as uncensored observations
    4. Are essential to allow calculation of the Kaplan Meier plot
    5. Are allocated to the baseline survival curve

  1. Survival Analysis: for a one unit change of an explanatory variable, the corresponding coefficient \(e^\beta\) represents:93
    1. baseline survival
    2. survival ratio
    3. baseline hazard
    4. hazard ratio

  1. In survival analysis, the closest interpretation of the value \(e^\beta\) is:94
    1. odds
    2. probability
    3. time to event
    4. relative risk
    5. odds ratio

  1. Let the event be death. If larger values of the explanatory variable are associated with higher likelihood of survival, the coefficient \((\beta)\) should be 95
    1. bigger than 1
    2. smaller than 1
    3. positive
    4. negative
    5. zero

  1. Let the event be death. If larger values of the variable are NOT associated with higher (or lower) likelihood of survival, the coefficient \((\beta)\) should be96
    1. bigger than 1
    2. smaller than 1
    3. positive
    4. negative
    5. zero

  1. BP violates the “linear HR” condition if:97
    1. the ln ratio of the hazard curves is not linear with respect to BP
    2. the ln ratio of the survival curves is not linear with respect to BP
    3. the effect of BP is to increase the hazard
    4. the effect of BP is to decrease the hazard
    5. there is no effect due to BP

  1. A Cox regression analysis:98
    1. Is used to analyze survival data when individuals in the study are followed for varying lengths of time.
    2. Can only be used when there are censored data
    3. Assumes that the relative hazard for a particular variable is always constant
    4. Uses the logrank statistic to compare two survival curves
    5. Relies on the condition that the explanatory variables (covariates) in the model are normally distributed.

  1. The effect of weight could violate PH if:99
    1. people of different weights are in control vs treatment group
    2. people tend to weigh less over time
    3. the hazard function for weight is not monotonic
    4. the hazard function changes as a function of weight which is also changing over time

  1. The effect of treatment could violate PH if:100
    1. the treatment has no effect
    2. the treatment produces short term benefits only
    3. the treatment effect interacts with a different variable, like gender
    4. there is more than one treatment group

  1. AIC, BIC, model validation, and stepwise regression are methods for101
    1. parameter estimation
    2. variable selection

  1. If \(\alpha = 0.05\), I would expect 5% of all hypotheses to be rejected.102
    1. TRUE
    2. FALSE

  1. Power is:103
    1. P(type I error)
    2. P(type II error)
    3. 1 – P(type I error)
    4. 1 – P(type II error)

type I = \(H_0\) true, but we reject
type II = \(H_0\) false, but we fail to reject
power = P(rejecting when \(H_0\) false)


  1. The p-value is104
    1. P(\(H_0\) is true | data)
    2. P(\(H_a\) is true | data)
    3. P(data | \(H_0\) is true)
    4. P(data | \(H_a\) is true)
    5. 1 – P(data | \(H_0\) is true)

RA Fisher (1929) >“… An observation is judged significant, if it would rarely have been produced, in the absence of a real cause of the kind we are seeking. It is a common practice to judge a result significant, if it is of such a magnitude that it would have been produced by chance not more frequently than once in twenty trials. This is an arbitrary, but convenient, level of significance for the practical investigator, but it does not mean that he allows himself to be deceived once in every twenty experiments. The test of significance only tells him what to ignore, namely all experiments in which significant results are not obtained. He should only claim that a phenomenon is experimentally demonstrable when he knows how to design an experiment so that it will rarely fail to give a significant result. Consequently, isolated significant results which he does not know how to reproduce are left in suspense pending further investigation.”


  1. For hypothesis testing, the problem of multiple comparisons (also known as the multiple testing problem) results from the increase in ________ that occurs when statistical tests are used repeatedly.105
    1. Type I errors
    2. Type II errors
    3. Null hypothesis
    4. Statistical hypothesis testing

  1. If \(H_0\) is true, the p-values should be distributed:106
    1. Uniformly (equal prob) on 0 to 1
    2. Uniformly on -1 to 1
    3. Unimodal on 0 to 1
    4. Skewed left on 0 to 1
    5. Skewed right on 0 to 1

  1. Given many many tests (presumably some are null and some are “true”), a good estimate of the number of null tests is:107
    1. (# p-values > 0.5) / 2
    2. (# p-values > 0.5) * 2
    3. (# p-values < 0.5) / 2
    4. (# p-values < 0.5) * 2

  1. What do I do if the adjusted p-value is bigger than 1?108
    1. Leave it unadjusted
    2. Assign the value of the previous (“smaller”) p-value
    3. Round it to 1
    4. Divide by 2

  1. With Holm’s method, what do I do if the (m+1)^th adjusted p-value is smaller than the m^th adjusted p-value?109
    1. Leave it unadjusted
    2. Assign the value of the m^th adjusted p-value to the (m+1)^th adjusted p-value
    3. Round it to 1
    4. Divide by 2

  1. The false discovery rate represents110
    1. the proportion of true discoveries out of the total tests
    2. the proportion of true discoveries out of the total discoveries
    3. the ratio of the number of true discoveries divided by the number of null discoveries
    4. the number of null discoveries out of the total tests
    5. the number of null discoveries out of the total discoveries

  1. FDR and FWER differ in that111
    1. FDR is a rate and FWER is a probability
    2. FDR controls the rate of false positives
    3. FWER controls the probability of getting a false positive
    4. some of the above
    5. all of the above

  1. Which multiple comparisons adjustment gives the highest power?112
    1. Bonferonni
    2. Holm
    3. Benjamini-Hochberg
    4. Storey (q-values)

Footnotes

  1. preferably d or e. maybe c on some of them.↩︎

  2. these are the topics we will be covering. Would be nice if you have heard of them.↩︎

    1. The distribution of the sample average (statistic) converges to a normal distribution
    ↩︎
    1. of data as or more extreme than the observed data given that the null hypothesis is true.
    ↩︎
    1. we don’t know the true standard deviation parameter
    ↩︎
    1. the p-value isn’t actually the probability of our data or more extreme if H0 is true.
    ↩︎
    1. \(n_2\)
    ↩︎
    1. \(n_2 \cdot \overline{y}_2\)
    ↩︎
    1. The residuals are normally distributed (which induces a., d., and e.). There is nothing in the technical conditions about the distribution of X (remember, X can be binary!).
    ↩︎
  3. FALSE. We can always minimize the sums of squares, regardless of whether or not the model is any good.↩︎

    1. so that the inference is valid (and also for fun). Note that d. so that the confidence level is right is also a correct answer because confidence intervals are all part of the “inference” paradigm.
    ↩︎
    1. due to estimation and average
    ↩︎
    1. happiness and longer life are correlated
    ↩︎
    1. FALSE, there is no reason that the statistic will equal the parameter.
    ↩︎
    1. FALSE, there is no reason that the statistic will equal the parameter.
    ↩︎
    1. decreases the variability of \(b_1\).
    ↩︎
    1. increases the variability of \(b_1\).
    ↩︎
    1. increases the variability of \(b_1\).
    ↩︎
    1. so that the technical conditions are met.
    ↩︎
    1. random allocation
    ↩︎
    1. 9 or more
    ↩︎
    1. 5.3 because (15/31)*11 = 5.3
    ↩︎
    1. Strong evidence that Botox is more effective than the placebo.
    ↩︎
    1. Close to 50% (the point estimate is 0.6)
    ↩︎
    1. P(NBA if 6’ tall) (cohort: cannot measure the probability of the explanatory variable given the response)
    ↩︎
    1. P(6’ tall if in the NBA) (case-control: cannot measure the probability of the response variable given a level of the explanatory variable)
    ↩︎
    1. both (cross-classification: can measure all the probabilities)
    ↩︎
    1. case-control (they selected based on people who had died or not)
    ↩︎
    1. cross-classification (they selected all uses of catheters)
    ↩︎
    1. the proportion of people in the population in each explanatory category (tbh, we can’t measure b either, but we can measure the proportion of people in each response group, separated by the explanatory variable)
    ↩︎
    1. the ratio of two proportions
    ↩︎
    1. which variable is called the explanatory does not change the value of the OR
    ↩︎
    1. Because if ln(p1/p2) is in the original interval, p1/p2 will be in the exponentiated interval.
    ↩︎
    1. because the \(\ln\) transformation makes the sampling distribution almost normal
    ↩︎
  4. The warm-up solutions and clicker questions are on the main course website. The HW solutions are on Canvas under Files.↩︎

    1. 0.5
    ↩︎
    1. \(\frac{1}{1+ e^{\beta_0+ \beta_1 x}}\)
    ↩︎
    1. \(e^{\beta_0+ \beta_1 x}\)
    ↩︎
    1. \(e^{-(\beta_0+ \beta_1 x)}\)
    ↩︎
    1. a non-linear function of X (which depends on X )
    ↩︎
    1. log-linear
    ↩︎
    1. 0.25
    ↩︎
    1. Bernoulli Y given X
    ↩︎
    1. \(\beta_0\) and \(\beta_1\)
    ↩︎
    1. Find the parameters which make the data most likely under the model.
    ↩︎
    1. Some of the above (a. It gives an principled approach for estimating the parameters. and b. The estimates are asymptotically normally distributed.)
    ↩︎
    1. \(L(b_0,b_1) \geq L(b_0,\beta_1=0)\) always
    ↩︎
    1. 4 parameter estimates: \(b_0, b_1, b_2, b_3\)
    ↩︎
    1. 7 parameter estimates: \(b_0, b_1, b_2, b_3, b_4, b_5, b_6\)
    ↩︎
    1. 3 (7 - 4 = 3)
    ↩︎
    1. 34
    ↩︎
    1. 33 (34 - 1 = 33)
    ↩︎
    1. 2 (4 - 2 = 2)
    ↩︎
    1. the relationship between X1 and P(success) changes for differing values of X2.
    ↩︎
    1. TRUE
    ↩︎
    1. avoid talking about main effects on their own
    ↩︎
    1. The interaction between variables 1 and 2. (probably… although there are many schools of thought on how to build models)
    ↩︎
    1. 3 (1 * (4-1) = 3)
    ↩︎
    1. The model predicts test data well
    ↩︎
    1. on the last model only.
    ↩︎
    1. overfitting
    ↩︎
    1. underfitting
    ↩︎
    1. FALSE. CV reduces the effect of overfitting, but at the end of the day, you are still building a model on the dataset at hand, and it is possible that you will overfit that dataet.
    ↩︎
    1. N1 = 10, N2 = 180, N3 = 20
    ↩︎
    1. My method achieves a test error lower than all previous methods! (Footnote: When variables are chosen so as to min CV error.)
    ↩︎
    1. blue because it is farther from the line y=x
    ↩︎
    1. False Pos Rate which we want low
    ↩︎
    1. A
    ↩︎
    1. B
    ↩︎
    1. C and D
    ↩︎
    1. TRUE
    ↩︎
    1. TRUE
    ↩︎
    1. FALSE
    ↩︎
    1. FALSE
    ↩︎
    1. FALSE
    ↩︎
    1. TRUE (the coefficient values are forced to be zero)
    ↩︎
    1. FALSE (the null model can exist within the full model because there is flexibility in the values of the coefficients)
    ↩︎
    1. There are many ways to find a good model. Also, c. there is no end to the fun.
    ↩︎
    1. positive
    ↩︎
    1. positive
    ↩︎
    1. positive
    ↩︎
    1. negative
    ↩︎
    1. underestimate the parameter
    ↩︎
    1. there is no censoring at time \(t_i\)
    ↩︎
    1. there are no deaths at time \(t_i\)
    ↩︎
    1. ~0.65
    ↩︎
    1. can’t tell because they cross (and also because d. the p-value is big)
    ↩︎
    1. Because the row and column totals are fixed.
    ↩︎
    1. log rank is more likely to reject \(H_0\) when \(H_0\) is false.
    ↩︎
    1. the instantaneous rate of the event
    ↩︎
    1. the last observation was censored. The reason the hazard is zero is because the width of the time interval is unknown, that is we don’t know when the last event time is.
    ↩︎
    1. Are assumed to have the same survival chances as uncensored observations
    ↩︎
    1. hazard ratio
    ↩︎
    1. relative risk
    ↩︎
    1. negative
    ↩︎
    1. zero
    ↩︎
    1. the ln ratio of the hazard curves is not linear with respect to BP
    ↩︎
    1. Is used to analyze survival data when individuals in the study are followed for varying lengths of time. and c. Assumes that the relative hazard for a particular variable is always constant
    ↩︎
    1. the hazard function changes as a function of weight which is also changing over time
    ↩︎
    1. the treatment produces short term benefits only
    ↩︎
    1. variable selection
    ↩︎
    1. FALSE, we’d expect 5% of all null hypotheses to be rejected
    ↩︎
    1. 1 – P(type II error)
    ↩︎
    1. P(data | \(H_0\) is true)
    ↩︎
    1. Type I errors
    ↩︎
    1. Uniformly (equal prob) on 0 to 1
    ↩︎
    1. (# p-values > 0.5) * 2
    ↩︎
    1. Round it to 1
    ↩︎
    1. Assign the value of the m^th adjusted p-value to the (m+1)^th adjusted p-value
    ↩︎
    1. the number of null discoveries out of the total discoveries
    ↩︎
    1. all of the above
    ↩︎
    1. Storey (q-values)
    ↩︎

Reuse