Effect of data missing at random on bias and other
Effect of data of missing based on a cutoff of the x value
As the slope of the regression line is the same across the domain of all x, we just have a smaller domain but still the same slope
Effect of data missing based on a cutoff of the y value
Error is represented by vertical distance between a point and the line.
small x values need a large positive error to meet threshold, thus as x increases, error term decreases on average –> omitted variable in ui that changes on avg. when x changes –> confounder
how to test if MAR believable
look at summary statistics of other variables and compare those with missing andn non-missing
What is ‘internal validity?
Estimate can be interpreted as a causal effect for the population that is used in the study
no issues (confounders, attenuation bias, bias due to y cutoffs, no simultanaeity/reverse causality, no bad control)
External validity
estimate is represenative of the effect for another population
nearly always an assumption, checked by creating estimates in various settings and checking if effects are comparable
Effect of data missing at random on bias and other
Effect of data of missing based on a cutoff of the x value
As the slope of the regression line is the same across the domain of all x, we just have a smaller domain but still the same slope
Effect of data missing based on a cutoff of the y value
Error is represented by vertical distance between a point and the line.
small x values need a large positive error to meet threshold, thus as x increases, error term decreases on average –> omitted variable in ui that changes on avg. when x changes –> confounder
What is ‘internal validity?
Estimate can be interpreted as a causal effect for the population that is used in the study
no issues (confounders, attenuation bias, bias due to y cutoffs, no simultanaeity/reverse causality, no bad control)
External validity
estimate is represenative of the effect for another population
nearly always an assumption, checked by creating estimates in various settings and checking if effects are comparable
What is standardising a variable?
standardising is a form of normalising where we
useful for when units cannot be easily understood
When standardiisng just x1, what is B1 interpreted as
𝛽1∗ is interpreted as “the average change in 𝑦 that is associated with 𝑥1 increasing by 1 standard deviation.”
standardising just y, what is B1 interpreted as
𝛽1∗ is interpreted as “the average number of standard deviations that 𝑦 changes by that is associated with 𝑥1 increasing by 1.”
standardising both x1 and y, what is B1 interpreted as?
“the average number of standard deviations that 𝑦 changes by that is associated with 𝑥1 increasing by 1 standard deviation.”
Do we have to subtract the mean and dividie by the standard deviation
No for the interpretation it is sufficient to divide by the standard deviation