Making Claims & Providing Evidence
Reporting Model Fit
Justified Decisions
Reporting Model Parameters
KahootR
There are multiple researcher degrees of freedom when it comes to conducting research and statistical analyses
Essentially, this means that the choice of the analysis, interpretation, and final model is up to you as the analyst!
So, you must decide which is the final/best model for your data for each analysis and argue/defend your case in your results section
What do each of these statements need?
"People who have synaesthesia are more creative than people who don't have synaesthesia."
"There are only two types of people in this world: those who can extrapolate from incomplete data."
"The second, more complex model was a better model than the first model with only one predictor."
What do each of these statements need? EVIDENCE!
"People who have synaesthesia are more creative than people who don't have synaesthesia."
"There are only two types of people in this world: those who can extrapolate from incomplete data."
"The second, more complex model was a better model than the first model with only one predictor."
Use your in-text statistical results as if they were citations!
The statistics you report in each sentence should provide evidence for the claim you make in that sentence, and the sentence should be a complete thought without the statistics:
Use your in-text statistical results as if they were citations!
The statistics you report in each sentence should provide evidence for the claim you make in that sentence, and the sentence should be a complete thought without the statistics:
NO: "The means were (mean = 135.30…"
YES: "The means indicated that the synaesthete group (M = 135.30, SD = 12.06) had a higher score for creativity than the non-synaesthete group (M = 105.78, SD = 32.90)."
When reporting numbers to provide evidence for a claim you're making, keep in mind:
Test statistics like t, F, and p are italicised
Always round to 2 decimal places, except for p which should be rounded to three
Look up the correct reporting format if you're not sure!
When reporting our overall model, we need to build a narrative clearly describing each step (in full sentences), with evidence to back up any claims, and justified decisions throughout:
Specifically, we need to include:
What models we've fit to our data (& how well they fit)
What model was better
The assumptions you checked, and the outcome of those checks
The main results from the better model
First we need to describe what models we've fit to our data:
NO: "The model being constructed is a dual-predictor multiple regression with OLS estimation. Predictors will be considered to be significant if the probability p of finding the b values by chance is less than 0.05. The two first predictor was beauty and had p = 3.13e-05. The second p-value was 0.00165 for nativeno."
First we need to describe what models we've fit to our data:
NO: "The model being constructed is a dual-predictor multiple regression with OLS estimation. Predictors will be considered to be significant if the probability p of finding the b values by chance is less than 0.05. The two first predictor was beauty and had p = 3.13e-05. The second p-value was 0.00165 for nativeno."
YES: "The first model investigated instructor beauty ratings as a predictor of teaching evaluations, and showed satisfactory model fit; (R2 = .04, F(1, 461) = 17.08, p < .001). The second model investigated instructor beauty ratings and native speakers of English as predictors of teaching evaluations (R2 = .06, F(2, 460) = 13.72, p < .001), showing significantly better model fit compared to the first model (R2change = .02, F(1, 460) = 10.02, p = .002)."
First we need to describe what models we've fit to our data:
NO: "The model being constructed is a dual-predictor multiple regression with OLS estimation. Predictors will be considered to be significant if the probability p of finding the b values by chance is less than 0.05. The two first predictor was beauty and had p = 3.13e-05. The second p-value was 0.00165 for nativeno."
YES: "The first model investigated instructor beauty ratings as a predictor of teaching evaluations, and showed satisfactory model fit; (R2 = .04, F(1, 461) = 17.08, p < .001). The second model investigated instructor beauty ratings and native speakers of English as predictors of teaching evaluations (R2 = .06, F(2, 460) = 13.72, p < .001), showing significantly better model fit compared to the first model (R2change = .02, F(1, 460) = 10.02, p = .002)."
After fitting our models, we want to check our assumptions to see if our model was biased, and decide on which final model to report
But there are fewer standardised formats for assumptions checks reporting
You should aim to explain the process of your decision-making clearly, step by step
So let's look at the assumptions first, and then think about what we did to check our model, and then let's add in the evidence...
We checked some plots of linearity, heteroscedasticity, normality of residuals, & influential cases
We checked outliers using standardised residuals
We checked some plots of linearity, heteroscedasticity, normality of residuals, & influential cases
We checked outliers using standardised residuals
We fit robust models as a sensitivity check to examine the pattern of results compared to our original model
We checked some plots of linearity, heteroscedasticity, normality of residuals, & influential cases
We checked outliers using standardised residuals
We fit robust models as a sensitivity check to examine the pattern of results compared to our original model
We decided which model to report
"Non-linearity and heteroscedasticity were checked using a scatterplot of the predicted values against the residuals. The plot showed no issues with non-linearity or heteroscedasticity in the data."
"A Q-Q plot of the standardized residuals indicated the residuals were fairly normally distributed, with deviations at the tails."
"Non-linearity and heteroscedasticity were checked using a scatterplot of the predicted values against the residuals. The plot showed no issues with non-linearity or heteroscedasticity in the data."
"A Q-Q plot of the standardized residuals indicated the residuals were fairly normally distributed, with deviations at the tails."
"Influential cases were checked using Cook’s distance with all values below 0.05 suggesting there were no influential cases in our data. The standardised residuals were inspected for outliers; all cases fell within the expected ranges."
"Non-linearity and heteroscedasticity were checked using a scatterplot of the predicted values against the residuals. The plot showed no issues with non-linearity or heteroscedasticity in the data."
"A Q-Q plot of the standardized residuals indicated the residuals were fairly normally distributed, with deviations at the tails."
"Influential cases were checked using Cook’s distance with all values below 0.05 suggesting there were no influential cases in our data. The standardised residuals were inspected for outliers; all cases fell within the expected ranges."
"Robust models were fit to the data as a sensitivity check, and showed the same pattern of results as the original model."
"Non-linearity and heteroscedasticity were checked using a scatterplot of the predicted values against the residuals. The plot showed no issues with non-linearity or heteroscedasticity in the data."
"A Q-Q plot of the standardized residuals indicated the residuals were fairly normally distributed, with deviations at the tails."
"Influential cases were checked using Cook’s distance with all values below 0.05 suggesting there were no influential cases in our data. The standardised residuals were inspected for outliers; all cases fell within the expected ranges."
"Robust models were fit to the data as a sensitivity check, and showed the same pattern of results as the original model."
"Therefore, the final model reported is the unadjusted model predicting teaching evaluation scores from instructor beauty ratings and whether the instructor is a native english speaker."
"Non-linearity and heteroscedasticity were checked using a scatterplot of the predicted values against the residuals. The plot showed no issues with non-linearity or heteroscedasticity in the data."
"A Q-Q plot of the standardized residuals indicated the residuals were fairly normally distributed, with deviations at the tails."
"Influential cases were checked using Cook’s distance with all values below 0.05 suggesting there were no influential cases in our data. The standardised residuals were inspected for outliers; all cases fell within the expected ranges."
"Robust models were fit to the data as a sensitivity check, and showed the same pattern of results as the original model."
"Therefore, the final model reported is the unadjusted model predicting teaching evaluation scores from instructor beauty ratings and whether the instructor is a native english speaker."
We should report the effect of each predictor in full, with statistics and a plain language summary, and should compare the standardised betas to evidence which predictor had a stronger relationship with our outcome, e.g.,
We should report the effect of each predictor in full, with statistics and a plain language summary, and should compare the standardised betas to evidence which predictor had a stronger relationship with our outcome, e.g.,
"In the final model, instructors' beauty ratings significantly predicted their teaching evaluation scores (b = 0.13, SE(b) = 0.03, t = 4.21, p < .001, 95% CI [0.07, 0.20])"
"The findings suggest that as instructors’ beauty scores increase by one point on a scale from one to five, their teaching evaluations increase by 0.13 points (scale: 1-10)."
We should report the effect of each predictor in full, with statistics and a plain language summary, and should compare the standardised betas to evidence which predictor had a stronger relationship with our outcome, e.g.,
"In the final model, instructors' beauty ratings significantly predicted their teaching evaluation scores (b = 0.13, SE(b) = 0.03, t = 4.21, p < .001, 95% CI [0.07, 0.20])"
"The findings suggest that as instructors’ beauty scores increase by one point on a scale from one to five, their teaching evaluations increase by 0.13 points (scale: 1-10)."
"The standardised estimates for instructor beauty and native English speaker suggest that whether the instructor is a native English speaker is a stronger predictor (B = -0.60, SE(B) = 0.19, 95% CI [-0.97, -0.23]) of teaching evaluations compared to instructor beauty (B = 0.19, SE(B) = 0.05, 95% CI [ 0.10, 0.28])"
We should report the effect of each predictor in full, with statistics and a plain language summary, and should compare the standardised betas to evidence which predictor had a stronger relationship with our outcome, e.g.,
"In the final model, instructors' beauty ratings significantly predicted their teaching evaluation scores (b = 0.13, SE(b) = 0.03, t = 4.21, p < .001, 95% CI [0.07, 0.20])"
"The findings suggest that as instructors’ beauty scores increase by one point on a scale from one to five, their teaching evaluations increase by 0.13 points (scale: 1-10)."
"The standardised estimates for instructor beauty and native English speaker suggest that whether the instructor is a native English speaker is a stronger predictor (B = -0.60, SE(B) = 0.19, 95% CI [-0.97, -0.23]) of teaching evaluations compared to instructor beauty (B = 0.19, SE(B) = 0.05, 95% CI [ 0.10, 0.28])"
We can also include an APA style table of our results...
Don't define any statistical concepts you would find in a statistics/psychology textbook
You should definitely explain the decisions & results specific to your study
You can assume that your audience are interested in & somewhat familiar with the field of research & know analysis techniques/stats terms
But they do not know any of the details of the study you have conducted and analysed & have no idea what your data look like:
Don't define any statistical concepts you would find in a statistics/psychology textbook
You should definitely explain the decisions & results specific to your study
You can assume that your audience are interested in & somewhat familiar with the field of research & know analysis techniques/stats terms
But they do not know any of the details of the study you have conducted and analysed & have no idea what your data look like:
If you're not sure whether you've given enough evidence or clearly justified a decision, for each claim you make just ask yourself...
Submit the easteR eggs you've found today on Canvas
Nominate your peers for the savioR award
Have a go at the Cloud ChallengRs
Give session feedback here 😀
Making Claims & Providing Evidence
Reporting Model Fit
Justified Decisions
Reporting Model Parameters
KahootR
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
s | Toggle scribble toolbox |
Esc | Back to slideshow |