Term
If a null hypothesis is true, but you reject it, you have made what type of error? |
|
Definition
|
|
Term
If a null hypothesis is not true, and you do not reject it, then you have made what type of error? |
|
Definition
|
|
Term
As alpha goes up, beta goes ? |
|
Definition
|
|
Term
As beta goes up, power goes ? |
|
Definition
down; Remember that power is 1 - beta |
|
|
Term
As alpha goes up, power goes ? |
|
Definition
|
|
Term
Choose: anova F-tests are always/sometimes one sided/two sided for determining a p-value. |
|
Definition
|
|
Term
What are the 5 assumptions for calculating ANOVA? |
|
Definition
Random samples-are required to be unbiased Indepedent samples If they're dependent (or repeated), you have to use repeated measures ANOVA. Independent observations w/in groups Normal distribution - (aka normal population) Equal variances (homogeneity) |
|
|
Term
When calculating ANOVA, if you have unequal variances, you have to the data. Name three ways that you would do this to your data. |
|
Definition
transform; log, natural log, square root |
|
|
Term
is the sum of squares (SS) What is SSB? What are other names for SSB? What is SSW? What are other names for SSW? Choose: The signal/noise should be > than the signal/noise, otherwise, we won't be able to detect the signal/noise. |
|
Definition
Variance; It is the variance b/w groups (it is the difference of each mean from the overall mean); SS(treatment) or SS(effects) or signal; It is the variance w/in groups; SS(error) or noise. signal-noise-signal |
|
|
Term
|
Definition
It is the mean square b/w groups |
|
|
Term
SSB = 43.21 SSW = 50.68 T or F: the error is greater than the treatment Why or why not? |
|
Definition
False, b/c you still have to divide by the df. The SSB will have a much smaller df (eg, 2) than the SSW (eg, 39). After dividing the SSB and SSW by their respective df's, you get MSB = 21.61 and MSW = 1.30. |
|
|
Term
If a comparison is not planned ahead of time, you must use even if only one post-hoc comparison was performed. However, you usually perform all comparisons. |
|
Definition
|
|
Term
If comparisons are selected a priori, what does c =? If comparisons are selected post hoc, what does c =? |
|
Definition
a priori: c = number of comparisons to be made post hoc: c = total number of possible comparisons |
|
|
Term
What is the point of the Bonferroni method? |
|
Definition
It's like an adjusted alpha, labeled alpha'. It increases the CL for each CI in order to ensure a specific overall CL. Table 11.4 shows how conservative the bonferroni method is. |
|
|
Term
When should you use bonferroni? When should you use tukey? |
|
Definition
The bonferroni method should be used when comparisons are selected a priori. The Tukey method should be used when ALL comparisons are selected - it doesn't matter if they're selected a priori or post hoc. If comparisons are selected post hoc, use the Tukey method. |
|
|
Term
The same three factors that affect statistical outcomes affect ANOVA results. What are they? |
|
Definition
Treatment effects (differences b/w means) Variance (SSB compared to SSW) # of subjects |
|
|
Term
What are the 4 assumptions of the regression model? |
|
Definition
NICL Normal distribution Independence (y observations are independent of each other) Constancy of variance (the sd of y is constant for each x) Linear relationship (as opposed to exponential, curvilinear, etc) |
|
|
Term
What is residual analysis used to test? |
|
Definition
The assumptions of linear regression. |
|
|
Term
Correlation analysis uses pearson, r. What are the 3 assumptions of pearson? |
|
Definition
1) Both the ind & dep variables must b continuous. 2) Dependent variable must be rational 3)No multicollinearity (meaning ind variables cant be related to other ind variables) |
|
|
Term
What are the assumptions of B? |
|
Definition
1) Each y observation is independent 2) All potential y values are normally distributed for each x value 3) The std deviation of y is constant for each x 4) The sampling distribution of b is normal |
|
|
Term
What are the 4 assumptions for r? |
|
Definition
1) Each y observation is independent 2) The sampling distribution of both x and y are normal. 3) The std deviation of y is constant for each x 4) The sampling distribution of r is normal. |
|
|
Term
If all population means are , the sample means will be very close to each other and close to the . The weighted square differences will be , and a F will be calculated. |
|
Definition
equal; grand mean; small; small |
|
|
Term
ANOVA T or F: Variance cannot be negative |
|
Definition
True. If it is, there is too much error. |
|
|
Term
If some population means are quite different, corresponding sample means will be than each other and the . Corresponding weighted squares will be , the and will be small, and the F will be . |
|
Definition
quite different; grand mean; large; SSB; MSB; large |
|
|
Term
As alpha increases, Fcrit . |
|
Definition
|
|
Term
When is one-way ANOVA used? |
|
Definition
When you need to compare more than two means. More specifically, when you have one factor w/ more than two levels of the factor (ie, Drug A, Drug B, and DrugC). |
|
|
Term
|
Definition
Error slippage occurs when you try to do a bunch of t-tests when you really should have done ANOVA. You'll think you have significant differences, but you really don't, b/c the alpha is slipping from 0.5 to 1. |
|
|
Term
|
Definition
It means group 3, observation 2. |
|
|
Term
If you had two or more independent variables, what type of ANOVA would you use? |
|
Definition
Factorial ANOVA. The different factors may be Drugs A, B, & C (1st factor), Age (second factor), & gender (third factor). |
|
|
Term
When would you use one-way ANOVA? Name a requirement of ANOVA. |
|
Definition
Use one-way ANOVA when you have one factor and more than two levels. The samples in the different levels must be independent of each other. |
|
|
Term
When do you use post-hoc analysis? |
|
Definition
When the null has been rejected and you're trying to find out which mean(s) are different. |
|
|
Term
When do you use repeated measures ANOVA? |
|
Definition
If measures are dependent (or repeated), you use a crossover design called repeated measures ANOVA. |
|
|
Term
What is the proper term used when your putting variances into groups? What is another name for variance? |
|
Definition
Partitioning Sums of Squares |
|
|
Term
SSB & SSW are weighted by the . |
|
Definition
|
|
Term
Choose: When p-Value is greater/lesser than or equal to alpha, H0 cannot be rejected. |
|
Definition
|
|
Term
Which is worse: a type I or type II error? |
|
Definition
|
|
Term
A type I error is aka A type II error is aka |
|
Definition
|
|
Term
Define alpha What does it mean when we set out alpha to 0.05? |
|
Definition
It is the probability that a difference does not exist when the study shows that a difference does exist. If a study was repeated 100 times, we would be willing to be wrong 5 times. We would conclude 5 times that a difference exists when in fact it does not. |
|
|
Term
T or F: A rejection region will be larger for a one-tail alpha than a two-tail alpha |
|
Definition
|
|
Term
|
Definition
It is the percent chance that a difference b/w experimental grps does exist even though the study concludes no difference. |
|
|
Term
T or F: We usually look at beta after an experiment, to find out why we DIDN'T reject the null. Beta is usually set at the beginning of an experiment. Where should it be set, and why do we set it? |
|
Definition
True Beta should be less than or equal to 0.2. We set beta, b/c we want a certain power (at least 80% or 0.80). Notice that this isn't as strict as alpha. Remember, that we are more willing to be wrong on a type II than a type I error. |
|
|
Term
What are 3 reasons for a type II error? Which is the most common? |
|
Definition
Better response rate in the comparator grp Increased variance Too few subjects (most common) |
|
|
Term
If type I & type II error are inversely related, how can we reduce the chance of a type I error w/o increasing the type II error? |
|
Definition
|
|
Term
What is power? How do you calculate power? |
|
Definition
It is the percent chance that if a difference b/w grps exists that your study will detect it statistically. power = 1 - beta |
|
|
Term
Name the 3 factors that affect power |
|
Definition
Size of treatment effect, alpha, and variability. |
|
|
Term
The size of treatment is aka |
|
Definition
|
|
Term
Power is affected by the "effect size". If there is a large effect, the difference is easy/difficult to detect and power is high/low. |
|
Definition
|
|
Term
Power is affected by the effect size. If there is a small effect, the difference will be easy/difficult to detect, type II error will be higher/lower, and power will be higher/lower. |
|
Definition
|
|
Term
A larger variance will result in a lower/higher power |
|
Definition
|
|
Term
Why is it that an increased sample size increases power? |
|
Definition
Because it increases df, and it controls variance (ie, SEM goes down, b/c SEM = s / sq root of n) |
|
|
Term
Why should power (and beta) be calculated a priori ? |
|
Definition
To determine the sample size needed for a study. |
|
|
Term
When does power not have to be calculated? |
|
Definition
When there is a significant difference. |
|
|
Term
T or F: Statistical significance does not equal clinical significance |
|
Definition
|
|
Term
When do you have to do post-hoc analysis? |
|
Definition
If the null is rejected, you have to do post-hoc analysis. Otoh, if the null accepted, the analysis is complete, b/c the variances are equal. |
|
|
Term
What is linear regression? |
|
Definition
To test for associations b/w variables. If independent values are given, we can use linear regression to predict the dependent variable. |
|
|
Term
Why is the simple regression equation different from the sample regression equation? |
|
Definition
B/c simple regression deals w/ population, which means error is included in the equation. Error is not included in the sample linear regression equation. |
|
|
Term
Least squares regression is often used to calculate linear regression. What is the least squares regression line used for? |
|
Definition
it sums the squared distance of the actual values to make them as close to what is predicted as possible. |
|
|
Term
If you see the least squares line equation, y = 560 + 0.14X, what does it mean? |
|
Definition
A 0.14 change in x will bring about a 1 unit change in Y. |
|
|
Term
|
Definition
It's the difference between the actual y-value and the estimated y-value. |
|
|
Term
What is a residual analysis used for? |
|
Definition
To test the assumptions of linear regression. It basically diagnoses problems w/ a model. |
|
|
Term
If you got a curve shaped residual scatter plot, would this be good or bad? Why? |
|
Definition
This would be bad. It would reveal that our data is not linear. We would have to transform the data. |
|
|
Term
What are the 2 problems w/ the funnel shaped residual scatter plot? |
|
Definition
1) The funnel reveals a problem w/ independence. 2) There is a constancy of variance. We have heteroscedasticity when we want homoscedasticity. |
|
|
Term
Why is the coefficient of determination important? The closer r2 is to , the better the regression model. |
|
Definition
It reveals how much variation is explained by the regression line. |
|
|
Term
What is the difference b/w the coefficient of determination and the correlation coefficient? |
|
Definition
The coefficient of determination is the variance accounted for by a whole model. It includes the y-variables, whereas the correlation coefficient doesn't. The correlation coefficient lets you know if 2 variables are associated. The correlation coefficient b/w dependent and independent variables will generally be less than the coefficient of determination, b/c the model, which has y-variables will account for more variance. |
|
|
Term
What is multiple linear regression and what makes it different than correlation? |
|
Definition
It's used when you want to fit multiple independent variables to one dependent variable. Fe, the dependent variable is a students fall semester GPA. The independent variables could be GPA, ACT, ethnicity, gender, age, etc. It's different than a correlation, b/c it has predictive abilities. Fe, it can predict a dependent variable from one or more independent variables. |
|
|
Term
Statistical significance is best evaluated w/ . Clinical significance is best evaluated w/ . |
|
Definition
correlation coefficient (r) coefficient of determination (r2) |
|
|