| Term 
 
        | If a null hypothesis is true, but you reject it, you have made what type of error? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | If a null hypothesis is not true, and you do not reject it, then you have made what type of error? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | As alpha goes up, beta goes         ? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | As beta goes up, power goes       ? |  | Definition 
 
        | down; Remember that power is 1 - beta |  | 
        |  | 
        
        | Term 
 
        | As alpha goes up, power goes        ? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Choose: anova F-tests are always/sometimes one sided/two sided for determining a p-value.   |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | What are the 5 assumptions for calculating ANOVA? |  | Definition 
 
        | Random samples-are required to be unbiased  Indepedent samples If they're dependent (or repeated), you have to use repeated measures ANOVA. Independent observations w/in groups  Normal distribution - (aka normal population) Equal variances (homogeneity) |  | 
        |  | 
        
        | Term 
 
        | When calculating ANOVA, if you have unequal variances, you have to                  the data.  Name three ways that you would do this to your data. |  | Definition 
 
        | transform; log, natural log, square root |  | 
        |  | 
        
        | Term 
 
        |                   is the sum of squares (SS)  What is SSB? What are other names for SSB? What is SSW? What are other names for SSW?   Choose: The signal/noise should be > than the signal/noise, otherwise, we won't be able to detect the signal/noise. |  | Definition 
 
        | Variance; It is the variance b/w groups (it is the difference of each mean from the overall mean); SS(treatment) or SS(effects) or signal; It is the variance w/in groups; SS(error) or noise.   signal-noise-signal |  | 
        |  | 
        
        | Term 
 | Definition 
 
        | It is the mean square b/w groups |  | 
        |  | 
        
        | Term 
 
        | SSB = 43.21 SSW = 50.68 T or F: the error is greater than the treatment Why or why not? |  | Definition 
 
        | False, b/c you still have to divide by the df. The SSB will have a much smaller df (eg, 2) than the SSW (eg, 39). After dividing the SSB and SSW by their respective df's, you get MSB = 21.61 and  MSW = 1.30.  |  | 
        |  | 
        
        | Term 
 
        | If a comparison is not planned ahead of time, you must use                even if only one post-hoc comparison was performed. However, you usually perform all comparisons. |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | If comparisons are selected a priori, what does c =? If comparisons are selected post hoc, what  does c =? |  | Definition 
 
        | a priori: c = number of comparisons to be made post hoc: c = total number of possible comparisons |  | 
        |  | 
        
        | Term 
 
        | What is the point of the Bonferroni method?     |  | Definition 
 
        | It's like an adjusted alpha, labeled alpha'. It increases the CL for each CI in order to ensure a specific overall CL.  Table 11.4 shows how conservative the bonferroni method is. |  | 
        |  | 
        
        | Term 
 
        | When should you use bonferroni? When should you use tukey? |  | Definition 
 
        | The bonferroni method should be used when comparisons are selected a priori. The Tukey method should be used when ALL comparisons are selected - it doesn't matter if they're selected a priori or post hoc. If comparisons are selected post hoc, use the Tukey method. |  | 
        |  | 
        
        | Term 
 
        | The same three factors that affect statistical outcomes affect ANOVA results. What are they? |  | Definition 
 
        | Treatment effects (differences b/w means) Variance (SSB compared to SSW) # of subjects |  | 
        |  | 
        
        | Term 
 
        | What are the 4 assumptions of the regression model? |  | Definition 
 
        | NICL  Normal distribution Independence (y observations are independent of each other) Constancy of variance (the sd of y is constant for each x) Linear relationship (as opposed to exponential, curvilinear, etc) |  | 
        |  | 
        
        | Term 
 
        | What is residual analysis used to test? |  | Definition 
 
        | The assumptions of linear regression. |  | 
        |  | 
        
        | Term 
 
        | Correlation analysis uses pearson, r. What are the 3 assumptions of pearson? |  | Definition 
 
        | 1) Both the ind & dep variables must b continuous. 2) Dependent variable must be rational 3)No multicollinearity (meaning ind variables cant be related to other ind variables) |  | 
        |  | 
        
        | Term 
 
        | What are the assumptions of B? |  | Definition 
 
        | 1) Each y observation is independent 2) All potential y values are normally distributed for each x value 3) The std deviation of y is constant for each x 4) The sampling distribution of b is normal |  | 
        |  | 
        
        | Term 
 
        | What are the 4 assumptions for r? |  | Definition 
 
        | 1) Each y observation is independent 2) The sampling distribution of both x and y are normal. 3) The std deviation of y is constant for each x 4) The sampling distribution of r is normal. |  | 
        |  | 
        
        | Term 
 
        | If all population means are              , the sample means will be very close to each other and close to the                 . The weighted square differences will be               , and a                    F will be calculated.     |  | Definition 
 
        | equal; grand mean; small; small |  | 
        |  | 
        
        | Term 
 
        | ANOVA T or F: Variance cannot be negative |  | Definition 
 
        | True. If it is, there is too much error. |  | 
        |  | 
        
        | Term 
 
        | If some population means are quite different, corresponding sample means will be                than each other and the               . Corresponding weighted squares will be                 , the                and                will be small, and the F will be               . |  | Definition 
 
        | quite different; grand mean; large; SSB; MSB; large |  | 
        |  | 
        
        | Term 
 
        | As alpha increases, Fcrit               . |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | When is one-way ANOVA used? |  | Definition 
 
        | When you need to compare more than two means. More specifically, when you have one factor w/ more than two levels of the factor (ie, Drug A, Drug B, and DrugC). |  | 
        |  | 
        
        | Term 
 | Definition 
 
        | Error slippage occurs when you try to do a bunch of t-tests when you really should have done ANOVA.  You'll think you have significant differences, but you really don't, b/c the alpha is slipping from 0.5 to 1. |  | 
        |  | 
        
        | Term 
 | Definition 
 
        | It means group 3, observation 2. |  | 
        |  | 
        
        | Term 
 
        | If you had two or more independent variables, what type of ANOVA would you use? |  | Definition 
 
        | Factorial ANOVA. The different factors may be Drugs A, B, & C (1st factor), Age (second factor), & gender (third factor). |  | 
        |  | 
        
        | Term 
 
        | When would you use one-way ANOVA? Name a requirement of ANOVA. |  | Definition 
 
        | Use one-way ANOVA when you have one factor and more than two levels. The samples in the different levels must be independent of each other. |  | 
        |  | 
        
        | Term 
 
        | When do you use post-hoc analysis? |  | Definition 
 
        | When the null has been rejected and you're trying to find out which mean(s) are different. |  | 
        |  | 
        
        | Term 
 
        | When do you use repeated measures ANOVA? |  | Definition 
 
        | If measures are dependent (or repeated), you use a crossover design called repeated measures ANOVA. |  | 
        |  | 
        
        | Term 
 
        | What is the proper term used when your putting variances into groups?   What is another name for variance? |  | Definition 
 
        | Partitioning     Sums of Squares |  | 
        |  | 
        
        | Term 
 
        | SSB & SSW are weighted by the                 . |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Choose: When p-Value is greater/lesser than or equal to alpha, H0 cannot be rejected. |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Which is worse: a type I or type II error? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | A type I error is aka                A type II error is aka                |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Define alpha     What does it mean when we set out alpha to 0.05? |  | Definition 
 
        | It is the probability that a difference does not exist when the study shows that a difference does exist.    If a study was repeated 100 times, we would be willing to be wrong 5 times. We would conclude 5 times that a difference exists when in fact it does not. |  | 
        |  | 
        
        | Term 
 
        | T or F: A rejection region will be larger for a one-tail alpha than a two-tail alpha |  | Definition 
 | 
        |  | 
        
        | Term 
 | Definition 
 
        | It is the percent chance that a difference b/w experimental grps does exist even though the study concludes no difference. |  | 
        |  | 
        
        | Term 
 
        | T or F: We usually look at beta after an experiment, to find out why we DIDN'T reject the null.   Beta is usually set at the beginning of an experiment. Where should it be set, and why do we set it?   |  | Definition 
 
        | True     Beta should be less than or equal to 0.2.  We set beta, b/c we want a certain power (at least 80% or 0.80). Notice that this isn't as strict as alpha. Remember, that we are more willing to be wrong on a type II than a type I error. |  | 
        |  | 
        
        | Term 
 
        | What are 3 reasons for a type II error?   Which is the most common? |  | Definition 
 
        | Better response rate in the comparator grp  Increased variance Too few subjects (most common) |  | 
        |  | 
        
        | Term 
 
        | If type I & type II error are inversely related, how can we reduce the chance of a type I error w/o increasing the type II error? |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | What is power?      How do you calculate power? |  | Definition 
 
        | It is the percent chance that if a difference b/w grps exists that your study will detect it statistically.     power = 1 - beta |  | 
        |  | 
        
        | Term 
 
        | Name the 3 factors that affect power |  | Definition 
 
        | Size of treatment effect, alpha, and variability. |  | 
        |  | 
        
        | Term 
 
        | The size of treatment is aka |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Power is affected by the "effect size". If there is a large effect, the difference is easy/difficult  to detect and power is high/low.   |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Power is affected by the effect size. If there is a small effect, the difference will be easy/difficult to detect, type II error will be higher/lower, and power will be higher/lower. |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | A larger variance will result in a lower/higher power |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | Why is it that an increased sample size increases power? |  | Definition 
 
        | Because it increases df, and it controls variance (ie, SEM goes down, b/c SEM = s / sq root of n) |  | 
        |  | 
        
        | Term 
 
        | Why should power (and beta) be calculated a priori ? |  | Definition 
 
        | To determine the sample size needed for a study. |  | 
        |  | 
        
        | Term 
 
        | When does power not have to be calculated? |  | Definition 
 
        | When there is a significant difference. |  | 
        |  | 
        
        | Term 
 
        | T or F: Statistical significance does not equal clinical significance |  | Definition 
 | 
        |  | 
        
        | Term 
 
        | When do you have to do post-hoc analysis? |  | Definition 
 
        | If the null is rejected, you have to do post-hoc analysis. Otoh, if the null accepted, the analysis is complete, b/c the variances are equal. |  | 
        |  | 
        
        | Term 
 
        | What is linear regression? |  | Definition 
 
        | To test for associations b/w variables. If independent values are given, we can use linear regression to predict the dependent variable. |  | 
        |  | 
        
        | Term 
 
        | Why is the simple regression equation different from the sample regression equation? |  | Definition 
 
        | B/c simple regression deals w/ population, which means error is included in the equation. Error is not included in the sample linear regression equation. |  | 
        |  | 
        
        | Term 
 
        | Least squares regression is often used to calculate linear regression.  What is the least squares regression line used for? |  | Definition 
 
        | it sums the squared distance of the actual values to make them as close to what is predicted as possible.  |  | 
        |  | 
        
        | Term 
 
        | If you see the least squares line equation,  y = 560 + 0.14X, what does it mean? |  | Definition 
 
        | A 0.14 change in x will bring about a 1 unit change in Y. |  | 
        |  | 
        
        | Term 
 | Definition 
 
        | It's the difference between the actual y-value and the estimated y-value. |  | 
        |  | 
        
        | Term 
 
        | What is a residual analysis used for? |  | Definition 
 
        | To test the assumptions of linear regression. It basically diagnoses problems w/ a model. |  | 
        |  | 
        
        | Term 
 
        | If you got a curve shaped residual scatter plot, would this be good or bad? Why? |  | Definition 
 
        | This would be bad. It would reveal that our data is not linear. We would have to transform the data. |  | 
        |  | 
        
        | Term 
 
        | What are the 2 problems w/ the funnel shaped residual scatter plot? |  | Definition 
 
        | 1) The funnel reveals a problem w/ independence. 2) There is a constancy of variance. We have heteroscedasticity when we want homoscedasticity. |  | 
        |  | 
        
        | Term 
 
        | Why is the coefficient of determination important?     The closer r2 is to           , the better the regression model. |  | Definition 
 
        | It reveals how much variation is explained by the regression line. |  | 
        |  | 
        
        | Term 
 
        | What is the difference b/w the coefficient of determination and the correlation coefficient? |  | Definition 
 
        | The coefficient of determination is the variance accounted for by a whole model. It includes the y-variables, whereas the correlation coefficient doesn't. The correlation coefficient lets you know if 2 variables are associated.  The correlation coefficient b/w dependent and independent variables will generally be less than the coefficient of determination, b/c the model, which has y-variables will account for more variance. |  | 
        |  | 
        
        | Term 
 
        | What is multiple linear regression and what makes it different than correlation? |  | Definition 
 
        | It's used when you want to fit multiple independent variables to one dependent variable.  Fe, the dependent variable is a students fall semester GPA. The independent variables could be GPA, ACT, ethnicity, gender, age, etc.  It's different than a correlation, b/c it has predictive abilities. Fe, it can predict a dependent variable from one or more independent variables. |  | 
        |  | 
        
        | Term 
 
        | Statistical significance is best evaluated w/                  .   Clinical significance is best evaluated w/                  . |  | Definition 
 
        | correlation coefficient (r)     coefficient of determination (r2) |  | 
        |  |