Term
|
Definition
Ways of categorizing. Dividing the data into categories. E.g. zip code. You don’t’ do anything with it, don’t manipulate it. Nonparametric. |
|
|
Term
|
Definition
Puts the data into rank order, but doesn’t tell you anything about the distance between data points. Nonparametric. |
|
|
Term
|
Definition
There is equal distance between the data points. E.g.: Fahrenheit temperature. There is no meaningful zero. You can do mathematical calculations, there are relative positions (e.g. if you add a constant to everyone the distribution stays the same). Parametric. Fits into a normal curve. |
|
|
Term
|
Definition
Equal distance between the data points and absolute zero. There is absence of a characteristic (e.g. zero inches of rainfall in a year). You can do addition, subtraction and multiplication and division. Parametric. Fits into a normal curve. |
|
|
Term
|
Definition
All measures of central tendency are in the same place. Curves are normal if they are symmetrical. Equal variability and distribution around the mean |
|
|
Term
Positively skewed distribution |
|
Definition
There is an outlier who has a much higher score than the rest which brings the mean much higher. It then makes the mean not a good measure of central tendency.
[image] |
|
|
Term
Negatively Skewed Distribution |
|
Definition
There is an outlier who has a much lower score than the rest and that brings the mean down. It then makes the mean not a good measure of central tendency |
|
|
Term
|
Definition
The middle score. With a skewed distribution, this is the best measure to look at for central tendency. (e.g. income) |
|
|
Term
|
Definition
How scores vary in a uniform way above and below the mean
· 68 % of the population is between one standard deviation above and below the mean
· 95 % of the population is between two standard deviations above and below the mean |
|
|
Term
|
Definition
The central limit theorem is based on a distribution of sample
means. We will take multiple samples of a population which will
vary around the mean. Eventually you will get a distribution of
sample means. As sample size increases, this distribution will
become a normal curve. |
|
|
Term
|
Definition
There is no difference between the treatment conditions (e.g. caffeine vs. no caffeine on test results) or accepted fact at the time (e.g. the world is flat). |
|
|
Term
|
Definition
Sometimes called the research hypothesis, experimental hypothesis): States that there is a difference between the treatment conditions. |
|
|
Term
|
Definition
I use this test if there is a difference but I am not specifying the direction (e.g. saying they will perform differently, but not better or worse). Non-directional hypothesis |
|
|
Term
|
Definition
You use this test when you are saying there is a difference and you are specifying the direction of the test- e.g. one group will perform better than the other. |
|
|
Term
|
Definition
Chance of detecting a real difference if it exists. |
|
|
Term
|
Definition
The probability of making a Type I error. This is your p value. You get to decide which significance you want to use: e.g. p < .05 = the probability of making a Type I error is less than 5 %. So you have a 95 % chance we are making the right decision. |
|
|
Term
|
Definition
When I reject the null hypothesis but there really is no difference (if I did then measure the whole population). You reject it when it’s really true. I say there’s a difference when there is not. Involves alpha. Reject when true is not type II. |
|
|
Term
|
Definition
I failed to reject the null when I should have. I say there is no difference between conditions, but when I measure everyone in the population I did find a difference. You cannot control for Type II error because you can never tell how many people from the population you will need to sample to find a difference. |
|
|
Term
Homogeneity of Variance Assumption |
|
Definition
When you are doing a statistical test, you have to be
comparing groups who have similar variability or else it
won’t work (e.g. you can’t compare a leptokurtic vs.
platykurtic). |
|
|
Term
Independence of observers assumption |
|
Definition
If I have two groups, there is not a factor that affects one group but not the other (e.g. hot room vs. cold room). There are similar conditions in both groups |
|
|
Term
|
Definition
The difference between two groups or two distributions
of scores. It is the appropriate test for testing difference
in means of two populations |
|
|
Term
|
Definition
This test is appropriate for testing differences in means for several groups to determine if they come from different populations, i.e. whether population means are equal or not. Just tells you whether there is a difference between any two, but doesn't tell you which. Have to do a post-hoc test |
|
|
Term
|
Definition
Use a MANOVA to control for Type I error. It is used
with more than one DV (e.g. several scores for each
subject). Every additional analysis done with an
ANOVA increases error (e.g. 3 ANOVAs at 0.05
= .15 error). MANOVA combines all three into one
analysis and reduces the error. |
|
|
Term
|
Definition
This test is used when a study involves multiple (two or more) independent variables. There may be a main effect and/or an interaction effect. |
|
|
Term
|
Definition
This test is used to see if there a significant difference in classifications according to certain categories (e.g. eye color). The Chi Square tells us: Is what we observe significantly different from what we expected? |
|
|
Term
Standard Error of the Estimate |
|
Definition
Your ability to make a prediction based on the correlation coefficient. The stronger the correlation between variables, the less error I will make. A perfect correlation (1.0) than the standard error of estimate = zero. The stronger relationship between the variables (correlation), the better predictive power, then less error you make. The smaller the standard error of estimate the closer the lines are to the regression line. |
|
|
Term
Correlation of determination |
|
Definition
This is the square of the correlation and it tells you how good the correlation is. (R2). So if correlation is .91 then .91 x .91 = .83 or 83 % shared variability or relationship between variables. |
|
|
Term
Structural Equation Modeling |
|
Definition
A special use of correlation coefficients and you develop a model ahead of time and based on correlations throughout the model you are implying a cause and effect relationships: Two classes- Unidirectional (Path Analysis) and Bidirectional (LISREL) Linear Structural Relation Analyses) |
|
|
Term
|
Definition
A type of structural equation modeling. It's a unidirectional Technique (going from point A to point B). Based on pairs of variables along the path we are implying there is a cause and effect relationship between variables. |
|
|
Term
|
Definition
A type of structural equation modeling that's bidirectional. The relationship goes in both directions between the variables. It allows you to go back and forth between the variables and allows you to identify latent variables. These are items that were not originally in the model, but you are able to identify them as a product of doing the analysis. |
|
|
Term
|
Definition
A technique based on a long term understanding of research in a given area. Given this knowledge, you make a trend prediction on difference between specific groups, predict the magnitude of these differences, and look at interactions and you will see nonlinear outcomes (e.g. “cubic” and “quadratic”). |
|
|
Term
|
Definition
A treatment design with one baseline, one treatment. |
|
|
Term
|
Definition
Two baselines, two treatment phases. Also called a reversal design, or withdrawal design. You take baseline data, then implement the treatment then withdrawal the treatment to see if they go back to their baseline data to see if they revert to their original behavior. Helps to show that your treatment is what really caused the change. Can be unethical |
|
|
Term
|
Definition
People will change their behavior because they know they are being studied. |
|
|
Term
|
Definition
A type of construct validity where you compare your measure to a well established measure and see how well they correlate. |
|
|
Term
|
Definition
This is used with divergent validity and you want a low coeeficient. Used when you compare your test with another construct- A low score shows a lack of convergent validity. |
|
|
Term
|
Definition
This is used with convergent validity, and you want a high coefficient. Compare your test with one of the same construct. You want a high score because that shows a lack of divergent validity. |
|
|
Term
|
Definition
This is the sum of the factor loadings in a factor analysis output. H2. It tells you about a specific predictor and it's relationship to each factor. |
|
|
Term
|
Definition
In factor analysis- This is when you rotate the factors and they remain uncorrelated. The communalities (H2), stay the same, but the factor loadings change. Axis stay at 90 degrees. |
|
|
Term
|
Definition
In factor analysis, you rotate the factors and they still have some correlation. Axis are not at 90 degrees (scissor effect). |
|
|
Term
|
Definition
A type of criterion related validity that is indicative of whether someone will develop a condition in the future. |
|
|
Term
|
Definition
A type of criterion-related validity where you are trying to establish whether someone has a certain disorder currently. |
|
|
Term
Criterion Related Validity |
|
Definition
This is the relationship between the criterion and the predictor. Can be predictive (future development), concurrent (detect current trains) or incremental (as in hiring decisions.) |
|
|
Term
|
Definition
How well are we measuring the construct of interest. Includes convergent and divergent validity. |
|
|
Term
|
Definition
How well does the test measure the subject matter itself. Primarily seen in academics. The test measures what you have learned, and it based on expert judgement. |
|
|
Term
|
Definition
This is a measure of inter-rater reliability. |
|
|
Term
Standard Error of Measurement |
|
Definition
This is based on a test's reliability coefficient and the obtained score. It is used to develop a confidence band where the person's true score falls. More reliable = smaller SEM. |
|
|
Term
Kuder Richardson Formula (KR20) |
|
Definition
This is a special version of Croenbach's alpha where all of the items are dichotemously scored (e.g. true/ false). It's a reliability formula. |
|
|
Term
Spearman Brown Prophecy Formula |
|
Definition
This formula is used to calculate split-half relability. It compares the two halves of the test against each other and then estimates reliability as if you had two versions of the original test. |
|
|
Term
|
Definition
This is when you compare the measuer against itself for reliability. |
|
|
Term
|
Definition
This is the result you get from a test-retest reliability. It is the reliability of scores over time. |
|
|
Term
|
Definition
A measure of how much a person's obtained score is made up of the true score.
0.90 = 90 % of performance is due to true ability, and 10 % is due to other things. |
|
|
Term
|
Definition
This is used to show the relationship between the difficulty of the item and the person's performance on the test overall. |
|
|
Term
|
Definition
The p value measures the difficulty of the items. P = the percentage of people who pass. For a maximum distribution, p = 50. So 50 people pass, and 50 people fail. For more serious conditions, you want a higher p level (e.g. detection of schizophrenia) |
|
|
Term
|
Definition
These tests have grade item difficulty. There is no time limit and the score is based on how many you can attain at a given level (e.g. Vocabulary). |
|
|
Term
|
Definition
These tests have to do with how many items you can complete in a given time period (e.g. Symbol Search). |
|
|
Term
|
Definition
Thi is experimenter expectancy. The experimenter will have the tendency to see things in their favor. You do a double blind study to correct for this. |
|
|
Term
|
Definition
Clues in the experimental setting that allow the participants to figure out the research hypothesis. As a result, the subjects change their behavior. |
|
|
Term
|
Definition
This is how well the resutls of your study and the sample generalize back to the population. |
|
|
Term
|
Definition
This is the purity of your experimental environment. It's the quality of your study. History and maturation effects are threats to internal validity. |
|
|
Term
|
Definition
A type of longitudinal design where you take people of different ages and study them over a period of time- e.g. 10 years. Eliminates cohort effects. |
|
|
Term
|
Definition
A type of longitudinal design where you take people of different ages and study them all at once. The problem is intergenerational (cohort) effects. |
|
|