Term
Reliability vs. validity:
|
|
Definition
Reliability refers to the confidence we can place on the measuring instrument to give us the same numeric value when the measurement is repeated on the same object. Validity on the other hand means that our measuring instrument actually measures the property it is supposed to measure. Reliability of an instrument does not warranty its validity."
|
|
|
Term
|
Definition
|
|
Term
Basic vs. applied science: |
|
Definition
Basic Science is concerned with the process of discovery. Basic scientists seek to discover new knowledge and information without the primary concern of how the knowledge they create might be used.
Applied Science takes information that already exists and utilizes it for the solution of an existing problem. |
|
|
Term
Four assumptions of science: |
|
Definition
1.True, physical universe exists 2.Universe is primarily orderly 3.The principles that define the functioning of the universe can be discovered
4.All ideas are tentative, potentially changed by new information |
|
|
Term
|
Definition
|
|
Term
Frequency, duration, and intervals methods: |
|
Definition
|
|
Term
Naturalistic observation: |
|
Definition
it is a method where the research very carefully observes and records some behavior or phenomenon, sometimes over a prolonged period, in its natural setting. |
|
|
Term
Experimental vs. non-experimental research |
|
Definition
|
|
Term
|
Definition
research conducted in a natural setting in which subjects typically do not know they are in an experiment |
|
|
Term
|
Definition
A base rate fallacy is committed when a person judges that an outcome will occur without considering prior knowledge of the probability that it will occur. |
|
|
Term
Random assignment to conditions (or groups) |
|
Definition
|
|
Term
The tyranny of the senses |
|
Definition
|
|
Term
Bryson and Hamblin’s (1988) study (MUM effect) |
|
Definition
|
|
Term
|
Definition
Correlational studies are used to look for relationships between variables. There are three possible results of a correlational study: a positive correlation, a negative correlation, and no correlation. The correlation coefficient is a measure of correlation strength and can range from –1.00 to +1.00. |
|
|
Term
Advantages and disadvantages of correlational studies |
|
Definition
Advantage 1. An advantage of the correlation method is that we can make predictions about things when we know about correlations. If two variables are correlated, we can predict one based on the other. For example, we know that SAT scores and college achievement are positively correlated. So when college admission officials want to predict who is likely to succeed at their schools, they will choose students with high SAT scores.
Disadvantage
1. The problem that most students have with the correlation method is remembering that correlation does not measure cause. Take a minute and chant to yourself: Correlation is not Causation! Correlation is not Causation! I always have my in-class students chant this, yet some still forget this very crucial principle. |
|
|
Term
Advantages and disadvantages of experiments |
|
Definition
|
|
Term
Independent variables vs. dependent variables |
|
Definition
|
|
Term
Experimenter-manipulated vs. experimenter-selected variables |
|
Definition
|
|
Term
Spallanzani’s (1785) experiment (dog insemination) |
|
Definition
|
|
Term
Lorge’s (1930) experiment (massed vs. distributed practice) |
|
Definition
|
|
Term
Asch’s (1952) experiment (presentation order) |
|
Definition
|
|
Term
Benbow and Stanley’s (1980) study (math gender difference) |
|
Definition
|
|
Term
Experimental vs. control groups |
|
Definition
|
|
Term
Supa, Cotzin, and Dallenbach’s (1944) experiment (facial vision) |
|
Definition
|
|
Term
Paul’s (1966) experiment (speech phobia) |
|
Definition
|
|
Term
|
Definition
bias introduced by an experimenter whose expectations about the outcome of the experiment can be subtly communicated to the participants in the experiment. |
|
|
Term
Single-blind vs. double-blind experiments |
|
Definition
In single-blind studies, only the patient is not told what is being given. In a double-blind study, only the pharmacist knows; the doctors, nurses, patients, and other health-care staff are not informed. If medically necessary, however, it is always possible to find out what the patient is taking. Document Actions |
|
|
Term
Rosenthal and Fode’s (1963) experiment (maze-bright rats) |
|
Definition
|
|
Term
|
Definition
Sometimes during an experiment, a participant might pick up on some clue or bias from the researcher, the situation, or something about the experiment that gives the participant and idea of what type of response the researcher is looking for. |
|
|
Term
|
Definition
|
|
Term
|
Definition
A statement of the procedures or ways in which a researcher is going to measure behaviors or qualities. |
|
|
Term
Factorial vs. non-factorial designs |
|
Definition
|
|
Term
Between-subjects, within-subjects, and “mixed” factorial designs |
|
Definition
|
|
Term
Levels of an independent variable |
|
Definition
|
|
Term
Possible outcomes of a 2 x 2 factorial design |
|
Definition
|
|
Term
Main effects and interactions |
|
Definition
|
|
Term
Bar graphs vs. line graphs vs. table of means |
|
Definition
|
|
Term
Chi’s (1978) experiment (chess) |
|
Definition
|
|
Term
Ex post facto experiment (i.e., a quasi-experiment) |
|
Definition
|
|
Term
|
Definition
some characteristics of people that can be measured or described but cannot be varied experimentally |
|
|
Term
|
Definition
scientific instrument used by psychologists; presents visual stimuli for brief exposures. |
|
|
Term
|
Definition
an analysis of the antecedents and consequences of s particular behavior; usually undertaken before the implementation of a behavioral treatment. |
|
|
Term
|
Definition
use experimental control by the careful manipulation of an independent variable on a small number of participant |
|
|
Term
|
Definition
ABA designs are a type of single-subject design. The purpose of this type of experimental design is to observe changes individual may exhibit as a result of the treatment. |
|
|
Term
Ayllon’s (1963) experiment (towel hoarding) |
|
Definition
|
|
Term
Identification of independent and dependent variables in abstracts |
|
Definition
|
|
Term
Identification of research designs in abstracts |
|
Definition
|
|
Term
|
Definition
Example 1: What is the Mean of these numbers?
6, 11, 7
Add the numbers: 6 + 11 + 7 = 24 Divide by how many numbers (there are 3 numbers): 24 / 3 = 8
The Mean is 8 |
|
|
Term
Determine median (precisely using graph) |
|
Definition
a measure of central tendency: the middle value when observations are ordered from least to most. |
|
|
Term
|
Definition
Example:
3, 7, 5, 13, 20, 23, 39, 23, 40, 23, 14, 12, 56, 23, 29
In order these numbers are:
3, 5, 7, 12, 13, 14, 20, 23, 23, 23, 23, 29, 39, 40, 56
This makes it easy to see which numbers appear most often.
In this case the mode is 23. |
|
|
Term
|
Definition
Ordering the test scores from least to greatest, we get: 73, 77, 84, 87, 89, 91, 94 highest - lowest = 94 - 73 = 21 Answer: The range of these test scores is 21 points. |
|
|
Term
Calculate variance and standard deviation |
|
Definition
Standard Deviation
The Standard Deviation is a measure of how spread out numbers are.
Its symbol is σ (the greek letter sigma)
The formula is easy: it is the square root of the Variance. So now you ask, "What is the Variance?"
Variance
The Variance is defined as:
The average of the squared differences from the Mean. |
|
|
Term
|
Definition
n statistics, the concept of the shape of the distribution refers to the shape of a probability distribution and it most often arises in questions of finding an appropriate distribution to use to model the statistical properties of a population, given a sample from that population. The shape of a distribution may be considered either descriptively, using terms such as "J-shaped", or numerically, using quantitative measures such as skewness and kurtosis. |
|
|
Term
|
Definition
In statistics, a z-score (or standard score) is used to compare means from different normally distributed sets of data. The actual score indicates how many standard deviations an observation is above or below the mean. How to Calculate a z-score in Statistics thumbnail Formula for Standard Score (z-score)
In statistics, a z-score (or standard score) is used to compare means from different |
|
|
Term
Mean and SD for a set of z scores |
|
Definition
z = \frac {X-\mu}{\sigma} (6)
X = ExperimentalValue
μ = Mean
σ = StandardDeviation |
|
|
Term
Convert a Z score into a raw score? |
|
Definition
|
|
Term
Draw a scatter diagram (scatterplot) |
|
Definition
|
|
Term
Positive vs. negative correlations |
|
Definition
|
|
Term
Strong vs. weak correlations |
|
Definition
|
|
Term
Linear vs. curvilinear relationships |
|
Definition
A linear relationship is different from a nonlinear/curvilinear relationship because a linear relationship has direct proportionality that causes the dependent variable to change when the independent variable changes, whereas a nonlinear/curvilinear relationship does not have proportionality between the dependent and independent variables (there is not consistent change). In other words, in a linear relationship all the points on the scatter diagram tends to lie near a straight line, while the points in a nonlinear/curvilinear relationship are depicted graphically by anything other than a straight line. |
|
|
Term
Possible values of a correlation coefficient |
|
Definition
|
|
Term
Computing the correlation coefficient |
|
Definition
|
|
Term
Interpreting correlations |
|
Definition
|
|
Term
Use regression equation to predict scores |
|
Definition
|
|
Term
|
Definition
|
|
Term
|
Definition
Inter-rater: Evaluates reliability among different participants
2. Test- Re-test: Evaluates reliability across time
3. Parallel Forms: Evaluates question that seek to assess the same construct with the out most consistency.
4. Internal Consistency: Evaluates individual questions in comparison with one another for their ability to give consistently appropriate results. |
|
|
Term
|
Definition
Face Validity: The content of the measure appears to reflect the construct being measured.
Content Validity: The content of the measured is linked to the universe of content that defines the construct.
Predictive Validity: Scores on the measure predict behavior on a criterion measured at a time in the future.
Concurrent Validity: Scores on the measure are related to a criterion measured at the same time (concurrently).
Convergent Validity: Scores on the measure are related to other measures of the same construct.
Discriminant Validity: Scores on the measure are not related to other measures that are theoretically different. |
|
|
Term
|
Definition
|
|
Term
|
Definition
|
|
Term
|
Definition
As we all know by now, psychologists like to control things -- in particular, we like to establish as much control as possible when conducting experiments. |
|
|
Term
|
Definition
A variable which is the common cause o f two things that may falsely appear to be in a causal relationship |
|
|
Term
Subject variable-subject variable confound |
|
Definition
|
|
Term
Five basic experimental paradigms |
|
Definition
what is to be observed and scrutinized
the kind of questions that are supposed to be asked and probed for
answers in relation to this subject how these questions are to be structured how the results of scientific investigations should be interpreted how is an experiment to be conducted, and what equipment is available to conduct the experiment. |
|
|
Term
|
Definition
Spontaneous healing, also called spontaneous remission or spontaneous regression, means an unexpected improvement or cure from a disease |
|
|
Term
|
Definition
The beneficial effect in a patient following a particular treatment that arises from the patient's expectations concerning the treatment rather than from the treatment itself. |
|
|
Term
Simon and Emmons (1956) experiment (sleep learning) |
|
Definition
|
|
Term
Experiment on perceptual defense |
|
Definition
|
|
Term
|
Definition
As its name suggests, the theory of one trial learning states that learning takes place in a single pairing of a response and stimulus and is not strengthened over time by repeated exposure to a stimulus. |
|
|
Term
Experiment briefs in Ch. 7 |
|
Definition
|
|
Term
|
Definition
a counterbalancing technique in which the treatment orders are randomized in block such all of the condition are present within the block |
|
|
Term
Random blocks technique (for matching) |
|
Definition
|
|
Term
|
Definition
refers to matching subjects |
|
|
Term
Potential problems associated with repeated measure design |
|
Definition
|
|
Term
Fatigue vs. practice effects |
|
Definition
1. Practice Effects – Occur when subjects get better at the task over time because of practice, so that they perform best in the later conditions.
2. Fatigue Effects – Occur when subjects get worse at the task over time because of fatigue. They might even quit trying and just “go through the motions |
|
|
Term
|
Definition
This is when having been tested under one condition affects how participants behave in another condition. There are many different kinds of carryover effects. Here are a few of the most important. |
|
|
Term
|
Definition
1. Practice Effects – Occur when subjects get better at the task over time because of practice, so that they perform best in the later conditions.
2. Fatigue Effects – Occur when subjects get worse at the task over time because of fatigue. They might even quit trying and just “go through the motions |
|
|
Term
|
Definition
|
|
Term
|
Definition
when the participant does not stay during the experiment. |
|
|
Term
Ethics case studies in Ch. 8 |
|
Definition
basically it is about how as a psychologist we must conduct ourselves as professionals |
|
|
Term
|
Definition
|
|
Term
Anonymity vs. confidentiality |
|
Definition
Confidentiality
Maintaining confidentiality of information collected from research participants means that only the investigator(s) or individuals of the research team can identify the responses of individual subjects; however, the researchers must make every effort to prevent anyone outside of the project from connecting individual subjects with their responses. Anonymity
Providing anonymity of information collected from research participants means that either the project does not collect identifying information of individual subjects (e.g., name, address, Email address, etc.), or the project cannot link individual responses with participants’ identities. A study should not collect identifying information of research participants unless it is essential to the study protocol. |
|
|
Term
|
Definition
it is important to treat animals right a during a experiment if you want your experiment to be valid |
|
|
Term
|
Definition
|
|
Term
|
Definition
is a phrase often used in law to indicate that the consent a person gives meets certain minimum standards. As a literal matter, in the absence of fraud, it is redundant. An informed consent can be said to have been given based upon a clear appreciation and understanding of the facts, implications, and future consequences of an action. |
|
|
Term
|
Definition
When a study or experiment ends, researchers are required to "debrief" participants. In a "debriefing" a researcher explains the purpose of the study, explains the use of deception (if any was used), encourages the participant to ask questions about the study, and allows the researcher to address any harm to the participant that may have resulted from their participation in the study. |
|
|
Term
|
Definition
is a committee that has been formally designated to approve, monitor, and review biomedical and behavioral research involving humans. They often conduct some form of risk-benefit analysis in an attempt to determine whether or not research should be done. |
|
|
Term
Major sections of an APA style research report |
|
Definition
|
|
Term
Citing sources according to APA style |
|
Definition
|
|
Term
34/14/2 approximation rule |
|
Definition
|
|
Term
|
Definition
|
|