Term
What are three things a experiment must be to be considered valid? |
|
Definition
- can be answered from seeing something
- can be repeated.
- is collaborative
|
|
|
Term
What are three biases in the thinking of Researchers? |
|
Definition
1.Confirmation Bias: we look for evidence that supports our view and ignore all else
2.Hindsight Bias: we can explain away anything
3.Availability Bias: We notice things that are easy to notice and not things that are hard to notice
-
- Example: when you buy are car and then you notice that car everywhere
|
|
|
Term
|
Definition
editor will send out to people in the field then send you info. Then published….hopefully
|
|
|
Term
What are the Three types of Research studies? |
|
Definition
Description: who what where when
Perdiction:(correlations) are they relate? With my knows can I find the unknown? What relates to what?
Explanation: (experioment) what causes? |
|
|
Term
What is a Quasi-experiment? |
|
Definition
missing one component of a true study
|
|
|
Term
Two Key things to have when collecting valuable data |
|
Definition
Reliability = Repeatability
Validity= are you really measuring/manipulating what you are trying to measure |
|
|
Term
What are two tyes of Validity? |
|
Definition
Internal validity: how well can you narrow down your causes // how many explanations can you come up with
External Validity: how well your findings generalize outside of your study (people, places, times, etc.)
|
|
|
Term
What 2 types of External Validity are there? |
|
Definition
High external validity= applies to all sorts of people and things Low external validity= only happens in rare circumstances |
|
|
Term
Inter-observer reliability (inter-rater)
|
|
Definition
a way to measure your reliability training someone else to look for them
|
|
|
Term
|
Definition
Procedure whereby a concept is defined solely in terms of the observable procedures used to produce and measure it.
|
|
|
Term
What are two Major Problems a Researcher should be aware of when conducting Research? |
|
Definition
Informed Consent:
- One of most important part
- Need to know what's going on what they'll be physically doing
Coercion vs. Incentive:
- The amount offered, most of the time you have to offer something but it needs to match the task
- "Indecent Proposal" the movie (a million dollars for one night)
- Institutional Review Board (IRB) to make sure that your research is done right. Everything goes through them.
|
|
|
Term
What are six threats to Validity? |
|
Definition
- Expectancy Effects (experimenter bias)
- Blind observers protect against expectancy
- Reactivity
- People change their behavior when they know their being watched
- Natural Observations (watch them in their environment) protects against reactivity
Response Bias
- Yea-saying
- Nay-Saying
-
- Memory Failure
- Inadequate Self-Knowledge
- How the items are delivered/phrased
|
|
|
Term
Three Causes of Reactivity? |
|
Definition
Demand Characteristics
- Scientist tell (hopefully un-intentionally) what your hypothesis is and the subjects will of course naturally react to give you the results you want
- Ex. A study on depression, new technique to control depression, one with old therapy, so they will do what you tell them they are supposed to suspect.
Social Desirability:
- How much you want to please society
Malingering:
- Someone is trying to make themselves look bad
|
|
|
Term
|
Definition
You respond to a positive attitude with a positive attitude. The same applies to negative. |
|
|
Term
Two types of unobtrusive Research? |
|
Definition
- Physical Traces
- Like the two examples before
- Archival Research
- Ex. Being someone seeing what color of jersey looses the most
- Content Analysis how do I tell what this bundle of info is really saying
|
|
|
Term
|
Definition
are you try you measure what you say you are going to measure
How do we see these
Physiological, nonverbal, verbal reports
Empirical = Observation |
|
|
Term
Construct (concept) Validity |
|
Definition
How well your getting at the abstract idea you are trying to find/prove // how well you are capturing the idea's behind your variables
- Ex; you do poorly on a test, but it doesn't really show how much you know.
|
|
|
Term
Two types of Internal Validity |
|
Definition
- High internal validity = only one cause
- Low internal validity= many ways to explain outcome
|
|
|