Term
| Which models (of the 4 we have covered in class)allow testing for Goodness of Fit in MARK? Why? |
|
Definition
Cormack Jolly Seber Model Dead Recovery Models
These models do not assume that the user knows with 100% certainty the fate of the marked individual. You test GOF to see if model assumptions have been violated by the data. You chat will tell you if there is over or underdispersion. |
|
|
Term
| Which models (of the 4 we have covered in class) do not allow for GOF testing? Why? |
|
Definition
Known Fate Nest Survival
Both of these models are known fate models, meaning the user is not uncertain about tag return or detectability when marked (although these assumptions are often violated). The GOF is considered unnecessary under these conditions. |
|
|
Term
|
Definition
| "The numeric expression of the expectation that an event will occur." The sum of all related outcomes to a single event is 1. |
|
|
Term
|
Definition
| "A product of probabilities that takes into account and is conditional on the distribution of the data." |
|
|
Term
| How are probability and likelihood related? |
|
Definition
| Probabilities are used to find the likelihood of an event taking place. Probability is absolute and does not take into account past events, where likelihood is conditional (related to past events). |
|
|
Term
| Define Induction and give an example. |
|
Definition
Induction is the use of past events to predict future, broad scale events. Ex: I like this apple pie, therefore all apple pies are good. |
|
|
Term
| Define Deduction an give an example. |
|
Definition
Deduction is using a general law or principle to make specific predictions (ie, null hypothesis testing). Ex: All apple pies are good, I will like this apple pie |
|
|
Term
| Define Retroduction and give an example. |
|
Definition
AKA Abductive reasoning. Making an assumption that event a led to event b without considering all possible alternatives. Similar to induction and often mistaken for it.
Ex: The lawn is wet. It must have rained last night. |
|
|
Term
| Compare and contrast induction, deduction, and retroduction. |
|
Definition
| Deduction is inferences made from hypothesis testing. It starts with the big picture to make assumptions about a very specific situation. Induction and retroduction are similar in that something learned from a specific situation is applied to a larger scale. However, induction is often referred to as correlative association and retroduction refers to post-analysis explanations. |
|
|
Term
| What are the assumptions of capture mark recapture? |
|
Definition
- no births/deaths, imm/emigration between trapping -Marks are permanent -Tags have minimal impact on indiv -Homogeneity of survival and capture prob -Enough time has passed that animals distribute through the landscape, but sort enough that the first assumption is not violated |
|
|
Term
| What is an important biological issue with CJS? |
|
Definition
| Only apparent survival is estimated. You can't really differentiate between animals that died and those that emigrated. |
|
|
Term
| What assumptions must be met to calculate probability and/or likelihood? |
|
Definition
| Independent and identically distributed data |
|
|
Term
| Why are binomial or log distributions used most often with likelihood? |
|
Definition
| Log-likelihoods are the basis for profile likelihood functions – unbiased estimators |
|
|
Term
|
Definition
Homogeneity of capture and survival probability for marked indiv w/in ea sampling period
Sampling periods are instantaneous (in reality they are very short periods) and recaptured animals are released immediately.
All emigration from the sampled area is permanent. |
|
|
Term
| List dead recovery assumptions |
|
Definition
1. Marked indiv are representative of the pop 2. Tags not lost 3. Tags are ID'd correctly and reported by hunters 4. Indiv fates are independent 5. Probability of survival and tag recovery homogenous |
|
|
Term
| List known fate assupmtions |
|
Definition
1. No tags lost 2. No radio tag failure (hahahahahaha) |
|
|
Term
| List nest survival assumptions |
|
Definition
-Nests are correctly aged when they are found -Nest fates are correctly determined -Nest discovery and check do not influence nest survival -Nest fates are independent (not always the case w social animals) -Homogeneity of daily nest survival rates (as nests age, more likely to be depredated so this assumptions is bs) |
|
|
Term
| What is P(data|null) and what is wrong with this? |
|
Definition
-The definition of r^2 -Also, the probability of the data given the null -It's fundamentally flawed, values are arbitrarily chosen for rejection thresholds - You can't compare models! |
|
|
Term
|
Definition
| The probability of the model given the data |
|
|
Term
|
Definition
| The top rated AICc/lower ranking AICc |
|
|
Term
| What is another term for multimodel inferencing? |
|
Definition
|
|
Term
| What is unconditional variance? |
|
Definition
-"Variance given model selection uncertainty" - also "a measure of repeatability based on model uncertainty" -Standard error for Beta -Accounts for model selection uncertainty -Used when AICc<.9 |
|
|
Term
| What is conditional variance? |
|
Definition
-variance given the top model (AICc >.9) -Does not incorporate model selection uncertainty - |
|
|
Term
| what d you do if your Chat is 2.6? |
|
Definition
Adjustments tabs -> chat -> enter 2.6 -> run QAICc will adjust accordingly |
|
|
Term
| What 3 parameters do you always report when estimating survival? |
|
Definition
| survival, st error, confidence interval |
|
|
Term
| How do you calculate AIC from RSS in an ANOVA table? |
|
Definition
| -2log(L)+2K = AIC L = -1/2n*log(sigma^2) sigma^2 = RSS/n |
|
|
Term
| What are the 3 most common approaches for general parameter estimation? |
|
Definition
| LS (least squares), ML (max likelihood), and Bayesian methods |
|
|
Term
| What is the domain of LS? |
|
Definition
| General linear models (regression, ANOVA) |
|
|
Term
|
Definition
| Maximum likelihood estimate - the value of the parameter that is most likely, given the data and model |
|
|
Term
| What are the advantages of using MLE? |
|
Definition
| unbiased, minimum variance, and normally distributed |
|
|
Term
| Name 4 ways you can check for fit of a model to the data |
|
Definition
| GOF, adj R^2, residual analyses, and checking for overdispersion in count data |
|
|
Term
| What is underfitting a model? |
|
Definition
| Not enough parameters - Some model structure is erroneously included in the residuals, leading to high bias and the illusion of high precision. |
|
|
Term
|
Definition
| Too many parameters - high level of uncertainty. Some residual variation treated as structural variation. |
|
|
Term
| What is a loose translation of Occam's razor? |
|
Definition
| To shave away all that is not needed |
|
|
Term
| What is a negative consequence of overfitting? |
|
Definition
Spurious effects! Imprecision! Madness! CHAOS!!!!!!!!
I've been studying too long. |
|
|
Term
| Describe Kullback-Leibler Information |
|
Definition
| I(f,g) - Info lost in the distance between full reality (f) and the model (g). We can only estimate K-L information, because we will ever know full reality (far out). WE are striving to minimize inefficiency - the quantitative measure from f to g. |
|
|
Term
| How are statistical principles link with information theory (specifically K-L information)? |
|
Definition
| Akaike used a property of logarithms to rewrite K-L theory and use that to derive AIC. He got rid of the true zero inherent to K-L info. This means we can rate models for best fit without having to know full reality (true zero). |
|
|
Term
| What three concepts did Akaike link? |
|
Definition
Boltzmann's entropy, K-L info, and maximum likelihood AKA thermodynamics, info theory, and statistics |
|
|
Term
| What is the the information criterion eqn used by Akaike to link thermodynamics, info theory, and statistics? |
|
Definition
AIC = -2log(L(Ohat)|data) + 2K
(L(Ohat)|data) = likelihood of the parameter, given the data |
|
|
Term
| Why multiply by -2 in the AIC eqn? |
|
Definition
| deviance penalized by 2K to correct for asymptotic bias |
|
|
Term
|
Definition
| Problem solving the lazy way (not based on rejection of all possible hypotheses). Examples of this method include using a rule of thumb, an educated guess, an intuitive judgment, stereotyping, or common sense. |
|
|
Term
| What is AICc and when do you use it? |
|
Definition
AIC for small sample sizes; it's AIC with an additional bias correction term to deal with too many paramters in relation to sample size.
AICc= AIC+(2K(K+1))/(n-K-1) |
|
|
Term
| Why shouldn't R^2 be used for formal model selection? When can it be use for model selection? |
|
Definition
| R^2 is a descriptive statistic and exaggerates the predictive ability of models fit to a given data set. In exploratory work, you may use R^2 to assess the worth of models (ranks mean nothing if all of the models suck) and direct future data collection. |
|
|
Term
| What is the crucial initial starting point for advancement in the life sciences? |
|
Definition
| A set of multiple working hypotheses defined a priori. Then using these you make a priori set of candidate models representing the hypotheses. |
|
|
Term
| Is the value of AICc important? |
|
Definition
| No, the differences between AICcs are more important - they can be related to information |
|
|
Term
| What is exp(-1/2 deltai)? |
|
Definition
| The L(Ohat|data, g) aka the likelihood of the parameter given the data and the model |
|
|
Term
|
Definition
| exp(-1/2 deltai)/(summed exp(-1/2 deltar)) |
|
|
Term
|
Definition
| The Prob(gi|data) - probabilty of the model given the data |
|
|
Term
|
Definition
Evidence Ratio L(gi|x)/L(gj|x) = wi/wj The likelihood of one event relative to another given the data |
|
|
Term
| What are the 3 main types of evidence? |
|
Definition
1) Model probabilities (the prob that model i is the K-L best model) wi and Bayesian posterior model probabilities 2) L(model|data) 3) Eij - provides empirical evidence of hypothesis i vs hypothesis j |
|
|
Term
| Name the pitfalls of information-theoretic methods |
|
Definition
1) Poor science question 2) Too many models 3) True model not in the set 4) Information-theoretic methods treated as a test 5) Poor modeling of hypotheses 6) Failure to consider aspects of model selection uncertainty 7) Failure to consider overdispersion in the data 8) Don't use post hoc data mining 9) GOF should be used with global model |
|
|
Term
|
Definition
| The relative likelihood of the model given the data |
|
|
Term
|
Definition
| a method of fitting a model to data by minimizing the squared differences between observed and predicted values |
|
|
Term
| Define maximum likelihood |
|
Definition
| A method of fitting a model to data by maximizing an explicit likelihood function. This function specifies the likelihood of unk parameters of the model given the model form and data. The parameter values are termed MLE (maximum likelihood estimators) |
|
|
Term
|
Definition
Naive occupancy = estimated number of sites occupied/ total sites
DOES NOT INLCUDE DETECTION PROB!!!! |
|
|
Term
|
Definition
| CJS time-dependent survival and recapture probs |
|
|
Term
| What does dead recovery diagram look like? |
|
Definition
|
|
Term
| What do K-L and delta AIC diamgrams look like? |
|
Definition
|
|
Term
| What does a known fate diagram look like? |
|
Definition
|
|
Term
| What does a mark resight diagram look like? |
|
Definition
|
|
Term
|
Definition
| Extent of occurrence and area of occupancy |
|
|
Term
| What kinds of data types are use for occupancy analysis? |
|
Definition
Museum/collection records Anecdotal reports (Ex: ebird) Biological surveys |
|
|
Term
| What is a major issue that must be accounted for in occupancy modeling? |
|
Definition
Probability of detection (false absences or present, but not detected) |
|
|
Term
| What is the difference between occupancy and use? |
|
Definition
| Occupancy is binary, used is not (used can mean sometimes). Occupied means ALWAYS physically present. |
|
|
Term
| Name the assumptions of occupancy modeling |
|
Definition
1) closed populations 2) P(occupation) constant 3) Detection among sites independent (unless modeled differently) |
|
|
Term
| What 2 parameters does PRESENCE calculate? |
|
Definition
Psi - proportion of sites occupied p - detection prob for a survey (dependent of occupancy model type)
A measure of variance is given for ea parameter |
|
|
Term
|
Definition
"present at a site and detected in at least one of k samples"
d= 1- (1-p)^k d = probability of detecting the species p = probability of detecting the species in a single survey k = # of surveys |
|
|
Term
| Why do you average models? |
|
Definition
| To account for model selection uncertainty. Even the lower-ranked models have valuable information |
|
|
Term
| What is Ybarhat (model averaged value for a parameter)? |
|
Definition
| the sum of wi*Yhat for all models |
|
|
Term
| The variance of beta is said to be _______ on the model. |
|
Definition
conditional
Given this model, one can obtain a numerical value for var(Beta) using LS or ML theory. |
|
|
Term
| Variance of a parameter is the sum of what? |
|
Definition
| Sampling variance given a model and variation due to model selection uncertainty |
|
|
Term
| What is an unconditional variance actually conditional on? |
|
Definition
| Unconditional variance is condition on the set of models considered (a weaker assumption than using just one model) |
|
|
Term
|
Definition
UNCONDITIONAL VARIANCE
What's 20 feet long and has 5 teeth?
The line for funnel cake at the TX state fair. |
|
|
Term
| CJS estimates 3 things. What are they? |
|
Definition
| Abundance, survival, and capture probability |
|
|
Term
| What is the main problem with CJS assumptions? |
|
Definition
| Under normal circumstances, these assumptions are violated |
|
|
Term
| Brownie model and Seber model are for what purposes? |
|
Definition
| Harvest recovery and found dead recovery, respectively |
|
|
Term
| What is the benefit of using information theory over null hypothesis testing? |
|
Definition
Ho testing finds p(data|null)
You can't compare models; models are not penalized for having more parameters
Information theory finds p(model|data)
K-L gives you the ability to test multiple hypotheses, compare evidence across models, rank models, and average them |
|
|
Term
| What is the importance of building a strong foundation for your science house? |
|
Definition
| You'll never know full reality (remember I(f,g)?), but you can collect great data and design the study well. |
|
|
Term
|
Definition
The value of your B parameter that is most liekly to occur, given your data. MLE is the link between null hyp and information theory.
[image] |
|
|
Term
| When can you use AIC and not AICc? |
|
Definition
|
|
Term
| What exactly does the (2K(K+1))/(n-K-1) do to AIC? |
|
Definition
Applies this idea:
[image] |
|
|