+ All Categories
Home > Documents > More on Maxent

More on Maxent

Date post: 06-Jan-2016
Category:
Upload: easter
View: 56 times
Download: 3 times
Share this document with a friend
Description:
More on Maxent. Env . Variable importance: Ma xent attempts to determine variable importance a couple ways: During modeling, Maxent asks how much “gain” occurs - PowerPoint PPT Presentation
Popular Tags:
33
More on Maxent Env. Variable importance: Ma xent attempts to determine variable importance a couple ways: During modeling, Maxent asks how much “gain” occurs During iterations, Maxent increases gain of the model by modifying coefficient for a single feature based on input environmental data Gain is related to information added by an environ. variable
Transcript
Page 1: More on  Maxent

More on MaxentEnv. Variable importance:

Ma xent attempts to determine variable importance a couple ways:

• During modeling, Maxent asks how much “gain” occurs

• During iterations, Maxent increases gain of the model by modifying coefficient for a single feature based on input environmental data

• Gain is related to information added by an environ. variable

• Maxent can then determine percent contribution of the env. variable to whole the model

Page 2: More on  Maxent

More on MaxentEnv. Variable importance caveats

• Percent contribution outputs don’t take into account covariances across environmental layers

• Variale importance is based on Maxent algorithm! It might be different using a different method!

• Interpret with caution

• Permutation importance a new measure.

Page 3: More on  Maxent

Env. Variable importance:

• Maxent attempts to determine variable importance a couple ways:

• jackknifing is another method to determine environmental layer importance

• It is a leave-one-layer out without replacement procedure

• How much “gain” occurs if we use individual layers or combinations?

• Which layer contributes the most gain when used individually?

Page 4: More on  Maxent

Env. Variable importance:

• comparing test and train variable importance is useful

• Note that precip6190_ann is the best predictor for test but not train when used alone

• Suggests that transferability of monthly values lower than annual values

• Note, length of red bar indicates gain using all variables

• If blue bar is shorter than red, corresponding loss of gain (explanation) when variable omitted.

Page 5: More on  Maxent

How to read these graphs

• Each graph shows the range of values for the pixels in the environment layers you used on the x-axis

• Probability of presence is shown on the y-axis (0 to 1)

• eg., tmax6190_annshows that for low tmaxs, probability of occurrence is ~1, which drops towards 0 around 22.5 C.

• Plots do not take into account correlation among variables

• Maxent produces a second graph with individual variables run separately

Page 6: More on  Maxent

New Maxent Goodies

The Explain Tool

• New in Maxent 3.3

• Shows for any point where the value is on all response curves

• Can be used to see how env. Variables matter in different areas.

• Haven’t had a chance to use this much

Page 7: More on  Maxent

How do we know when models are “realistic” or, better put, of “reasonable quality”, or even

better put… “valid”?

Page 8: More on  Maxent

Two Types of Error in Distributional Predictions

Actual geographic distribution

Page 9: More on  Maxent

Two Types of Error in Distributional Predictions

Predicted geographic distribution

Page 10: More on  Maxent

Two Types of Error in Distributional Predictions

Actual geographic distribution

Predicted geographic distribution

Page 11: More on  Maxent

Two Types of Error in Distributional Predictions

Actual geographic distribution

Predicted geographic distribution

Overprediction,or Commission

Underprediction,or Omission

Page 12: More on  Maxent

Two Types of Error in Distributional Predictions

Objective: To Minimize Both Forms of Error

Page 13: More on  Maxent

Two Types of Error in Distributional Predictions

Objective: To Minimize Both Forms of Error

Page 14: More on  Maxent

There are at least two strategies to do so:- Collect New Data- Split your data into two sets, one set used in training, one set used in model testing.

To evaluate model quality you need to:1. Generate an ‘independent’ set of data

Page 15: More on  Maxent

2. Generate a model with the training data

Page 16: More on  Maxent

3. Quantify error components with a confusion matrix by utilizing the testing data

dcPredicted Absent

baPredicted Present

Actually Absent

Actually Present

a & d = correct predictions b = commission error

(false positives, overprediction)

c = omission error (false negatives,

underprediction)

Page 17: More on  Maxent

Measuring Omission Redux

• Omission error is “easy” to measure if you have a test and training datset

• Training dataset to create model

• Test dataset to verify if suitable habitats include the “pixels” that also contain the test locations.

• If yes, then omission error is low.

Page 18: More on  Maxent

Measuring Commission Error Redux

• Measuring commission error is much trickier• We don’t know anything about true absences, because we

only collected presence data• How to measure commission error?• In Maxent, the commission error is measured in reference to

the “background” (all pixels)• And we are therefore distinguishing presence from “random”

rather than presence from absence

Page 19: More on  Maxent

High OmissionLowCommission

Zero OmissionHigh Commission

We know the model getsthis wrong. How? Explainin terms of omission, commission and in terms of true/false presence/absence

This one probably getssomething wrong too.Also explain….

Zero OmissionNo Commission Overfitting

And this one too…

EXAMPLE MODELS THAT YOU MIGHT NOT BELIEVE

Page 20: More on  Maxent

Some stochastic algorithms (e.g. GARP) produce different models with the same input data. Good models find minimization between commission/omission error. So can find those models.

Commission Index (% of area predicted present)

Om

issi

on E

rror

(% o

f oc

curr

ence

poi

nts

outs

ide

the

pred

icte

d ar

ea)

0 100

100

For species with a fair number of

occurrence data this is a typical curve

Page 21: More on  Maxent

Commission Index (% of area predicted present)

Om

issi

on E

rror

(% o

f occ

urre

nce

poin

ts o

utsi

de th

e pr

edic

ted

area

)

0 100

100

Distribution of a species in an area

High OmissionLow Commission

Zero OmissionHigh CommissionZero Omission

No Commission Overfitting

Page 22: More on  Maxent

The question now is, which of these models are good and which ones are bad?

Commission Index (% of area predicted present)

Om

issi

on E

rror

(% o

f occ

urre

nce

poin

ts o

utsi

de th

e pr

edic

ted

area

)

0 100

100Models with high omission error are

clearly bad(not capturing

environment of known occurrences)

Page 23: More on  Maxent

The question now is, which of these models are good and which ones are bad?

Commission Index (% of area predicted present)

Om

issi

on E

rror

(% o

f occ

urre

nce

poin

ts o

utsi

de th

e pr

edic

ted

area

)

0 100

100

overprediction

Region of the best models

overfitting Median

Page 24: More on  Maxent

The following discussion made a big assumption: That models results are binary – either suitable or unsuitable.

HOWEVER…

Page 25: More on  Maxent

SOME TOOLS PRODUCE CONTINUOUS MEASURESOF SUITABLE on a scale from 0 (unsuitable) to 100 (really suitable)

Like Maxent… the tools we’ll use.

Page 26: More on  Maxent

So how to Threshold ?(eg convert a continuous map into a binary one)

• Lots of potential thresholds to choose. Some of the most common are:– Fixed (eg. all maxent values 10-100 are suitable, all 0-10

are not. Note … arbitrary)

– Lowest presence threshold (threshold that requires the lowest possible suitable area that still includes all training occurrence data points)

– Sensitivity-specificity equality (where true positive and true negative fraction are equal)

Page 27: More on  Maxent

HOW TO READ SOME OF THE MAXENT OUTPUTS PART 2 – RECEIVER OPERATOR

CURVES.

WHAT THEY ARE, AND WHAT THEY MEAN.

AS BEST AS I CAN REMEMBER (and explain)

Page 28: More on  Maxent

 

actual value

  p n total

prediction outcome

p'True

Positive

FalsePositive

(commission)P'

n'False

Negative(omission)

TrueNegative

N'

total P N

- We can caculate the true positive rate (TPR) as TP/P - We can calculate the false positive rate (FPR) as FP/N- We can calculate accuracy (ACC) = (TP + TN) / (P + N)

Remember, we have test data to use to find errors of commission and omission

to see how the model performs

Page 29: More on  Maxent

A B C C'

TPR = 0.63 TPR = 0.77 TPR = 0.24 TPR = 0.88

FPR = 0.28 FPR = 0.77 FPR = 0.88 FPR = 0.24

ACC = 0.68 ACC = 0.50 ACC = 0.18 ACC = 0.82

TP=63 FP=28 91

FN=37 TN=72 109

100 100 200

TP=77 FP=77 154

FN=23 TN=23 46

100 100 200

TP=24 FP=88 112

FN=76 TN=12 88

100 100 200

TP=88 FP=24 112

FN=12 TN=76 88

100 100 200

TPR is also called sensitivity1-FPR is also called specificity

Page 30: More on  Maxent

WE CAN PLOT THE TPR against 1-FPR:This is called a Receiver-Operator Characteristic Curve

- Comes from information theory

Page 31: More on  Maxent

ACC VALUES ARE GREAT WHEN YOU HAVE THRESHOLDED YOUR MAXENT RESULT. WHAT ABOUT VALUES OVER MULTIPLE THRESHOLDS?

- You calculate your ACC value at all thresholds- low thresholds overpredict (high commission errors)- high thresholds underpredict (high omission errors)

- you get a curve of under to overprediction – this is the ROC curve- the area under the curve is a good indicator of model performance

- You want AUC values close to 1

Page 32: More on  Maxent

- Another graph showing how omission increases with increasing threshold. - Background (kind of pseudoabsences) go the other direction since you are predicting more absence as you increase threshold- You are looking at specificity versus sensitivity across thresholds

Page 33: More on  Maxent

HOW TO READ SOME OF THE MAXENT OUTPUTS PART 2 – RECEIVER OPERATOR

CURVES.

WHAT THEY ARE, AND WHAT THEY MEAN.

People often report AUC values as measures of model performance and you can too. With caveats:

1. AUCs vary dependening whether a species is widespread or narrowly distributed

2. The choice of the “geographic window of extent” for modeling matters

3. AUCs are typically inflated due to spatial autocorrelation

4. So interpret “good” model results with caution.


Recommended