Processing math: 100%
+ - 0:00:00
Notes for current slide
Notes for next slide

What makes a good multispecies model?

Goal 4 Keynote
Multispecies Modeling Applications Workshop, June 2021

Sarah Gaichas
Northeast Fisheries Science Center

1 / 24

What makes a good model? (NRC, 2007), p 108

Is the model based on generally accepted science and computational methods?

Does it work, that is, does it fulfill its designated task or serve its intended purpose?

Does its behavior approximate that observed in the system being modeled?

Fisheries models range from single and multispecies models to full ecosystem models

2 / 24

Model life cycle

(NRC, 2007)

What makes a good model?

  • Differs by life stage

  • Each builds on the next

  • Common themes

3 / 24

Best practices 1: Why do we need a model? Identify the problem

whybuild

4 / 24

Best practices 1: Why do we need a model? Identify the problem

whybuild

whyindeed

4 / 24

Problem identification ↔ Appropriateness of model for problem

Define the problem, and why this model is appropriate.

  • What processes need to be modeled to address the problem?

  • What focal species, spatial, and temporal resolution are needed to address the problem?

Example problem (ICES, 2019; ICES, 2021):

Multiple predators feed on multiple fished prey stocks within a particular ecosystem. There is a need to include predation mortality within assessments for individual prey stocks in that ecosystem.

Model(s) must estimate predation mortality at age, and to provide M-at-age time series for herring, cod, whiting, haddock, sprat, sandeel, etc. Spatial scale is at the stock level and temporal resolution is annual, starting at a stock-specific year and going to the present.

Model(s) must provide this output and sensitivity to model assumptions in the output M at age must be evaluated.

5 / 24

Problem identification ↔ Appropriateness of model for problem

A different example problem (ICES WKIRISH, Bentley, et al. (2021)):

The aim with the Irish Sea Ecopath is to use the model to “fine tune” the quota advice within the predefined EU Fmsy ranges. In “good” conditions you could fish at the top of the range, in “poor” conditions you should fish lower in the range. The range has already been evaluated as giving good yield while still being precautionary, so this should be fine for ICES to use in advice, so any reviewers should have this in mind.

For the Irish Sea EwE model, key outputs will be used to determine where the catch advice should be within the MSY range for each species. Therefore, outputs defining ecosystem conditions and both ecosystem and species productivity under the prevailing conditions are most important.

6 / 24

Best Practices 2: Is the scientific basis of the model sound?

Previously used models often skip this step, reasoning that if the science was sound previously it still is.

(Discuss...)

ICES WGSAM has provided model framework reviews for the LeMans ensemble (2016), FLBEIA (2017), and a multispecies state-space model (2017). “Constructed Model” evaluation issues and best practices apply here:

  1. Spatial and temporal resolution
  2. Algorithm choices
  3. Assumptions (scientific basis, computational infrastructure; adequacy of conceptual model)
  4. Data availability/software tools
  5. Quality assurance/quality control (code testing)
  6. Test scenarios (for example, stability; Kaplan, et al. (2016)):
    1. All functional groups persist in an unfished unperturbed run
    2. Model stabilizes for the last ~20 years of an unfished, unperturbed 80-100 year run
  7. Corroboration with observations
  8. Uncertainty/sensitivity analysis
  9. Peer review

Note linkages between problem identification and scientific basis

7 / 24

Best Practices 3: Is the input data quality and parameterization sufficient for the problem?

See above defining the problem. Which datasets are adequate, which could be improved, and which are missing?

8 / 24

Best Practices 3: Is the input data quality and parameterization sufficient for the problem?

See above defining the problem. Which datasets are adequate, which could be improved, and which are missing?

Show the input data as a simple chart: beginning and end of time series, gaps, different length of time series, spatial resolution of data.

Baltic Gadget sprat data inputs from ICES WGSAM 2019 Baltic Gadget cod data inputs from ICES WGSAM 2019

8 / 24

Best Practices 3: Is the input data quality and parameterization sufficient for the problem?

Give information on input data pedigree/quality, reference for where it comes from, whether it is survey data or comes from other model output, whether confidence intervals or other uncertainty measures are available and used in the model.

9 / 24

Best Practices 3: Is the input data quality and parameterization sufficient for the problem?

Give information on input data pedigree/quality, reference for where it comes from, whether it is survey data or comes from other model output, whether confidence intervals or other uncertainty measures are available and used in the model.

Categorize the assumptions behind modeled ecological or biological processes. Emphasize those related to species interactions (predation, competition), environmental pressures, and also fleet dynamics if needed to address the problem. If the model is spatial, how do these processes happen in space?

Baltic Gadget conceptual model from ICES WGSAM 2019

Baltic Gadget timing conceptual model from ICES WGSAM 2019

Is the parameterization consistent with scientific knowledge (e.g. (PREBAL) diagnostics (Link, 2010) for general relationships across trophic levels, sizes, etc).

9 / 24

Best Practices 4: Does model output compare well with observations? (Kaplan, et al., 2016)

Characterize the reference dataset used for comparisons. Has the data been used to construct this model? Is the reference dataset from another model? Describe referece data source(s).

  1. Define the hindcast time period where agreement with other data sources or assessments is needed. Error ranges of model output and reference datasets are needed for comparison.
  2. Focal species should match biomass and catch trends over the hindcast time period. For full system models, species comprising a majority of biomass should also match general hindcast trends. Suggested tests include modeling efficiency, RMSE, etc. (Sterman, 1984; Stow, et al., 2009; Allen, et al., 2009; Lehuta, et al., 2013; Lehuta, et al., 2016; Olsen, et al., 2016)
  3. Patterns of temporal variability should be captured as appropriate to the problem (emergent or forced with e.g. recruitment time series)
  4. Productivity for focal species (or groups totaling ~80% of system biomass in full system models) should qualitatively match life history expectations (prebal diagnostics)
  5. Natural mortality decreases with age for majority of groups
  6. Age and length structure qualitatively matches expectations for majority of groups
  7. Diet predicted qualitatively matches empirical diet comp for majority of groups
  8. Spatial distribution of outputs match reference datasets for spatial models (most important if output required at spatial resolution of model, comment if a match in aggregate but not at higher resolution)
  9. Ecosystem indicators (relationship between abundance and body size, pelagic to demersal, Large Fish Indicator) match reference data if needed for problem
10 / 24

Intermission: Why include species interactions? ...and why fit criteria alone are not sufficient

Ignore predation at your peril: results from multispecies state-space modeling (Trijoulet, et al., 2020)

Ignoring trophic interactions that occur in marine ecosystems induces bias in stock assessment outputs and results in low model predictive ability with subsequently biased reference points.

VanessaPaper

EM1: multispecies state space

EM2: multispecies, no process error

EM3: single sp. state space, constant M

EM4: single sp. state space, age-varying M

note difference in scale of bias for single species!

modcomp

11 / 24

This is an important paper both because it demonstrates the importance of addressing strong species interactions, and it shows that measures of fit do not indicate good model predictive performance. Ignoring process error caused bias, but much smaller than ignoring species interactions. See also Vanessa's earlier paper evaluating diet data interactions with multispecies models

Best Practices 5: Uncertainty

Has uncertainty been assessed in the output of interest? Has sensitivity analysis been performed and how does it affect those outputs?

12 / 24

Best Practices 5: Uncertainty

Has uncertainty been assessed in the output of interest? Has sensitivity analysis been performed and how does it affect those outputs?

Show estimates of uncertainty in the output quantity of interest. Uncertainty analysis estimating confidence intervals is ideal. Otherwise list key sources of uncertainty and expected bounds on outputs based on those (possibly from sensitivity analysis)--i.e. design sensitivity analysis to approximate uncertainty analysis.

12 / 24

Best Practices 5: Uncertainty

Has uncertainty been assessed in the output of interest? Has sensitivity analysis been performed and how does it affect those outputs?

Show estimates of uncertainty in the output quantity of interest. Uncertainty analysis estimating confidence intervals is ideal. Otherwise list key sources of uncertainty and expected bounds on outputs based on those (possibly from sensitivity analysis)--i.e. design sensitivity analysis to approximate uncertainty analysis.

Specific analyses:

  1. Retrospective analysis (5 year peel of all input data, example)
  2. Forecast uncertainty: remove last 3-5 years of survey index only to see how well the model works in forecast mode, given the catch that actually happened.
  3. Sensitivity to stomach data and other key or low-confidence data sources
  4. Sensitivity to key parameters: consumption rates, residual mortality (M1, M0)
  5. Sensitivity to initial conditions

Simpler investigation of uncertainty can be appropriate for complex models with long runtimes (Kaplan, et al., 2016).

Retain multiple parameterizations that meet the above criteria to allow scenario testing across a range of parameterizations. A simple method uses bounding (e.g. base, low bound, and high bound productivity scenarios; (Saltelli, et al., 2010).

12 / 24

Best Practices 6: Peer review

What did they point out and have issues been addressed?

Review of constructed models should have evaluated spatial and temporal resolution, algorithm choices, data availability and software tools, quality assurance/quality control of code, and test scenarios.

Bentley, et al. (2021) conclude that continuous peer review throughout a process is itself a best practice.

Also, being part of a management process improves likelihood that the model will address a management problem of interest, and that the model will be used; see also Townsend, et al. (2019)

How about a real life problem?

13 / 24

Are any Atlantic herring harvest control rules good for both fisheries and predators?

Harvest control rules are:

  • plans for changing fishing based on stock status
  • pre-determined

"Which harvest control rules best consider herring's role as forage?"

  • DESIGN a harvest control rule (HCR):
    • balancing fishing benefits and ecological services
    • addressing diverse stakeholder interests
  • TRANSPARENTLY within management time frame!
14 / 24

Problem identification: New England Herring MSE

What is the goal of the model?

  • Evaluate effects of herring harvest strategies
    • on herring
    • on herring fisheries
    • on herring predators

Who will use it?

  • Modelers, in support of New England Fishery Management Council

What types of decsions will it support?

  • Decision on harvest strategy → strategic advice

What data are available to support the model?

  • Herring assessment inputs
  • Groundfish diet data
  • Groundfish assessment inputs
  • Seabird diet/population data
15 / 24

Appropriateness of model for problem

How do changes in herring populations affect predators?

Availble models:

Multispecies catch at age (Curti, et al., 2013)

Multispecies catch at length (Gaichas, et al., 2017)

Multispecies production (Gamble, et al., 2009)

16 / 24

Appropriateness of model for problem

How do changes in herring populations affect predators?

Availble models:

Multispecies catch at age (Curti, et al., 2013)

Multispecies catch at length (Gaichas, et al., 2017)

Multispecies production (Gamble, et al., 2009)

all model predation mortality but not prey effects on predators!

16 / 24

Appropriateness of model for problem

How do changes in herring populations affect predators?

Availble models:

Multispecies catch at age (Curti, et al., 2013)

Multispecies catch at length (Gaichas, et al., 2017)

Multispecies production (Gamble, et al., 2009)

all model predation mortality but not prey effects on predators!

Food web (Link, et al., 2008; Lucey, et al., 2020)

models 2 way interactions but aggregated species groups and fishery

16 / 24

MICE for MSE: Addressing stakeholder objectives (Deroba, et al., 2018)

The Dream1 Convert the effects of control rules on 4 user groups to dollars:

  1. Users of landed herring (Demand)
    • Lobster industry, aquariums
  2. Herring harvesters (Supply)
  3. Direct users of herring in the ocean (not people)
    • Terns and Whales
    • Striped Bass, Dogfish
  4. Indirect users of herring in the ocean (people, Derived Demand)
    • Bird- and whale-watchers
    • Recreational and Commercial Fishing

The Reality

  • parallel development of herring, predator, economic models
  • limited predator responses, economic effects

1 Credit: Min-Yang Lee

17 / 24

Develop a harvest control rule considering herring's role as prey

Species interactions:

  • Bottom up only: Predators with herring dependent dynamics
  • NO predator feedback to herring
  • Alternative Herring operating models with high and low M

(Also, done in ~ 6 months)

ternpoptrend

Time constraints forced:

  • selection of predators with previous modeling and readily available data
  • selection of single strongest herring-predator relationship
  • models ignoring high variance in prey-predator relationships
18 / 24

In general, if support for a relationship between herring and predator recruitment was evident, it was modeled as a predator recruitment multiplier based on the herring population Ny relative to a specified threshold Nthresh:

ˉRPy+a=RPy+aγ(Ny/Nthresh)(γ1)+(Ny/Nthresh)

where γ > 1 links herring population size relative to the threshold level to predator recruitment.

If a relationship between predator growth and herring population size was evident, annual changes in growth were modeled by modifying either the Ford-Walford intercept αPy or slope ρPy:

BPy+1=SPy(αPyNPy+FwslopeBPy)+αPyRPy+1 or

BPy+1=SPy(FwintNPy+ρPyBPy)+FwintRPy+1

where either αPy or ρPy are defined for a predator using herring population parameters.

Finally, herring population size Ny could be related to predator survival using an annual multiplier on constant predator annual natural mortality v:

vy=ve(NyNF=0)δ

where 0 < δ <1 links herring population size to predator survival.

Uncertainty bracketed by herring population and assessment assumptions

Managing tradeoffs under uncertainty: What control rules give us 90% of everything we want?

Tern productivity 1.0 > 90% of the time, Herring biomass > 90% of SSBmsy, and Fishery yield > 90% of MSY

All three of the conditions on the left, AND fishery closures (F=0) less than 1% of the time

19 / 24

Peer review, March 2017

Report

The Panel agreed that the NEFSC technical team constructed a series of models (Atlantic herring, predator, and economic) appropriate for evaluating ABC control rules for the Atlantic herring fishery in the context of herring’s role as a forage fish. The Panel detailed areas of strength and areas for improvement in the MSE workshop process, modeling, and synthesis. The Panel concluded that the data, methods, and results of the MSE are sufficient for the Council to use when identifying and analyzing a range of ABC control rule alternatives for the Atlantic Herring Fishery Management Plan. Overall, the Panel concluded that the Atlantic herring MSE represents the best available science at this time for evaluating the performance of herring control rules and their potential impact on key predators.

But, predator models were simple and may not capture all important effects, combination of approaches may be better, integrate impacts of predators on herring, consider predator's alternative prey, include process error in predator models.

20 / 24

Epilogue: Herring MSE food web modeling supplemental results

  • Tradeoffs between forage groups and mixed impacts to predators apparent when multiple species and full predator prey interaction feedbacks can be included
  • Rpath Ecosense functions evaluate parameter uncertainty within a scenario

  • Now we have MSE closed loop possibilities in Rpath (Lucey, et al., 2021)

  • Can implement HCRs with predator prey interactions

21 / 24

Conclusions: What makes a good multispecies model?

  1. Identify the problem
    • Does it require modeling multispecies dynamics?
    • What is the output of interest?
  2. Scientific basis: are model assumptions appropriate for the problem?
  3. Input data and parameterization: sufficient for the problem?
    • Identify data sources, series length, quality, gaps
    • Identify key assumptions related to the output of interest
  4. Does model output compare well with observations?
  5. Has uncertainty been assessed in the output of interest?
  6. Peer review
22 / 24

References

Allen, J. I. et al. (2009). "A multivariate approach to model skill assessment". En. In: Journal of Marine Systems. Skill assessment for coupled biological/physical models of marine systems 76.1, pp. 83-94. ISSN: 0924-7963. DOI: 10.1016/j.jmarsys.2008.05.009. URL: http://www.sciencedirect.com/science/article/pii/S0924796308001152 (visited on Nov. 15, 2019).

Bentley, J. W. et al. (2021). "Refining Fisheries Advice With Stock-Specific Ecosystem Information". In: Frontiers in Marine Science 8. Publisher: Frontiers. ISSN: 2296-7745. DOI: 10.3389/fmars.2021.602072. URL: https://www.frontiersin.org/articles/10.3389/fmars.2021.602072/full (visited on Jun. 04, 2021).

Curti, K. L. et al. (2013). "Evaluating the performance of a multispecies statistical catch-at-age model". En. In: Canadian Journal of Fisheries and Aquatic Sciences 70.3, pp. 470-484. ISSN: 0706-652X, 1205-7533. DOI: 10.1139/cjfas-2012-0229. URL: http://www.nrcresearchpress.com/doi/abs/10.1139/cjfas-2012-0229 (visited on Jan. 13, 2016).

Deroba, J. J. et al. (2018). "The dream and the reality: meeting decision-making time frames while incorporating ecosystem and economic models into management strategy evaluation". In: Canadian Journal of Fisheries and Aquatic Sciences. ISSN: 0706-652X. DOI: 10.1139/cjfas-2018-0128. URL: http://www.nrcresearchpress.com/doi/10.1139/cjfas-2018-0128 (visited on Jul. 20, 2018).

Gaichas, S. K. et al. (2017). "Combining stock, multispecies, and ecosystem level fishery objectives within an operational management procedure: simulations to start the conversation". In: ICES Journal of Marine Science 74.2, pp. 552-565. ISSN: 1054-3139. DOI: 10.1093/icesjms/fsw119. URL: https://academic.oup.com/icesjms/article/74/2/552/2669545/Combining-stock-multispecies-and-ecosystem-level (visited on Oct. 18, 2017).

Gamble, R. J. et al. (2009). "Analyzing the tradeoffs among ecological and fishing effects on an example fish community: A multispecies (fisheries) production model". En. In: Ecological Modelling 220.19, pp. 2570-2582. ISSN: 03043800. DOI: 10.1016/j.ecolmodel.2009.06.022. URL: http://linkinghub.elsevier.com/retrieve/pii/S0304380009003998 (visited on Oct. 13, 2016).

ICES (2019). "Working Group on Multispecies Assessment Methods (WGSAM)". . Publisher: ICES. DOI: 10.17895/ICES.PUB.5758. URL: http://www.ices.dk/sites/pub/Publication Reports/Forms/DispForm.aspx?ID=36455 (visited on Jun. 07, 2021).

ICES (2021). "Working Group on Multispecies Assessment Methods (WGSAM)". En. . Publisher: ICES. DOI: 10.17895/ICES.PUB.7695. URL: http://www.ices.dk/sites/pub/Publication Reports/Forms/DispForm.aspx?ID=37316 (visited on Jun. 07, 2021).

Kaplan, I. C. et al. (2016). "A guinea pig's tale: learning to review end-to-end marine ecosystem models for management applications". En. In: ICES Journal of Marine Science 73.7, pp. 1715-1724. ISSN: 1054-3139. DOI: 10.1093/icesjms/fsw047. URL: https://academic.oup.com/icesjms/article/73/7/1715/2458746 (visited on Aug. 30, 2019).

Lehuta, S. et al. (2016). "Reconciling complex system models and fisheries advice: Practical examples and leads". En. In: Aquatic Living Resources 29.2, p. 208. ISSN: 0990-7440, 1765-2952. DOI: 10.1051/alr/2016022. URL: https://www.alr-journal.org/articles/alr/abs/2016/02/alr150069/alr150069.html (visited on Nov. 15, 2019).

Lehuta, S. et al. (2013). "Selection and validation of a complex fishery model using an uncertainty hierarchy". En. In: Fisheries Research 143, pp. 57-66. ISSN: 0165-7836. DOI: 10.1016/j.fishres.2013.01.008. URL: http://www.sciencedirect.com/science/article/pii/S0165783613000118 (visited on Nov. 15, 2019).

Link, J. S. (2010). "Adding rigor to ecological network models by evaluating a set of pre-balance diagnostics: A plea for PREBAL". En. In: Ecological Modelling 221.12, pp. 1580-1591. ISSN: 0304-3800. DOI: 10.1016/j.ecolmodel.2010.03.012. URL: http://www.sciencedirect.com/science/article/pii/S0304380010001468 (visited on Nov. 20, 2019).

Link, J. et al. (2008). "The Northeast U.S. continental shelf Energy Modeling and Analysis exercise (EMAX): Ecological network model development and basic ecosystem metrics". In: Journal of Marine Systems 74.1–2, pp. 453-474. ISSN: 0924-7963. DOI: 10.1016/j.jmarsys.2008.03.007. URL: http://www.sciencedirect.com/science/article/pii/S0924796308000560 (visited on Jan. 13, 2016).

Lucey, S. M. et al. (2021). "Evaluating fishery management strategies using an ecosystem model as an operating model". En. In: Fisheries Research 234, p. 105780. ISSN: 0165-7836. DOI: 10.1016/j.fishres.2020.105780. URL: http://www.sciencedirect.com/science/article/pii/S0165783620302976 (visited on Dec. 09, 2020).

Lucey, S. M. et al. (2020). "Conducting reproducible ecosystem modeling using the open source mass balance model Rpath". En. In: Ecological Modelling 427, p. 109057. ISSN: 0304-3800. DOI: 10.1016/j.ecolmodel.2020.109057. URL: http://www.sciencedirect.com/science/article/pii/S0304380020301290 (visited on Apr. 27, 2020).

NRC (2007). "Chapter 4. Model Evaluation". En. In: Models in Environmental Regulatory Decision Making. Washington D.C.: The National Academies Press, pp. 104-169. DOI: 10.17226/11972. URL: https://www.nap.edu/read/11972/chapter/6 (visited on Aug. 29, 2019).

Olsen, E. et al. (2016). "Ecosystem Model Skill Assessment. Yes We Can!" En. In: PLOS ONE 11.1. Ed. by C. N. Bianchi, p. e0146467. ISSN: 1932-6203. DOI: 10.1371/journal.pone.0146467. URL: http://dx.plos.org/10.1371/journal.pone.0146467 (visited on Jan. 08, 2016).

Saltelli, A. et al. (2010). "How to avoid a perfunctory sensitivity analysis". En. In: Environmental Modelling & Software 25.12, pp. 1508-1517. ISSN: 13648152. DOI: 10.1016/j.envsoft.2010.04.012. URL: https://linkinghub.elsevier.com/retrieve/pii/S1364815210001180 (visited on Nov. 15, 2019).

Sterman, J. D. (1984). "Appropriate summary statistics for evaluating the historical fit of system dynamics models". In: Dynamica 10 Part II Winter (1984), pp. 51-66. URL: https://www.systemdynamics.org/assets/dynamica/102/4.pdf (visited on Nov. 15, 2019).

Stow, C. A. et al. (2009). "Skill assessment for coupled biological/physical models of marine systems". En. In: Journal of Marine Systems. Skill assessment for coupled biological/physical models of marine systems 76.1, pp. 4-15. ISSN: 0924-7963. DOI: 10.1016/j.jmarsys.2008.03.011. URL: http://www.sciencedirect.com/science/article/pii/S0924796308001103 (visited on Nov. 15, 2019).

Townsend, H. et al. (2019). "Progress on Implementing Ecosystem-Based Fisheries Management in the United States Through the Use of Ecosystem Models and Analysis". In: Frontiers in Marine Science 6. ISSN: 2296-7745. DOI: 10.3389/fmars.2019.00641. URL: https://www.frontiersin.org/articles/10.3389/fmars.2019.00641/full (visited on Nov. 19, 2019).

Trijoulet, V. et al. (2020). "Performance of a state-space multispecies model: What are the consequences of ignoring predation and process errors in stock assessments?" En. In: Journal of Applied Ecology n/a.n/a. ISSN: 1365-2664. DOI: 10.1111/1365-2664.13515. URL: https://besjournals.onlinelibrary.wiley.com/doi/abs/10.1111/1365-2664.13515 (visited on Dec. 04, 2019).

23 / 24

ms-keyrun project objectives

EBFM Objective 1: what happens with all the species in the region under a certain management regime?

  • Model needs to minimally
    • Include all the species in the region
    • Implement the alternative management regimes
  • Define "what happens"
    • Increase/decrease relative to base?
    • Status relative to reference points? (model needs to estimate reference points)
    • Details on distribution, demographics needed?

EBFM Objective 2: how well do multispecies models perform for assessment?

  • Which models are better at what (size/age structured and biomass dynamics)
    • Define "better"
    • Define "what" key outputs are most important, what are we assessing
  • What data are necessary?
  • Which species actually need to be included? Use Rpath to id
  • Need a fleet based model? Add econ?

The above includes fits to historical data for all the models

Overall goal: tools vetted and ready to go with any of further questions from the Council

What makes a good model? (NRC, 2007), p 108

Is the model based on generally accepted science and computational methods?

Does it work, that is, does it fulfill its designated task or serve its intended purpose?

Does its behavior approximate that observed in the system being modeled?

Fisheries models range from single and multispecies models to full ecosystem models

2 / 24
Paused

Help

Keyboard shortcuts

, , Pg Up, k Go to previous slide
, , Pg Dn, Space, j Go to next slide
Home Go to first slide
End Go to last slide
Number + Return Go to specific slide
b / m / f Toggle blackout / mirrored / fullscreen mode
c Clone slideshow
p Toggle presenter mode
t Restart the presentation timer
?, h Toggle this help
Esc Back to slideshow