+ - 0:00:00
Notes for current slide
Notes for next slide

Challenging assessment models
with realistic complexity

Atlantis Summit 2022 Day 3 Keynote

Sarah Gaichas
Northeast Fisheries Science Center

Many thanks to:
(atlantisom development): Christine Stawitz, Kelli Johnson, Alexander Keth, Allan Hicks, Sean Lucey, Emma Hodgson, Gavin Fay
(Atlantis): Isaac Kaplan, Cecilie Hanson, Beth Fulton
(atlantisom use): Bai Li, Alphonso Perez Rodriguez, Howard Townsend
(skill assessment design): Patrick Lynch

1 / 27

Outline

  • Management decisions and models

  • Assessment model testing

    • What do models need to do?
    • Challenges with what models need to do
    • Addressing challenges with Atlantis
  • Skill assessment with Atlantis "data"

    1. R package atlantisom
    2. Single species assessment
    3. Multispecies assessment
    4. Other models? Ensembles?
  • Discussion: what's next

    • Automation of other model inputs!
    • Atlantis output verification
    • New outputs?
    • Best practices for
      • dataset creation
      • parameter estimation

poseidon

2 / 27

We do a lot of assessments

NOAAlogo

3 / 27

With a wide range of models

4 / 27

How do we know they are right?

  • Fits to historical data (hindcast)

  • Influence of data over time (retrospective diagnostics)

  • Keep as simple and focused as possible

  • Simulation testing

But, what if

data are noisy?

we need to model complex interactions?

conditions change over time?

5 / 27

Fisheries: what do we need to know?                               What do models need to estimate?

How many fish can be caught sustainably?

  • How many are there right now?
  • How many were there historically?
  • How productive are they (growth, reproduction)?
  • How many are caught right now?
  • How many were caught historically?
  • Current biomass overall and at age/size
  • Reconstruct past population dynamics
  • Individual and population growth
  • How many should be caught next year?
  • How certain are we?
6 / 27

Fisheries: what do we need to know?                               What do models need to estimate?

How many fish can be caught sustainably?

  • How many are there right now?
  • How many were there historically?
  • How productive are they (growth, reproduction)?
  • How many are caught right now?
  • How many were caught historically?
  • Current biomass overall and at age/size
  • Reconstruct past population dynamics
  • Individual and population growth
  • How many should be caught next year?
  • How certain are we?

"New" questions and challenges, broader management objectives

  • What supports their productivity?
  • What does their productivity support, besides fishing?
  • How do they interact with other fish, fisheries, marine animals?
  • How do environmental changes affect them?
  • What is their ecological, economic, and social value to people?
6 / 27

Risks to meeting fishery management objectives

Climate icon made by EDAB       Wind icon made by EDAB

Hydrography icon made by EDAB       Phytoplankon icon made by EDAB       Forage fish icon made by EDAB       Apex predators icon made by EDAB       Other human uses icon made by EDAB

7 / 27

Risks: Climate change indicators in US Mid-Atlantic

Indicators: ocean currents, bottom and surface temperature, marine heatwaves

 
 
     
 
 
   
 
   
   
 
   
 

8 / 27

A marine heatwave is a warming event that lasts for five or more days with sea surface temperatures above the 90th percentile of the historical daily climatology (1982-2011).

The stock assessment community is well aware of this

  • Changing climate and ocean conditions → Shifting species distributions, changing productivity

  • Needs:

    • Improve our ability to project global change impacts in the ecosystems around the world
    • Test the performance of stock assessments to these impacts
    • Design assessment methods that perform well despite these impacts

Climate-Ready Management, Karp et al 2019 Climateready

9 / 27

"Both our model predictions and the observations reside in a halo of uncertainty and the true state of the system is assumed to be unknown, but lie within the observational uncertainty (Fig. 1a). A model starts to have skill when the observational and predictive uncertainty halos overlap, in the ideal case the halos overlap completely (Fig. 1b). Thus, skill assessment requires a set of quantitative metrics and procedures for comparing model output with observational data in a manner appropriate to the particular application."

Skill assessment with ecological interactions... fit criteria alone are not sufficient

Ignore predation at your peril: results from multispecies state-space modeling Trijoulet et al. 2020

Ignoring trophic interactions that occur in marine ecosystems induces bias in stock assessment outputs and results in low model predictive ability with subsequently biased reference points.

VanessaPaper

EM1: multispecies state space

EM2: multispecies, no process error

EM3: single sp. state space, constant M

EM4: single sp. state space, age-varying M

note difference in scale of bias for single species!

modcomp

11 / 27

This is an important paper both because it demonstrates the importance of addressing strong species interactions, and it shows that measures of fit do not indicate good model predictive performance. Ignoring process error caused bias, but much smaller than ignoring species interactions. See also Vanessa's earlier paper evaluating diet data interactions with multispecies models

Virtual worlds with adequate complexity: end-to-end ecosystem models

Atlantis modeling framework: Fulton et al. 2011, Fulton and Smith 2004

Norwegian-Barents Sea

Hansen et al. 2016, 2018

NOBA scale 70%

Building on global change projections: Hodgson et al. 2018, Olsen et al. 2018

12 / 27

Design: Ecosystem model scenario (climate and fishing)

13 / 27
  • Recruitment variability in the operating model

  • Specify uncertainty in assessment inputs using atlantisom

sardinerec scale 100%

Designing skill assessment with Atlantis

poseidon

fragile ecosystems robust assessments

14 / 27

Started at the 2015 Atlantis Summit atlantisom intro

atlantisom get started

15 / 27

atlantisom workflow: get "truth"

  • locate files
  • run om_init
  • select species
  • run om_species
16 / 27

atlantisom workflow: get "data"

  • specify surveys
    • (can now have many, file for each)
    • area/time subsetting
    • efficiency (q) by species
    • selectivity by species
    • biological sample size
    • index cv
    • length at age cv
    • max size bin
  • specify survey diet sampling (new)
    • requires detaileddietcheck.txt
    • diet sampling parameters
  • specify fishery
    • area/time subsetting
    • biological sample size
    • catch cv
  • run om_index
  • run om_comps
  • run om_diet
  • environmental data functions too
17 / 27

Perfect information (one Season) NOBA fall survey 1

Survey with catchability and selectivity NOBA fall survey 2

18 / 27

capelin lengths halibut lengths

capelin ages halibut ages

19 / 27

true diets

seasonal survey diets

20 / 27

Testing a simple "sardine" assessment, CC Atlantis Kaplan et al. 2021

Kaplan et al Fig 2

Will revisit with newer CC model; issues with different growth than assumed in SS setup?

workinprogress

21 / 27

Cod assessment based on NOBA Atlantis (Li, WIP)

https://github.com/Bai-Li-NOAA/poseidon-dev/blob/nobacod/NOBA_cod_files/README.MD

Conversion from SAM to SS successful

Fitting to NOBA data more problematic

22 / 27

Multispecies assessment based on NOBA Atlantis (Townsend et al, WIP)

Stepwise development process of self fitting, fitting to atlantis output, then skill assessment using atlantis output

Profiles for estimated parameters; but what to compare K values to?

Can test model diagnostic tools as well Using simulated data in mskeyrun package, available to all

23 / 27

P.S. What else could we test?

xkcd_ensemble_model_2x

24 / 27

Multispecies production model ensemble assessment--use Atlantis instead

Hydra OM setup

ensemble results

25 / 27

Difficulties so far

Atlantis related

  • Understanding Atlantis outputs (much improved with documentation since 2015)
  • Reconciling different Atlantis outputs, which to use?
  • Calculations correct? attempted M, per capita consumption

Skill assessment related

  • Running stock assessment models is difficult to automate. A lot of decisions are made by iterative running and diagnostic checks.
  • Generating input parameters for models that are consistent with Atlantis can be time consuming (Atlantis is a little too realistic...)
    • Fit vonB length at age models to atlantisom output for input to length based model, not all converge (!)
    • What is M (see above)
26 / 27
  • atlantisom is using outputs not often used in other applications

    • I don't run Atlantis so putting print statements in code not an option
    • could be more efficient with targeted group work

    • should we expect numbers in one output to match those in others?

    • diet comp from detailed file matches diet comp in simpler output
    • catch in numbers not always matching between standard and annual age outputs
    • YOY output
    • ... others that have been encountered

    • estimating per capita consumption from detaileddiet.txt results in lower numbers than expected

    • still can't get reasonable mortality estimates from outputs--understand this is an issue

Atlantis and assessment model skill assessment: Thank you and Discussion

Fisheries stock assessment and ecosystem modeling continue to develop. Can we build assessments to keep pace with climate?

Interested in your thoughts:

  • Is this a good use of Atlantis model outputs?
  • Any obvious errors in atlantisom setup?
  • How to improve atlantisom?
    • Functions to write specific model inputs in progress (SS included, more complex models in own packages/repositories)
    • New package with skill assessment functions?
    • Update documentation and vingnettes
    • Integrate/update with other R Atlantis tools
    • Other?
  • What are best practices to use Atlantis model outputs in skill assessment?
    • Spatial and temporal scale that is most appropriate
    • Calculating assessment model input parameters from Atlantis outputs
    • Storing Atlantis outputs for others to use (large file size)
    • Other?
27 / 27

Outline

  • Management decisions and models

  • Assessment model testing

    • What do models need to do?
    • Challenges with what models need to do
    • Addressing challenges with Atlantis
  • Skill assessment with Atlantis "data"

    1. R package atlantisom
    2. Single species assessment
    3. Multispecies assessment
    4. Other models? Ensembles?
  • Discussion: what's next

    • Automation of other model inputs!
    • Atlantis output verification
    • New outputs?
    • Best practices for
      • dataset creation
      • parameter estimation

poseidon

2 / 27
Paused

Help

Keyboard shortcuts

, , Pg Up, k Go to previous slide
, , Pg Dn, Space, j Go to next slide
Home Go to first slide
End Go to last slide
Number + Return Go to specific slide
b / m / f Toggle blackout / mirrored / fullscreen mode
c Clone slideshow
p Toggle presenter mode
t Restart the presentation timer
?, h Toggle this help
Esc Back to slideshow