Skip to content

Qualitative tests for mesa and mesa-frames benchmarks #106

@adamamer20

Description

@adamamer20

Comparability between mesa and mesa-frames models are crucial for validating the correctness of model implementations.
To achieve this given the same seed and initial conditions, we need:

  1. A compatibility layer that transforms a mesa.Model into a mesa-frames.ModelDF (this can be implemented relatively easily, I have already done some work on a local branch)
  2. Consistent randomized operations (also including shuffling of agents) across frameworks. This is more challenging because of the different random number generators (mesa uses random.Random and mesa-frames uses numpy.random.Generator) and different shuffling (mesa uses random.Random.shuffle and mesa-frames uses the native DFs shuffle operation). Maybe with an appropriate decorator, we could substitute random operations for mesa-frames models at runtime, at the cost of performance but gaining the reproducibility.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementImprovements to existing features or performance.examplesadditions or modifications to sample models

    Type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions