Our mission is to help computational modelers develop, document, and share their computational models in accordance with community standards and good open science and software engineering practices. Model authors can publish their model source code in the Computational Model Library with narrative documentation as well as metadata that supports open science and emerging norms that facilitate software citation, computational reproducibility / frictionless reuse, and interoperability. Model authors can also request private peer review of their computational models. Models that pass peer review receive a DOI once published.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and feel free to contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 4 of 4 results irrigation games clear search
The model includes different formulations how agents make decisions in irrigation games and this is compared with empirical data. The aim is to test different theoretical models, especially explaining effect of communication.
Comparing 7 alternative models of human behavior and assess their performance on a high resolution dataset based on individual behavior performance in laboratory experiments.
Irrigation game calibrated on experimental data
This model can be used to explore under which conditions agents behave as observed in field experiments on irrigation games.