Our mission is to help computational modelers develop, document, and share their computational models in accordance with community standards and good open science and software engineering practices. Model authors can publish their model source code in the Computational Model Library with narrative documentation as well as metadata that supports open science and emerging norms that facilitate software citation, computational reproducibility / frictionless reuse, and interoperability. Model authors can also request private peer review of their computational models. Models that pass peer review receive a DOI once published.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and feel free to contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 10 of 62 results adaptive clear search
This model is an extension of the Artificial Long House Valley (ALHV) model developed by the authors (Swedlund et al. 2016; Warren and Sattenspiel 2020). The ALHV model simulates the population dynamics of individuals within the Long House Valley of Arizona from AD 800 to 1350. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. The present version of the model incorporates features of the ALHV model including realistic age-specific fertility and mortality and, in addition, it adds the Black Mesa environment and population, as well as additional methods to allow migration between the two regions.
As is the case for previous versions of the ALHV model as well as the Artificial Anasazi (AA) model from which the ALHV model was derived (Axtell et al. 2002; Janssen 2009), this version makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original AA model to estimate annual maize productivity of various agricultural zones within the Long House Valley. A new environment and associated methods have been developed for Black Mesa. Productivity estimates from both regions are used to determine suitable locations for households and farms during each year of the simulation.
This model extends the original Artifical Anasazi (AA) model to include individual agents, who vary in age and sex, and are aggregated into households. This allows more realistic simulations of population dynamics within the Long House Valley of Arizona from AD 800 to 1350 than are possible in the original model. The parts of this model that are directly derived from the AA model are based on Janssen’s 1999 Netlogo implementation of the model; the code for all extensions and adaptations in the model described here (the Artificial Long House Valley (ALHV) model) have been written by the authors. The AA model included only ideal and homogeneous “individuals” who do not participate in the population processes (e.g., birth and death)–these processes were assumed to act on entire households only. The ALHV model incorporates actual individual agents and all demographic processes affect these individuals. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. Thus, the ALHV model is a combination of individual processes (birth and death) and household-level processes (e.g., finding suitable agriculture plots).
As is the case for the AA model, the ALHV model makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original model (from Janssen’s Netlogo implementation) to estimate annual maize productivity of various agricultural zones within the valley. These estimates are used to determine suitable locations for households and farms during each year of the simulation.
Flibs’NLogo implements in NetLogo modelling environment, a genetic algorithm whose purpose is evolving a perfect predictor from a pool of digital creatures constituted by finite automata or flibs (finite living blobs) that are the agents of the model. The project is based on the structure described by Alexander K. Dewdney in “Exploring the field of genetic algorithms in a primordial computer sea full of flibs” from the vintage Scientific American column “Computer Recreations”
As Dewdney summarized: “Flibs […] attempt to predict changes in their environment. In the primordial computer soup, during each generation, the best predictor crosses chromosomes with a randomly selected flib. Increasingly accurate predictors evolve until a perfect one emerges. A flib […] has a finite number of states, and for each signal it receives (a 0 or a 1) it sends a signal and enters a new state. The signal sent by a flib during each cycle of operation is its prediction of the next signal to be received from the environment”
Provided is a landscape of properties where pastoralists make decisions how much livestock they put on their property and how much to suppress fire from occuring. Rangelands can be grass dominated, or unproductive shrubb dominated. Overgrazing and fire suppresion lead to shrub dominated landscapes. What management strategies evolve, and how is this impacted by policies?
The model is discussed in Introduction to Agent-Based Modeling by Marco Janssen. For more information see https://intro2abm.com/.
Modeling an economy with stable macro signals, that works as a benchmark for studying the effects of the agent activities, e.g. extortion, at the service of the elaboration of public policies..
…
This NetLogo model illustrates the cultural evolution of pro-environmental behaviour patterns. It illustrates how collective behaviour patterns evolve from interactions between agents and agents (in a social network) as well as agents and the affordances (action opportunities provided by the environment) within a niche. More specifically, the cultural evolution of behaviour patterns is understood in this model as a product of:
EffLab was built to support the study of the efficiency of agents in an evolving complex adaptive system. In particular:
- There is a definition of efficiency used in ecology, and an analogous definition widely used in business. In ecological studies it is called EROEI (energy returned on energy invested), or, more briefly, EROI (pronounced E-Roy). In business it is called ROI (dollars returned on dollars invested).
- In addition, there is the more well-known definition of efficiency first described by Sadi Carnot, and widely used by engineers. It is usually represented by the Greek letter ‘h’ (pronounced as ETA). These two measures of efficiency bear a peculiar relationship to each other: EROI = 1 / ( 1 - ETA )
In EffLab, blind seekers wander through a forest looking for energy-rich food. In this multi-generational world, they live and reproduce, or die, depending on whether they can find food more effectively than their contemporaries. Data is collected to measure their efficiency as they evolve more effective search patterns.
…
Investigate spatial adaptive behaviors of narco-trafficking networks in response to various counterdrug interdiction strategies within the cocaine transit zone of Central America and associated maritime areas. Through the novel application of the ‘complex adaptive systems’ paradigm, we implement a potentially transformative coupled agent-based and interdiction optimization modeling approach to compellingly demonstrate: (a) how current efforts to disrupt narco-trafficking networks are in fact making them more widespread, resilient, and economically powerful; (b) the potential for alternative interdiction approaches to weaken and contain traffickers.
A model for simulating the evolution of individual’s preferences, incliding adaptive agents “falsifying” -as public opinions- their own preferences. It was builded to describe, explore, experiment and understand how simple heuristics can modulate global opinion dynamics. So far two mechanisms are implemented: a version of Festiguer’s reduction of cognitive disonance, and a version of Goffman’s impression management. In certain social contexts -minority, social rank presure- some models agents can “fake” its public opinion while keeping internally the oposite preference, but after a number of rounds following this falsifying behaviour pattern, a coherence principle can change the real or internal preferences close to that expressed in public.
We develop an IBM that predicts how interactions between elephants, poachers, and law enforcement affect poaching levels within a virtual protected area. The model is theoretical at this stage and is not meant to provide a realistic depiction of poaching, but instead to demonstrate how IBMs can expand upon the existing modelling work done in this field, and to provide a framework for future research. The model could be further developed into a useful management support tool to predict the outcomes of various poaching mitigation strategies at real-world locations. The model was implemented in NetLogo version 6.1.0.
We first compared a scenario in which poachers have prescribed, non-adaptive decision-making and move randomly across the landscape, to one in which poachers adaptively respond to their memories of elephant locations and where other poachers have been caught by law enforcement. We then compare a situation in which ranger effort is distributed unevenly across the protected area to one in which rangers patrol by adaptively following elephant matriarchal herds.
Displaying 10 of 62 results adaptive clear search