CoMSES Net maintains cyberinfrastructure to foster FAIR data principles for access to and (re)use of computational models. Model authors can publish their model code in the Computational Model Library with documentation, metadata, and data dependencies and support these FAIR data principles as well as best practices for software citation. Model authors can also request that their model code be peer reviewed to receive a DOI. All users of models published in the library must cite model authors when they use and benefit from their code.
CoMSES Net also maintains a curated database of over 7500 publications of agent-based and individual based models with additional metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Cultural group selection model used to evaluate the conditions for agents to evolve who have other-regarding preferences in making decisions in public good games.
ICARUS is a multi-agent compliance inspection model (ICARUS - Inspecting Compliance to mAny RUleS). The model is applicable to environments where an inspection agency, via centrally coordinated inspections, examines compliance in organizations which must comply with multiple provisions (rules). The model (ICARUS) contains 3 types of agents: entities, inspection agency and inspectors / inspections. ICARUS describes a repeated, simultaneous, non-cooperative game of pure competition. Agents have imperfect, incomplete, asymmetric information. Entities in each move (tick) choose a pure strategy (comply/violate) for each rule, depending on their own subjective assessment of the probability of the inspection. The Inspection Agency carries out the given inspection strategy.
A more detailed description of the model is available in the .nlogo file.
Full description of the model (in line with the ODD+D protocol) and the analysis of the model (including verification, validation and sensitivity analysis) can be found in the attached documentation.
This model aims to explore how gambling-like behavior can emerge in loot box spending within gaming communities. A loot box is a purchasable mystery box that randomly awards the player a series of in-game items. Since the contents of the box are largely up to chance, many players can fall into a compulsion loop of purchasing, as the fear of missing out and belief in the gambler’s fallacy allow one to rationalize repeated purchases, especially when one compares their own luck to others. To simulate this behavior, this model generates players in different network structures to observe how factors such as network connectivity, a player’s internal decision making strategy, or even common manipulations games use these days may influence a player’s transactions.
ThomondSim is a simulation of the political and economic landscape of the medieval kingdom of Thomond, southwestern Ireland, between 1276 and 1318.
Its goal is to analyze how deteriorating environmental and economic conditions caused by the Little Ice Age (LIA), the Great European Famine of 1315-1322, and wars between England and Scotland affected the outcomes of a local war involving Gaelic and English aristocratic lineages.
This ABM attempts to model both the effects of devastation on the human environment and the modus operandi of late-medieval war and diplomacy.
The model is the digital counterpart of the science discovery board game The Triumphs of Turlough. Its procedures closely correspond to the game’s mechanics, to the point that ToT can be considered an interactive, analog version of this ABM.
This paper introduces an experimental and exploratory approach, combining game theory and Genetic Algorithms to create a model to simulate evolutionary economic learning. The objective of this paper is to document the implementation of a genetic algorithm as a simulator for economic learning, then analyze how strategic behavior affects the evolution towards optimal outcomes, departing from different starting points and potentially transforming conflict into harmonious scenarios. For this purpose, the introduced construct aimed at allowing for the evaluation of different strategy selection methods and game types. 144 unique 2x2 games, and three distinct strategy selection rules: Nash equilibrium, Hurwicz rule and a Random selection method were used in this study. The particularity of this paper is that rather than changing the strategies themselves or player-specific features, the introduced genetic algorithm changes the games based on the player payoffs. The outcome indicated optimal player scenarios for both The Nash equilibrium and Hurwicz rules strategies, the first being the best performing strategy. The random selection method fails to converge to optimal values in most of the populations, acting as a control feature and reinforcing that strategic behavior is necessary for the evolutionary learning process. We documented also two additional observations. First, the games are often transformed in such a way that agents can coordinate their strategies to achieve a stable optimal equilibrium. And second, we observed the mutation of the populations of games into sets of fewer (repeating) isomorphic games featuring strong characteristics of previous games.
This model simulates the Hawk-Dove game as first described by John Maynard Smith, and further elaborated by Richard Dawkins in “The Selfish Gene”. In the game, two strategies, Hawks and Doves, compete against each other, and themselves, for reproductive benefits. A third strategy can be introduced, Retaliators, which act like either Hawks or Doves, depending on the context.
In our model, individual agents are distributed over a two-dimensional square lattice. The agents play the prisoner’s dilemma game with their neighbors, imitate the highest strategy, and then migrate to empty sites based on their tag preference.
The purpose of the model is to collect information on human decision-making in the context of coalition formation games. The model uses a human-in-the-loop approach, and a single human is involved in each trial. All other agents are controlled by the ABMSCORE algorithm (Vernon-Bido and Collins 2020), which is an extension of the algorithm created by Collins and Frydenlund (2018). The glove game, a standard cooperative game, is used as the model scenario.
The intent of the game is to collection information on the human players behavior and how that compares to the computerized agents behavior. The final coalition structure of the game is compared to an ideal output (the core of the games).
The three-day participatory workshop organized by the TISSS Lab had 20 participants who were academics in different career stages ranging from university student to professor. For each of the five games, the participants had to move between tables according to some pre-specified rules. After the workshop both the participant’s perception of the games’ complexities and the participants’ satisfaction with the games were recorded.
In order to obtain additional objective measures for the games’ complexities, these games were also simulated using this simulation model here. Therefore, the simulation model is an as-accurate-as-possible reproduction of the workshop games: it has 20 participants moving between 5 different tables. The rules that specify who moves when vary from game to game. Just to get an idea, Game 3 has the rule: “move if you’re sitting next to someone who is waring white or no socks”.
An exact description of the workshop games and the associated simulation models can be found in the paper “The relation between perceived complexity and happiness with decision situations: searching for objective measures in social simulation games”.