Our mission is to help computational modelers at all levels engage in the establishment and adoption of community standards and good practices for developing and sharing computational models. Model authors can freely publish their model source code in the Computational Model Library alongside narrative documentation, open science metadata, and other emerging norms that facilitate software citation, archival, interoperability, and reuse. Model authors can also request that their model code be peer reviewed to receive a DOI.
All users of models published in the library must cite model authors when they use and benefit from their code.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with additional detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 4 of 4 results coherence clear
We present a socio-epistemic model of science inspired by the existing literature on opinion dynamics. In this model, we embed the agents (or scientists) into social networks - e.g., we link those who work in the same institutions. And we place them into a regular lattice - each representing a unique mental model. Thus, the global environment describes networks of concepts connected based on their similarity. For instance, we may interpret the neighbor lattices as two equivalent models, except one does not include a causal path between two variables.
Agents interact with one another and move across the epistemic lattices. In other words, we allow the agents to explore or travel across the mental models. However, we constrain their movements based on absorptive capacity and cognitive coherence. Namely, in each round, an agent picks a focal point - e.g., one of their colleagues - and will move towards it. But the agents’ ability to move and speed depends on how far apart they are from the focal point - and if their new position is cognitive/logic consistent.
Therefore, we propose an analytical model that examines the connection between agents’ accumulated knowledge, social learning, and the span of attitudes towards mental models in an artificial society. While we rely on the example from the General Theory of Relativity renaissance, our goal is to observe what determines the creation and diffusion of mental models. We offer quantitative and inductive research, which collects data from an artificial environment to elaborate generalized theories about the evolution of science.
A model for simulating the evolution of individual’s preferences, incliding adaptive agents “falsifying” -as public opinions- their own preferences. It was builded to describe, explore, experiment and understand how simple heuristics can modulate global opinion dynamics. So far two mechanisms are implemented: a version of Festiguer’s reduction of cognitive disonance, and a version of Goffman’s impression management. In certain social contexts -minority, social rank presure- some models agents can “fake” its public opinion while keeping internally the oposite preference, but after a number of rounds following this falsifying behaviour pattern, a coherence principle can change the real or internal preferences close to that expressed in public.
The model formalizes a situation where agents embedded in different types of networks (random, small world and scale free networks) interact with their neighbors and express an opinion that is the result of different mechanisms: a coherence mechanism, in which agents try to stick to their previously expressed opinions; an assessment mechanism, in which agents consider available external information on the topic; and a social influence mechanism, in which agents tend to approach their neighbor’s opinions.
This is a model of coherency based belief within a dynamic network of individuals. Described in an invited talk on workshop on Coherence, Berlin, 9th July 2016.