Our mission is to help computational modelers develop, document, and share their computational models in accordance with community standards and good open science and software engineering practices. Model authors can publish their model source code in the Computational Model Library with narrative documentation as well as metadata that supports open science and emerging norms that facilitate software citation, computational reproducibility / frictionless reuse, and interoperability. Model authors can also request private peer review of their computational models. Models that pass peer review receive a DOI once published.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and feel free to contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 10 of 63 results for "Merlin Radbruch" clear search
This is an agent-based model with two types of agents: customers and insurers. Insurers are price-takers who choose how much to spend on their service quality, and customers evaluate insurers based on premium, brand preference, and their perceived service quality. Customers are also connected in a small-world network and may share their opinions with their network.
The ABM contains two types of agents: insurers and customers. These act within the environment of a motor insurance market. At each simulation, the model undergoes the following steps:
The largely dominant meritocratic paradigm of highly competitive Western cultures is rooted on the belief that success is due mainly, if not exclusively, to personal qualities such as talent, intelligence, skills, smartness, efforts, willfulness, hard work or risk taking. Sometimes, we are willing to admit that a certain degree of luck could also play a role in achieving significant material success. But, as a matter of fact, it is rather common to underestimate the importance of external forces in individual successful stories. It is very well known that intelligence (or, more in general, talent and personal qualities) exhibits a Gaussian distribution among the population, whereas the distribution of wealth - often considered a proxy of success - follows typically a power law (Pareto law), with a large majority of poor people and a very small number of billionaires. Such a discrepancy between a Normal distribution of inputs, with a typical scale (the average talent or intelligence), and the scale invariant distribution of outputs, suggests that some hidden ingredient is at work behind the scenes. In a recent paper, with the help of this very simple agent-based model realized with NetLogo, we suggest that such an ingredient is just randomness. In particular, we show that, if it is true that some degree of talent is necessary to be successful in life, almost never the most talented people reach the highest peaks of success, being overtaken by mediocre but sensibly luckier individuals. As to our knowledge, this counterintuitive result - although implicitly suggested between the lines in a vast literature - is quantified here for the first time. It sheds new light on the effectiveness of assessing merit on the basis of the reached level of success and underlines the risks of distributing excessive honors or resources to people who, at the end of the day, could have been simply luckier than others. With the help of this model, several policy hypotheses are also addressed and compared to show the most efficient strategies for public funding of research in order to improve meritocracy, diversity and innovation.
This model computes the guaranteed viability kernel of a model describing the evolution of a population submitted to successive floods.
The population is described by its wealth and its adaptation rate to floods, the control are information campaigns that have a cost but increase the adaptation rate and the expected successive floods belong to given set defined by the maximal high and the minimal time between two floods.
This model extends the original Artifical Anasazi (AA) model to include individual agents, who vary in age and sex, and are aggregated into households. This allows more realistic simulations of population dynamics within the Long House Valley of Arizona from AD 800 to 1350 than are possible in the original model. The parts of this model that are directly derived from the AA model are based on Janssen’s 1999 Netlogo implementation of the model; the code for all extensions and adaptations in the model described here (the Artificial Long House Valley (ALHV) model) have been written by the authors. The AA model included only ideal and homogeneous “individuals” who do not participate in the population processes (e.g., birth and death)–these processes were assumed to act on entire households only. The ALHV model incorporates actual individual agents and all demographic processes affect these individuals. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. Thus, the ALHV model is a combination of individual processes (birth and death) and household-level processes (e.g., finding suitable agriculture plots).
As is the case for the AA model, the ALHV model makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original model (from Janssen’s Netlogo implementation) to estimate annual maize productivity of various agricultural zones within the valley. These estimates are used to determine suitable locations for households and farms during each year of the simulation.
This model is an extension of the Artificial Long House Valley (ALHV) model developed by the authors (Swedlund et al. 2016; Warren and Sattenspiel 2020). The ALHV model simulates the population dynamics of individuals within the Long House Valley of Arizona from AD 800 to 1350. Individuals are aggregated into households that participate in annual agricultural and demographic cycles. The present version of the model incorporates features of the ALHV model including realistic age-specific fertility and mortality and, in addition, it adds the Black Mesa environment and population, as well as additional methods to allow migration between the two regions.
As is the case for previous versions of the ALHV model as well as the Artificial Anasazi (AA) model from which the ALHV model was derived (Axtell et al. 2002; Janssen 2009), this version makes use of detailed archaeological and paleoenvironmental data from the Long House Valley and the adjacent areas in Arizona. It also uses the same methods as the original AA model to estimate annual maize productivity of various agricultural zones within the Long House Valley. A new environment and associated methods have been developed for Black Mesa. Productivity estimates from both regions are used to determine suitable locations for households and farms during each year of the simulation.
The model simulates agents behaviour in wine market parallel trading systems: auctions, OTC and Liv-ex. Models are written in JAVA and use MASON framework. To run a simulation download source files with additional src folder with sobol.csv file. In WineSimulation.java set RESULTS_FOLDER parameter. Uses following external libraries mason19..jar, opencsv.jar, commons-lang3-3.5.jar and commons-math3-3.6.1.jar.
This model simulates the behaviour of the agents in 3 wine markets parallel trading systems: Liv-ex, Auctions and additionally OTC market (finally not used). Behavioural aspects (impatience) is additionally modeled. This is an extention of parallel trading systems model with technical trading (momentum and contrarian) and noise trading.
The purpose of this model is to explore the importance of geographic factors to the settlement choices of early Neolithic agriculturalists. In the model, each agriculturalist spreads to one of the best locations within a modeler specified radius. The best location is determined by choosing either one factor such as elevation or slope; or by ranking geographic factors in order of importance.
The goal of the paper is to propose an abstract but formalised model of how Schwartz higher order values may influence individual decisions on sharing an individual effort among alternative economic activities. Subsequently, individual decisions are aggregated into the total (collective) economic output, taking into account interactions between the agents. In particular, we explore the relationship between individual higher order values: Self–Enhancement, Self–Transcendence, Openness to Change, and Conservation – measured according to Schwartz’s universal human values theory – and individual and collective economic performance, by means of a theoretical agent based model. Furthermore, based on empirical observations, Openness to Change (measured by the population average in the case of collective output) is positively associated with individual and collective output. These relations are negative for Conservation. Self-Enhancement is positively associated with individual output but negatively with collective output. In case of Self–Transcendence, this effect is opposite. The model provides the potential explanations, in terms of individual and population differences in: propensity for management, willingness to change, and skills (measured by an educational level) for the empirically observed relations between Schwartz higher order values and individual and collective output. We directly calibrate the micro–level of the model using data from the ninth round of the European Social Survey (ESS9) and present the results of numerical simulations.
This paper investigates the impact of agents' trading decisions on market liquidity and transactional efficiency in markets for illiquid (hard-to-trade) assets. Drawing on a unique order book dataset from the fine wine exchange Liv-ex, we offer novel insights into liquidity dynamics in illiquid markets. Using an agent-based framework, we assess the adequacy of conventional liquidity measures in capturing market liquidity and transactional efficiency. Our main findings reveal that conventional liquidity measures, such as the number of bids, asks, new bids and new asks, may not accurately represent overall transactional efficiency. Instead, volume (measured by the number of trades) and relative spread measures may be more appropriate indicators of liquidity within the context of illiquid markets. Furthermore, our simulations demonstrate that a greater number of traders participating in the market correlates with an increased efficiency in trade execution, while wider trader-set margins may decrease the transactional efficiency. Interestingly, the trading period of the agents appears to have a significant impact on trade execution. This suggests that granting market participants additional time for trading (for example, through the support of automated trading systems) can enhance transactional efficiency within illiquid markets. These insights offer practical implications for market participants and policymakers aiming to optimise market functioning and liquidity.
Displaying 10 of 63 results for "Merlin Radbruch" clear search