Replies: 36 comments 39 replies
-
|
We did talk about performance optimization in the past, and a Mesa performance optimization project might be interesting this year. Mesa's scalability to millions of agents depends on efficient core operations in AgentSet, spatial grids, and event scheduling. This project could systematically identify and address performance bottlenecks across the library. The first phase involves comprehensive profiling of Mesa's example models (Boltzmann Wealth, Schelling, Wolf-Sheep, Flocking) using tools like cProfile, py-spy, and memory_profiler to create a performance baseline and identify hotspots. Likely candidates include AgentSet operations ( The second phase could explore optimization strategies: expanding NumPy vectorization for batch agent operations, restructuring data layouts for cache efficiency, and evaluating Rust acceleration via PyO3 for compute-intensive components like spatial indexing, large-scale shuffling, and event queue management. Rust is particularly promising for operations with clear data boundaries (grids, coordinate math) where Python object overhead can be avoided. Deliverables could include a reproducible low-level benchmarking suite, documented performance improvements with before/after comparisons, and potentially a @Ben-geo @adamamer20 I'm also curious if there are lessons or techniques of mesa-frames transferable to the main library. |
Beta Was this translation helpful? Give feedback.
-
|
These project ideas look great! I'm really interested in getting involved with Mesa for GSoC 2026. The Behavioral Framework project really caught my attention. I find it fascinating how individual agents making their own decisions can lead to complex emergent behaviors in the system. What excites me most is the challenge of taking these theoretical behavioral models - things like BDI or needs-based architectures - and turning them into something practical that people can actually use. It's basically building the "brain" for agents, which is pretty cool.
One question: which existing Mesa models would you recommend looking at to see how people currently work around things like time-consuming tasks, competing priorities, or continuous state changes? I'd love to understand the current pain points from real examples. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH, exciting list! Building on the "Behavioral Framework" and "Performance" ideas, I'd like to propose a distinct direction for 2026 (or potential experiments in 2025): "Generative Agents Integration" (LLM-driven Behavior). While the Behavioral Framework (BDI/GOAP) is excellent for deterministic, rule-based complexity, there is a growing need for agents that can reason dynamically via LLMs (inspired by the "Generative Agents" architecture by Park et al.). Idea: Create a
I believe this complements the traditional Behavioral Framework by offering a "Probabilistic" alternative for scenarios where defining explicit rules is too difficult (e.g., natural language negotiation between agents or simulating social dynamics). I've already started experimenting with this using the new Mesa 3.0 syntax and ensuring strict JSON outputs. I would love to explore this further as a potential project area. |
Beta Was this translation helpful? Give feedback.
-
|
Thanks for pointing that out! Yes, I have explored My proposal is essentially to modernize and expand that initiative for the Mesa 3.0 era. Specifically, I aim to:
I’d love to take |
Beta Was this translation helpful? Give feedback.
-
|
I have two more: Making clean-sheet designs of how we run models in experimental setups and how we collect data. Mesa's current Reimagining Model Execution and Experimental setupThe fundamental problem with The goal isn't to design a specific solution but to map the territory: What capabilities should Mesa provide natively versus enable through documented patterns? Where are the natural architectural boundaries between experiment specification, execution strategy, and result aggregation? How do we balance simplicity for beginners running local parameter sweeps against power users orchestrating thousands of replications across HPC resources? What standards exist for experiment metadata and provenance that Mesa should adopt? The project could start with a comprehensive requirements analysis, ecosystem survey, user research findings, and architectural principles that can guide Mesa 4.0's design. And if time allows, a (proof of concept) implementation.
Rethinking Data Collection and Management
The research phase should investigate: What data patterns do ABM researchers need to capture, and how do current workflows break down? What can we learn from how climate models, computational biology, and other simulation-heavy domains handle output management? Where does the impedance mismatch lie between ABM data patterns and standard scientific data formats? How do users currently bridge Mesa outputs to their analysis tools, and what friction exists? What performance characteristics matter most—collection overhead during simulation, storage efficiency, or query speed during analysis? The contributor should survey existing solutions, prototype integration patterns with ecosystem tools, and identify fundamental architectural questions: Should Mesa embrace lazy evaluation and streaming? How can we handle the "collect everything vs. collect strategically" tension? What's the right balance between flexibility and performance? The outcome should include a clear problem taxonomy, evaluation of how existing tools address (or don't address) ABM-specific needs, user requirements gathered from community research, and architectural recommendations for Mesa 4.0. Both projects emphasize discovery over delivery: understanding the landscape, learning from successes and failures elsewhere, and establishing principled foundations rather than rushing to implementation. One we know what we need to build and which tools we have to build it, the implementation becomes relatively easy.
|
Beta Was this translation helpful? Give feedback.
-
|
Since there is huge interest in doing things with ML/DL/RL/AI, here is something that might be interesting: A concrete research idea is to build a “Mesa Inference” layer that turns any Mesa model into a calibratable scientific simulator: users specify a parameter prior, an observation model (what macro data they have), and an extraction function that maps each simulation run into a vector of macro summaries (or a learned embedding). Mesa already provides a natural home for this through its core model loop, data collection, and analysis workflow, so the research contribution is primarily about standardizing the simulator ↔ inference boundary. The immediate goal is posterior inference over micro-parameters given macro observations—i.e., identifying which micro behaviors are consistent with the observed macro phenomenon—using modern likelihood-free inference methods. This directly aligns with the Mesa ecosystem and its focus on reproducible ABM experimentation. Methodologically, the first integration target could be simulation-based inference (SBI) as a default backend. This would involve a Mesa-to-SBI adapter that exposes a function of the form A second, more exploratory research track is to support gradient-assisted calibration for differentiable or partially differentiable ABMs implemented in Mesa. The idea is to provide an alternative execution mode for models whose agent transitions or aggregation steps are differentiable (or can be relaxed), enabling gradient-based optimization and sensitivity analysis alongside traditional simulation. The deliverable could be a “differentiable Mesa kernel” for a restricted subset of schedulers and spaces, coupled to automatic differentiation frameworks, with seamless fallback to standard stochastic simulation for general models. This direction builds on prior demonstrations that differentiable ABMs can significantly accelerate calibration, such as Kotthoff’s work in JASSS on gradient-based calibration of ABMs (Kotthoff, 2022), and would position Mesa at the frontier of scientific machine learning for agent-based modeling. Finally, the framework should explicitly support partitioning variance between hypothesized structure and unexplained effects, enabling a form of hypothesis evaluation rather than mere fitting. This can be done via nested model comparisons: (i) ABM-only predictions of macro outcomes, (ii) ABM plus a flexible residual model that captures systematic deviations, and (iii) a purely black-box benchmark. The relative out-of-sample performance of these models quantifies how much variance is explained by the ABM’s micro-level hypothesis versus residual structure. Within an SBI or ABC framework, this analysis naturally incorporates parameter uncertainty and allows posterior predictive checks to diagnose model misspecification. Reporting cross-validated error reductions, posterior predictive discrepancies, and mechanism-level ablations would make the explanatory contribution of the ABM explicit and measurable (see |
Beta Was this translation helpful? Give feedback.
-
|
Hi everyone, I'm Briony! |
Beta Was this translation helpful? Give feedback.
-
|
Hi everyone I am Joel Jose, I have prior experience working with CNN-based models and I’m particularly interested in how AI systems perceive visual information. I’m exploring the idea of introducing a minimal perception abstraction for Mesa agents. The concept would start very small: for example, agents could receive simple visual inputs, like colors or local spatial patterns, converted into numerical observations. Over time and through incremental development, agents could learn to interpret more complex observations, potentially evolving toward recognizing structured patterns or small images. The goal would be to implement this as a modular, optional extension that doesn’t affect Mesa’s core scheduler or deterministic execution, while allowing experimentation with perception-based learning in agent-based simulations. I’d love feedback from the maintainers on whether this direction aligns with Mesa’s road map or if there are existing discussions or abstractions I could build on. |
Beta Was this translation helpful? Give feedback.
-
Reviving Mesa-examplesAfter moving our core examples to the main repo (see #2330 (comment) and #2364), mesa-examples got a bit abandoned. We stopped extensive testing in the main Mesa CI (we do that on our core examples now). Defining, reviving and formalizing Mesa-examples could be really useful. A proposal on this topic could answer (one or more) questions like:
This is my initial take, any other visions/ideas are welcome on this! |
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH, thanks for laying out these ideas so clearly — especially the repeated emphasis on building models first before committing to new abstractions. One direction I’d like to explore (as a contributor, not as a framework proposal yet) is using new example models as a diagnostic tool for the Behavioral Framework discussion. Concretely, instead of starting from BDI / task systems / APIs, I plan to: design a small but non-trivial example model that is not currently present in Mesa, implement it using only existing Mesa primitives, and use that experience to surface where current patterns feel natural vs where behavior logic becomes awkward or overly tangled. The type of behavior I have in mind is adaptive / experience-based decision-making within a single run: agents choose between alternatives (e.g. risky vs safe actions), remember negative outcomes locally, and adjust future decisions based on that memory. This is intentionally below a full behavioral framework: no new APIs, no new abstractions, just a clean example + README that documents the pain points encountered. My goal is to contribute this as a new example, and only then reflect on what it teaches us about states, decisions, action duration, or task structure — feeding that insight back into the broader Behavioral Framework discussion. If this example-first, evidence-driven approach is useful, I’d be happy to iterate further and expand it based on feedback. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
|
Hi everyone I'm Liyue. Based on my research needs and the recent community discussions on Scenarios (#3176), I'm exploring the idea of mesa-calibration. It aims to provide a standardized interface for automated parameter optimization and validation against empirical data. I have posted the technical proposal in a separate discussion here: #3207 I would love to hear any feedback or feature requests from the community! |
Beta Was this translation helpful? Give feedback.
-
Modernizing Mesa-Geo: Refactor and Visualization OverhaulGoal: Refactor Mesa-Geo to align with Mesa's modern architecture, moving from object-based raster storage to high-performance vectorized Phase 1: Core Data Refactor (
|
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH, thanks for starting and actively guiding this discussion. I really appreciate the repeated emphasis on understanding real modeling pain points by building and extending concrete Mesa models before introducing new abstractions. I plan to spend the coming weeks working through existing examples (such as Wolf–Sheep, Schelling, and Sugarscape), experimenting with small extensions or new example models that stress areas like adaptive decision-making, memory, and multi-step actions. The goal is to document where current Mesa primitives feel expressive versus where behavior logic becomes awkward or overly tangled. I also find the Mesa Inference direction very exciting, particularly the idea of formalizing the simulator–inference boundary and supporting SBI/ABC workflows for calibrating ABMs against observed macro data. I’ll aim to contribute incrementally through examples, documentation, or small PRs, and share concrete findings back into this discussion rather than proposing abstractions prematurely. Looking forward to learning from and contributing to the Mesa community. |
Beta Was this translation helpful? Give feedback.
-
|
Hey @EwoutH, the project ideas and descriptions have been really informative. I’ve used Mesa agents before, and the meta-agent project really caught my attention. I’ve gotten started as instructed in the document, and I was wondering if there is a separate discussion thread for the meta-agent project. Also, are there any current issues I could work on? |
Beta Was this translation helpful? Give feedback.
-
|
Hi everyone, I’m planning to submit a GSoC 2026 proposal for the Mesa-LLM iteration to push to production project. I’m a recent Computer Science graduate with experience in AI/ML systems, large language models, and agent-based AI architectures. Recently, I’ve been working on an Agentic AI system where LLMs interact with external tools and APIs to perform multi-step reasoning tasks, which sparked my interest in exploring how similar approaches could be applied to agent-based simulations using Mesa. I’ve started going through the Mesa documentation and exploring the mesa-llm repository to understand how LLM reasoning, memory handling, and API integrations are currently implemented. Before finalizing my proposal, I wanted to ask: • Are there specific issues in the mesa-llm repository that would be good entry points for contributors? Looking forward to learning more about the Mesa ecosystem and contributing. Thanks! |
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH I’ve started exploring the mesa-llm repository to better understand how LLM-driven agents integrate with the Mesa framework. While going through the repository and documentation, I noticed a small issue in the README structure and opened a PR to improve it: The PR fixes a duplicated section and improves the readability of the supported LLM providers section. Next, I’m planning to run and analyze the mesa-llm examples to better understand how reasoning, memory, and communication modules interact with Mesa agents. I’m particularly interested in the “Mesa-LLM iteration to push to production” project and exploring areas like improving documentation, testing coverage, and modularization of LLM tools. Looking forward to contributing more and learning from the community! |
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH and @quaquel! I'm Nouman Hameed, a BSCS student from Pakistan. I've been following the discussion on the Behavioral Framework. Since I have a strong background in Scikit-learn, I'm interested in implementing 'Learning Agents' that use classic ML models (like Random Forest or SVM) to drive decision-making. My goal is to show how we can replace bulky 'if-else' logic in models like Sugarscape or Wolf-Sheep with trained Sklearn estimators. I’ve already started looking at the mesa-examples to identify where these ML baselines would be most effective. Does this 'ML-as-a-Brain' approach align with what you're looking for in the Behavioral Framework project? |
Beta Was this translation helpful? Give feedback.
-
|
Hi @EwoutH and @quaquel , I've spent the past few days building up to the project from first principles: 5 models in Mesa 3.x (Schelling → Wolf-Sheep → Wealth/NetworkGrid → BDI prototype → Q-learning RL agent), with a running log of 15 pain points encountered along the way. The BDI work includes a reusable BDIAgent base class with a full perceive → deliberate → execute cycle. I have pushed it from my own repo and added it in to the format requested in the forked GSoC Learning Space. Key finding while researching: mesa.experimental.actions already covers part of pain point #9 (multi-step plans). My proposal angle is to build the BDI deliberation layer on top of the Actions system rather than replacing it — beliefs, desire selection, and replanning logic that integrates with what's already there. I've also opened two discussions on pain points not yet tracked: partial observability in grid spaces, and IntentionQueue semantics. Happy to hear whether those are in or out of scope for this project. |
Beta Was this translation helpful? Give feedback.
-
|
Hi! I'm Yogin, a 2nd year undergrad interested in the mesa-examples revival project for GSoC 2026. I spent a few days exploring the examples repo and noticed there are zero continuous space examples (the README literally says "No user examples available yet" for that category). So I went ahead and built one — a Brownian particle diffusion model using mesa.experimental.continuous_space. Just opened a PR: The model has particles doing random walks with soft repulsion so they spread out naturally, and colours them based on local density using a plasma colormap. Pretty simple but it shows the full setup: ContinuousSpaceAgent, SolaraViz, DataCollector, Scenario params. For the actual GSoC project I'm thinking about focusing on:
Happy to discuss scope. Is there a particular part of the revival that's highest priority from the team's side right now? |
Beta Was this translation helpful? Give feedback.
-
|
Hi! I'm Alaa, a CS student from Egypt at New Mansoura University, specializing in Artificial Intelligence. I found Mesa through the GSoC 2026 website and I'm really interested in the Mesa Examples Revival project. |
Beta Was this translation helpful? Give feedback.
-
|
Update: I successfully ran the forest_fire example locally with Mesa 3.5.1! I can see the fire spreading and tested different tree density parameters. Looking forward to exploring more examples. |
Beta Was this translation helpful? Give feedback.
-
|
Thank you for the guidance! I have already explored the documentation and
ran the forest_fire example locally. I found two issues: matplotlib is
missing from the dependencies, and the code produces deprecation warnings
about Mesa 4.0. I am now reviewing more examples to document similar issues
before writing my proposal.
…On Tue, Mar 24, 2026 at 3:53 AM Nitesh verma ***@***.***> wrote:
I think it would be helpful to explore the documentation and review a few
example implementations before starting your contribution.
—
Reply to this email directly, view it on GitHub
<#2927?email_source=notifications&email_token=B2ZA4SK3EVCYSSZ2HU5VWML4SHTC5A5CNFSNUABIM5UWIORPF5TWS5BNNB2WEL2ENFZWG5LTONUW63SDN5WW2ZLOOQXTCNRSHA2TKMZQUZZGKYLTN5XKOY3PNVWWK3TUUVSXMZLOOSWGM33PORSXEX3DNRUWG2Y#discussioncomment-16285530>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/B2ZA4SIZS3MAAKG35LJVC2T4SHTC5AVCNFSM6AAAAACOEGUB6GVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTMMRYGU2TGMA>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
|
Hey everyone! I'm Meer. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @colinfrisch , @jackiekazil Hi, I’ve been exploring Mesa-LLM and identified two user-facing issues that seem important for usability and scalability. First, LLM outputs are recomputed on every run, which makes repeated simulations slow for local models, costly for API-based models, and difficult to reproduce or reuse offline. Second, there’s currently limited visibility and control over LLM usage (calls, tokens, latency, cost), which can make larger simulations harder to manage and potentially expensive. To address these, I’m thinking of working on two related features: My goal is to keep both solutions simple and practical so they improve developer experience without adding too much complexity. I’d appreciate your thoughts on whether this direction aligns with the project goals. I would like to structure my proposal around two core execution-layer issues in Mesa-LLM: the lack of reusable record/cache/replay support for LLM outputs, and the lack of built-in usage tracking and budget controls. I think these are both real and impactful problems that affect reproducibility, local inference workflows, cost-awareness, and overall usability. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @jackiekazil, @colinfrisch, and @EwoutH! I am Nazeef Danladi Adamu, a postgraduate student from Nigeria applying for the Mesa-LLM Stabilization project (175 hours). My background is in building AI orchestration systems. I recently built research-dossier-app, a full-stack tool that coordinates multiple LLM capabilities through a directive-execution architecture. The core challenge in that project was making LLM calls reliable enough to embed inside an automated pipeline where a silent hang or an unexpected crash breaks everything downstream. That is exactly what this project addresses at the framework level, which is why it caught my attention immediately. Before writing my proposal I read through the mesa-llm 0.3.0 source code and found five concrete issues:
I posted detailed findings with proposed fixes in mesa-llm/discussions/269. My full proposal covers all five stabilization issues, a Mesa 4.0 migration path with version-conditional imports, and example models drawn from scientific literature that demonstrate LLM agents alongside standard agents. Happy to answer any questions or share the proposal draft before the deadline. Thank you for the work on this project. |
Beta Was this translation helpful? Give feedback.
-
|
Hi, I'm Farseen Shaikh — applying for the Behavioral Framework project. I started with the recommended starter task: a needs-based Wolf-Sheep extension with continuous internal states (hunger, fear, fatigue) driving priority-based decisions. PR: mesa/mesa-examples#456. Friction points documented in #2538. Plan: build 2–3 models (needs-based ✓, RL-based next, optionally BDI), document friction, compare with GAMA's gymnasium wrapper, and propose improvements justified by implementation experience. |
Beta Was this translation helpful? Give feedback.
-
|
Hi everyone! I'm Yadnyesh Awasarkar, a student from India interested in the Mesa Examples Revival project for GSoC 2026. |
Beta Was this translation helpful? Give feedback.
-
|
All we appreciate all the submissions for GSoC. We are now in review mode. I am going to close this thread. @EwoutH feel free to reopen if you think we still need it open. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @jackiekazil, @colinfrisch, and @EwoutH! I am Nazeef Danladi Adamu, a postgraduate student from Nigeria applying for the Mesa-LLM Stabilization project (175 hours). My background is in building AI orchestration systems. I recently built research-dossier-app, a full-stack tool that coordinates multiple LLM capabilities through a directive-execution architecture. The core challenge in that project was making LLM calls reliable enough to embed inside an automated pipeline where a silent hang or an unexpected crash breaks everything downstream. That is exactly what this project addresses at the framework level, which is why it caught my attention immediately. Before writing my proposal I read through the mesa-llm 0.3.0 source code and found five concrete issues:
I posted detailed findings with proposed fixes in mesa-llm/discussions/269. My full proposal covers all five stabilization issues, a Mesa 4.0 migration path with version-conditional imports, and example models drawn from scientific literature that demonstrate LLM agents alongside standard agents. Happy to answer any questions or share the proposal draft before the deadline. Thank you for the work on this project. |
Beta Was this translation helpful? Give feedback.




Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
The idea list with committed mentors is prepared here. See also the Mesa's Google Summer of Code 2026 guide.
Let's start the discussion on 2026 GSoC ideas!
2025 ideas can be found here. Ones leftover from last year:
mesa-geo(https://github.com/projectmesa/mesa-geo) package directly into the core Mesa library as amesa.geomodule, resolving compatibility issues arising from their separate evolution and simplifying dependency management. By leveraging Mesa's new experimental cell and continuous space architectures, the project will create a unified spatial modeling framework that supports GIS functionality, coordinate transformations, and standard file formats like GeoJSON within a consistent API. The consolidation aims to make advanced geospatial modeling a first-class feature, ensuring that property layers and spatial visualizations work seamlessly across all Mesa projects.What more ideas and ambitions do we have?
Beta Was this translation helpful? Give feedback.
All reactions