Toolkit: Complex Systems Toolkit.

Author: Dr. Stuart Grey, SFHEA (University of Glasgow).

Topic: Student created interactive simulation for complex sociotechnical systems.

Title: LLM-driven interactive simulation for complex sociotechnical systems.

Resource type: Teaching activity.

Relevant disciplines: Any.

Keywords: Artificial Intelligence; Large Language Model; Sociotechnical systems; Ethics; Modelling or simulation; Emergence.

Licensing: This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. It is based upon the author’s article “Enhancing Ethical Reasoning in Engineering Education through Student-Created Interactive Ethical Scenarios Using Generative AI,” 2025 IEEE Global Engineering Education Conference (EDUCON), London, United Kingdom, 2025, pp. 1-5, doi: 10.1109/EDUCON62633.2025.11016531. 

Downloads: 

Related INCOSE Competencies: Toolkit resources are designed to be applicable to any engineering discipline, but educators might find it useful to understand their alignment to competencies outlined by the International Council on Systems Engineering (INCOSE). The INCOSE Competency Framework provides a set of 37 competencies for Systems Engineering within a tailorable framework that provides guidance for practitioners and stakeholders to identify knowledge, skills, abilities and behaviours crucial to Systems Engineering effectiveness. A free spreadsheet version of the framework can be downloaded.

This resource relates to the Systems Thinking, Life Cycle, Configuration Management, Requirements Definition, Verification, and Validation INCOSE Competencies. 

AHEP mapping: This resource addresses several of the themes from the UK’s Accreditation of Higher Education Programmes fourth edition (AHEP4):  Analytical Tools and Techniques (critical to the ability to model and solve problems), and Integrated / Systems Approach (essential to the solution of broadly-defined problems). In addition, this resource addresses AHEP themes of Ethics and Communication. 

Education level: Intermediate.

 

Learners have the opportunity to: 

Teachers have the opportunity to: 

 

Overview:

This resource enables engineering students to create, run, and debug a textbased, interactive simulation of a complex sociotechnical system using a Large Language Model (LLM). It is intentionally flexible and may be delivered as a multisession studio activity (including assessment) or used solely as a compact assessment.

  

Purpose and use:

In both modes, students design a robust text prompt, test it with a user, document changes, and submit auditable artefacts that evidence learning. The key activity is interrogating their own thinking on how complex systems should be modelled by making judgements as to how their game does and does not capture the system dynamics. 

 

Why and how: 

The approach aims to give students hands-on experience in putting systems thinking into practice. Concepts such as stakeholders, feedback loops, delays, uncertainty, and emergent behaviour can be implemented and interrogated without heavy tooling.  

The submission is a text LLM prompt with tracked changes, which allows students to demonstrate system design and debugging, produce transparent process evidence, and scale to large cohorts with minimal infrastructure. 

 

Delivery options at a glance:

Audience Undergraduate Years 2–4 and taught MSc, any engineering discipline 
Modes Studio activity (3–5×2 h + independent study) or Assessmentonly (promptonly; 1–2×2 h + 4–6 h)
Teams 3–4 students (solo permitted for assessmentonly) 
Assessment Portfolio (studio) or promptpluschangelog (assessmentonly) 
Platforms Institutional Copilot licences successful; encourage exploration of free tools (students record model/version)

 

Materials and software:

 

Delivery modes:

Mode A — Studio activity (3–5 sessions) 

Mode B — Assessmentonly (promptonly; 1–2 sessions) 

In both modes, module leaders may supply a predefined scenario(s) to standardise scope and simplify marking. A readytouse example is provided in Appendix C. 

 

Assessment:

Studio portfolio — rubric (suggested weighting):

Criterion  D–E 
Complexity modelling  Clear boundary; rich stakeholders; ≥4 correct loops; delays explicit; coherent KPIs  Mostly sound  Basic map  Superficial  25 
Simulation design and prompt quality  Consistent state logic; visible feedbacks/delays; nonlinearity; negative choices allowed with consequences; clear commands  Mostly coherent  Playable but brittle  Confusing/linear  25 
Debugging evidence  Systematic playtests; clear issue → fix → retest artefacts  Some iteration  Minimal  None  20 
Insight and reflection  Deep analysis of emergence, tradeoffs, equity, uncertainty, and LLM limits  Good  Descriptive  Vague  20 
Communication and referencing  Clear, concise, correct Harvard referencing  Minor issues  Adequate  Disorganised  10 

 

Assessment‑only (prompt‑only) — compact rubric: 

 

Scenario options: 

Students may propose their own topic or the module leader may supply a predefined scenario. Options suited to UK engineering contexts include: 

 

Appendix A — Prompt template (simulation + debugready): 

Title: Complex Systems Simulator — [Scenario] 

Purpose: Run a turnbased interactive simulation of a complex sociotechnical system. Track named state variables, apply feedback and delays, and let the player’s decisions drive nonlinear outcomes. 

Setup: 

  1) Offer three roles (distinct authority/constraints). 

  2) Introduce 3–5 NPCs with clear goals and plausible interventions. 

  3) Show a dashboard of [STATE_VARIABLES] each turn with short context. 

State rules: 

Commands: status, talk [npc], inspect [asset], implement [policy], pilot [intervention], advance time, review log. 

Debug commands (for testing): trace on/off (print update logic), why (state which loops/delays drove the change), show variables (print current state table), revert (roll back one turn), reseed (slight exogenous shock). 

Realism and ethics: Allow all plausible actions and report consequences neutrally. If unsafe in the real world, refuse, propose safer alternatives, and continue with plausible systemic effects. 

LLM pitfalls to avoid: Do not invent new variables; ask clarifying questions rather than guessing; keep outputs concise; summarise trajectory every five turns. 

Begin: Greet the player, state the scenario, ask for a role, and wait. 

 

Appendix B — Debugging and playtest checklist: 

Functional coherence 

Robustness 

User experience and clarity 

Report 

 

Appendix C — Predefined scenario (Urban Heatwave Response, UK city): 

Boundary: One UK local authority area during the July–August heatwave period. Focus on public health, energy demand, and community resilience. 

Roles: (1) Local Authority Resilience Lead; (2) NHS Trust Capacity Manager; (3) Distribution Network Operator (DNO) Duty Engineer. 

Stakeholders: Residents (with a focus on vulnerable groups), care homes, schools, SMEs, DNO, local NHS Trust, emergency services, voluntary/community groups, Met Office (for alerts), and local media. 

State variables (examples): Heathealth alert level (0–4); Emergency Department occupancy (%); Electricity demand/capacity (% of peak); Indoor temperature exceedance hours (hrs > 27 °C); Public trust (0–100); Budget (£); Equity index (0–100). 

Events/shocks: Red heat alert; substation fault; procurement delay; misinformation spike on social media; transport disruption; community centre cooling failure. 

KPIs and stop conditions: Heatrelated admissions; unserved energy; cost variance; equity gap across wards. Stop if alert level 4 persists >3 days, budget overspends >10%, or trust <25. 

Notes for assessors: Using a standard, predefined scenario simplifies marking and ensures comparable complexity across teams, while still allowing for diverse strategies and outcomes. 

 

Any views, thoughts, and opinions expressed herein are solely that of the author(s) and do not necessarily reflect the views, opinions, policies, or position of the Engineering Professors’ Council or the Toolkit sponsors and supporters.  

Let us know what you think of our website