Limits to Control
Can we keep enough control over AI? If systems are developed to be more and more autonomous, this is no longer a given. It's a hypothesis, calling for serious investigation.
Even alignment relies on control. Researchers build mechanisms to control AI's impacts in line with human values.
So how must a control mechanism operate? What limits its capacity to track and correct all the AI signals/effects? Does it provide enough stability, or give way eventually to runaway impacts?
We explore these questions in a new field: Limits to Control.
About the workshop
This in-person workshop is meant to facilitate deep collaboration. We're bringing together researchers to map the territory of AI control limitations – to understand the dynamics, patterns, and impossibilities of control. We are thrilled to welcome Roman Yampolskiy, Anders Sandberg, Forrest Landry, and other researchers working on control limitations.
This workshop aims to:
- Build common knowledge among researchers on the limits to AI control.
- Facilitate high-fidelity discussions through whiteboard sessions and collaborative problem-solving.
- Identify and clarify viable directions in control research by establishing boundaries on controllability.
- Formalize and elevate this critical research topic.
Dates & location
- Date: Wednesday, June 11 - Friday, June 13, 2025
- Location: University of Louisville, Kentucky, USA.
Detailed logistical information will be provided to confirmed participants.
Agenda
We aim to strike a balance between structured sharing and messy exploration – we believe this is where the best ideas tend to emerge. Over the three days, we will do:
- Talks: Researchers present their current work, keeping it under an hour.
- Discussions: We break out into groups to discuss specific questions and bounce ideas around.
- Regrouping: We come back into one room to synthesize what came out of the discussions.
- Next steps: On the final day, we'll plan further development of this research agenda.
Sessions
Sessions will include:
- Anders Sandberg's talk on theoretical limits to control: "Do any of them actually tell us anything?"
- Forrest Landry's whiteboard chat on an overlooked dynamic: "Virtual machines in recursive feedback"
- Richard Everheart's logical framework for AI alignment, aiming to refine foundational understanding.
- Thibaud Veron's session on a framework and engineered toy models that illustrate control dynamics.
- Will Petillo's session on better communication and narrative framings for AI safety/control concepts.
(More details on specific talks and activities will be added as confirmed.)
Proceedings, post-workshop outputs
This section will be updated after the workshop with summaries, key insights, and any public materials generated.
Potential outputs may include:
- Summary report of discussions, key agreements, and open questions
- Notes or photos from sessions
- Links to research papers, blog posts, or pre-prints influenced by the workshop
- A refined list of open research questions in the "limits to control" domain
- Presentations or slide decks (if speakers consent to public sharing)
Join us
This workshop is for researchers actively working on or deeply interested in the theoretical and practical limits of AI control. Do you wish to contribute to these focused discussions? Email Orpheus at o@horizonomega.org to express your interest.
Costs & funding: Participants are generally expected to cover their own travel and accommodation. We can reimburse only some whose research is not yet funded. The workshop has a grant offer from Survival and Flourishing Fund.
To prepare: Read work by participants you are curious to chat with. Then we share some understanding already going in. Most of our time will be in collaborative discussions, so consider where you could bring in specific problems or concepts.
Suggested reading
Writings by participating researchers:
Papers:
- On the Controllability of Artificial Intelligence: An Analysis of Limitations, by Roman Yampolskiy
- AI: Unexplainable, Unpredictable, Uncontrollable, by Roman Yampolskiy
- Impossibility Results in AI, by Roman Yampolskiy
- [Forthcoming paper on control limits] by Anders Sandberg, Aybars Kocoglu, Thibaud Veron
- [Forthcoming paper on a logical framework] by Richard Everheart
Essays:
- Lenses of Control, by Will Petillo
- The Control Problem: Unsolved or Unsolvable?, by Remmelt Ellen
- Control as a Causative Feedback Process, by Forrest Landry
- An Exploration of AGI Uncontainability, by Forrest Landry
- On Error Detection, by Forrest Landry
Sorted roughly by ease of reading. To clarify an argument, do reach out to authors. Authors value questions!
Organizing team
This event is hosted by HΩ, and organized by:
- Orpheus Lummis (HΩ)
- Remmelt Ellen
- Thibaud Veron
Contact
For inquiries regarding the workshop, please contact Orpheus at o@horizonomega.org.