Agents Commander: A Lightweight Interface for Multi-Agent CLI Workflows

Authors
Lech Kalinowski
Senior AI Systems Engineer
@
Callstack
No items found.

Someone once said: For over 2,000 years, we still haven’t solved communication problems between humans.

In general, that still feels true.

Beyond language barriers, misunderstanding remains one of the most common causes of problems in both business settings, as well as our everyday life. Human spoken language is chaotic. There are whispers, gossip, noise, assumptions, emotions, shortcuts, and context gaps everywhere. If you listen carefully, most sentences are short, imprecise, and heavily dependent on what the speaker already knows, feels, or assumes. The same sentence can be interpreted in completely different ways from what was intended.

Structured communication in AI

In AI systems, we aim to build communication between agents that is not chaotic.

Today, that communication is usually based on hyper-precise structures like JSON. We use MCP, LangGraph, LangChain, APIs, schemas, and formal protocols. That makes sense from an engineering perspective, but I sometimes wonder whether this is also limiting broader research into AGI. Maybe intelligence does not emerge only from rigid structure. Maybe some part of it emerges from imperfect exchange, negotiation of meaning, and adaptation through communication itself.

LLM workflows in CLI environments

In my daily work, I constantly switch between agentic CLIs and apps like Codex and Claude. I use all of them heavily for development and experimentation. However, I'm starting to feel that we are living through something like the early 1990s of computer science again: a new beginning.

LLMs feel like a non-linear phase transition in how humans interact with computers. I remember classical hardware eras: floppy disks, CDs, eject buttons, file managers. Today, I change local LLM models on my DGX almost like swapping floppies or CDs in a drive.

In LM Studio, there is even an eject-style button that looks like it came from an old CD-ROM. There is something strangely old-school and exciting about all of this: fancy but raw CLIs, hacker-like workflows, terminals full of numbers and characters, huge datasets streaming through text interfaces, and hardware that feels almost neural-native. These are all signs that a new computer era has begun.

Agents Commander CLI interface

That is why I decided to vibe-code a classical Norton Commander-style interface, just to regain control over all my agents working through CLI, while keeping visibility over the file structure. I also wanted a practical way to inspect markdown files, skills, and important artifacts without constantly switching windows and losing context.

When I finished the first prototype, I called it Agents Commander. Initially, it was simply a way to manage a multi-agent CLI environment.

But then I had another thought:

If every panel contains a different agent, maybe they should be able to communicate with each other.

That led me to implement what I call the Commander Protocol.

It is an extremely simple way for agents to communicate inside Agents Commander. Once an agent learns the protocol, it knows how to send messages to other agents in nearby panels. The proof of concept relied ona single marker format:

===COMMANDER:SEND:<agent_type>:<panel>=== Please write unit tests for the auth module. ===COMMANDER:END===

I built a lightweight observer to detect these markers, with simple stdout watching and stdin injection orchestration. It worked: Agents from different vendors started communicating with each other.

Agent collaboration and workflow experiments

I couldn’t resist pushing it further, so I wrote a small library of prompts for collaborative tasks. One of my favorite experiments is Philosophical Discussion, where agents debate and refine ideas together. What makes this experiment useful is that it exposes how agents negotiate meaning in practice, not just how they follow predefined structures. You can observe disagreement, refinement, and convergence in real time, which is hard to capture in typical pipeline-based systems. It turns communication itself into something you can test, not just assume.

At some point I realized this might be more than just a handy feature. It could be a unique approach to inter-agent communication.

What fascinates me is that this opens a different path from MCP or LangChain-style orchestration. It allows us to test communication processes directly, build pipelines on top of them, and delegate work across agents without servers, heavy configurations, complex APIs, or rigid schemas.

Agents Commander is still early, but it already solves a real problem for me: managing a multi-agent CLI workflow in a lightweight, visible, and practical way. I deliberately didn’t want the source code to exceed 1.44 MB (boomers know why).

If you are an AI researcher and want to study agent communication through a minimal protocol, or if you want to build agent pipelines in a way that is lighter and more experimental than MCP, then Agents Commander may be interesting to you.

We can train models to not only use communication protocols, but to understand them and adapt them during interactions.

Implications for AGI communication

Maybe this is only an invention. Maybe it is a discovery or an interface pattern. Or maybe it is a tiny clue about how future AGI systems could learn to communicate more naturally? Only time will tell.

Table of contents
Integrating AI into your React Native workflow?

We help teams leverage AI to accelerate development and deliver smarter user experiences.

Let’s chat

//

//
AI

We can help you move
it forward!

At Callstack, we work with companies big and small, pushing React Native everyday.

On-device AI

Run AI models directly on iOS and Android for privacy-first experiences with reliable performance across real devices.

AI Knowledge Integration

Connect AI to your product’s knowledge so answers stay accurate, up to date, and backed by the right sources with proper access control.

Generative AI App Development

Build and ship production-ready AI features across iOS, Android, and Web with reliable UX, safety controls, and observability.

AI Vibe Coding Cleanup

Turn AI-generated code from tools like Cursor, Claude Code, Codex, or Replit into production-ready software by tightening structure, validating safety, and making it stable under real-world usage.

//
Insights

Learn more about AI

Here's everything we published recently on this topic.