A workflow I find useful is to have multiple CLI agents running in different Tmux panes and have one consult/delegate to another using my Tmux-CLI [1] tool + skill. Advantage of this is that the agents’ work is fully visible and I can intervene as needed.
Thank you so much for sharing Denis! I definitely believe in the that as the world start switching from single agent to agentic teams where each agent does have specific capabilities. do you know of any benchmarks that covers collaborative agents ?
Why make it a vscode extension if the point of these 3 tools is a cli interface? Meaning most of the people I know use these tools without VSCode. Is VSC required?
> Meaning most of the people I know use these tools without VSCode.
I guess it depends?
You can usually count on Claude Code or Codex or Gemini CLI to support the model features the best, but sometimes having a consistent UI across all of them is also nice - be it another CLI tool like OpenCode (that was a bit buggy for me when it came to copying text), or maybe Cline/RooCode/KiloCode inside of VSC, so you don't also have to install a custom editor like Cursor but can use your pre-existing VSC setup.
Okay, that was a bit of a run on sentence, but it's nice to be able to work on some context and then to switch between different models inline: "Hey Sonnet, please look at the work of the previous model up until this point and validate its findings about the cause of this bug."
I'd also love it if I could hook up some of those models (especially what Cerebras Code offers) with autocomplete so I wouldn't need Copilot either, but most of the plugins that try to do that are pretty buggy or broken (e.g. Continue.dev). KiloCode also added autocomplete, but it doesn't seem to work with BYOK.
https://github.com/just-every/code "Every Code - push frontier AI to it limits. A fork of the Codex CLI with validation, automation, browser integration, multi-agents, theming, and much more. Orchestrate agents from OpenAI, Claude, Gemini or any provider." Apache 2.0 ; Community fork;
My similar workflow within Claude Code when it gets stuck is to have it consult Gemini. Works either through Gemini CLI or the API. Surprisingly powerful pattern because I've just found that Gemini is still ahead of Opus in architectural reasoning and figuring out difficult bugs. https://github.com/raine/consult-llm-mcp
Interesting indeed but would it behave the same as Claude code or will it have its own behavior, I think the system prompt is one of the key things that differentiate every agent
Will give it a look indeed, I think one of the challenges with the MCP approach is that the context need to be passed and that would add to the overhead of the main agent. Is that right?
Don’t quote me, but I think the other methods rely on passing general detail/commands and file paths to Gemini to avoid the context overhead you’re thinking about.
I have been using it for some time and it getting better and better with time in many cases it’s giving better output than other tools the comparison is great feature too keep up the good work
> Is multi-agent collaboration actually useful or am I just solving my own niche problem?
I often write with Claude, and at work we have Gemini code reviews on GitHub; definitely these two catch different things. I'd be excited to have them working together in parallel in a nice interface.
If our ops team gives this a thumbs-up security wise I'll be excited to try it out when back at work.
This is very useful! I frequently copy the response of one model and ask another to review it and I have seen really good results with that approach.
Can you also include Cursor CLI for the brainstorming? This would allow someone to unlock brainstorming with just one CLI since it allows to use multiple models.
Will do. I was thinking of also making the LLMs configurable across the agents. I saw a post from the founder of openrouter that you can use DeepSeek with Claude code and was thinking of making it possible to use more LLMs across agents
Regarding DeepMyst. In the future will offer “optionally” the ability to use smart context where the context will be automatically optimized such that you won’t hit the context window limit “ basically no need for compact” and you would get much higher usage limits because the number of tokens needed will be reduced by up to 80% so you would be able to achieve with a 20 USD claude plan the same as the Pro plan
I think the main difference is that Mysti consults with agents rather than the underlying LLM and in the future potentially the agents can switch LLMs as well
Have you tried executing multiple agents on one model with modified prompts and have them try to reach consensus?
That may solve the original problem of paying for three different models.
That sounds like it could get expensive?
A workflow I find useful is to have multiple CLI agents running in different Tmux panes and have one consult/delegate to another using my Tmux-CLI [1] tool + skill. Advantage of this is that the agents’ work is fully visible and I can intervene as needed.
[1] https://github.com/pchalasani/claude-code-tools?tab=readme-o...
I will look it up indeed
Multi agent collaboration is quite likely the future. All agents have blind spots, collaboration is how they are offset.
You may want to study [1] - this is the latest thinking on agent collaboration from Google.
[1] https://www.linkedin.com/posts/shubhamsaboo_we-just-ran-the-...
Thank you so much for sharing Denis! I definitely believe in the that as the world start switching from single agent to agentic teams where each agent does have specific capabilities. do you know of any benchmarks that covers collaborative agents ?
Why make it a vscode extension if the point of these 3 tools is a cli interface? Meaning most of the people I know use these tools without VSCode. Is VSC required?
> Meaning most of the people I know use these tools without VSCode.
I guess it depends?
You can usually count on Claude Code or Codex or Gemini CLI to support the model features the best, but sometimes having a consistent UI across all of them is also nice - be it another CLI tool like OpenCode (that was a bit buggy for me when it came to copying text), or maybe Cline/RooCode/KiloCode inside of VSC, so you don't also have to install a custom editor like Cursor but can use your pre-existing VSC setup.
Okay, that was a bit of a run on sentence, but it's nice to be able to work on some context and then to switch between different models inline: "Hey Sonnet, please look at the work of the previous model up until this point and validate its findings about the cause of this bug."
I'd also love it if I could hook up some of those models (especially what Cerebras Code offers) with autocomplete so I wouldn't need Copilot either, but most of the plugins that try to do that are pretty buggy or broken (e.g. Continue.dev). KiloCode also added autocomplete, but it doesn't seem to work with BYOK.
Very true, I like the fact that I can now use them with a consistent UI, shared context and ability to brainstorm
Will definitely try to add those features in a future release as well
That’s a great idea! I can make it a CLI too
Anyone knows of something similar but for terminal?
Update:
I've already found a solution based on a comment, and modified it a bit.
Inside claude code i've made a new agent that uses the MCP gemini through https://github.com/raine/consult-llm-mcp. this seems to work!
Claude code:
Now let me launch the Gemini MCP specialist to build the backend monitoring server:
gemini-mcp-specialist(Build monitoring backend server) ⎿ Running PreToolUse hook…
https://github.com/just-every/code "Every Code - push frontier AI to it limits. A fork of the Codex CLI with validation, automation, browser integration, multi-agents, theming, and much more. Orchestrate agents from OpenAI, Claude, Gemini or any provider." Apache 2.0 ; Community fork;
When you say orchestrate agents then what it would do? Would it allow the same context across agents and can I make agents brainstorm?
My similar workflow within Claude Code when it gets stuck is to have it consult Gemini. Works either through Gemini CLI or the API. Surprisingly powerful pattern because I've just found that Gemini is still ahead of Opus in architectural reasoning and figuring out difficult bugs. https://github.com/raine/consult-llm-mcp
This is one of the reasons I actually built it but wanted to make it more generalized to work with any agent and on the same context without switching
I like this solution that you can ask Gemini
Any other ideas that you think would make it more powerful?
Perhaps that you can tell it to "use gemini for task x, claude for task y" as sub-agents.
How about adding the ability to tag an agent. for example:
@gemini could you review the code and then provide a summary to @claude?
@claude can you write the classes based on an architectural review by @codex
What do you think? Does that make sense ?
http://opencode.ai/
Interesting indeed but would it behave the same as Claude code or will it have its own behavior, I think the system prompt is one of the key things that differentiate every agent
I do not understand your question. Even in Claude code you have access to multiple models. You can have one critique the other.
I can make it for the terminal if that would be helpful, what do you think?
Pal MCP (formerly Zen) is pretty awesome.
https://github.com/BeehiveInnovations/pal-mcp-server
Will give it a look indeed, I think one of the challenges with the MCP approach is that the context need to be passed and that would add to the overhead of the main agent. Is that right?
The CLINK command will spawn separate CLI.
Don’t quote me, but I think the other methods rely on passing general detail/commands and file paths to Gemini to avoid the context overhead you’re thinking about.
I have been using it for some time and it getting better and better with time in many cases it’s giving better output than other tools the comparison is great feature too keep up the good work
> Is multi-agent collaboration actually useful or am I just solving my own niche problem?
I often write with Claude, and at work we have Gemini code reviews on GitHub; definitely these two catch different things. I'd be excited to have them working together in parallel in a nice interface.
If our ops team gives this a thumbs-up security wise I'll be excited to try it out when back at work.
Would love to hear your feedback! Please let me know if I can make it any better or if there is anything that would make it very useful
This is very useful! I frequently copy the response of one model and ask another to review it and I have seen really good results with that approach.
Can you also include Cursor CLI for the brainstorming? This would allow someone to unlock brainstorming with just one CLI since it allows to use multiple models.
I’m planning to add Cursor and Cline in the next major release, will try to get in out in Jan
Please also add qwen cli support
Will do. I was thinking of also making the LLMs configurable across the agents. I saw a post from the founder of openrouter that you can use DeepSeek with Claude code and was thinking of making it possible to use more LLMs across agents
> Would love feedback on the brainstorm mode. Is multi-agent collaboration actually useful or am I just solving my own niche problem?
If it's solving even your own niche problem, it is actually useful though right? Kind of a "yes or yes" question.
True and hearing feedback is always helpful and helps validate if it is a common problem or not
Any benchmarks? For example vs a single model?
It would be great if the community can run some benchmarks and post it on the repo, planning to do that sometime in Jan
Why limit to 2 agents? I typically use all 3.
Planning to make it work without that limit, did that to avoid complexity but contributions are welcome
I think once I add cursor and cline then will also try to make it work with any number of agents
licensing with BSL when basically every month the AI world is changing is not a smart decision.
> licensing with BSL when basically every month the AI world is changing is not a smart decision
This turned me off as well. Especially with no published pricing and a link to a site that is not about this product.
At minimum, publish pricing.
Regarding DeepMyst. In the future will offer “optionally” the ability to use smart context where the context will be automatically optimized such that you won’t hit the context window limit “ basically no need for compact” and you would get much higher usage limits because the number of tokens needed will be reduced by up to 80% so you would be able to achieve with a 20 USD claude plan the same as the Pro plan
It is free and open source. Will make it MIT
Thinking of switching to MIT, what do you think? Is there any other license you would recommend ?
AGPL, it requires anyone who creates a derivative to publish the code of said derivative.
This reminds me a lot of eye2.ai, but outside of coding
I will check it out indeed. What is common between the two?
I guess both consult multiple llms and draw conclusion from them to cover blindspots
I think the main difference is that Mysti consults with agents rather than the underlying LLM and in the future potentially the agents can switch LLMs as well