Quick feedback about your demo video: I generally quite liked it and it really helped me understand Swarm. Two thoughts:
- you lament the chat interface, but the first 1m30s of the video I only see the chat interface
- your research task is LLM/AI related. There were moments where I found this slightly confusing and I wasn't sure if I was reading about Swarm itself or just its own research. Would recommend something non-LLM related and more generally applicable for the demo video.
I think this is really neat. You should probably take it as a compliment that the biggest criticisms so far are about the website landing page. ;)
I like canvases in general, and I especially like them for mentally organizing and referring to this sort of broad work. (Honestly, I think zoomable canvases would make a better window manager in general, but I digress)
One small piece of friction: My default mouse-based ways of dragging the canvas around (that work in most canvases like Figma) aren't working. I saw that you had a tutorial, and I have learned to hold space now, but I prefer the "hold middle mouse button to drag my canvas view around".
I've got a couple of research tasks running now, and my current open questions as a very new user are:
1) How easy will it be to store the outputs into a Github repository.
2) How easy will it be to refer back to this later?
3) Can I build upon it manually or automatically?
4) Can I (securely) share it with someone else for them to see and build upon it?
5) Can I do something "locally" with it? Not necessarily the model, but my preferred interface for LLMs at this point is Claude Code. Could I have a Claude Code instance running in one of these boxes somehow?
6) What if I want to do private stuff with it and don't like the traffic going through Spine's servers? Could I pay them for the interface, but bring my own keys? (Related: Can I self host somehow?)
7) When this is done, each artifact it found (screenshot, webpage, etc), is going to be helpful. The data-hoarder in me wants to make sure I can search these later. Heck, if I could do that, this would become my preferred "web browser". (But again, I digress.)
Really appreciate the detailed feedback and questions! And yes, we'll take the website criticism as a compliment :)
Good callout on the canvas navigation, we'll look into middle mouse button support.
To answer your questions:
1) GitHub integration is on our roadmap. Right now you can export outputs manually but we want to make this seamless.
2) All your canvases are saved and you can search them by name in your dashboard. We're also working on a dedicated section for deliverables across canvases.
3) Yes to both! You can manually add or edit blocks, or kick off new agent runs that build on existing work.
4) You can currently only share public links of your canvas to others (but you can make it private at any point). We are testing out a teams feature which allows you to share canvases with members on your team securely. Beyond that, we are working on adding roles and email-based sharing controls which is in our roadmap.
5) Claude Code in a block is a really interesting idea. We don't support that today but we're thinking about computer use and coding workflows.
6)BYOK (bring your own keys) is something we've heard interest in and are considering. Self-hosting isn't available right now, though we do support private deployments for enterprise customers if that's ever relevant.
7) Love the 'preferred web browser' framing. Right now you can search canvases but searchable artifacts across canvases is definitely where we want to head.
Thanks for giving it a real spin, this kind of feedback is incredibly valuable.
> And yes, we'll take the website criticism as a compliment :)
ugh. guys. come on. stop celebrating at the 1 yard line. people are telling you they didnt even look at the product becacuse your landing page was so bad. you wasted your launch HN linking directly to it, ofc thats the first thing people are going to give feedback on. fix it right now you still have time.
Got some great results for a rather broad domain in the first pass.
HN is going to tend towards negative/constructive feedback, for me the only issue is that the mouse interaction is a bit wonky. Took me a minute to realize that i could select different mouse modes. With that I'd say I'd echo TheTaytay's comment about mouse interaction and for me generating docx (which was the output of my agents, haven't even explored explicitly asking for something else) creates a bit of a barrier to use the content for me. Markdown or even HTML would be helpful.
But these are just minor nits, love the concept and great execution.
It might just be me, but this interface is the first time I felt the desire to interact with long-running agents even though I use chat interfaces all day long. Maybe it was the demo video on the landing page which was compelling with its examples. Maybe it was the feeling that I could see what was going on because I would be on a canvas. Nicely done!
Off to keep iterating on the prototype app I started...
Calling it a 'canvas' makes me think that this tool is about AI agents doing some kind of collaborative drawing. Looking at the vid though, it seems more like an environment for visually organizing and managing agentic work (which seems very cool, and quite a bit more than just a canvas).
I didn’t read the post, I checked out the website just like 99% of the people will do.
Simple advice, if you are selling a product with a selling point of being visual, show it on your website. Not in a YouTube video but actual screenshots, short cut 10 sec video/gif
I like the overall idea and presentation. In trying it out, I hit the token cap before my trial task was able to complete and show me the end result. I'm sure your free-tier token costs are non-trivial but it was definitely a bummer that I couldn't even see one initial run's output to decide if I wanted to pay.
I decided to gamble the one month fee to let it continue, but the payment defaulting to annual was jarring. I can see it lets you advertise a lower price but that only made me more tempted to leave altogether when I saw the price go up on the final screen.
Congrats on the launch! Meta comment, but I just ain't reading all of the above. You need to be able to explain this in about 20% the number of words or you'll lose people, especially VC.
My advice is to start with "Spine Swarm solves _____" then how, then why you're different. 3 short paragraphs, preferably 1-2 sentences each.
I have not seen my final report yet for my query around machinist.com. However I would like to say my initial impression is very positive. At least in terms of the digestion of my somewhat nebulous request. I like the way your app was able to burrow down to pain points I have experienced and am trying to work out in terms of product market fit for the domain. I look forward to exploring more and giving you more feedback when I see the final report. I will also add that I am looking forward to using your product to explore other opportunities that I'm sure are out there in this age of AI.
I'm completely sold on the canvas layer. Embracing non linearity is such a boon when you're on the ideas stage. When you have verified it though, moving it to another medium (a document, presentation or just code) is often the best choice.
Do you see the canvases created with Spine as "one off" that you discard when you have got your deliverable, or as something living that you keep around?
I'm building a side project for running SQL on a canvas (kavla.dev), so I'm thinking about canvas workflows all the time!
Thanks! Great question. We see canvases as living workspaces, you can revisit, iterate on, and build on them over time.
But the deliverables (docs, slides, code) are first-class outputs you can export and use independently. So it works both ways depending on the workflow.
Kavla looks cool, canvas-based SQL is a great use case for this kind of thinking!
I had to read this text in order to understand what this tool does, because I could not know from the website (without watching a video). You should use Spine to improve your website. ;-)
"I make AI output lots of stuff" is not an intrinsically valuable thing. I can run the same thing on Claude in research mode and get a report with cited sources in a more digestable format on my phone. What's the eval here on if any of this is good? Is it even possible to test (ie, you cant really AB test startup ideas)?
Great question. The core of Spine is coordinating multiple specialized agents across multiple models, using the canvas to store and pass context selectively so each agent works with exactly what it needs.
In the demo video you shared (yt link) how many credits did that whole project take? What is the prices to fix elements of it (for example of you dislike a minor aspect of the generated spreadsheet do follow up instructions utilize only the narrow subset of agents that has been demoed to that subtask, or does it create new agents who have to create new context in the narrow follow up task?)
Credits are consumed by the blocks that get generated, not by the agents themselves. Some blocks are cheaper than others. A simple prompt or image block is a single model call, while browser use or deliverable blocks like documents and spreadsheets run models in a loop and cost more. Blocks also cost more when they have more blocks connected to them (more input tokens).
In the demo video I shared, the task cost about ~7,000 credits since it ran around 10 BrowserUse blocks and produced multiple deliverables.
If you want to fix a specific block (or set of blocks), you can select them and the chat will scope itself to primarily work on those. In that case fewer blocks run, so it's cheaper.
What does it mean to say 30,000 monthly credits and 1500 daily refresh credits? If my project takes 7000 credits (the way your demo does) then does that mean I couldn’t actually do it on the lowest available pricing plan because I couldn’t use 7000 credits in one run? If this is the case, what am abysmal pricing model!
The daily refresh isn't a cap on usage, it's additional credits you get each day (resets to 1,500 nightly regardless of use).
You can use your full 30k balance in a single run if needed. The daily refresh just tops you back up over time so you're not waiting for a monthly reset.
Interesting idea, I wanted to see an example of the agents working on a canvas when I opened your page. I saw nothing of the sort. Sorry, but immediate fail.
This may be too harsh, but you need to make it immediately clear to someone today why they can't just have Claude Code one shot your app!
I read "AI agents that collaborate on a visual canvas" and I thought it was a shared canvas (as in an image) that virtual agents could contribute to, sort of like an image-only Moltbook.
Just as a tiny first piece of feedback, the main marketing website is very hard to understand or grok without a demo of how the tool works. Even just the quick YouTube video that you added in your post here, if embedded, would make a difference.
There are so many "agentic tools" out there that it's really hard to see what differentiates this just based on the website.
excuse my memory at this point, arent there like a 100 of these posted on HN every month that all have something to do with multi agent collaboration that support 1000 models?
Why do I need a canvas to visualize the work that the agents are doing? I don't want to see their thought process, I just want the end product like how ChatGPT or Claude currently work.
That is definitely a valid way of using Spine as well. You can just work in the chat and consume the deliverables similar to how you would in other tools.
The canvas helps when you want to trace back why an output wasn't what you expected, or if you're curious to dig deeper.
Even beyond auditability, the canvas also helps agents do better work: they can generate in parallel, explore branches, and pass context to each other in a structured way (especially useful for longer-running tasks).
whoa congrats on the launch. lol I launched my visual canvas for agents today too. I went in a more of a collaborative canvas IDE, agent orchestration direction. But very cool to see your take on it
Rather than just finding a way to link your own product, why don't you do the rest of us favor and provide a comparison at least, so it becomes a tiny bit informative instead of just spammy?
Nothing wrong with sharing your own stuff, but at least contribute something back to the submission you're commenting on.
We've only partially explored this so far, but it's a great suggestion.
The canvas architecture naturally supports this kind of loop since agents can already read and build on each other's outputs — so the plumbing is there, it's more about building the right orchestration on top. Definitely something we're exploring.
Fair point, we should be more upfront about the sign-up step. Given that tasks are long-running and token-intensive, we do need an auth barrier to protect against abuse, but we can definitely do a better job signaling that before you hit the canvas.
Or, just show us in an animated GIF how the product works in practice. Then, should we somehow find benefit in a visual representation of a swarm's workflow, we could sign up rather than having to, unintuitively, scroll down to watch a YouTube video.
How am I supposed to get anything out of this? Consider that agents are going to get faster and run more and more tasks in parallel. This is not manageable for a human to follow in real time. I can barely keep up with one agent in real-time, let alone a swarm.
What I could see being useful is if you monitored the agents and notified me when one is in the middle of something that deserves my attention.
This is a fair point, we are exploring progressive disclosure on the canvas to better utilize the space and make the key artifacts more readily visible. We do have other panels (the chat, task and deliverable) that have alternate views of what the agent did and the key deliverables.
Beyond human auditability, the canvas helps the agents do a better job by generating in parallel, exploring branches and passing context to each other in a structured way.
Spot on. The persistence layer is a huge part of what makes the canvas work.
For failures, we handle it at multiple levels: first, standard retries and fallbacks to alternate models/providers. If that fails, the agents look for alternate approaches to accomplish the same task (e.g. falling back to web search instead of browser use).
For completeness, you can also manually re-run or edit individual blocks if they fail (though the agents may or may not consider this depending on where they are in their flow).
Great framing. You're right that context fragility is a big part of it. The canvas helps because each block maintains its own context explicitly, and connected blocks pass context between blocks without polluting the agents' context windows.
On conflict resolution, the synthesizer block can see all upstream outputs, so it has full visibility into any divergence. It does surface contradictions to the user, though this is something we're constantly improving.
- you lament the chat interface, but the first 1m30s of the video I only see the chat interface
- your research task is LLM/AI related. There were moments where I found this slightly confusing and I wasn't sure if I was reading about Swarm itself or just its own research. Would recommend something non-LLM related and more generally applicable for the demo video.
Very cool!
I like canvases in general, and I especially like them for mentally organizing and referring to this sort of broad work. (Honestly, I think zoomable canvases would make a better window manager in general, but I digress)
One small piece of friction: My default mouse-based ways of dragging the canvas around (that work in most canvases like Figma) aren't working. I saw that you had a tutorial, and I have learned to hold space now, but I prefer the "hold middle mouse button to drag my canvas view around".
I've got a couple of research tasks running now, and my current open questions as a very new user are: 1) How easy will it be to store the outputs into a Github repository. 2) How easy will it be to refer back to this later? 3) Can I build upon it manually or automatically? 4) Can I (securely) share it with someone else for them to see and build upon it? 5) Can I do something "locally" with it? Not necessarily the model, but my preferred interface for LLMs at this point is Claude Code. Could I have a Claude Code instance running in one of these boxes somehow? 6) What if I want to do private stuff with it and don't like the traffic going through Spine's servers? Could I pay them for the interface, but bring my own keys? (Related: Can I self host somehow?) 7) When this is done, each artifact it found (screenshot, webpage, etc), is going to be helpful. The data-hoarder in me wants to make sure I can search these later. Heck, if I could do that, this would become my preferred "web browser". (But again, I digress.)
Good callout on the canvas navigation, we'll look into middle mouse button support.
To answer your questions: 1) GitHub integration is on our roadmap. Right now you can export outputs manually but we want to make this seamless. 2) All your canvases are saved and you can search them by name in your dashboard. We're also working on a dedicated section for deliverables across canvases. 3) Yes to both! You can manually add or edit blocks, or kick off new agent runs that build on existing work. 4) You can currently only share public links of your canvas to others (but you can make it private at any point). We are testing out a teams feature which allows you to share canvases with members on your team securely. Beyond that, we are working on adding roles and email-based sharing controls which is in our roadmap. 5) Claude Code in a block is a really interesting idea. We don't support that today but we're thinking about computer use and coding workflows. 6)BYOK (bring your own keys) is something we've heard interest in and are considering. Self-hosting isn't available right now, though we do support private deployments for enterprise customers if that's ever relevant. 7) Love the 'preferred web browser' framing. Right now you can search canvases but searchable artifacts across canvases is definitely where we want to head.
Thanks for giving it a real spin, this kind of feedback is incredibly valuable.
ugh. guys. come on. stop celebrating at the 1 yard line. people are telling you they didnt even look at the product becacuse your landing page was so bad. you wasted your launch HN linking directly to it, ofc thats the first thing people are going to give feedback on. fix it right now you still have time.
HN is going to tend towards negative/constructive feedback, for me the only issue is that the mouse interaction is a bit wonky. Took me a minute to realize that i could select different mouse modes. With that I'd say I'd echo TheTaytay's comment about mouse interaction and for me generating docx (which was the output of my agents, haven't even explored explicitly asking for something else) creates a bit of a barrier to use the content for me. Markdown or even HTML would be helpful.
But these are just minor nits, love the concept and great execution.
Off to keep iterating on the prototype app I started...
Simple advice, if you are selling a product with a selling point of being visual, show it on your website. Not in a YouTube video but actual screenshots, short cut 10 sec video/gif
I decided to gamble the one month fee to let it continue, but the payment defaulting to annual was jarring. I can see it lets you advertise a lower price but that only made me more tempted to leave altogether when I saw the price go up on the final screen.
My advice is to start with "Spine Swarm solves _____" then how, then why you're different. 3 short paragraphs, preferably 1-2 sentences each.
I'm completely sold on the canvas layer. Embracing non linearity is such a boon when you're on the ideas stage. When you have verified it though, moving it to another medium (a document, presentation or just code) is often the best choice.
Do you see the canvases created with Spine as "one off" that you discard when you have got your deliverable, or as something living that you keep around?
I'm building a side project for running SQL on a canvas (kavla.dev), so I'm thinking about canvas workflows all the time!
But the deliverables (docs, slides, code) are first-class outputs you can export and use independently. So it works both ways depending on the workflow.
Kavla looks cool, canvas-based SQL is a great use case for this kind of thinking!
On the eval side, we ran Spine Swarm against GAIA Level 3 and Google DeepMind's DeepSearchQA and hit #1 on both.Full writeup: https://blog.getspine.ai/spine-swarm-hits-1-on-gaia-level-3-...
In the demo video I shared, the task cost about ~7,000 credits since it ran around 10 BrowserUse blocks and produced multiple deliverables.
If you want to fix a specific block (or set of blocks), you can select them and the chat will scope itself to primarily work on those. In that case fewer blocks run, so it's cheaper.
You can use your full 30k balance in a single run if needed. The daily refresh just tops you back up over time so you're not waiting for a monthly reset.
This may be too harsh, but you need to make it immediately clear to someone today why they can't just have Claude Code one shot your app!
There are so many "agentic tools" out there that it's really hard to see what differentiates this just based on the website.
The canvas helps when you want to trace back why an output wasn't what you expected, or if you're curious to dig deeper.
Even beyond auditability, the canvas also helps agents do better work: they can generate in parallel, explore branches, and pass context to each other in a structured way (especially useful for longer-running tasks).
https://getmesa.dev is mine
Nothing wrong with sharing your own stuff, but at least contribute something back to the submission you're commenting on.
The canvas architecture naturally supports this kind of loop since agents can already read and build on each other's outputs — so the plumbing is there, it's more about building the right orchestration on top. Definitely something we're exploring.
e: 'be' to 'we'; oops.
How am I supposed to get anything out of this? Consider that agents are going to get faster and run more and more tasks in parallel. This is not manageable for a human to follow in real time. I can barely keep up with one agent in real-time, let alone a swarm.
What I could see being useful is if you monitored the agents and notified me when one is in the middle of something that deserves my attention.
Beyond human auditability, the canvas helps the agents do a better job by generating in parallel, exploring branches and passing context to each other in a structured way.
For failures, we handle it at multiple levels: first, standard retries and fallbacks to alternate models/providers. If that fails, the agents look for alternate approaches to accomplish the same task (e.g. falling back to web search instead of browser use).
For completeness, you can also manually re-run or edit individual blocks if they fail (though the agents may or may not consider this depending on where they are in their flow).
On conflict resolution, the synthesizer block can see all upstream outputs, so it has full visibility into any divergence. It does surface contradictions to the user, though this is something we're constantly improving.
https://news.ycombinator.com/newsguidelines.html