Built it myself. The filing structure is somewhat consistent enough to parse programmatically, then an LLM handles the unstructured parts (with a lot of trial and error)
>If instead of a copy-pasting spree, or setting up a whateverClaw, the user might just click a button in Chrome, it could be actually useful. (Consider a dozen such buttons.)
isn't this basically just putting a decision tree on top of the llms?
We have our doubts about this. Can you share your code or product?
Anecdotally, my mistakes and lack of understanding exponentiate the more I try to parallelize.
As I said in the neighboring comment, for vibe coding side projects and prototypes for work I just merge and iterate. It works out more than it doesn’t. For anything bigger at work I cannot share as I’m at Apple.
But you have to keep it in your head, and remember all stuff at the same time. How is it possible to track, and do reviews one after another? Or are these pretty long running agents?
I’m not sure what you mean by keep it in your head? I know all of the parts the agents are working on. It’ll often be a mix between bigger tasks (some large refactor, new feature, etc) and small tasks (little bug fixes).
For prototyping I just merge. I don’t bother to review the code. For anything more important than I am reviewing the code and going back and forth. Basically there’s a queue of stuff demanding my attention, and I just serially go through them.
What’s also been really helpful to me is /simplify and similar code review skills (I have my own). That alone takes an agent a while to parse through everything it’s done and self reviews. It catches quite a lot itself this way.
>I’m not sure what you mean by keep it in your head?
If the project I work on is large enough, it takes me some time to get everything I need to understand for review into the short term memory. If it's small enough, it's less of a problem for me.
If you've ever been in a meeting with multiple L8's arguing over features, you should be able to estimate how much each hour of that meeting is costing the org.
how do you justify the compute investment for something like nemotron ? especially if all the labs are willing to pay for those same GPU clusters for inference or training runs?
> One can envision a world in which OpenAI pays chefs money to cook while ChatGPT watches—narrating their thought process, tasting the dishes, and describing the results. This information could be used for general-purpose training, but it might also be packaged as a “book”, “course”, or “partner” someone could ask for.
So we're speed running the idea of AI Facebook friends and creating a new para(ai)social relationship
because it's a separate context window, it makes the model bigger, that space is not accessible to the "user".
And the "language understanding" basically had to be done twice because it's a separate input to the transformer so you can't just toss a pile of text in there and say "figure it out".
so we are currently in the era of one giant context window.
reply