I built Sandclaw (https://github.com/qix/sandclaw) to test an idea of having a human in the loop on any write path that the agent might do. The core agent has access to read anything, but other than to the LLM provider that information is securely firewalled off.
Over time I can loosen the firewall, and allow safer actions (i.e. adding a task to my todo list is always safe), and perhaps build smarter checks to allow more autonomy. But the `DEFAULT DENY` to sending/deleting emails, browsing untrusted websites, or running shell commands.
Is there any project that actually does this? I've been searching but it looks like all the options focus on letting the agent handle when to ask, which feels... problematic.
I wanted to try and see if it was possible to get somewhat decent engine based on signs of AI writing from wikipedia themselves.
It runs entirely in the browser via WASM. Added instructions to further add more ways to figure out as we find out more about LLMs.
Would love feedback!!, especially: - cases where it completely fails - patterns you think are stronger indicators - whether this approach is even worth pursuing vs proper ML models, considering this is a classical classification task, I just don't like the idea of detection of a black box with another black box.
I have connect to IB, but it always make mistake over and over again. I hope to perfect, but it always make mistake, and I need it working OK so I can recover loss.
Where I looking I see no code for this. Anyones have code for this?