Does this need depth data capture as well? The “casual captures” makes it seem like it only needs images, but apparently they are using depth data as well
I think it does use depth data from parameters in docs: python infer_shape.py --input_pkl <sample.pkl> (possibly achievable using software like MapAnything). I believe CUDA only.
Yeah they confirm that at the bottom of the linked page
> Furthermore, by leveraging tools like MapAnything to generate metric points, ShapeR can even produce metric 3D shapes from monocular images without retraining.
No affiliation and didn't get around to use it yet, but clawdbot[1] does WhatsApp, Telegram, Slack, Discord, Signal, iMessage, Microsoft Teams, and WebChat.
I wrote an mcp for claude code to talk to you (and receive messages back) over slack, though it has the caveat of being unable to initiate a session. It was pretty easy to do with claude but probably not the best approach.
Why aren't you using the tools we already have: ansible, salt, chef, puppet, bcfg2, cfengine... every one of which was designed to do systems administration at scale.
I mean, both, but in this case I'm saying "don't use it to access any kind of production resource", with a side order of "don't rely on simple sandboxing (e.g. command patterns) to prevent things like database deletions".
How do you provide db access? For example, to access an RDS db, you have to connect from within the AWS/EC2 environment, which means either providing the agent ssh access to a server, from which it can run psql, or creating a tunnel
Additionally, with multiple apps/dbs, that means having to do the setup multiple times. It would be nice to be able to only configure the agent instead of all the apps/dbs/servers
This is probably the safest thing to do, also the most time consuming
It would be nice to just be able to solve it through instructions to the agent, instead of having to apply all the other things for each application/server/database that I'd like to give it access to
The restrictions have to be enforced by the non-LLM deterministic control logics (in the OS/database/software, or the agent's control plane). It cannot be just verbal instructions and you expect the LLM not to generate certain sequences of tokens.
What I imagine is you might instruct an agent to help you set up the restrictions for various systems to reduce the toil. But you should still review what the agent is going to do and make sure nothing stupid is done (like: using regexes to filter out restricted commands).
Shouldn't you already be using low privilege accounts for stuff like gathering information about prod?
Overprivileged accounts is a huge anti-pattern for humans too. People make mistakes. Insider threats happen. Part of ops is making it so users don't have privileges to do damage without appropriate authorization.
Yeah but this is like exposing `sudo eval $input` as a web service and asking the clients to please, please, not do anything bad.
Can create scripts or use stuff like Nix, Terraform, Ansible or whatever to automate the provisioning of restricted read only accounts for your servers and DBs.
I don't know; I've never done something like that. If no one else answers, you can always ask Claude itself (or another chatbot). This kind of thing seems tricky to get right, so be careful!
If you control the ssh server it can be configured to only allow what you want. Certainly tedious but I would consider it worth while as it stands with agents being well, agentic.
Would love something like this for Fusion 360. Being able to just prompt the UI to create or edit objects. It would be cool if (like with coding agents in which you can add context using @filepath), you could use the mouse to click/select context objects for the prompt to execute with
Also, can it run on Apple silicon?
reply