Hacker Newsnew | past | comments | ask | show | jobs | submit | more ezyang's commentslogin

But only Claude Desktop gets flat $20 pricing from Claude Pro lol


And with Claude Code, it's typically not $0.08... it's more like $0.50, $5.00 just for a roll LOL. Variable rewards gambling addiction? Definitely...


I was thinking cursor pricing. It becomes a whole different ballgame when you plug these tools into the providers API and pay by the token. Suddenly you really start evaluating how much value you are actually getting out of the tool!


If you'd done it with MCP it would only have cost you $20 and you would still have had the rest of the month to use your Claude Pro sub :P


This is true! :)


You can try (self promo) https://github.com/ezyang/codemcp . https://github.com/rusiaaman/wcgw is also quite popular, although they allow unrestricted shell access (that's why it's named wcgw lol).


Thanks! Does codemcp support having the server on another machine? Maybe communicate over ssh?


Not built in, you'll have to use something like https://github.com/sparfenyuk/mcp-proxy


There's also some fundamental limitations to the Desktop MCP experience that are probably never getting fixed; Claude Code can spin off subagents and play around with the context, I assume that Claude Desktop's form factor is basically going to stay the way it is until the end of time lol.


IMO, the big problem with Aider is that it's not agentic. This is good because it means costs are down, but most of the edit-test-fix loop magic in coding agents comes from the agent loop.


There's a few coding MCPs out there. I have also written one (codemcp) and the pitch for mine is that it DOESN'T provide a bash tool by default and checkpoints your filesystem edits every change in Git, so that it's all about feeling comfortable with letting the agent run to completion and then only inspect the final result. The oldest one in the space, I think, is wcgw.


The llama source code in the original repo has been updated for llama 2: https://github.com/facebookresearch/llama


do you know if llama.cpp will work out of the box or do we need to wait for the code to be updated?


https://github.com/ggerganov/llama.cpp/issues/2262

Likely needs to be updated

Edit: Only the case for the 34B and 70B models. 7B and 13B run as-is.

You can download the GGML model already

https://huggingface.co/TheBloke/Llama-2-7B-GGML

https://huggingface.co/TheBloke/Llama-2-13B-GGML


If you want to try this in Python, you can use https://github.com/ezyang/expecttest which I wrote to do expect tests in PyTorch.


My go-to library in python for this is:

https://pypi.org/project/pytest-regressions/

It's a bit different in that it'll save the expected to a different file... IMHO that's usually nicer because the test result is usually big and having it separated makes more sense.

When rerunning it's possible to run pytest with '--force-regen' and then check the git diff to see if all the changes were expected.


It's especially bad for rr, since it doesn't otherwise have any reason to talk to the Internet (I see people mentioning Firefox telemetry, but you know, Firefox is a browser, you expect it to talk to the net).

The best I can think of is to incentivize it other ways; e.g., telemetry only for bug reporting, or a "you ping us, we give you a nice hat" or something.


Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: