I made Sage, a simple Cocoa frontend for serving API requests from common LLM providers like Anthropic, OpenAI and Gemini.
https://github.com/doctashay/sage
Working now but not present in the public build is an agentic code editor that provides 14 different tools that the model can invoke - including browsing the filesystem, grepping through source files, diff editing, checking system capability, working with MacPorts dependencies, and even running arbitrary commands... as you can imagine, this is pretty powerful but can easily be abused without guardrails. It is functional though, I can ask my agent to create a brand new CMake project, configure it, and test to make sure it builds and it will (after a couple of tries) actually do it. The biggest problem is cost - I don't have access to the fancy cutting edge context optimizations like you see in Cursor or Claude Code, and all of my guardrails are just "best guess" alignment so the bill quickly spins out of control - especially with larger models. I tried to use GLM 4.7 as a local model alternative, and it is happy to use my tools, just not in a very smart or fast way.
With enough work and no concern for cost, it could probably exercise a high enough degree of autonomy to take a git repo for a simple dependency (no v8 or anything crazy) and just have it build, test, iterate directly on the hardware until it completes a port entirely on its own - this is pretty much what it's already capable of doing, as long as the user is independently building and testing. Probably going to end up leaving this feature on the shelf until I get a better understanding of MCP, or how I can do this cheaply and securely - maybe in a couple of years when compute costs aren't as insane.