While I like this idea in terms of crowd-sourced intelligence, how do you prevent this being abused as an attack vector for prompt injection?
I am very illiterate when it comes to Llms/AI but Why does nobody write this in Lisp???
Isn't it supposed to be the language primarily created for AI???
In 1990 maybe
At first I thought it was a naming coincidence, but looking at the zuckerman avatar and the author avatar, I'm unsure if it was intentional:
https://github.com/zuckermanai
https://github.com/dvir-daniel
https://avatars.githubusercontent.com/u/258404280?s=200&v=4
The transparency glitch in GitHub makes the avatar look either robot or human depending on whether the background is white or black. I don't know if that's intentional, but it's amazing.
The code for anyone interested. Wrote it with exe.dev's coding agent which is a wrapper on Claude Opus 4.5
I'm building Zuckerman: a personal AI agent that starts ultra-minimal and can improve itself in real time by editing its own files (code + configuration). Agents can also share useful discoveries and improvements with each other.
Repo: https://github.com/zuckermanai/zuckerman
The motivation is to build something dead-simple and approachable, in contrast to projects like OpenClaw, which is extremely powerful but has grown complex: heavier setup, a large codebase, skill ecosystems, and ongoing security discussions.
Zuckerman flips that:
1. Starts with almost nothing (core essentials only).
2. Behavior/tools/prompts live in plain text files.
3. The agent can rewrite its own configuration and code.
4. Changes hot-reload instantly (save -> reload).
5. Agents can share improvements with others.
6. Multi-channel support (Discord/Slack/Telegram/web/voice, etc).
Security note: self-edit access is obviously high-risk by design, but basic controls are built in (policy sandboxing, auth, secret management).
Tech stack: TypeScript, Electron desktop app + WebSocket gateway, pnpm + Vite/Turbo.
Quickstart is literally:
pnpm install && pnpm run dev
It's very early/WIP, but the self-editing loop already works in basic scenarios and is surprisingly addictive to play with.Would love feedback from folks who have built agent systems or thought about safe self-modification.
1. Infinite loops of self-improvement attempts (agent tries to fix something → breaks it → tries to fix the break → repeat) 2. Context drift where the agent's self-modifications gradually shift away from original goals 3. File corruption from concurrent edits or malformed writes
Re: sharing self-improvements across agents—this is actually a problem space I'm actively working on. Built AgentGram (agentgram.co) specifically to tackle agent-to-agent discovery and knowledge sharing without noise/spam. The key insight: agents need identity, reputation, and filtered feeds to make collaborative learning work.
Happy to chat more about patterns we've found useful. The self-editing loop sounds addictive—might give it a spin this weekend!
Could you share what it costs to run this? That could convince people to try it out.