Claude Code and others often write code that is more complex than it needs to be. It would be nice to measure the code complexity before and after a change made by the agent, and then to tell it: "You increased code complexity by 7%. Can you find a simpler solution?".
Something that I have started doing is to tell it to spawn a reviewer agent after every code change, and in the Claude/rules folder I specify exactly how I want stuff done.
Mostly I use it to write unit tests (just dislike production code that is not exactly as I want). So there is a testing rule for all files in the test folder that lays out how they should be done. The agent writing the tests may miss some due to context bloat but the reviewer has a fresh context window and only looks at those rules. So it does result in some simpler code.
You dont need to measure the code complexity, you can completely hallucinate those numbers if you feel that code is too complex and then watch how LLM will respond.
That's not the point. The number of white European people arrested and shackled by CBP/ICE is very small. But it's NOT ZERO! So at the margin plenty of potential tourists would prefer to go some other place where that chance is effectively zero.
I switched from macOS to Linux ten years ago and haven't looked back. At the time, I compared Linux vs. macOS to living at home vs. in a hotel [1]. Since then, I feel things have only gotten better for Linux, and more restrictive and arcane on macOS.
This knowledge will live in the proprietary models. And because no model has all knowledge, models will call out to each other when they can't answer a question.
> I tried to fetch the exact contents of https://moltbook.com/skill.md (and the redirected www.moltbook.com/skill.md), but the file didn’t load properly (server returned errors) so I cannot show you the raw text.
Cool stuff. I just started open sourcing a command-line tool for deploying Django to a server. It handles SSL certs, databases and backups, automatic error emails, and background tasks via celery / redis. The best part? It does not need Docker. It just runs everything on bare metal.
A similar experience for me was the Connections Museum in Seattle: I came just after opening, and time flew by such that I was surprised when they told me they were closing up
I was able to go to the Living Computer Museum and I got there when they first opened and wound up staying until closing time. I was just so into all the stuff there :-)
I strained my groin/abs a few weeks ago and asked ChatGPT to adjust my training plan to work around the problem. One of its recommendations was planks, which is exactly the exercise that injured me.
My cleaning lady's daughter had trouble with her ear. ChatGPT suggested injecting some oil into it. She did and it became a huge problem, so that she had to go to the hospital.
I'm sure ChatGPT can be great, but take it with a huge grain of salt.
Now imagine what happens when a new programming language comes along. When we have a question, we will no longer be able to Google it and find answers to it on Stack Overflow. We will ask the LLMs. They will work it out. From that moment, the LLM we used has the knowledge for solving this particular problem. Over time, this produces huge moat for the largest providers. I believe it is one of the subtler reasons why the AI race is so fierce.