We've definitely faced the review bottleneck on the open-source project that I maintain. ( github.com/robotmcp/ros-mcp-server )
> Three architectural principles are gaining consensus: "token efficiency" as a design constraint (structuring code to minimize the context an AI model needs for any given task), explicit over implicit everywhere (explicit types, explicit error handling, explicit interfaces), and co-location of related code. These principles aren't new, but AI has given them renewed urgency.
I found the architecture section of his article (which is pretty far down) to be the most interesting.
The rules and guidelines for good code have not changed. Don't change that for the sake of today's Ai and limitations