AI is changing the programming world, which has been evolving for several years. I could talk about how the emerging practice of using AI-driven coders increases speed and reduces costs, but there are some downsides that many fail to see.
Again, the question is not “Can we?” It’s “Should we?” Let’s go over a few core concerns.
What AI can’t do
AI-driven coders learn from existing code repositories. They often need a more contextual understanding of the code generated. They produce code that works but may need help to comprehend or maintain. This hinders developers’ control over their software and often causes mistakes when fixing or changing applications.
Moreover, the generated code must meet style conventions or best practices and include appropriate error handling. This can make debugging, maintenance, and collaboration difficult.
Remember that AI-driven code generation focuses on learning from existing code patterns to generate net-new code. Generative AI coders have a “monkey see, monkey do” approach to development, whereas the coding approaches are learned from the vast amount of code used as training data.
This approach is helpful for repetitive or standard tasks, which is much of what developers do, but enterprises may require more creativity and innovation for complex or unique problems. Using generative AI code can limit the potential for novel solutions and hinder the development of truly innovative applications.
Not sure if you’ve looked out there, but innovation is lacking. We seem to be building the same things over and over again.
My biggest concern is that code could be more efficient and optimized for the platform the application is deployed on. It takes sound engineering practices to understand how to optimize processors, memory, and storage management.
I think that many people will generate and deploy an application without understanding how it could leverage resources in a more optimized way. We end up with applications that are more expensive to run and have a much larger carbon footprint.
The shame is that, in most cases, just the fact that the application works is good enough for many. The applications operate for years, waste a great deal of money, and fail to return the optimal value to the business. “Oh, well,” people say, “it works, doesn’t it?”
Another scary aspect of AI-driven development is that many security vulnerabilities are left within the application and go unnoticed until the postmortem after a breach. Again, we need human engineering to spot and fix those, albeit some helpful AI-driven scanning tools can be practical.
By removing humans from the development process, which many organizations are looking to do, we sacrifice the understanding needed to create practical applications. The appropriate answer is to find a balance between the value of AI in terms of speed and cost and the fact that many human skills still need to be involved. I fear that we won’t understand that until it’s too late.