Hey everyone,
We're diving into Amazon Q Developer and our team is looking to track how much code is generated by the AI compared to what our developers write manually. This isn't about policing anyone but rather gathering metrics for internal analysis of productivity and efficiency. We're particularly interested in:
- Methods to differentiate between code produced by Q and that written by humans.
- Tools that can help us measure how much of our codebase is AI-generated versus manually coded.
- Best practices for tracking the impact of AI code generation on our productivity overall.
We've explored the built-in analytics features of Amazon Q, but it seems like they focus more on usage metrics instead of tracking specific code contributions. Here are some specific questions I have:
1. Are there any built-in tools in Amazon Q Developer for tracking which parts of the code are generated?
2. Are there any tools available that could analyze our existing code to identify sections that might be AI-generated?
3. How are other teams managing this tracking for their metrics and compliance?
We're mainly working with Python and JavaScript, so recommendations suited for those languages would be great.
Looking forward to hearing how others handle this! Thanks in advance!
5 Answers
Your management might want to rethink the focus. Instead of tracking AI versus human code, it could be more beneficial to monitor classic engineering metrics like cycle time or defect rates. These metrics really highlight the impact of AI tools on overall performance, rather than just spotting the source of code.
In my experience, counting code lines can be misleading. Instead, evaluate the time spent on tickets and see if AI tools help speed up ticket completion. That’s a clearer measurement of efficiency.
Honestly, it's pretty tough to effectively distinguish what's AI-generated versus human-written code. One major challenge is the nature of how AI works—it often produces code that looks just like what a developer would write. Even if you tweak code that was generated by an AI, it's hard to claim credit for that work since it’s all mixed together. Plus, judging contributions by lines of code doesn’t reflect true productivity!
I agree! It's often more about the quality and functionality of the code rather than just the origin. Sure, recognizing contributions is important, but we shouldn't get too hung up on who created what.
If you're looking for tracking options, I’d suggest just reaching out to Amazon Q’s support. They may provide some insights or features that aren't fully documented yet. It's worth a shot!
Have you tried instructing Q to prefix all the generated names? Just a quirky thought, but it might help identify AI code segments!

Exactly! If the performance shows improvement, that's what really matters in the end, not the source of the lines.