This study examines the security performance of generative artificial intelligence (AI) tools of ChatGPT, Copilot, and Gemini within software development workflows. Through static and dynamic code analysis, security vulnerabilities in web application login code generated by these tools were systematically evaluated. Results indicate that while AI models offer efficiency in code generation, they also introduce varying levels of security risk. Copilot exhibited the highest cumulative risk with multiple high-level vulnerabilities, while ChatGPT demonstrated a lower risk profile. Gemini produced relatively optimized code but contained critical security flaws that require manual review. The most common vulnerabilities across all models were insecure design and security logging and monitoring failures, indicating a systemic issue in AI-generated code. The findings emphasize that generic prompts focusing on security are insufficient and that developers must use specific, security-oriented prompts, such as applying secure-by-design principles and implementing OWASP Top Ten protections. This study contributes to the growing body of literature addressing the security implications of integrating AI into software development, highlighting the importance of human oversight and carefully crafted prompts to mitigate potential risks.
Generative AI ChatGPT Copilot Gemini Software Security Static Code Analysis Dynamic Code Analysis
Primary Language | English |
---|---|
Subjects | Computer Software |
Journal Section | Research Articles |
Authors | |
Publication Date | August 31, 2025 |
Submission Date | June 16, 2025 |
Acceptance Date | August 9, 2025 |
Published in Issue | Year 2025 Volume: 11 Issue: 2 |