- AI IP for Cybersecurity monitoring - Smart Monitor
- ARC EV Processors are fully programmable and configurable IP cores that are optimized for embedded vision applications
- Enhanced Neural Processing Unit for safety providing 32,768 MACs/cycle of performance for AI applications
- EV74 processor IP for AI vision applications with 4 vector processing units
- EV7x Vision Processors
- EV7xFS Vision Processors for Functional Safety
- More Products...
IP-SOC DAYS 2025 IP-SOC DAYS 2024 IP-SOC DAYS 2023 IP-SOC DAYS 2022 IP-SOC DAYS 2021 IP-SOC 2024 IP-SOC 2023 IP-SOC 2022 IP-SOC 2021
|
|||||||
![]() |
|

Has AI Finally Destroyed the Need for Software Testing?
- See the 2025 Best Edge AI Processor IP at the Embedded Vision Summit
- Cadence Unveils Millennium M2000 Supercomputer with NVIDIA Blackwell Systems to Transform AI-Driven Silicon, Systems and Drug Design
- Cadence Accelerates Physical AI Applications with Tensilica NeuroEdge 130 AI Co-Processor
- Kyocera Licenses Quadric's Chimera GPNPU AI Processor IP
- The future of AI runs on the GPU
- Versatile Whitebox 1G Ethernet PHY IP Core with BroadR-Reach™ for Connected Automotive and Industrial Systems (May. 19, 2025)
- Codasip: Toward Custom, Safe, Secure RISC-V Compute Cores (May. 19, 2025)
- Semidynamics: From RISC-V with AI to AI with RISC-V (May. 19, 2025)
- InPsytech Joins Samsung SAFE™ IP Partner Program for Excellence in ONFI and UCIe IP Solutions (May. 19, 2025)
- Perforce Partners with Siemens for Software-Defined, AI-Powered, Silicon-Enabled Design (May. 16, 2025)
- See Latest News>>
AI is making it easier than ever to generate code, but harder to maintain quality, well-engineered software. So, does AI-generated code actually reduce the need for testing, or does it demand more rigorous validation than ever before?
csrwire.com, Apr. 16, 2025 –
The explosion of AI tools that can generate code has opened the door to any Tom, Dick, and Harry who wants to develop software. Making coding more accessible and faster sounds like a great thing. ‘Expert’ coding buddies whipping up function, classes, or even whole app in the blink of an eye seems like a dream scenario.
If only that was true.
While AI-generated code speeds things up, it can also turn your codebase into a complete mess. Unfortunately, copy-pasting code and moving on without refactoring or reviewing is becoming more common. This trend has been observed in the analysis of 211 million lines of open-source code, including major projects like VSCode, where copy-pasting has surged while code refactoring has plummeted.
The AI Code Quality Crisis
More or faster code doesn’t mean better code. Far from it. The output often creates more technical debt and bugs, while making it increasingly harder to maintain. There is obvious appeal to be able to create applications in minutes, but there are still significant risks when relying on AI-assisted programming tools:
- Code redundancy increases – Developers can often use AI-generated code without understanding or optimizing it.
- Refactoring takes a back seat – Codebases become bloated as engineers add AI-generated code snippets without restructuring existing logic.
- Bugs are hard to spot – Bugs introduced by AI can be difficult to detect because developers trust the generated code to be correct.
- Vulnerabilities are introduced – AI may inadvertently generate insecure code, increasing security risk and exposing applications to cyber threats.
There’s no doubt that AI tools for coding will improve over time with larger context windows, tried and tested prompting, and better training data. But right now, they cannot and should not replace sound engineering principles.
The Case for More Testing, Not Less
Writing software isn’t just about producing code (and tons of it, quickly). It’s about maintaining it, optimizing it, and ensuring it functions as expected in real-world scenarios. Testing has always been a necessary evil, which when given the option to reduce or stop it, it must be appealing. However, there are reasons why testing needs to still happen. Acting as quality guardrails, software testing should not be overlooked. Yes, AI can produce code ridiculously quickly. Some of it more than adequate, but to soley rely on it, with blind faith, is where things can go awry:
AI doesn’t guarantee correctness
AI-generated code might be syntactically correct, but it can remain logically flawed. As a result, you would still need comprehensive testing to catch these subtle errors. This might be fine for a basic app with minimal lines of code, but when you consider multiple integrated systems, the job at hand has just got increasingly more difficult.