- AI IP for Cybersecurity monitoring - Smart Monitor
- ARC EV Processors are fully programmable and configurable IP cores that are optimized for embedded vision applications
- Enhanced Neural Processing Unit for safety providing 32,768 MACs/cycle of performance for AI applications
- EV74 processor IP for AI vision applications with 4 vector processing units
- EV7x Vision Processors
- EV7xFS Vision Processors for Functional Safety
- More Products...
IP-SOC DAYS 2025 IP-SOC DAYS 2024 IP-SOC DAYS 2023 IP-SOC DAYS 2022 IP-SOC DAYS 2021 IP-SOC 2024 IP-SOC 2023 IP-SOC 2022 IP-SOC 2021
|
|||||||
![]() |
|

SoCs Get a Helping Hand from AI Platform FlexGen
- See the 2025 Best Edge AI Processor IP at the Embedded Vision Summit
- ComputeRAM in AI accelerators: An LLM case study
- Cadence Unveils Millennium M2000 Supercomputer with NVIDIA Blackwell Systems to Transform AI-Driven Silicon, Systems and Drug Design
- Cadence Accelerates Physical AI Applications with Tensilica NeuroEdge 130 AI Co-Processor
- Kyocera Licenses Quadric's Chimera GPNPU AI Processor IP
- NVIDIA Unveils NVLink Fusion for Industry to Build Semi-Custom AI Infrastructure With NVIDIA Partner Ecosystem (May. 20, 2025)
- sureCore extends its sureFIT design service to include custom memory solutions for AI applications (May. 20, 2025)
- Versatile Whitebox 1G Ethernet PHY IP Core with BroadR-Reach™ for Connected Automotive and Industrial Systems (May. 19, 2025)
- Codasip: Toward Custom, Safe, Secure RISC-V Compute Cores (May. 19, 2025)
- Semidynamics: From RISC-V with AI to AI with RISC-V (May. 19, 2025)
- See Latest News>>
Mar. 24, 2025 –
By Gary Hilson, EETimes | March 21, 2025
FlexGen, a network-on-chip (NoC) interconnect IP, is aiming to accelerate SoC creation by leveraging AI.
Developed by Arteris Inc., FlexGen promises to deliver a 10× productivity boost while reducing design iterations and the time required to develop. In a briefing with EE Times, Arteris CMO Michal Siwinski said FlexGen achieves up to a 30% reduction in wire length to lower power use, as well as up to 10% reduction in latency that results in improved performance in SoC and chiplet designs.
“It will actually do the process of generating that network on chip, generating that interconnect and it will actually generate however many interconnects are required to meet the requirement,” Siwinski said.
Using AI and machine learning for chip design isn’t new, Siwinski said, but often the productivity boosts come at the cost of performance or power. FlexGen is able to reduce the number of elements on the critical path of a project, while also improving power/performance metrics, which he said is unique.