Breaking the Memory Wall: How d-Matrix Is Redefining AI Inference with Chiplets
By Maurizio Di Paolo Emilio, embedded.com | April 23, 2025
As AI workloads push the limits of performance, power efficiency, and memory bandwidth, chiplets are rapidly emerging as the architectural solution of choice. In this interview, Sree Ganesan, Vice President of Product at d-Matrix, dives deep into how its pioneering chiplet-based platform is revolutionizing AI inference. From solving the memory wall with Digital In-Memory Computing (DIMC) to enabling seamless multi-chiplet communication via custom interconnects, d-Matrix reveals how its innovations are unlocking 10x faster token generation, 3x better energy efficiency, and a scalable roadmap for generative AI.
To read the full article, click here
Related Chiplet
- DPIQ Tx PICs
- IMDD Tx PICs
- Near-Packaged Optics (NPO) Chiplet Solution
- High Performance Droplet
- Interconnect Chiplet
Related News
- EdgeCortix Awarded New 3 Billion Yen NEDO Project to Develop Advanced Energy-Efficient AI Chiplet for Edge Inference and Learning
- D-Matrix Targets Fast LLM Inference for ‘Real World Scenarios’
- Ayar Labs Unveils World's First UCIe Optical Chiplet for AI Scale-Up Architectures
- Wave Photonics Acquires Phoelex Chiplet IP to Power Datacoms & AI Growth
Latest News
- NLM Photonics Samples Silicon Organic Hybrid PICs Manufactured at GlobalFoundries
- Avalanche Technology and NHanced Semiconductors Leverage Advanced 2.5D Integration to Bring Optimal SWaP and Reliability to Rad-Hard FPGAs
- Open EU Foundry status granted to innovative chiplet facility
- Siluxtek and GlobalFoundries Forge a Deep Strategic Partnership to Mass-Produce 200G/Lane Silicon Photonic Receiver Chips, Paving the Way for the Industrial Revolution of AI Computing Interconnects
- Lightwave Logic High-Speed Modulator Platform Now Available in GDS Factory PDK for GlobalFoundries Silicon Photonics Platform