Side-by-side comparison of AI visibility scores, market position, and capabilities
d-Matrix builds in-memory AI inference accelerator chips (Corsair) that deliver 10x faster inference at 3x lower cost than GPU-based systems; raised $275M Series C at a $2B valuation in November 2025; Raptor chip due 2026.
d-Matrix is a Santa Clara-based AI semiconductor company founded in 2019, developing purpose-built inference accelerator hardware that challenges Nvidia''s dominance in AI compute. Its flagship Corsair inference accelerator card uses in-memory computing (IMC) architecture — performing computations directly inside the memory arrays rather than moving data between separate processing and memory units. This eliminates the "memory wall" bottleneck that limits GPU-based inference performance for large language models and generative AI workloads, enabling what d-Matrix claims is 10x faster inference, 3x lower cost, and 3–5x better energy efficiency versus GPU systems.
Fremont CA semiconductor etch and deposition (NASDAQ: LRCX) $14.9B FY2024 revenue; 3D NAND/HBM etch leader, 40%+ plasma etch share, $5B+ services revenue competing with Applied Materials and Tokyo Electron.
Lam Research Corporation is a Fremont, California-based semiconductor equipment company — publicly traded on the NASDAQ (NASDAQ: LRCX) as an S&P 500 Information Technology component — designing and manufacturing etch and deposition systems critical for semiconductor chip fabrication, providing products across plasma etch (removing material layers with precision), chemical vapor deposition (CVD — depositing thin films on wafers), atomic layer deposition (ALD — depositing single atomic layers with Angstrom-level precision), and related services through approximately 17,000 employees worldwide. In fiscal year 2024 (ending June 2024), Lam Research reported revenues of $14.9 billion, with strong revenue recovery driven by semiconductor industry capex expansion (NAND flash memory producers resuming equipment orders after the 2022-2023 memory market downturn, and DRAM producers expanding capacity for HBM — High Bandwidth Memory — required in NVIDIA AI GPU packages). CEO Tim Archer has positioned Lam Research as an "advanced process technology" partner rather than a pure equipment vendor: Lam's ALD-Select, VECTOR deposition, and Kiyo etch systems are co-developed with leading chipmakers (TSMC, Samsung, SK Hynix, Micron) for specific process nodes — creating application-specific systems optimized for 3nm logic, 1-alpha DRAM, and 200+ layer 3D NAND that require Lam's process understanding rather than generic equipment. Lam Research's Global Customer Support (GCS) organization provides equipment maintenance, spare parts, and process consulting services — generating $5+ billion annually in recurring service revenue that is less cyclical than equipment capital expenditure.
Monitor how your brand performs across ChatGPT, Gemini, Perplexity, Claude, and Grok daily.