CONNECT WITH US
Mar 27
In-depth: Google TurboQuant cuts LLM memory 6x, resets AI inference cost curve

Google has introduced TurboQuant, a compression algorithm that reduces large language model (LLM) memory usage by at least 6x while boosting performance, targeting one of AI's most persistent bottlenecks: memory. The breakthrough lowers inference costs and expands deployment across cloud and edge environments.

Samsung Electronics is facing a looming labor strike in May as its memory and foundry businesses take off, marking a more complex challenge than the initial 2024 walkout. The upcoming strike reflects significant changes in industry conditions and union size, highlighting Samsung's structural difficulties with labor issues amid evolving laws and its broad business portfolio.
Chang Wah Technology (CWTC) approved a plan to build a new factory in Weihai, Shandong. It confirmed leadership changes as it presses on with overseas expansion to boost production capacity and operational growth. The board's extraordinary meeting authorized a tentative CNY1 billion (US$145.00 million) investment through a subsidiary.
Innodisk told attendees at the 2026 AI EXPO that effective AI deployment requires more than raw computing power; it depends on tight integration between software and hardware, and on selecting components tailored to specific environments. The company argued that edge AI has progressed from image recognition and language models to autonomous learning and decision-making.
Amid the rapid development of generative artificial intelligence (GenAI) and large language models (LLM), global demand for high-performance computing (HPC) continues to rise. Memory module maker Adata Technology announced a US$3 million investment in the Series A funding round of artificial intelligence (AI) computing infrastructure provider KonstTech (Konst).

SEMICON China 2026 spotlighted the scale of the AI investment boom, with Handel Jones, CEO of International Business Strategies (IBS), estimating that global AI and data center capital expenditure has surged from about US$110 billion in 2020 to roughly US$600 billion in 2026.

espite a surge in demand driven by generative artificial intelligence, the fundamental economics of the memory industry remain largely intact. While high-bandwidth memory (HBM) has created a premium segment, the broader market continues to operate on standardized, high-volume production rather than structural product differentiation.

The surge in artificial intelligence demand is pushing the semiconductor industry toward a supply crunch, as Elon Musk outlined a plan to build what he described as the largest chip manufacturing effort in history.

Samsung Electronics' next-generation high-bandwidth memory (HBM4) program is entering a customer validation phase, according to NewDaily, as major technology firms reportedly begin on-site audits of its production lines.

On March 27, the US International Trade Commission (ITC) launched an investigation into memory chip imports by SK Hynix Inc. and KIOXIA Holdings Corporation following a patent complaint filed by MonolithIC 3D Inc., according to a notice published by the agency.
Elon Musk, CEO of Tesla, announced in mid-March that he would launch the Terafab megafab project within seven days, and on March 21 confirmed that the facility will be built in Austin, Texas. While many remain skeptical, industry forum SemiWiki offers an alternative perspective. Fifteen years ago, Musk acquired the Fremont production line in California from General Motors and Toyota, transforming it from a struggling facility into a cornerstone of Tesla's electric vehicle production.

OmniVision Group said on March 20 it will invest CNY1 billion (US$145 million) in Rong Semiconductor (Ningbo) Co. (RongSemi) via a capital increase, taking a 5.88% stake based on a CNY4 billion funding round. The move targets tighter upstream integration, aiming to secure wafer capacity and improve supply chain resilience.