AI

ByteDance Boosts AI Spend to $29.4B as Chip Costs Soar

ByteDance has increased its 2026 AI infrastructure budget to $29.4 billion, a 25% jump from earlier plans, as rising memory chip costs and U.S. export restrictions reshape its strategy. The company is balancing domestic chip development with offshore computing deals to secure access to cutting-edge AI hardware. This move reflects broader trends in Big Tech, where AI capex is surging despite soaring component prices.

ByteDance has revised its 2026 AI infrastructure budget upward by 25%, allocating over 200 billion yuan ($29.4 billion) to hardware, semiconductors, and cloud capacity. The increase reflects both the company’s expanding AI ambitions and the impact of skyrocketing memory chip prices, which have nearly doubled in some segments this year.

Why the Budget Jumped

The initial 2026 budget, set at 160 billion yuan in late 2025, was split between advanced semiconductors (80 billion yuan) and AI processors (85 billion yuan), with roughly 100 billion yuan earmarked for Nvidia chips pending U.S. export approvals. However, industry-wide price surges in DRAM and other components forced a recalibration. TrendForce reported DRAM contract prices rose 95% quarter-over-quarter in Q1 2026, with further increases of 58% to 63% projected for Q2.

ByteDance is not alone in facing these cost pressures. Microsoft attributed $25 billion of its record $190 billion 2026 capex budget to higher memory and component costs, while Meta raised its full-year capex range to $125–145 billion. Collectively, major U.S. tech firms announced $725 billion in AI capital expenditure for 2026, a 77% increase from 2025.

Domestic Chips vs. Offshore Workarounds

A key shift in ByteDance’s revised budget is a larger allocation toward domestic AI chips, reducing reliance on U.S.-controlled supply chains. The company is developing its own AI inference chip in partnership with Samsung, targeting production of 100,000 units in 2026 with plans to scale to 350,000. Its in-house chip design team, now numbering around 1,000 engineers, has reportedly created a processor that matches the efficiency of Nvidia’s China-market H20 chip at a lower cost.

Simultaneously, ByteDance is pursuing offshore computing deals to access restricted hardware. Through a partnership with a Southeast Asian cloud provider, the company has deployed 36,000 Nvidia B200 Blackwell chips in Malaysia, representing a $2.5 billion investment in AI research capacity outside China.

What This Means for AI Infrastructure

ByteDance’s $29.4 billion bet underscores three trends:

  1. Cost inflation: Memory chip price spikes are forcing even cash-rich firms to revise budgets upward.
  2. Geopolitical hedging: Companies are diversifying supply chains through domestic chip development and offshore data center deals.
  3. Long-term AI prioritization: Despite near-term cost pressures, ByteDance is doubling down on AI as a core competitive advantage across its portfolio, including TikTok, e-commerce, and enterprise cloud services.

The company’s dual strategy—building domestic silicon while securing offshore capacity—mirrors moves by other global tech firms navigating U.S.-China tensions. For now, the willingness to spend nearly $30 billion signals confidence that AI infrastructure will remain a critical differentiator in the years ahead.

Bottom Line

ByteDance’s budget revision is a microcosm of the broader AI infrastructure race: higher costs, geopolitical maneuvering, and a relentless focus on hardware as the foundation for future growth. While the $29.4 billion figure is eye-catching, the real story lies in how the company is adapting to constraints—balancing domestic innovation with offshore workarounds to keep its AI ambitions on track.

Similar Articles

More articles like this

AI 2 min

OpenAI Unveils Advanced Voice Models

OpenAI has released three new audio models through its Realtime API, enabling more intelligent and multilingual voice-powered applications. The models, GPT-Realtime-2, GPT-Realtime-Translate, and GPT-Realtime-Whisper, offer advanced reasoning, translation, and transcription capabilities. These models are designed to make voice interactions more natural and effective, with potential applications in customer service, language learning, and more. Early adopters have reported significant improvements in call success rates and word error rates using these models.

AI 3 min

Instagram Drops End-to-End Encryption for DMs on May 8 — Here's What Changes

Meta will strip end-to-end encryption from Instagram direct messages on May 8, 2026, ending a feature it began testing in 2021. The company says few users opted in, but critics argue the feature was deliberately buried. Users who enabled encrypted chats must download their data before the deadline or switch to WhatsApp for continued encryption.

AI 4 min

Airbnb’s AI Now Writes 60% of Its Engineers’ Code—What It Means for Tech Teams

Airbnb revealed that AI now generates nearly 60% of its engineers’ code, doubling the industry average and accelerating feature development. The shift has also slashed customer support costs, with AI resolving 40% of issues autonomously. CEO Brian Chesky warns that traditional management roles are becoming obsolete, urging leaders to engage directly with work rather than overseeing teams. The trend extends beyond Airbnb, with companies like Coinbase and Block flattening org structures to adapt.

AI 2 min

Microsoft Integrates GPT-5.5 Instant into 365 Copilot

Microsoft has announced the integration of OpenAI's GPT-5.5 Instant model into Microsoft 365 Copilot and Copilot Studio. This upgrade replaces the previous GPT-5.3 Instant model and brings improved accuracy, context handling, and a 'smart-switching' capability. The new model is designed to provide quicker, clearer, and more accurate responses to user queries. With this integration, Microsoft aims to enhance the AI capabilities of its 365 Copilot platform and compete with Google's Gemini in the enterprise AI market.

AI 3 min

Google to let job candidates use Gemini AI in software engineering interviews

Google is piloting a program that lets software engineering candidates use its Gemini AI assistant during a portion of the interview process. The move, reported by Business Insider based on an internal document, aims to reflect how engineers actually work with AI tools. The AI-assisted round will assess prompt engineering, output validation, and debugging skills rather than pure memorization. The pilot begins in the second half of 2026 for select U.S. teams, with broader interview changes including a technical design discussion and an open-ended engineering challenge.

AI 3 min

Microsoft Accelerates Push to Kill Passwords by 2027

Microsoft has announced a comprehensive set of updates to eliminate passwords as the default sign-in method across its ecosystem. New enterprise and consumer passkey features, including cross-device sync and biometric recovery, go live in May 2026. The company reports 99.6% of its own users now use phishing-resistant authentication. Security questions will be removed from Entra ID in January 2027.