Huawei is set to unveil a new technological solution aimed at reducing China’s reliance on High Bandwidth Memory (HBM) chips for Artificial Intelligence (AI) inference at the 2025 Financial AI Reasoning Application Forum on August 12.
AI inference, the process where models use their knowledge to deliver accurate outputs efficiently, relies heavily on HBM chips due to their lower latency and higher memory bandwidth compared to traditional memory. This facilitates faster data processing and improved performance for large language models.
However, due to US restrictions, Huawei has faced limitations in accessing HBM chips. In response, the company has developed a proprietary solution designed to circumvent this dependency. This new technology is expected to not only lessen China’s and Huawei’s reliance on imported HBM AI chips but also significantly boost the inference performance of large-scale AI models within the country.
The development is seen as a strategic move to strengthen China’s domestic AI inference ecosystem. Huawei has been actively seeking self-developed technological integrations to expand its AI business within China, aiming to reduce its reliance on foreign goods. While the specific details of the new solution remain undisclosed, further information is anticipated to be revealed at the upcoming forum.




