More

    Bye bye Nvidia? Chinese cloud providers aggressively cut down AI inference costs by using Huawei’s controversial accelerators and DeepSeek’s tech




    • DeepSeek’s V3 and R1 models are available through Huawei’s Ascend cloud service
    • They are powered by the Ascend 910x accelerators banned in the US, EU and UK
    • The pricing is much lower than offered by Azure and AWS who have started trialing DeepSeek

    DeepSeek recently massively unsettled global markets with the launch of its open reasoning LLM, which was built and trained for a fraction of the cost of models from much larger US competitors, although OpenAI has since accused DeepSeek’s developers of using its models to train theirs.

    A new paper had claimed DeepSeek’s V3 LLM was trained on a cluster of just 2,048 Nvidia H800 GPUs – crippled versions of the H100 designed to comply with US export restrictions to China. Rumors around DeepSeek’s newer reasoning model, R1, suggest it may have been trained on as many as 50,000 Nvidia “Hopper” GPUs, including H100, H800, and the newer H20, although DeepSeek hasn’t – and likely won’t – confirm this. If true, it raises serious questions about China’s access to advanced AI hardware despite ongoing trade restrictions, although it’s no secret there’s a thriving black market for advanced Nvidia AI hardware there.

    https://cdn.mos.cms.futurecdn.net/RFHhebaafuggVviVNzSsy-1200-80.jpg



    Source link
    waynewilliams@onmail.com (Wayne Williams)

    Latest articles

    spot_imgspot_img

    Related articles

    Leave a reply

    Please enter your comment!
    Please enter your name here

    spot_imgspot_img