Artificial Intelligence

Out of context: Reply #1678

  • Started
  • Last post
  • 2,578 Responses
  • jonny_quest_lives0

    "DeepSeek claimed it used just 2,048 Nvidia H800s and $5.6mn to train a model with 671bn parameters, a fraction of what OpenAI and Google spent to train comparably sized models.

    Ritwik Gupta, AI policy researcher at the University of California, Berkeley, said DeepSeek’s recent model releases demonstrate that “there is no moat when it comes to AI capabilities”.

    “The first person to train models has to expend lots of resources to get there,” he said. “But the second mover can get there cheaper and more quickly.”

    Gupta added that China had a much larger talent pool of systems engineers than the US who understand how to get the best use of computing resources to train and run models more cheaply.

    Industry insiders say that even though DeepSeek has shown impressive results with limited resources, it remains an open question whether it can continue to be competitive as the industry evolves.

    Returns at High-Flyer, its big backer, lagged behind in 2024, which one person close to Liang blamed on the founder’s attention being mostly focused on DeepSeek.
    "

    https://www.ft.com/content/747a7…

View thread