• A_A@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    16 hours ago

    The basic model of DeepSeek-R1 14B was already groundbreaking since it reached the level of GPT-1o. But this does much better by bring it to the level of GPT-4o

    Authors are from :

    1 - Lightning Rod Labs (USA)


    https://www.lightningrod.ai/about

    2 - (UK)

    London School of Economics and Political Science


    Machine learning is still developing very fast.
    “We used 8, H100 GPUs, for training.”
    Huge amounts of processing power are not required.