• AggressivelyPassive@feddit.de
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    7 months ago

    That’s maybe because we’ve reached the limits of what the current architecture of models can achieve on the current architecture of GPUs.

    To create significantly better models without having a fundamentally new approach, you have to increase the model size. And if all accelerators accessible to you only offer, say, 24gb, you can’t grow infinitely. At least not within a reasonable timeframe.