• pinguinu [any]@lemmygrad.ml
    link
    fedilink
    arrow-up
    5
    ·
    4 days ago

    The Mac Studio they talk about running the 685b model on costs 12 000 dollaridoos (after 20% taxes here). I get that it’s a consumer device and that it draws less power, but at that point you’d just get a server for less. The power consumption is an outlier since it has everything on chip

    • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
      link
      fedilink
      arrow-up
      8
      ·
      4 days ago

      I think the key part is that you can run these large scale models cheaply in terms of energy cost. The price of hardware will inevitably come down going forward, but now we know that there is no fundamental blocker for running models efficiently.

      • pinguinu [any]@lemmygrad.ml
        link
        fedilink
        arrow-up
        2
        ·
        4 days ago

        I generally agree, but given how niche a powerful SoC like this is, I doubt it matters right now (<5 years). I understand it proves a point, but I wager there’s still a long-ish way to see power-efficient hardware like this available for cheaper (which will most likely come from China natively)

        • ☆ Yσɠƚԋσʂ ☆@lemmygrad.mlOP
          link
          fedilink
          arrow-up
          6
          ·
          4 days ago

          Yeah, a 5 year or so timeline before we see SoC design becomes dominant is a good guess. There are other interesting ideas like analog chips that have potential to drastically cut power usage for neural networks as well. Next few years will be interesting to watch.

    • CriticalResist8@lemmygrad.ml
      link
      fedilink
      arrow-up
      4
      ·
      3 days ago

      But imagine what you’ll be able to run it on in four more months. But yeah, it’s stretching the definition of consumer hardware a bit.