• MuchPineapples@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    1 year ago

    So it will be locally hosted on the phone? I seriously doubt it will be very useful in offline only mode. Even relativity small language models (7B or 13B) struggle on even desktop pc’s if you don’t have a high end graphics card with 12+ GB of vram. Analyzing can be relatively fast, but generating will be terribly slow, especially images.

    Edit: So after some reading, the snapdragon gen 3 has some impressive specs, but can someone explain how a phone can generate fast ai content while a PC needs let’s say 24GB’s of vram? I get the phone has an ai-specialized chip, but you still need to load the model into memory.

    • _s10e@feddit.de
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Did they say, it would run locally?

      They said it would be available on their upcoming flagship product, sure, that’s marketing. To me, this does not imply that the heavy-lifting is on the phone’s hardware.

      (And maybe a special-purpose AI, which requires significantly smaller models, can run on the phone, which is a high-end mobile computer.)

    • ijeff@lemdro.idOPM
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      If you don’t need an incredible amount of creativity, quantization can get it to work with relatively little RAM.