• Omega_Haxors@lemmy.ml
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    1 year ago

    It’s not that impressive, they’re just generating ChatGPT prompts then feeding those into Stable Diffusion. It’s like peak AI grifting.

    • Omega_Haxors@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      1 year ago

      They’re hiding it behind academic language, but when you actually know how the underlying tech works, that’s what they’re doing.

    • figarin@lemmy.ml
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Could you explain why this is a grift? The produced images in the example look scarily close to the original, whatever method they use to go from brainwaves -> image.

      • Omega_Haxors@lemmy.ml
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        Because predictive texting and stable diffusion are both EXTREMELY easy to fudge data for. If you’ve ever used either you’ll realize it’s extremely hard to get it to do what you want and extremely easy to get it to do what it wants. All those high quality art renders you see are always texas sharpshoots where they try a bunch of random crap to see what looks good then say “Look at what it can do!” when it does.

        • figarin@lemmy.ml
          link
          fedilink
          arrow-up
          2
          ·
          1 year ago

          But where exactly is the “fudging” happening?

          the subjects were shown an image different from the 1,200, and their brain activity was measured under the fMRI 30 minutes to an hour later while asked to imagine what kind of image they had seen. Inputting the records, the neural signal translator then created score charts. The charts were input into another generative AI program in order to reconstruct the image, undergoing a 500-step revision process.

          This sounds pretty straightforward. Even if the methods involve “fudging” and “throwing random crap at the wall”, what matters in the end is the accuracy of the results, as long as there’s no human-in-the-middle tweaking anything during each prediction.