Greg Rutkowski, a digital artist known for his surreal style, opposes AI art but his name and style have been frequently used by AI art generators without his consent. In response, Stable Diffusion removed his work from their dataset in version 2.0. However, the community has now created a tool to emulate Rutkowski’s style against his wishes using a LoRA model. While some argue this is unethical, others justify it since Rutkowski’s art has already been widely used in Stable Diffusion 1.5. The debate highlights the blurry line between innovation and infringement in the emerging field of AI art.

    • falsem@kbin.social
      link
      fedilink
      arrow-up
      0
      ·
      1 year ago

      If I look at someone’s paintings, then paint something in a similar style did I steal their work? Or did I take inspiration from it?

      • Pulse@dormi.zone
        link
        fedilink
        arrow-up
        0
        ·
        1 year ago

        No, you used it to inform your style.

        You didn’t drop his art on to a screenprinter, smash someone else’s art on top, then try to sell t-shirts.

        Trying to compare any of this to how one, individual, human learns is such a wildly inaccurate way to justify stealing a someone’s else’s work product.

        • falsem@kbin.social
          link
          fedilink
          arrow-up
          0
          ·
          1 year ago

          If it works correctly it’s not a screenprinter, it’s something unique as the output.

          • Pulse@dormi.zone
            link
            fedilink
            arrow-up
            0
            ·
            1 year ago

            The fact that folks can identify the source of various parts of the output, and that intact watermarks have shown up, shows that it doesn’t work like you think it does.

            • jarfil@beehaw.org
              link
              fedilink
              arrow-up
              0
              ·
              1 year ago

              Does that mean the AI is not smart enough to remove watermarks, or that it’s so smart it can reproduce them?

              • nickwitha_k (he/him)@lemmy.sdf.org
                link
                fedilink
                arrow-up
                0
                ·
                1 year ago

                LLMs and directly related technologies are not AI and possess no intelligence or capability to comprehend, despite the hype. So, they are absolutely the former, though it’s rather like a bandwagon sort of thing (x number of reference images had a watermark, so that’s what the generated image should have).

                • jarfil@beehaw.org
                  link
                  fedilink
                  arrow-up
                  0
                  ·
                  1 year ago

                  LLMs […] no intelligence or capability to comprehend

                  That’s debatable. LLMs have shown emergent behaviors aside from what was trained, and they seem to be capable of comprehending relationships between all sorts of tokens, including multi-modal ones.

                  Anyway, Stable diffusion is not an LLM, it’s more of a “neural network hallucination machine” with some cool hallucinations, that sometimes happen to be really close to some or parts of the input data. It still needs to be “smart” enough to decompose the original data into enough and the right patterns, that it can reconstruct part of the original from the patterns alone.

                  • nickwitha_k (he/him)@lemmy.sdf.org
                    link
                    fedilink
                    arrow-up
                    0
                    ·
                    1 year ago

                    Thanks for the clarification!

                    LLMs have indeed shown interesting behaviors but, from my experience with the technology and how it works, I would say that any claims of intelligence being possessed by a system that is only an LLM would be suspect and require extraordinary evidence to prove that it is not mistaken anthropomorphizing.