• criitz@reddthat.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    1 year ago

    That’s what I mean. We elevate the human thought process as if what we come up with is more valid than what a (future) computer could think up. But is it?

    So if a computer synthesizing Shakespeare is stealing, maybe so is a human doing it. But maybe then we could never create anything at all. And if we must not be blocked from it, must a machine?

    • merc@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      So if a computer synthesizing Shakespeare is stealing

      Copyright infringement is never stealing. But, as to whether it’s infringing copyright, the difference is that current laws were designed based on human capabilities. If memorizing hundreds of books word for word was a typical human ability, copyright would probably look very different. Instead, normal humans are only capable of memorizing short passages, but they’re capable of spotting patterns, understanding rhythms, and so-on.

      The human brain contains something like 100 billion neurons, and many of them are dedicated to things like hearing, seeing, eating, walking, sex, etc. Only a tiny fraction are available for a task like learning to write like Shakespeare or Stephen King. GPT-4 contains about 2 trillion parameters, and every one of them is dedicated to “writing”. So, we have to think differently about whether what it’s storing is “fair” when it comes to infringing someone’s copyright.

      Personally, I think copyright is currently more harmful than helpful, so I like that LLMs are challenging the system. OTOH, I can understand how it’s upsetting for an artist or a writer to see that SALAMI can reproduce their stuff almost exactly, or produce something in their style so well that it effectively makes them obsolete.