Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
×
AI Technology

Thousands of Artists Allege Midjourney Used Their Work To Train AI Software (theguardian.com) 182

An anonymous reader shares a report: Since the emergence of Midjourney and other image generators, artists have been watching and wondering whether AI is a great opportunity or an existential threat. Now, after a list of 16,000 names emerged of artists whose work Midjourney had allegedly used to train its AI -- including Bridget Riley, Damien Hirst, Rachel Whiteread, Tracey Emin, David Hockney and Anish Kapoor -- the art world has issued a call to arms against the technologists. British artists have contacted US lawyers to discuss joining a class action against Midjourney and other AI firms, while others have told the Observer that they may bring their own legal action in the UK.

"What we need to do is come together," said Tim Flach, president of the Association of Photographers and an internationally acclaimed photographer whose name is on the list. "This public showing of this list of names is a great catalyst for artists to come together and challenge it. I personally would be up for doing that." The 24-page list of names forms Exhibit J in a class action brought by 10 American artists in California against Midjourney, Stability AI, Runway AI and DeviantArt. Matthew Butterick, one of the lawyers representing the artists, said: "We've had interest from artists around the world, including the UK."

This discussion has been archived. No new comments can be posted.

Thousands of Artists Allege Midjourney Used Their Work To Train AI Software

Comments Filter:
  • by MikeDataLink ( 536925 ) on Wednesday January 24, 2024 @12:10PM (#64184697) Homepage Journal

    Are they going to sue my brain because it stores a copy of their artwork and used it for inspiration in a derivative work?

    • Are they going to sue my brain because it stores a copy of their artwork and used it for inspiration in a derivative work?

      Well yeah, but your lone brain gets fatigued before too long and then it ain't worth shit. Plus you'll charge us for your time and we're really cheap. Especially after the passage of time!

    • by Petersko ( 564140 ) on Wednesday January 24, 2024 @12:22PM (#64184727)

      Your brain does not store a copy of their artwork. Your brain keeps an incomplete memory - even less complete than the models do. Pedantry aside... Considering the time it would take for you to create that derivative work, the fact that you can't jump to proficiency quickly, and the high likelihood of you not actually being up to the task, even if they could pursue you for it, I doubt they would.

      • "Your brain keeps an incomplete memory - even less complete than the models do"

        In a large model there is typically less than one byte per training image.

        • The question is more complicated (and fascinating) than that. There's been quite a lot of work that demonstrates you can pull "ghosts" of the training data out. And the more times the image entered the training cycle, the clearer the ghost. No, you can't pull bit for bit out, and the success rate is low... but it demonstrates that the "less than one byte" claim per training image is actually a misdirection.

          • Just as a followup: https://www.kaspersky.com/blog... [kaspersky.com]

            It's cool stuff. Nerdy.

          • by timeOday ( 582209 ) on Wednesday January 24, 2024 @12:48PM (#64184855)
            The "claim" (or rather, fact) is less than one byte per image on average. A more popular image will be seen more times and leave a stronger impression - but mathematically that also means it is pushing down the number of bits devoted to unique information from less popular works.

            So, yeah, it's going to learn a pretty solid conception of the Mona Lisa. Not unlike a person.

            • I am referring to a single source image as "an image" because I am a nerd and specificity matters to me.

              Multiple different images of the same work can produce what you are talking about, but it will NEVER produce a pixel perfect reproduction of ANY of the source images. It may produce something recognizable, but that is not the claim.

            • by AmiMoJo ( 196126 ) on Thursday January 25, 2024 @06:42AM (#64186708) Homepage Journal

              People can get in trouble for this too. There is currently a lawsuit going on over a tattoo of a famous jazz player. The photographer alleges that the tattoo artist copied his work, with the same composition and pose etc. The tattoo artist is saying she used his photo as a reference, but made enough changes that it qualifies as transformative and not infringing.

              That's basically the issue with AI. Is the output different enough from the original to qualify as a separate, non-infringing work. Google just published a video about their new video generation AI, and in it you can see Iron Man in his standard Disney favourite arm-outstretched pose, so the tattoo case seems extremely relevant.

              Outside of the US the situation is similar in many countries. Some years ago a photographer in the UK won a copyright case against another photographer who copied his composition of a bus driving over a bridge in London.

      • Your brain does not store a copy of their artwork.

        Please correct me if I'm wrong, but I believe these AI systems, do NOT in fact, store the images.

        They pretty much only use them to set weights.

        • Kinda. I mean yes, but it turns out that in many cases there is enough information in the weights that original data can be pretty accurately reproduced.

          • As the image generators begin with noise and pattern match out of the noise, you will never get an exact reproduction of the original image. It's statistically possible, but you are talking about millions of values having to line up exactly right.

    • Are they going to sue my brain because it stores a copy of their artwork and used it for inspiration in a derivative work?

      This is important: "artificial Intelligence" is not intelligent . It is pattern-matching and replication software. There is no "intelligence" involved.

      When you say "the human brain does the same thing"-- really? You know how the human brain works? So, publish in a peer reviewed journal and you will be the most famous brain scientist in the world, because in fact no one knows how your brain works.

      • This is important: "artificial Intelligence" is not intelligent . It is pattern-matching and replication software. There is no "intelligence" involved.

        When I view AI images I see quite a bit of intelligence. If I ask the machine to draw a lake with mountains, trees and a castle the machine produces something with convincing GI, lighting, shadows, and reflections. In the water you can see realistic specular reflections of the castle and surrounding trees and mountains. These systems don't have fancy libraries full of material properties, ray tracers or fancy lighting / illumination systems. Nobody programed these systems to do any of that. They learne

        • This is important: "artificial Intelligence" is not intelligent . It is pattern-matching and replication software. There is no "intelligence" involved.

          When I view AI images I see quite a bit of intelligence. If I ask the machine to draw a lake with mountains, trees and a castle the machine produces something with convincing GI, lighting, shadows, and reflections.

          Producing the appearance of intelligence (enough to fool you) is not the same as actually being intellilgent.

          ...
          While AI image generators almost entirely lack any sort of high level understanding of scenes

          Exactly. It doesn't have any understanding of what it's producing a picture of; it doesn't even know that there exists a real world with actual physical objects in it. It is producing an image matching what it calculates the pattern for images described by the words "lake with mountains and trees" should be.

    • Are they going to sue my brain because it stores a copy of their artwork and used it for inspiration in a derivative work?

      If you memorize how to draw Bart Simpson and then create a cartoon whose main character has yellow skin, spikey hair, on roller skates, and shouts things like, "Bite my shirt!" Then you will have a big problem.

      Copyright violations don't have to be verbatim. They just have to be close enough. And what constitutes "close enough" is very fluid from court to court. Look at Harry Potter, for instance. It is a spitting image of Star Wars, yet has never been hit by Disney for copyright violations.

    • by dpille ( 547949 )
      Are they going to sue my brain because it stores a copy of their artwork and used it for inspiration in a derivative work?

      YES. People that don't know what a derivative work is should probably stay out of discussions on copyright. They won't sue you over your brain, they'll sue you for having "used it for... a derivative work."
    • no, the human brain doesn't.

      if you want to draw traditionally and well, you need to do a set of very boring exercises like shading platonic solids under different light conditions, learn about human surface anatomy and their landmarks, learn about perspective, and then after countless practice drawing from life ... you may start getting somewhere that is if you are lucky with said strategy.

      looking (as in deeply studying) other people's artwork and photographs non-stop will only turn you into a pretty medioc

  • by Petersko ( 564140 ) on Wednesday January 24, 2024 @12:14PM (#64184705)

    I've been following discussions on here, in musician's forums, and sometimes on author blogs. There's usually a mishmash of responses that break into a few camps. "That's no different than how people learn", "If it can be reached on the internet it's fair game", "No it isn't fair game, copyright holds"... etc.. People want to shove it cleanly into a nicely packaged dichotomy - especially those who believe that AI "training" is the same as human "learning". Problem is, it's not. This is a new thing. It has parallels, but it's very fundamentally different, and it now touches copyright and fair use in ways never considered.

    So this is what has to happen - it has to be negotiated. In courts, in public opinion, in policy... the world will need to come to a new paradigm, under new consideration. Precedent should be a very minor player in this one - we shouldn't leap to demanding "which old way is best". I think "learning licenses" is a dodgy compromise... but at least it's a compromise.

    • Are there any articles or source papers that discuss how this is fundamentally different, and how it touches fair use differently? I'm not taking a position on whether your statement is right or wrong. I'm looking for more information and discussion on *how* using copyrighted works for AI training is not fair use and is different than fair use of those same materials when training humans.
      • by Calydor ( 739835 )

        The problem as I see it is that a human, no matter how skilled and which artists they have learned to imitate, is limited by time. We need to eat and sleep, and we can only draw so fast.

        AI does not need to eat or sleep (at least, not in ways that affect productivity) and can churn out such mass amounts of content that no human can ever dream of competing.

        • Is that an argument that applies to all technology? Think of all the humans that lost their jobs when calculators were invented (the job title for employees who performed those calculations was literally calculator, adding insult to injury). No human can ever dream to compete with a [machine] calculator. But I think it's fine to keep [machine] calculators around to the benefit of society, even if no humans can compete with them. Is there another argument that goes to fair use?
    • by WDot ( 1286728 ) on Wednesday January 24, 2024 @12:37PM (#64184807)
      This might be a reason to update copyright law. Generative AI is clearly not the same as printing a copy of a book and selling it without paying the author, or putting up a pirated files website and stuffing it with work you didn’t create. If one wants to “steal” an artist’s work, there are much easier methods than poking and prodding an AI until it serves up an image sufficiently close to one in its training set.

      Also, I would argue that the artists and writers really contributed very little to the final product. No artist or writer ever created their work with the intent that it would be used to inform an algorithm’s understanding of vision or language. If none of these AI companies ever existed, none of these artists or writers would ever collaborate to produce these models out of their pooled creative output. The real invention here was the machine learning algorithms and infrastructure that produced this model. Somebody’s deviantart portfolio of anime elf paintings, that they posted publicly online for free for people to look at, is only valuable insomuch as that it slightly inched the AI’s understanding of anime elf paintings in one direction or another.

      One other thing that annoys me about these lawsuits, is that they only exist now, after the technology has been developed to a level of maturity that there are viable profit-generating businesses to be made. For years academics have been publishing papers and showing off demos based on the same level of flagrant “copyright infringement,” and the public’s response has been “oh that’s neat.” Now that it can become a profitable product, the knives come out and everybody wants their piece.
      • by taustin ( 171655 )

        This might be a reason to update copyright law.

        That is the only possible end game, but it has a journey through the courts to get there. Once the Supreme Court rules, Congress has to decide if and how to update Title 17 to account for it.

        There's plenty of historical examples of how this works.

      • Re: (Score:2, Informative)

        by Anonymous Coward

        This might be a reason to update copyright law.

        Everyone needs to be very careful what they wish for, because you are absolutely 100% correct.

        Copyright law only protects distribution and public performance, neither of which have anything to do with training.
        (Note I mean actual *training* here, I am not referring to "AI" that spits out reproductions)

        In the US this would require a constitutional amendment, which is hard (for a reason), and these complainers KNOW it's hard, thus the endless failed lawsuits.

        Yet the last times we did update copyright law were

    • Even if it's the same as how the brain works, if I 'generated' mickey mouse, Disney could still sue me.

      • Yes. They couldn't sue the pen and paper you used to create the infringing image. They can't sue the tools, only the people that used them, and even then only if they used the image in a way that infringed upon the copyrights. If they hung their drawing of Mickey in their house for their private enjoyment, there's no infringement.

      • Even if it's the same as how the brain works, if I 'generated' mickey mouse, Disney could still sue me.

        Exactly this. The training, outputs, and uses are three separate issues that need to be dealt with individually. There should be no restrictions whatsoever on training because all they're doing is building a tool that in no way resembles the source material. If the tool is intended for illicit use or fails to take adequate precautions to preclude illicit use (e.g. "secure" phones marketed as cartels), we already have laws and precedent that cover those sorts of situations, so you go after the product, not t

    • by taustin ( 171655 )

      The fundamental definition of technology is that which amplifies the effects of a person's labor, what the military calls a "force multiplier." Where at one time, 95%+ of humanity had to spend their lives growing food to feed the other 5%, now the percentages are reversed - and everyone whose ancestors farmed had to find new work. A century or two ago, every village had a blacksmith, because you needed one. Now, we all use industrially produced steel products every day of our lives, and most of us have neve

  • by rsilvergun ( 571051 ) on Wednesday January 24, 2024 @12:28PM (#64184773)
    and it has almost incalculable value to replace coders, artists, writers and anyone who makes content of any kind.

    It really doesn't matter what the artists say or do here. There's just too much money involved. The laws will be changed to accommodate the owner class. Most likely using the courts as was done when Uber, Lyft, et al wanted to redefine what it means to be an employee. That's why folks like the Koch Brothers (now singular) spent 40 years packing the courts with pro-corporate judges while we were busy worrying about gay frogs and whatnot.
    • I completely agree. AI is poised to eventually do everything better than humans and we have no idea how far away we are from that reality. Once it hits, there will be little demand for goods or services provided by humans and I just don't see how capitalism fits into that world. Billions of people may suddenly have no income, not to mention any sense of purpose. We could transition to something other than capitalism, but I imagine the few people that own AI (and pretty much all other resources they've a
  • What is the basis in copyright law for asserting rights over being influenced by or learning from works whether the thing doing the learning be a computer or a person? Copyright protects the reproduction and performance of works and derivatives. It isn't a patent. It isn't an exclusive grant of authority over use, styles, information...etc.

    • I think this will be argued by attorneys for years. I'm sure it will be a way they will be earning their keep for a long time to come, with laws shifting one way or another.

      In any case, I'm sure billions to trillions will be given to the legal eagles to sort this one out, similar to how MP3s went from completely illegal to thashed out... but on a much larger scale. Of course, while Western companies battle for rights, I'm sure there will be LLMs from countries that don't care about IP vacuuming up any ima

    • by HBI ( 10338492 )

      US Constitution: Article I, Section 8, Clause 8: [The Congress shall have Power . . . ] To promote the Progress of Science and useful Arts, by securing for limited Times to Authors and Inventors the exclusive Right to their respective Writings and Discoveries.

      As language goes, that's pretty cut and dried.

      • US Constitution: Article I, Section 8, Clause 8: [The Congress shall have Power . . . ] To promote the Progress of Science and useful Arts, by securing for limited Times to Authors and Inventors the exclusive Right to their respective Writings and Discoveries.

        You are going to have to be a lot more specific than that to provide an answer responsive to my question. "What is the basis in copyright law for asserting rights over being influenced by or learning from works whether the thing doing the learning be a computer or a person?"

        As language goes, that's pretty cut and dried.

        You didn't even bother to quote any applicable copyright law.

      • Ah, but that is only the vague first cut of law: it all hinges upon the definition of "the exclusive Right".
        This gets elaborated upon in 17 U.S. Code 107 - Limitations on exclusive rights: Fair use:
        "Notwithstanding the provisions of sections 106 and 106A, the fair use of a copyrighted work, including such use by reproduction in copies or phonorecords or by any other means specified by that section, for purposes such as criticism, comment, news reporting, teaching (including multiple copies for classroom us

    • by taustin ( 171655 )

      The basis of the copyright claims is that what is being produced is derivative, which requires permission. The defense is that it's not, it's transformative, which does not.

      Same argument as the Google book scanning [wikipedia.org] lawsuit (which Google won at trial).

      It will be several years before the Supreme Court hears it, and they will, and several more while Congress dithers over whether or not to change the law in light of that ruling. Both sides make good enough arguments to keep it from being simple.

      • The basis of the copyright claims is that what is being produced is derivative, which requires permission.

        It will be several years before the Supreme Court hears it, and they will, and several more while Congress dithers over whether or not to change the law in light of that ruling. Both sides make good enough arguments to keep it from being simple.

        Personally I think this aspect is a lost cause because AI models themselves are clearly transformative works. What I think will continue to happen is what has been happening. A claim is made, the judge asks for evidence to support that claim, none is forthcoming.

        The more realistic legal questions have to do with whether model outputs from companies offering AI as a service outputting content deemed to be derivative constitutes a fixed copy, whether it is protected by fair use or whether it constitutes an

        • by taustin ( 171655 )

          I'm inclined to agree, but as things stand right now, unless there's a major change at the Library of Congress, only Congress can provide copyright protection for AI art. Whether or not they will is probably the biggest obstacle to widespread adoption on the AI art.

    • As far as I'm concerned, the training doesn't really matter. What matters is the output. If someone uses it to generate a fairly accurate reproduction of a copyrighted work, that's copyright infringement. It's an arbitrary distinction, left up to the courts or a jury.

      If you use it to make a pastiche, or critique of that work, that's fair use and it's fine.

      It's not the input that matters in copyright, it's the output.

  • The reason all of this keeps coming up is that these AI art generator sites seem to be promising one thing, but doing something very different in reality.

    The "claim" is that they ingest all of this human artwork and then "learn" from it so they're able to synthesize their own unique works of art based on the collective knowledge gained from processing the sum total of existing art it scraped or had fed into it.

    The reality? It seems to simply match up existing art to a user's requests for the type of art th

  • 100s of artist butthurt. Millions of people benefit. None of your original work is infringed on, And you still have the right to sell it in its (un)original form.

    If you were locked in a box your whole life and some how spontaneously generated your magnificent art, then you can talk.

    • oh, because girls really dig "prompt engineers"

      the embarrassment,
      of not being able to draw another human being well,
      with simple paper and pencil, while many of us still can code.

      once again: art is not learnt via osmosis.
      it is practice and theory. drawing from nature.

      the wide ignorance is astounding.

  • From now on, every one of them is BANNED from ALL art galleries. Nor may they own ANY books with pictures in them. Would hate to think that they'd infringe on some other artists' intellectual property by incorporating a color palette or scene they liked into their work.

  • ...by seeing the work of others

  • People say, "But they made a copy in order to train their model."
    About that: my browser makes a copy of images to the local cache when I visit websites.
    If making a local copy of something online is copyright infringement, then we're all guilty.

  • Then don't publish it for the world to see. This is essentially a bunch of artists complaining the work they published to be seen by others was seen by others. Are they going to sue anyone that draws inspiration from their work?

Air pollution is really making us pay through the nose.

Working...