Generative AI doesn’t copy art, it ‘clones’ the artisans — cheaply – R&A IT Strategy & Architecture

There is much discussion on what Generative AI is going to mean for jobs. The best contribution I read during the last couple of months came from Brian Merchant, author of Blood in the Machine (which is the title of his substack as well as the book he wrote about the Luddites). Brian draws insightful comparisons between what happens today and what happened during the early Industrial Revolution.

The fact that GenAI results may be shoddy is often used as a critique by people saying “they cannot replace humans”. But what Brian pointed out is that GenAI doesn’t need to be particularly good to replace labour (see here and here). The first machines that replaced artisans produced pretty shoddy stuff too. But it was ‘good enough’ to replace them.

In fact, what the technology opened was the possibility of ‘cheap. And that is ‘cheap’ in both meanings. It was this ‘cheap’ that disrupted the labour market, really good machines came much, much later. And this echoes what some companies are doing now, slashing creative stuff like writing and graphics in marketing and replacing it with mostly shoddy, boring, cheap results, we can make a lot of. As a result, wages in those areas are going down, the remaining people have far less enjoyable jobs.

So, what kind of jobs are at risk? Well, jobs where cheap, shoddy work can have real value. So, shoddy graphic art. Shoddy text descriptions of products.

And by the way, people, coding is not such a job! Code must be reliable and GOOD. Information Technology itself is extremely brittle, after all (so ‘GenAI creating better GenAI’, is not a realistic trajectory). If I still was a software engineer, I would not worry about being replaced. I would worry that a few years from now your work in part will consist of having to repair and refactoring lots of brittle, shoddy GenAI-generated code.

But there is also a fundamental difference with the beginning of the Industrial Revolution

Generative AI is trained on (hopefully good/correct/worthwhile) output created by humans. It stores information about the token/pixel patterns of that output in a volume of parameters. With those parameters it can create new patterns based on a combination of the statistical relations in the old patterns and the throwing of the dice. Sometimes there is so much information about a certain data in the parameters that it can actually recreate the training data. When we like that effect it is called ‘memorisation’. When we dislike it it is called ‘data leakage’. But the wanted and unwanted version are one and the same thing. The unwanted version now leads to for instance The New York Times suing OpenAI on a copyright breach when they used the NYT website to train GPT. It is interesting to note that neither NYT nor OpenAI wanted this to happen. But OpenAI was unable to prevent it. OpenAI might be able to find a solution for a few sources such as NYT (e.g. a special model overtrained on a certain source that recognises plagiarism), but it is hard to see how they can completely solve this.

But there is a different version of this problem that isn’t based on copyright but on trademark. See this example (from that previous post on memorisation):

None of the images are original art. There is technically no copyright breach here. But you immediately know that there is a problem. No way Disney will let you generate your own Pixar-style animation or Donald Duck stories. And there are many more examples. For the most popular styles there now may be band-aids, but the problem runs much deeper.

I realised recently that the kind of shift here is not so much copying creations but ‘copying’ (‘cloning’) creators. By using the creations, GenAI is ‘sort-of-cloning the (collective) creator(s)’ without consent. This — while innovative — is also a form of theft, as none of these creators gave permission to be ‘sort-of-cloned’ that way.

Our intellectual property rights were not designed for what GenAI does. Even if your work is not copied, your creativity, your style, etc. — in a ‘cheap’ way — is. This is indeed a bit like what happened at the start of the Industrial Revolution. But… the big difference is that weaving (pottery etc.) machines were independent inventions (the machines might have been based on understanding the principles of the art, but they were not ‘trained’ on the actual output of the artisans.

What we might therefore need is an extension of ‘intellectual property rights’ to ‘intellectuality rights‘ (bypassing the potential solution to state that you ‘own yourself’ — but maybe that is even an effective shortcut).

Now, GenAI is trained on a large collection of ‘the output of artisan skill’ and it depends on the complex interplay of training data, parameter volume (more on that another time) how much of an individual skill is getting ‘cloned’. The goal of the GenAI people is not have recognisable ‘cloning’ at all, they want to use all those skills to make an ‘independent skilled agent’. And let’s be fair: humans are mostly copying machines too. The whole enterprise rests on that assumption, namely that the technique in the end is — like with humans learning — resulting in independent ‘intelligence’. Assessments In how far that is attainable vary. We might as well get some sort of shoddy average approximation without actual intelligence (e.g. understanding your own output — see below)

Anyway, legally, none of this will be easy. Years of popcorn, legal careers and all that.

Post Scriptum

I generally do not use GenAI to create my visuals. I find the overabundance of GenAI art (e.g. on the last conference I went to) quite boring and irritating. But in this case, I wanted an illustration that shows the ‘cloning of the artist, not the artwork’. So I went to ChatGPT to create an example (assuming it was multimodal these days). To prevent running into a filter, I wanted something in the style of Dutch Golden Age painter Rembrandt van Rijn. So, I asked it

and ChatGPT replied:

I see what ChatGPT is doing here. It is creating a prompt to hand to the drawing model. And indeed it continued with

With such a good prompt, I expected a great result. This is what I got:

I must say, ChatGPT has acquired a real sense of humour. Or it is a nice example of completely not understanding your own output.

(It’s the output of python code using matplotlib and numpy)

So, I took ChatGPT’s nice prompt and gave it to another engine (don’t recall which one) and it produced the one on the right at the top of the page. Of course, zoom in and you’ll notice GenAI’s ‘cheap’ (but fair: both the robot and the man in the painting in the background have weird hands):

It’s shoddy. It’s ugly. It’s cheap. But it does the job. And that is what’s going to be exploited. We can count on it.

This article is part of the ChatGPT and Friends Collection. In case you found any term or phrase here unclear or confusing (e.g, I can understand that most people do not immediately know what a ‘token’ is (and it is relevant to understand that), that ‘context’ in LLMs is whatever has gone before in human (prompt) and LLM (reply) generated text, before producing the next token), you can probably find a clear explanation there.

Must Read

Related Articles