this is actually a really debatable argument. If you’re buying it first hand, from somebody trying to make money, yes it could arguably be unethical, but if you’re buying it second hand, i.e. someone who just doesn’t want it anymore, you could make the argument that it’s an ethically net positive transaction. Since they no longer want the diamond, and wish to have money instead, and you wish to have the diamond, and less money. Everybody wins in 2nd hand deals, weirdly enough.
Setting aside “there is no ethical consumption under capitalism”, which is a debatable for another time:
I don’t totally agree with your assessment of 2nd hand sales: it’s not ethical positive, at best it’s ethically neutral, because demand trickles up the market. I could go into this more, but ultimately it it’s irrelevant:
The 2nd hand LLM market doesn’t work like that because LLMs are sold as a service. The LLM producers take a cut from all LLM resellers.
You could make a case that self hosting a free open source LLM like OLlama is ok, but that’s not how most LLMs are distributed.
In this analogy, using the diamond does use it up. In the sense that none else can use that diamond concurrently. If someone else wants a diamond, more children must die.
This is different from the trained AI model, which can concurrently be used by everyone at the same time, at very little extra cost.
Even if the diamond mine owners stop mining, it’s unethical to buy their stockpile of blood diamonds.
Also, there is a cost besides electricity - the theft of artist’s work is inherent to the use of the model, not just in the training. The artist is not being compensated whenever an AI generates art in their style, and they may in fact lose their job or have their compensation reduced due to artificial supply.
Finally, this is an analogy, it’s not perfect. Picking apart incidental parts of the analogy doesn’t really prove anything. Use an analogy to explain a problem, but don’t pick apart an analogy as though you’re picking apart the problem.
and they may in fact lose their job or have their compensation reduced due to artificial supply.
highly doubt. Any artists that do lose their job are probably mostly ok with it anyway, since it’s most likely going to be graphical drivel anyway. In fields like media theres a different argument to be made, but even then it’s iffy sometimes. Also i don’t think this would be considered artificial supply, it would be artificially insisted demand instead no? Or perhaps an inelastic demand side expectation.
Although, it would be nice to have some actual concrete data on artists and job prospects in relation to AI. Unfortunately it’s probably too early to tell right now, since we’re just out of the Luddite reactionary phase, who knows.
Any artists that do lose their job are probably mostly ok with it anyway, since it’s most likely going to be graphical drivel anyway.
Replace “artist” and “graphical”, and you just described most jobs. I don’t think most people are ok losing their jobs even if those jobs aren’t especially inherently rewarding; they’re getting paid for their area of training. They’re not just gonna be able to find a new job because in this hypothetical, the demand for it (a living human doing the work) is gone.
I consider this an increase in supply because it’s an increase in the potential supply. Productivity increases (which is what this is) mean you can make more, which drives down the price, which means that artists get paid less (or just get replaced).
Remember: if you 10x the productivity of an employee, that typically doesn’t mean you produce 10x the product, it typically means you need 1/10th the employees. That payroll saving goes right into the pockets of execs.
Also wrt luddites, they weren’t wrong. It did absolutely demolish their industry and devastate the workers. It’s just that the textile industry was only a small part of the economy, and there were other industries who could absorb the displaced workers after they got retrained.
LLMs threaten almost every industry, so there is a greater worker impact and fewer places for displaced workers to go. Also now workers are responsible for bearing the costs of their own retraining, unlike back in the day of the luddites.
A very large amount of those dug up diamonds end up as “industrial diamonds.” Because they are far from gemstone quality. And they definitely get used up. I have used up my share of them as cutting tools when I was a toolmaker.
I know the analogy isn’t a perfect fit for LLMs in general. Analogies never come close to describing the entire thing they’re analogs of, they don’t need to.
It doesn’t matter because this is a suitable analogy for the argument. This is how analogies work.
The argument is that because the harm has already been done, it’s fine to use LLMs.
That same argument can be made for blood diamonds, and it’s untrue for exactly the same reason:
Because buying the use of LLMs (which is mostly how they’re used, even if you pay in data instead of money) is funding the continued harmful actions of the producer.
I can’t believe I have to explain how analogies are used to a grown ass adult.
No, using an already-trained model doesn’t “use up” the model in exactly the same way that pirating a movie doesn’t steal anything from Hollywood.
Use a diamond doesn’t “use up” that diamond.
And yet, it’s still unethical to buy already mined blood diamonds from people who continue to mine more blood diamonds. Funny thing about that, huh
this is actually a really debatable argument. If you’re buying it first hand, from somebody trying to make money, yes it could arguably be unethical, but if you’re buying it second hand, i.e. someone who just doesn’t want it anymore, you could make the argument that it’s an ethically net positive transaction. Since they no longer want the diamond, and wish to have money instead, and you wish to have the diamond, and less money. Everybody wins in 2nd hand deals, weirdly enough.
Setting aside “there is no ethical consumption under capitalism”, which is a debatable for another time:
I don’t totally agree with your assessment of 2nd hand sales: it’s not ethical positive, at best it’s ethically neutral, because demand trickles up the market. I could go into this more, but ultimately it it’s irrelevant:
The 2nd hand LLM market doesn’t work like that because LLMs are sold as a service. The LLM producers take a cut from all LLM resellers.
You could make a case that self hosting a free open source LLM like OLlama is ok, but that’s not how most LLMs are distributed.
In this analogy, using the diamond does use it up. In the sense that none else can use that diamond concurrently. If someone else wants a diamond, more children must die.
This is different from the trained AI model, which can concurrently be used by everyone at the same time, at very little extra cost.
Even if the diamond mine owners stop mining, it’s unethical to buy their stockpile of blood diamonds.
Also, there is a cost besides electricity - the theft of artist’s work is inherent to the use of the model, not just in the training. The artist is not being compensated whenever an AI generates art in their style, and they may in fact lose their job or have their compensation reduced due to artificial supply.
Finally, this is an analogy, it’s not perfect. Picking apart incidental parts of the analogy doesn’t really prove anything. Use an analogy to explain a problem, but don’t pick apart an analogy as though you’re picking apart the problem.
highly doubt. Any artists that do lose their job are probably mostly ok with it anyway, since it’s most likely going to be graphical drivel anyway. In fields like media theres a different argument to be made, but even then it’s iffy sometimes. Also i don’t think this would be considered artificial supply, it would be artificially insisted demand instead no? Or perhaps an inelastic demand side expectation.
Although, it would be nice to have some actual concrete data on artists and job prospects in relation to AI. Unfortunately it’s probably too early to tell right now, since we’re just out of the Luddite reactionary phase, who knows.
Replace “artist” and “graphical”, and you just described most jobs. I don’t think most people are ok losing their jobs even if those jobs aren’t especially inherently rewarding; they’re getting paid for their area of training. They’re not just gonna be able to find a new job because in this hypothetical, the demand for it (a living human doing the work) is gone.
I consider this an increase in supply because it’s an increase in the potential supply. Productivity increases (which is what this is) mean you can make more, which drives down the price, which means that artists get paid less (or just get replaced).
Remember: if you 10x the productivity of an employee, that typically doesn’t mean you produce 10x the product, it typically means you need 1/10th the employees. That payroll saving goes right into the pockets of execs.
Also wrt luddites, they weren’t wrong. It did absolutely demolish their industry and devastate the workers. It’s just that the textile industry was only a small part of the economy, and there were other industries who could absorb the displaced workers after they got retrained.
LLMs threaten almost every industry, so there is a greater worker impact and fewer places for displaced workers to go. Also now workers are responsible for bearing the costs of their own retraining, unlike back in the day of the luddites.
A very large amount of those dug up diamonds end up as “industrial diamonds.” Because they are far from gemstone quality. And they definitely get used up. I have used up my share of them as cutting tools when I was a toolmaker.
Ok cool, but this is an analogy. Why are you defending the use of AI by megacorps by objecting to irrelevant parts of an analogy on technicality?
It’s a bad analogy and just plain wrong fact. Do better.
You’re insufferable.
I know the analogy isn’t a perfect fit for LLMs in general. Analogies never come close to describing the entire thing they’re analogs of, they don’t need to.
It doesn’t matter because this is a suitable analogy for the argument. This is how analogies work.
The argument is that because the harm has already been done, it’s fine to use LLMs.
That same argument can be made for blood diamonds, and it’s untrue for exactly the same reason:
Because buying the use of LLMs (which is mostly how they’re used, even if you pay in data instead of money) is funding the continued harmful actions of the producer.
I can’t believe I have to explain how analogies are used to a grown ass adult.