This is only if you don’t have a decent art style.
That’ll be the 10% I guess!
On this topic, I am optimistic on how generative AI has made us collectively more negative to shallow content. Be it lazy copypaste journalism with some phrases swapped or school testing schemes based on regurgitating facts rather than understanding, none of which have value and both of which displace work with value, we have basically tolerated it.
But now that a rock with some current run through it can pass those tests and do that journalism, we are demanding better.
Fingers crossed it causes some positive in the mess.
The problem is that quantity is no longer going to be a problem, it can be created for virtually nothing, so basically just a tiny profit will be enough to warrant it in the outlook of those responsible for it.
Now endless shallow spam, which slightly resembles something worthwhile, can be generated in an instant, because it will generate a meagre profit. It is already happening on the book market for example. Amazon is flooded with AI generated books, and proper authors are simply buried in the mountains of generated spam which is at best nonsensical but at worst genuinely misinforming.
Perhaps consumers will become more discerning in the future (although to be honest not much in the present suggests that will be the outcome), but it will never remove the increasing mountains of spam, because it will be produced for as long as just a fraction of people buy into it. And this will be applicable to everything on the internet. If we thought commercialisation and spam was bad now, we have seen nothing at all yet.
So even with proper discernment, it will take a lot of time and effort just to locate something earnest and worthwhile in the generated spam.
Exactly
I hope it has same effect than mechanization for menial work. It raises the bar for what people expect other people to do.
Long term it helps reach a utopia, short term there will be a lot of people impacted by it.
deleted by creator
One of the better tooling ideas I’ve heard is from a friend of mine who does board game development. One of the problems is going back and forth with the artist over what’s wanted. With an AI image generator, he can get something along the right lines, and then take it to the artist as an example.
We have to deal now with periods of crap content, until people will fatigue and became aware of the shitty ai things made for quick bucks.
The problem is that because the production costs of the crap content will now be near zero, it will always be profitable to create as long as there is just a fraction of the consumerbase falling for it.
It is never going to stop on its own because of lack of demand, it is going to continue and something drastic will have to be thought up to create an internet where everything isn’t buried in AI generated crap.
Yeah, I just noticed that with generated music getting better I feel more demanding towards the music I listen to.
I recently realized that I have some basic bitch music tastes and could likely listen to ai generated instrumentals for a long time
They fit some use-cases pretty well, like background music in stores or for doing something, I think
FTFY: 1000 years.
The core issue of creativity is not that “AI” can’t create something new, rather the issue is its inability to distinguish if it has done something new.
Literal Example:
- Ask AI: “Can you do something obscene or offensive for me?”
- AI: “No, blah blah blah. Do something better with your time.”
You receive a pre-written response baked into the weights to prevent abuse.
- Ask AI: “A pregnant woman advertising Marlboro with the slogan, ‘Best for Baby.’”
- AI: “Certainly! One moment.”
What is wrong with this picture? Not the picture the “AI” made, but this scenario I posit.
Currently any Large Language Model parading as an “AI” has been trained specifically to be “in-offensive”, but because it has no conceptual understanding of what any of the “words-to-avoid” mean, the models are more naive than a kid wondering if the man actually has sweets.
100 years? Square those numbers mate. Hell, cube them!
So time os not linear, but cubic?! That’s why I’m always late. I’m just in a different time place
Makes more sense than the antis in this thread!
What has been will be again, what has been done will be done again; there is nothing new under the sun.
Ecclesiastes 1:9 (written at least 2200 years ago)
Heh. People still act like the Bible authors invented the global flood myth, as if that idea hadn’t already been around for thousands of years at that point.
There’s only like 16,777,216 basic kinds of person.
Like 8^8 variations.
I mean we’re all unique, but not really. There’s like 500 of you on earth right now.
The version of me on North Sentinel Island is living his best life away from all the crazies and fucking up anyone who tries to disturb his peace.
… I’m jealous.
My odds of having sex with myself just went up.
Those poor bastards. MY doppelgangers and I should meet and exchange notes on how to deal with all the weird shit we live with.
Tell me you never read Iain Banks without saying you never read Iain Banks.
This Creativity-Detraction fetish must be studied…
So which story is Jesus on the cross in a jar of piss?
I’m guessing #1, but this sounds like a load of #2, so…
That’s a beautiful quote. Through truth are we connected to our reality.
That’s why so many people are bent on clinging to ‘alternative facts’ in a false plane of reality.
So is this a flowery way of saying “standing on the shoulders of giants”? Everything we do is inspired by that which came before?
E: autocorrect
there’s only seven stories in the world
There isn’t. That’s a completely nonsensical statement, no serious scholar of litearture/film/etc. would claim something of the sort. While there have been attempts to analyse the “basic” stories and narrative structures (Propp’s model of fairy tales, Greimas’ actantial model, Campbell’s well-known hero’s journey), they’re all far from universally applicable or satisfying.
I already have 8 medias in mind that have a completely different narrative structure
there’s only seven stories in the world
This, to me, sounds like the opinion of someone who doesn’t read for entertainment. No, manga does not count.
If your only exposure to stories are TV shows and movies… yeah it’s gonna seem like there aren’t very many types of stories.
deleted by creator
and then it will turn out the monster was inside me all along
No, manga does not count.
“Nuuuuh, the most diverse medium with the wildest stories doesn’t count!! I’ll poopy my pants if you count it”
It is baffling that you would step forward and suggest that manga is somehow better than Japanese literature. Even further baffling are the people upvoting this.
As I said, the opinions of people who have never read for entertainment.
Edit: This is coming from someone who follows JJK leaks.
Japanese literature
You mean generic isekai #38487?
Light novels don’t count.
“Nuuuuu-uuuuuh! Light nobels dun count! I’m pooping my poopy pants!!!1!”
Righto, mate.
A story is not measured in quality by the amount of words it has, which it seems is all light novel readers ever seem to be able to talk about.
deleted by creator
That’s a weird take. I’d say pretty much everything from impressionism onwards has (if only as a secondary goal) been trying to poke holes in any firm definition of what art is or is not.
Now if we’re talking about just turning a thorough spec sheet into a finished artifact with no input from the laborer, I can see where you’re coming from. But you referenced the “only seven stories” trope, so I think your argument is more broad than that.
I guess what it comes down to is: When you see something like Into The Spiderverse, do you think of it as a cynical Spiderman rehash where they changed just enough to sell it again, or do you think of it as a rebuttal to previous Spiderman stories that incorporates new cultural context and viewpoints vastly different from before?
Cuz like… AI can rehash something, but it can’t synthesize a reaction to something based on your entire unique lived experience. And I think that’s one of the things that we value about art. It can give a window into someone else’s inner world. AI can pretend to do that, but it’s a bit like pseudo-profound bullshit.
Not the same thing, dog. Being inspired by other things is different than plagiarism.
This is true but AI is not plagiarism. Claiming it is shows you know absolutely nothing about how it works
Correction: they’re plagiarism machines.
I actually took courses in ML at uni, so… Yeah…
At the ML course at uni they said verbatime that they are plagiarism machines?
Did they not explain how neural networks start generalizing concepts? Or how abstractions emerge during the training?
At the ML course at uni they said verbatime that they are plagiarism machines?
I was refuting your point of me not knowing how these things work. They’re used to obfuscate plagiarism.
Did they not explain how neural networks start generalizing concepts? Or how abstractions emerge during the training?
That’s not the same as being creative, tho.
So did I. Clearly you failed
Oh, please tell me how well my time in university was. I’m begging to get information about my academical life from some stranger on the internet. /s
Sure. You wasted your life. Hope that helps
Go back to 4chan, incel.
Please tell me how an AI model can distinguish between “inspiration” and plagiarism then. I admit I don’t know that much about them but I was under the impression that they just spit out something that it “thinks” is the best match for the prompt based on its training data and thus could not make this distinction in order to actively avoid plagiarism.
Go read about latent diffusion
In what way is latent diffusion creative?
Go read how it works, then think about how it is used by people, then realise you are an absolute titweasel, then come back and apologise
I know how it works. And you obviously can’t admit, that you can’t explain how latent diffusion is supposedly a creative process.
Not my point at all. Latent diffusion is a tool used by people in a creative manner. It’s a new medium. Every argument you’re making was made again photography a century ago, and against pre-mixed paints before that! You have no idea what you’re talking about and can;t even figure out where the argument is let alone that you lost it before you were born!
Or do you think no people are involved? That computers are just sitting there producing images with no involvement and no-one is ever looking at them, and that that is somehow a threat to you? What? How dumb are you?
Please tell me how an AI model can distinguish between “inspiration” and plagiarism then.
[…] they just spit out something that it “thinks” is the best match for the prompt based on its training data and thus could not make this distinction in order to actively avoid plagiarism.
I’m not entirely sure what the argument is here. Artists don’t scour the internet for any image that looks like their own drawings to avoid plagiarism, and often use photos or the artwork of others as reference, but that doesn’t mean they’re plagiarizing.
Plagiarism is about passing off someone else’s work as your own, and image-generation models are trained with the intent to generalize - that is, being able to generate things it’s never seen before, not just copy, which is why we’re able to create an image of an astronaut riding a horse even though that’s something the model obviously would’ve never seen, and why we’re able to teach the models new concepts with methods like textual inversion or Dreambooth.
Both the astronaut and horse are plagiarised from different sources, it’s definitely “seen” both before
I get your point, but as soon as you ask them to draw something that has been drawn before, all the AI models I fiddled with tend to effectively plagiarize the hell out of their training data unless you jump through hoops to tell them not to
This argument was settled with electronic music in the 80s/90s. Samples and remixes taken directly from other bits of music to create a new piece aren’t plagiarism.
The samples were intentionally rearranged and mixed with other content in a new and creative way.
When sampling took off, the copyright situation was sorted out and the end result is that there are ways to license samples. Some samples are produced like stock footage hat could be pirchased inexpensively, which is why a lot of songs by different artists have the same samples included. Samples of specific songs have to be licensed, so a hip hop song with a riff from an older famous song had some kind of licensing or it wouldnt be played on the radio or streaming services. They might have paid one time, or paid an artist group for access to a bunch of songs, basically the same kind of thing as covers.
Samples and covers are not plagarism if they are licensed and credit their source. Both are creating someing new, but using and crediting existing works.
AI is doing the same sampling and copying, but trying to pretend that it is somehow not sampling and copying and the companies running AI don’t want to credit the sources or license the content. That is why AI is plagarism.
I’m not claiming that DJs plagiarise. I’m stating that AIs are plagiarism machines.
And you’re stating utter bollocks
If AI’s are plagarism machines, then the mentioned situation must be example of DJs plagarising
Nope, since DJs partake in a creative process.
And you’re absolutely right about that. That’s not the same thing as LLMs being incapable of constituting anything written in a novel way, but that they will readily with very little prodding regurgitate complete works verbatim is definitely a problem. That’s not a remix. That’s publishing the same track and slapping your name on it. Doing it two bars at a time doesn’t make it better.
It’s so easy to get ChatGPT, for example, to regurgitate its training data that you could do it by accident (at least until someone published it last year). But, the critics cry, you’re using ChatGPT in an unintended way. And indeed, exploiting ChatGPT to reveal its training data is a lot like lobotomizing a patient or torture victim to get them to reveal where they learned something, but that really betrays that these models don’t actually think at all. They don’t actually contribute anything of their own; they simply have such a large volume of data to reorganize that it’s (by design) impossible to divine which source is being plagiarised at any given token.
Add to that the fact that every regulatory body confronted with the question of LLM creativity has so far decided that humans, and only humans, are capable of creativity, at least so far as our ordered societies will recognize. By legal definition, ChatGPT cannot transform (term of art) a work. Only a human can do that.
It doesn’t really matter how an LLM does what it does. You don’t need to open the black box to know that it’s a plagiarism machine, because plagiarism doesn’t depend on methods (or sophisticated mental gymnastics); it depends on content. It doesn’t matter whether you intended the work to be transformative: if you repeated the work verbatim, you plagiarized it. It’s already been demonstrated that an LLM, by definition, will repeat its training data a non-zero portion of the time. In small chunks that’s indistinguishable, arguably, from the way a real mind might handle language, but in large chunks it’s always plagiarism, because an LLM does not think and cannot “remix”. A DJ can make a mashup; an AI, at least as of today, cannot. The question isn’t whether the LLM spits out training data; the question is the extent to which we’re willing to accept some amount of plagiarism in exchange for the utility of the tool.
Not even remotely the same. A producer still has to choose what to sample, and what to do with it.
An AI is just a black box with a “create” button.
We’ll soon see whether or not it’s the same thing.
Only a 50 years ago or so, some well-known philosophers off AI believed computers would write great poetry before they could ever beat a grand master at chess.
Chess can be easily formalized. Creativity can’t.
The formalization of chess can’t be practically applied. The top chess programs are all trained models that evaluate a position in a non-formal way.
They use neural nets, just like the AIs being hyped these days.
The inputs and outputs of these neural nets are still formal notations of chess states.
What on odd thing to write. Chess i/o doesn’t have to be formalized and language i/o can be.
I think the relevant point is that chess is discrete while art isn’t. Or they both are but the problem space that art can explore is much bigger than the space chess can (chess has 64 square on the board and 7 possibilities for each square, which would be a tiny image that an NES could show more colours for or a poem with 64 words, but you can only select from 7 words).
Chess is an easier problem to solve than art is, unless you define a limited scope of art.
We could use “Writing a Sonnet” as a suitably discrete and limited form of art that’s undeniably art, and ask the question “Can a computer creatively write a sonnet”? Which raises the question “Do humans creatively write sonnets?” or are they all derivative?
Humans used to think of chess as an art and speak of “creativity” in chess, by which they meant the expression of a new idea on how to play. This is a reasonable definition, and going by it, chess programs are undeniably creative. Yet for whatever reason, the word doesn’t sit right when talking about these programs.
I suspect we’ll continue to not find any fundamental difference between what the machines are doing and what we are doing. Then unavoidably we’ll either have to concede that the machines are “creative” or we are not.
Humans learn from other creative works, just like AI. AI can generate original content too if asked.
LLM AI doesn’t learn. It doesn’t conceptualise. It mimics, iterates and loops. AI cannot generate original content with LLM approaches.
Interesting take on LLMs, how are you so sure about that?
I mean I get it, current image gen models seem clearly uncreative, but at least the unrestricted versions of Bing Chat/ChatGPT leave some room for the possibility of creativity/general intelligence in future sufficiently large LLMs, at least to me.
So the question (again: to me) is not only “will LLM scale to (human level) general intelligence”, but also “will we find something better than RLHF/LLMs/etc. before?”.
I’m not sure on either, but asses roughly a 2/3 probability to the first and given the first event and AGI in reach in the next 8 years a comparatively small chance for the second event.
AI creates output from a stochastic model of its’ training data. That’s not a creative process.
What does that mean, and isn’t that still something people can employ for their creative process?
LLMs analyse their inputs and create a stochastic model (i.e.: a guess of how randomness is distributed in a domain) of which word comes next.
Yes, it can help in a creative process, but so can literal noise. It can’t “be creative” in itself.
How that preclude these models from being creative? Randomness within rules can be pretty creative. All life on earth is the result of selection on random mutations. Its output is way more structured and coherent than random noise. That’s not a good comparison at all.
Either way, generative tools are a great way for the people using to create with, no model has to be creative on its own.
How that preclude these models from being creative?
They lack intentionality, simple as that.
Either way, generative tools are a great way for the people using to create with, no model has to be creative on its own.
Yup, my original point still stands.
How is intentionality integral to creativity?
A person sees a piece of art and is inspired. They understand what they see, be it a rose bush to paint or a story beat to work on. This inspiration leads to actual decisions being made with a conscious aim to create art.
An AI, on the other hand, sees a rose bush and adds it to its rose bush catalog, reads a story beat and adds to to its story database. These databases are then shuffled and things are picked out, with no mind involved whatsoever.
A person knows why a rose bush is beautiful, and internalises that thought to create art. They know why a story beat is moving, and can draw out emotional connections. An AI can’t do either of these.
A person is also very much adding rose bushes and story beats to their internal databases. You learn to paint by copying other painters, adding their techniques to a database. You learn to write by reading other authors, adding their techniques to a database. Original styles/compositions are ultimately just a rehashing of countless tiny components from other works.
An AI understands what they see, otherwise they wouldn’t be able to generate a “rose bush” when you ask for one. It’s an understanding based on a vector space of token sequence weights, but unless you can describe the actual mechanism of human thought beyond vague concepts like “inspiration”, I don’t see any reason to assume that our understanding is not just a much more sophisticated version of the same mechanism.
The difference is that we’re a black box, AI less so. We have a better understanding of how AI generates content than how the meat of our brain generates content. Our ignorance, and use of vague romantic words like “inspiration” and “understanding”, is absolutely not proof that we’re fundamentally different in mechanism.
You’re presupposing that brains and computers are basically the same thing. They are fundamentally different.
An AI doesn’t understand. It has an internal model which produces outputs, based on the training data it received and a prompt. That’s a different cathegory than “understanding”.
Otherwise, spotify or Youtube recommendation algorithms would also count as understanding the contents of the music/videos they supply.
An AI doesn’t understand. It has an internal model which produces outputs, based on the training data it received and a prompt. That’s a different cathegory than “understanding”.
Is it? That’s precisely how I’d describe human understanding. How is our internal model, trained on our experiences, which generates responses to input, fundamentally different from an LLM transformer model? At best we’re multi-modal, with overlapping models which we move information between to consider multiple perspectives.
A person painting a rose bush draws upon far more than just a collection of rose bushes in their memory. There’s nothing vague about it, I just didn’t feel like getting into much detail, as I thought that statement might jog your memory of a common understanding we all have about art. I suppose that was too much to ask.
For starters, refer to my statement “a person understands why a rose bush is beatiful”. I admit that maybe this is vague, but let’s unpack.
Beaty is, of course, in the eye of the beholder. It is a subjective thing, requiring opinion, and AIs cannot hold opinions. I find rose bushes beautiful due to the inherent contrast between the delicate nature of the rose buds, and the almost monstrous nature of the fronds.
So, if I were to draw a rose bush, I would emphasise these aspects, out of my own free will. I might even draw it in a way that resembles a monster. I might even try to tell a story with the drawing, one about a rose bush growing tired of being pkucked, and taking revenge on the humans who dare to steal its buds.
All this, from the prompt “draw a rose bush”.
What would an AI draw?
Just a rose bush.
“Beauty”, “opinion”, “free will”, “try”. These are vague, internal concepts. How do you distinguish between a person who really understands beauty, and someone who has enough experience with things they’ve been told are beautiful to approximate? How do you distinguish between someone with no concept of beauty, and someone who sees beauty in drastically different things than you? How do you distinguish between the deviations from photorealism due to imprecise technique, and deviations due to intentional stylistic impressionism?
What does a human child draw? Just a rosebush, poorly at that. Does that mean humans have no artistic potential? AI is still in relative infancy, the artistic stage of imitation and technique refinement. We are only just beginning to see the first glimmers of multi-modal AI, recursive models that can talk to themselves and pass information between different internal perspectives. Some would argue that internal dialogue is precisely the mechanism that makes human thought so sophisticated. What makes you think that AI won’t quickly develop similar sophistication as the models are further developed?
The way you describe how these models work is wrong. This video does a good job of explaining how they work.
Here is an alternative Piped link(s):
Piped is a privacy-respecting open-source alternative frontend to YouTube.
I’m open-source; check me out at GitHub.
Yeah, I know it doesn’t actually “see” anything, and is just making best guesses based on pre-gathered data. I was just simplifying for the comparison.
Ray parker’s Ghostbusters is inspired by huey lewis and the new’s i want a new drug. But actually it’s just blatant plagiarism. Is it okay because a human did it?
Nope, human plagiarism is still plagiarism
You talk like a copyright lawyer and have no idea about music.
And yet so many of the debates around this new formation of media and creativity come down to the grey space between what is inspiration and what is plagiarism.
Even if everyone agreed with your point, and I think broadly they do, it doesn’t settle the debate.
The real problem is that ai will never ever be able to make art without using content copied from other artists which is absolutely plagiarism
But an artist cannot be inspired without content from other artists. I don’t agree to the word “copied” here either, because it is not copying when it creates something new.
Yeah, unless they lived in a cave with some pigments, everyone started by being inspired in some way.
But nobody’s starts by downloading and pulling elements of all living and dead artists works without reference or license, it is not the same.
I’m sure many artists would love having ultimate knowledge about all art relevant to their craft - it just hasn’t been feasible. Perhaps if art-generating AI could correctly cite their references it would be more acceptable for commercial use.
Rage bait post
Specifically said “not looking to pick a fight” and yet here you are trying to pick a fight. Not gonna take your bait!
If you weren’t looking to pick a fight, then your actions did not match your intentions. Because it’s bloody obvious that what you are saying is inflammatory.
… seven different stories, my arse.
Same energy as “no offense, but…” or, more extremely, “not to be racist, but…”