Technology
This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.
Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.
Rules:
1: All Lemmy rules apply
2: Do not post low effort posts
3: NEVER post naziped*gore stuff
4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.
5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)
6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist
7: crypto related posts, unless essential, are disallowed
view the rest of the comments
This needs to be hammered into techbro's heads until they shut the fuck up about the so-called "AI" revolution.
I see it an more an inability to analyze, evaluate, and edit. A lot of "creativity" in the world of musical composition is putting together existing elements and seeing what happens. Any composer from pop to the very avant-garde, is influenced and sometimes even borrow from their predecessors (it's why copyright law is so complex in music).
It's the ability to make judgements, does this sound good/interesting, does this have value, would anyone want to listen to this, and adjust accordingly that will lead to something original and great. Humans are so good at this, we might be making edits before the notes hit the page (Brainstorming). This AI clearly wasn't. And deciding on value, seems wildly complex for modern day computers. Humans can agree on it (if you like Rock, but hate country for example).
So in the end, they are "creative" but in a monkey-typewritter situation, but who is going to sort through the billions of songs like this to find the one masterpiece?
One of the overlooked aspects of generative AI is that effectively by definition generative models can also be classifiers.
So let's say you were Spotify and you fed into an AI all the songs as well as the individual user engagement metadata for all those songs.
You'd end up with a model that would be pretty good at effectively predicting the success of a given song on Spotify.
So now you can pair a purely generative model with the classifier, so you spit out song after song but only move on to promoting it if the classifier thinks there's a high likelihood of it being a hit.
Within five years systems like what I described above will be in place for a number of major creative platforms, and will be a major profit center for the services sitting on audience metadata for engagement with creative works.
Right, the trick will be quantifying what is 'likely to be a hit', which if we're honest, has already been done.
Also, neural networks and other evolutionary algorithms can inject random perturbations/mutations to the system which, operate a bit like uninformed creativity (something like banging on a piano and hearing something interesting that's worth pursuing). So, while not 'inspired' or 'soulful' as we would generally think of it, these algorithms are capable of being creative In some sense. But it would need to be recognized as 'good' by someone or something...and back to your point.
What you described in your second paragraph is basically how image generation AI works.
Starting from random noise and gradually moving towards the version a classifier identifies as best matching the prompt.
Plenty of humans make those judgements about their own creations. And plenty of them get a shock when they release their creations to the masses and don't get the praise that they expected.
I believe that's vital to the creative process, but yeah, I basically agree.