This will only increase.
Not every such images has a distinct mark that denotes it as AI-generated. They could be mistaken for real photograph or real work of (digital) art by a human. Especially by an algorithm.
I also understand a lot of today's cutting-edge models are trained on images scraped from the web. Not sure what curation happens but it cannot be foolproof.
Will future AI models that generate "realistic" images feed on this as input and generate images that mimic some of these attributes -- creating some kind of feedback loop that will eco for generations of models?
Has anyone already thought of such issues -- not just with images but with AI-generated text, data, music, etc.
Curious to know what is the thinking of this group here.
Especially with text there's an arms race to make undetectable AI text for blogspam and similar purposes. It's going to end up like carbon dating: once nuclear weapons were used in the atmosphere, everything ended up contaminated and had to be accounted for. https://www.radiocarbon.com/carbon-dating-bomb-carbon.htm
The future will include humans claiming AI art as their own, possibly touched up a bit, and AIs claiming human art as their own.
I was tinkering with Stable Diffusion yesterday to come up with ideas for an apartment interior design. These aren't even cherry picked, you can generate images like this, one every 10 seconds or so, for as long as you like:
All of these beautiful things have been degraded by the inauthentic, focus-group, advertising data harvesting machines of mega-corporate greed. Unique websites and blogs are drowned out into oblivion, unprofitable and hidden by the SEO Gods of Google, funneling you into their own products and advertising pathways. Forums bled out into Reddit, which is now an astroturfed corporate dream world where advertisers can masquerade as real users and corporate appointed moderators funnel all conversation into the optimum advertising framework--deleting anything that could harm reddit's shareholder pool of giant corporations and governments. Video games went from novel, artistic experiments, to hyper-optimized addiction machines built to drain the time, money, and drive from their young audience. Even memes, with all their raw vulgarity and juvenile silliness, have been coopted by corporations trying to bend this new form of expression to their advertising goals.
First, content online was authentic and human. Then, big tech started trimming and censoring and funneling and optimizing it into something less real... less human, but far more ripe for advertising revenue and data collection. Now, we are entering the stage of AI-generated content. Articles written by algorithm, art created by machine, bots filling up the whole internet with noise. The level of distrust, paranoia and questioning of reality that users will experience online in the coming years will be unparalleled. Is this image real? Is this person I'm talking to a bot? Is this artwork human made?
Which brings me back to my main point. Authenticity will be the new luxury. And the builders of tomorrow who figure out how to curate authentic online communities and experiences will be the winners in this content war.
Although, to some extent I wonder how much it matters. If we're creating images using AI tools, and then sharing the best results, doesn't that become valid training data? In some sense are we supervising the learning?
Everything will become a copy of a copy of a copy until everything just sounds the same and looks like a parody of itself lacking the soul that made it attractive in the first place.
https://research.google/pubs/pub43146/
Abstract: Machine learning offers a fantastically powerful toolkit for building complex systems quickly. This paper argues that it is dangerous to think of these quick wins as coming for free. Using the framework of technical debt, we note that it is remarkably easy to incur massive ongoing maintenance costs at the system level when applying machine learning. The goal of this paper is highlight several machine learning specific risk factors and design patterns to be avoided or refactored where possible. These include boundary erosion, entanglement, hidden feedback loops, undeclared consumers, data dependencies, changes in the external world, and a variety of system-level anti-patterns.
1. The actual pollution happens in culture (our imaginary) and, as history shows, censorship (cultural via cancellation, or legal via politization of the issue) is not a moral nor practical solution. Then, high culture and filtering technology to the rescue?
2. Images are just the start as Murphy's Law ensures we'll face this same problem for every categorizable piece of knowledge you can think of using in an AI artifact (music, patterns of movements, speech recognition, behavior recognition, art recognition, etc)
For example Dall-e has the pixels watermark (bottom-right), and I assume there's a possibility of an indicator that may by hidden in the data itself. One could also exclude common meme formats and their derivatives. Then there's the option of mapping to produced content via hashing a la Shazam, or have a discriminator component etc.
But you're right, it's not trivial. I just don't think it's too big of a deal.
However, there is some evidence that NNs currently are somewhat limited by the availability of high quality data,[0] however I'm not sure this is really a problem because neural nets already accomplish amazing things, so one might not need that much data to get something useful (perhaps at the expensive of more compute, but so what; e.g. analog computing might give some 1000x speedup anyhow).
[0] https://www.lesswrong.com/posts/6Fpvch8RR29qLEWNH/chinchilla...
Researchers are finding ways to identify the tale-tell markers that currently give them away, but yes, for the neophyte this is going to be a real issue on what can I trust.
However, the great thing is that you will always have data...the challenge will then become how well do I TRUST my predictions, which I believe will spur some very interesting algorithms such as anomaly detection (i.e., the RGB distributions, spatial-markers, etc are way too distorted if I compare metadata from other pictures of this type).
Secondly, there is still a big human selection process going on. Only the most interesting and coherent images will find their way onto the public internet. In fact, if you can automatically detect that these images are AI, then they can serve as an additional training signal to help teach the AI which of its outputs are most likely to delight the human.
Perhaps by a naive algorithm. I'm fairly sure it's fairly easy to train a neural network to recognize current generation AI generated images. And probably for quite a while longer.
Btw, if you want to create realistic images, it's fairly easy to create guaranteed pristine data: just take a video camera and create some footage.
Perhaps there will even be a market for such pristine data.
Now, if you want to create art and train on human artists' output, that might perhaps get harder in the future.
https://www.youtube.com/watch?v=Q6Fuxkinhug
Starring Malkovich, as, Malkovich.
There is sort of a fixed-point on this stuff that creates a Nash point. Every relevant move is a move for some advantage (from megacorp copyright laundering to aspiring influencer content output) and that competition tends to wash out roughly where you started.
This would mean it's always likely to be filterable.
And if not, it's arguable that this pollution becomes an asset. It would be high quality synthetic training data which is commonly intentionally used.
It would also be possible to look for the metadata that accompanies photos taken on phones etc. and weigh that more highly.
Then again, and again, and again, until we end up in the 10th dimension of AI surrealism.
This is not in denigration of human artists, photographers, etc
I think it'll drive artists to make better/different work (just like the advent/adoption of the vanishing point changed art ~500-1000y ago)
Is it the free floating images on the internet or is it the data that we keep on the servers of the BigTech?
Jesting aside, if the resulting image is good enough to publish and tag, then it's good enough to put back into a training set.
I also assume there could be a market for AI trained only on "human produced art" the same way there's a market for organic vegetables.
If a billion humans take 50 photos like that, and spend fifteen minutes of their life to do so, we will have almost as much data as the Laion database, but for door knobs. The photo workers will be paid something like 0.00001 dollar for a picture, by the users of the deep learning algorithms.
The payment method is called blockchain and bitcoin if you have heard of such a thing. Bitcoin, the money of information will enable a marketplace of information, in which the better the information, the more the producer is paid. Bitcoin bsv, can support almost a million transactions per second as of today, and every year the tps is increasing tenfold.