We learned this approach the hard way after HN got inundated by topic avalanches in the past (e.g. Snowden in 2013). It works fairly well, but it requires us to see the posts, so if you notice a follow-up submission that isn't singing particularly well for its supper, we'd appreciate a heads-up at hn@ycombinator.com so we can downweight it.
Past explanations, if anyone wants more, can be found by scrolling back through these links:
'significant new information' is key when there's a major ongoing topic - https://hn.algolia.com/?dateRange=all&page=0&prefix=false&so...
we tend to downweight follow-ups unless they're particularly good - https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...
repetition is bad for curiosity - https://hn.algolia.com/?dateRange=all&page=0&prefix=false&so...
diffs are what's interesting - https://hn.algolia.com/?dateRange=all&page=0&prefix=false&so...
I'm especially bored with the "Will AI replace artists?" breathlessness. It's tedious. Of course not. These tools are each like an artist in and of themselves. Did Picasso replace Rembrandt? Did Warhol replace Picasso? Nah. No AI will create what I'd create, or what Takashi Murakami would create, or any given high school kid with a paintbrush.
If all you think art is is being able to complete a commission, then sure, you'd probably see this as replacing what you think the purpose of an artist is. But you'd be wrong, and wrong in such a boring way that it's not worth debating.
Even if these AIs started generating their own ideas, it would be amazing and fascinating and worth celebrating, but it's not replacing anything. Just adding something new and cool into the world.
Think about everything else that was a thing everyone posted about and frame it the same way. "what are other use cases are there for the greater collective good?" And you'll see how silly this sounds
It is basically visualization of words and expressions, and I expect we will see it woven into many parts of our daily life in the near future.
It will not put artists out of business, although it may significantly impact stock photo services. Everywhere you look - any place that has a photo or illustration - may soon be using generated images. I even expect meta versions which can take large amounts of text (such as an entire paragraph or essay) and generate images which seem to fit the topic or feel.
This tech will give non-artists the ability to express and visualize ideas (for fun or profit), and it will be a boon to professional content creators who labor over image collection and selection.
Heck, at some point the images will be generated on the fly by our browsers for content or even ads.
But recently it is booming. The reason is not related to DALL·E, but Stable Diffusion.
Stable Diffusion is 1) open sourced 2) the model is small and easy enough to run on people's computer. As a result, it can embed in a lot of tools and services, so people are getting excited and start building ecosystem and business around it. When people can create art without effort, how to streamline the creative process and dealing with giant amount of arts are going to be the problems that worth solving.
As for use cases, I can see it either replacing a lot of artists or being one of the most important tool for artists, or both. Also, it might flood the Internet with its creations in the following years so be prepared.
That said, the excitement is understandable. AI/ML have been hyped a lot and there have also been a lot of unresolvable ethical questions. So there's a certain sense of relief at something objectively cool and fun, as well as recapturing the 'wow' factor of what's possible with computers. Often the latter is reserved for a game or video demonstration of what can be achieved with skill + a great toolchain + new hardware (or newly affordable or tiny hardware). It's not so often that it applies to tools, or that new tools so dramatically lower the skill barriers to creating high quality work. Since humans are so visually driven and graphics have always been the fancy chocolate of computer technology, such impressive new tools hit at a visceral level for many people.
I think the biggest (but less obvious) use case is that this in combination with other tools like GPT-3 is suddenly very close to AGI. Up to now AI in computer vision has been about recognizing objects, entities, or behaviors for narrow tasks like driving or surveillance. Suddenly language models, scene decomposition, and visual feedback are good enough that we're close to having real conversations with computers: you can show your computer a picture and talk about how it has people and animals, what clothes the people are wearing, whether they're on a boat or in a field, what they're doing and so on. This is not that far off from sitting with a child and looking at a picture book together.
I mind them just about everywhere else. I have a hard time to describe it and I seem to be somewhat alone in it, but the images coming out of these algorithms generally tend to trigger some revulsion reactions in me. It's possibly just an uncanny valley relative, but sometimes that revulsion hits notes of true, disgusting horror. The images tend to smell too much like nightmares for my comfort. That simile is about the best I can explain it, I think, they all smell too much like nightmares to me.
I've dropped/blocked people and bots on social media for posting lots of generated images without CWs. I've threatened to drop more. I hate how much Facebook's algorithm tends to surface posts with photos over other content in general, but especially in this moment where many of those photos are these generated things that give me the creeps. There's no good way to set a filter to never show me a DALL-E/Midjourney/Stable Diffusion image ever again.
TL;DR: I don't mind conversing about these algorithms in HN, but please stop posting the images themselves.
A slightly higher order application of the DALL-E/SD/MidJourney primitives.
Now after having read, I think it is good to have one odd post here and there as there are always some people not in the know, but I agree we are seeing too many of those.
Generative art seems to make such identification much more difficult. Such tech can hardly be stopped from helping to image the full range of human fantasy. CSE and every other form of crime and taboo will be included. The volume of such media will increase by orders of magnitude. Whether they are or should be legal is a fascinating topic, but it matters little to the advertisers that bankroll business models like Twitter's.
I can't even guess if this results in more or less strictness around such fantasy content.
Saying that this is uninteresting because art does not advance the collective good is a pretty narrow view of humanity.
So: what are you asking, exactly?
Per the guidelines[1], you can flag a post if you'd like.
Personally, the generative art part is pretty exciting to me. It's a good way to come up with new ideas. I don't see that usefulness going away.
Actually this might not be for the greater collective good. It might be bad.
It could radically change how humans produce and perceive art.
I just wish there were more technical discussions around prompt engineering, so far most of the articles focus on particularly good examples authors got out of DALL-E, MJ, SD.
No, there's quite a few of us. https://imgur.com/a/9gXjJ61
The progress demonstrated in these models, particularly SD img2img is quite spectacular, and fully warrants the hype, IMHO.
I find corruption of the human form, in general, in lousy taste and wish people would cut that out.
Just my $0.02
So I'm watching it in amazed fear.
I'm not tired of it, as we often have "deluges" of posts about current events.
It'll die down.
Good native ad team though.