What does HackerNews think of clip-interrogator?

Image to prompt with BLIP and CLIP

Language: Python

It's image -> words, the inverse of stable diffusion.

see: https://github.com/pharmapsychotic/clip-interrogator

P.S. It's strange to me that this is a big deal, there are plenty of libraries for this stuff afaik, but I don't pay attention to licenses personally.

People have been using this to amuse themselves by putting in selfies and seeing the AI comments, which are often quite rude (or else calls them "popular on reddit" or "a discord mod").

It's unusable for me though, haven't been able to get in the queue once in the last few days.

You might have better luck with Colab: https://github.com/pharmapsychotic/clip-interrogator

Context being that another AI art model (StableDiffusion) knows the names of many popular artists and can create images that sort of kind of look like their work. This terrified a lot of artists on Twitter who've now gone around harassing AI developers and claiming they're plagiarists, then simultaneously posting things like "umm this is all uncreative and ugly collages" and "this is going to take all our jobs".

Oddly, the main instigator turned out to be a "Pokemon in real life" fanartist who didn't notice he's already a professional plagiarist.

> So in that context, saying “horses stuck around when the automobile came” is true, but if you went up to a painter and said “hey, within your lifetime painting will see a 90% decline, stop being taught formally, disappear from daily life or awareness”.

The issue with this claim isn't automation replacing artists (though I don't think that will happen either due to Jevons' paradox) - it's just that AI generated images don't replace paintings because they aren't paintings! Print shops already exist and may have replaced you though.

> I’ve had a lot of struggles with this. I have a specific image in my head, I’m trying to prompt for it, and the AI just does not want to do it. The most trouble that I’ve had so far has been with trying to get a tavern running across the plain with chicken legs.

There's a general unfixable problem here, which is it's hard to be aligned with silly prompts without giving you silly output for "normal" prompts. That's also why they're complaining the model output has too safe composition - the developers are lucky they even got it to do that, it's better than random blobs of color and body parts like older models would generate.

But the picture they want probably is hiding somewhere in Midjourney's latent space; it's just a matter of finding a prompt that recreates it.

One way to do that could be to sketch the picture you want some other way and run it through a reverse image-to-text notebook like https://github.com/pharmapsychotic/clip-interrogator.

Another would be prompting it with your sketch so that you can get an image in its "house style" - which doesn't seem very appealing for most models, but Midjourney has a pretty strong one.