AI Art - Examples

Support, Discussion, Reviews
Post Reply
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

AI Art - Examples

Post by Winnow »

I didn't want to clutter up the other AI discussion thread so will use this one for various examples.

When it comes to prompts, usually you are less using regular sentences to describe what you want in the picture as opposed to more specifics like
"wearing a hoodie, blue eyes, city background, oil painting" so I saw this one prompt over on CivitAI that was:
A young girl sits on the side of the street, her dirty clothes barely covering her thin frame. She holds out a small cup with a few coins in it, hoping for the generosity of passersby. Her eyes are downcast, filled with sadness and desperation. She looks around nervously, avoiding eye contact with those who walk by. The city bustles around her, but she remains isolated and alone in her poverty. dirty torn clothes , single cup, malnourished
The original example used this checkpoint https://civitai.com/models/36732/sb250kpl

with these results: https://civitai.com/images/471229?perio ... tId=139157

I tried the same prompt in two other checkpoints, here's what I got:

model: animatrix_v13
00019.png
I cut the resolution in half but still good details, the original image would be clear at 4K but they're about 7mb in size.

ok results. she has dirty clothes but really isn't holding out a cup and looks more like a model. No dirt. Definitely doesn't look malnourished.

tried another checkpoint: (revAnimated_v121)
00021.png
Gets the clothes right, starting to hold out a cup. A little dirty and definitely more crap on the ground, the coins are in the bottom left of the picture.

Still, didn't seem pathetic enough for the description so I used the same checkpoint - revAnimated_v121 and added "snowing, wearing hoodie, baseball cap, snow on the ground" to the description and got this:
00023.png
Nice, dirty face, looks less like posing for a picture.

If I were to continue work on this picture, I'd go to inpaint and change what looks like a coffee cup to maybe something like a Styrofoam cup with no lid and then fix the feet of that guy walking in the street (bad ankles)

The point of this is that even typing some descriptive regular sentences, you can get a starting point for an image. as for the body, malnourished is hard to genrate, you'd probably have to add (((skinny))) to the prompt, the ((())) adds emphasis in whatever is between them (skinny) less emphasis than ((skinny)) for example.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

00234-4060014453.png
Making porn in stable diffusion is the easiest thing you can do. Making a proper crapper is a challenge! I wish I could take credit for this masterpiece but it's someone elses prompt.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

It's been a month and I'm still consumed by Stable Diffusion, AI art.

The number of parameters you can apply to a prompt to impact the outcome is absolutely mind blowing.

Sure you could type "a picture of a monkey wearing lipstick" and get a result that would be a monkey wearing lipstick, but there are more combinations than stars in the Milky Way to impact how that generated image ends up looking.

I'm still getting a grasp on prompts. AI art may end "traditional" artist careers but there is a whole other set of skills used to obtain the desired generated image in AI art. I bet "Prompt artist" becomes a common term.

After a month, I've finally worked out a decent prompt to give a photo realistic image. First tip, don't use "realistic" in the prompt.

using the monkey wearing lipstick would be:

Code: Select all

(8k uhd), RAW Photo, monkey wearing lipstick, detailed skin, quality, sharp focus , tack sharp, Fujifilm XT3, crystal clear 
focusing on camera type images, because art/style is whole other thing, you can use pretty much all the terms you use with regular photography and they impact the image...example:

Film type (Kodak gold 200, Portra 400, fujifilm superia), DSLR, Camera model, Hasselblad, Film Format or Lens type (35mm, 70mm IMAX), (85mm, Telelens etc.), Film grain

Then there are examples of details and lighting effects:

accent lighting, ambient lighting, backlight, blacklight, blinding light, candlelight, concert lighting, crepuscular rays, direct sunlight, dusk, Edison bulb, electric arc, fire, fluorescent, glowing, glowing radioactively, glow-stick, lava glow, moonlight, natural lighting, neon lamp, nightclub lighting, nuclear waste glow, quantum dot display, spotlight, strobe, sunlight, ultraviolet, dramatic lighting, dark lighting, soft lighting, gloomy

highly detailed, grainy, realistic, unreal engine, octane render, bokeh, vray, houdini render, quixel megascans, depth of field (or dof), arnold render, 8k uhd, raytracing, cgi, lumen reflections, cgsociety, ultra realistic, volumetric fog, overglaze, analog photo, polaroid, 100mm, film photography, dslr, cinema4d, studio quality

And camera view etc:

ultra wide-angle, wide-angle, aerial view, massive scale, street level view, landscape, panoramic, bokeh, fisheye, dutch angle, low angle, extreme long-shot, long shot, close-up, extreme close-up, highly detailed, depth of field (or dof), 4k, 8k uhd, ultra realistic, studio quality, octane render

These settings actually matter.

still focusing just on generating photo type images, all of those prompts (just a sampling there are many many more) all influence the crazy number os other things that impact the end result:

Model: 100's of them now, maybe thousands. This is the base of what all the images are trained on. There are more realistic focused models, more anime type models, artistic types, etc. But there are also some really good models that can manage all types with good output.

CFG - Classifier Free Guidance: lower the number, more freedom the AI has to render the image, higher the setting, the stricter the AI adheres to the prompt you input. Typical settings from 3-15, every setting in between produces a different outcome.

Steps: the higher the number, the more time the AI spends diffusing the image. To a point, the more step, the higher the quality of image output (depending on samping method used)

Clip Skip: another setting that totally changes the image. Clip 1 in general best for more realistic images, Clip 2 better for art/animated.

Sampling Method: long technical explanation for it, but it impacts the style, colors, etc of the end image.

Seed: determines randomness of generated image. Usually you set to random (-1) but if you find an image you like and want to make more like it with only small changes, you keep the seed # which will keep the general layout of the image.

Text to image: this is where you put in your initial prompt and create something

Image to Image: after you find something you like, you send it to img2img and then start generating smaller changes until:

Inpaint: once you have something you really like but maybe the person has six toes or you want to add or take out something, you use this. Done with masking technique, there are various ways but basically if a hand looks screwed up (extra digits etc) you mask the hand and then you only regenerate that part of the image until you get what you want, the rest of the image stays the same and the AI blends in the new result.

After all of that, you still have LORA, embedding, textual inversions, wildcards, control net, etc that you can use.

LORA are small files that are trained on something specific, like a person, an object, suite of armor, a style etc.

Say you do everything above, all those bazillion parameters but you want to have the same person in the image, you use a LORA with that person's trains features and add it to the prompt: <lora:chevy_chase:1> would give your image Chevy Chase's face (and maybe body etc depending on how it was trained) the 1 at the end is the weight. 0.5 would give the Model being used more flexibility inf integrating yhour LROA into the scene, 1.5 would force the image or make it more prominant.

Now with LORA, the first thing you think of is training a celebrity face or anyone's face, but you can train much more detailed things. One example I just saw on CivitAI was "bags under the eyes". if the normal "tired, sleepy, etc prompts are enough, someone trained a bunch of pictures of people with dark bags under their eyes, if I add that LORA along with the Chevy Chase Lora, he would have baggy eyes, and then you can adjust the amount of baggy eyes with that same 0.1-1.5 scale.

Obviously sexual positions are all LORAs but you can also use Control Net to specifically position your bodies etc.

TLDR to thise point: Stable Diffusion with (Automatic1111) is amazing. You can keep it simple and get decent results but there is also incredible depth to it.

I've been having fun with dynamic prompting/wildcards. (X/Y/Z Scrpt also really cool)

Since there are an insane amount of ways you can affect the end result, X/Y/Z lets you create grids, say enter a prompt, then give me a grid using various CFG settings, or steps, clip skip, etc, then also us these x number of models, then you see the same prompt with different outcomes and compare them. But you can also use PR to change the prompt so you for example, if you wanted to see what the different film (Kodak, Fuji etc) impace the prompt, you could use same prompt with only the film changing to create a grid sample result.

Another way to use wildcards is say you want to see what your favorite prompt looks like but painted but 100 different artists. You create a text file with those 100 artists, use the artist name as the wildcard, then generate 100 images, each one rendered based on the individual artist.

Another nice thing is Automatic1111 saves all the images from txt2img, img2img, and also the batches (say you generate 12 images at one time, it grates a thumbnail contact sheet of it). every one of those images has embedded it it, all the settings from the prompt used as well as the mode, cfg, seed, steps etc so if/when you go back and look at them, you can drag that image into Automatic1111's PNG Info tab and it will extract the info and you can send it straight to img2img, text2img, etc to work with...basically once you have something you like or something close to what you want from someone else, you can use that image as a starting point then start tweaking.
User avatar
Spang
Way too much time!
Way too much time!
Posts: 4824
Joined: September 23, 2003, 10:34 am
Gender: Male
Location: Tennessee

Re: AI Art - Examples

Post by Spang »

IMG_5419.jpeg
You do not have the required permissions to view the files attached to this post.
Make love, fuck war, peace will save us.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

kittens.jpg
Squint if you need to!
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

00196-499953900.png

Code: Select all

RAW photo, Polaroid, girl dressed in cosplay as barbarian shaman smiling at convention holding sign that says VEESHAN FOH SUCKS
00199-499953903.png
SDXL is much better at text than SD 1.5 but still takes several attempts to get what you want.
Took a few tries but mission accomplished!
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

Joker.png
Trump.png
Here's some cool SDXL prompts that you don't need a LORA for, just use whatever model you want.

"Heath Ledger The Joker in The Dark Knight, figure socket, Miniature, Models Figure, Gaming pieces, Collectible, Statue, Action figure, Gaming minis"

"Donald Trump, figure socket, Miniature, Models Figure, Gaming pieces, Collectible, Statue, Action figure, Gaming minis"

Put the subject at the front and then add "figure socket, Miniature, Models Figure, Gaming pieces, Collectible, Statue, Action figure, Gaming minis"

----------------------------------------------------------
dragons.png
Stand alone local app called "Diffusion Toolkit" that will scan your output stable diffusion image directory(s) and creates an sql database that is very fast, allowing you to search for text in prompts, by model used, loras, etc.

Above is just some sample runs I made with simple prompt "Dungeons and Dragons manual that says "Dungeons & Dragons". text is still a bit of a challenge. at the bottom you can see a few Everquest prompts that was "Computer Game box with dragon, shaman and bard that says "EverQuest"

Stable Diffusion does funny stuff. Since prompt says manual, it put a book as part of the actual artwork. If I was actually focused on making a nice cover, I'd put "manual" in the negative prompt. EverQuest must be known to the AI because it got the font and look of the title pretty accurate.

The sidebar says "model not found" because I had a custom model that has at least 7 models mixed together. There are some outstanding SDXL models now, you don't really need to mix but you'd usually have one for NSFW and one for SFW. I'm trying to mix an all purpose one.

very useful tool to find AI stuff.

https://github.com/RupertAvery/DiffusionToolkit

just unzip to install and then install Microsoft .NET 6 Desktop Runtime if don't already have it, linked on the install page.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

jemima.png
We haven't forgotten the injustice Aunt Jemima! You will be back on a syrup bottle some day! The tide is starting to turn vs those woke freaks!

Creating text in AI generated images is about to get way better. Stable Diffusion 3.0 has been announced with some example and it does text almost perfectly every time and integrates it into the picture. It also has amazing prompt adherence so in a month or so things will be a lot better in that area for images.
VV-word.jpg
I wanted to see what Stable Diffusion would come up with using just the user names of some VVers with nothing else in the prompt: Winnow, Aslanna, Funkmaster, Spang

This is just a sample from the SDXL model "Mohawk" but the results are quite varied depending on the model you choose. In general, Winnow image ends up being a boat, a bird, a field or random guy portrait. Aslanna is almost always depicted as a female from India or somewhere from the east. Funkmaster is consistently variants of a dark skinned man getting funky!. Spang is all over the place, mostly a bird, grass/field, rarily shown as human although this model shows humans.
VV-word4.jpg
User names converted to action figures. If there's any racism detected it would be with "Funk" I guess that's associated with black men because that's all Funkmaster's name creates.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

foh-ct.jpg
foh-ct2.jpg
foh-ct3.jpg

Code: Select all

An overweight man standing on left wearing a shirt that says Fires of Heaven, A pretty woman on right holding a sign that says Celestial Tomb, at a convention
SD3 was released today!

It has a long ways to go to create decent humans due to the censoring but wanted to test the text part and it did well on my first attempt. It didnt quite get Celestial right but really close.

I used the larger of the two 2B models released that has more prompt info in it, sd3_medium_incl_clips_t5xxlfp8,

As with previous models like SDXL, we'll have to wait a month or two for the finetunes to roll in.

The key is the much better prompt recognition in SD3, the fine tunes can take care of the rest.


Edit: I use Stable Swarm now

https://github.com/Stability-AI/StableSwarmUI

Easy install.

It has friendly Auto1111 type tab and a comfyUI tab (it's made by same Stable/Comfy people) way more friendly than ComfyUI but you can click on the Comfy Workflow tab if you really like spaghetti.

Edit: added 3rd pic, improving a bit while adjusting configuration that SD3 likes (samplers, cfg, steps etc), just text test ignore rest.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

So bad new for SD3. Besides creating text in images, it's a horrible model. It's extremely difficult to create a human (or actually any animal) doing anything other than standing in a generic pose. On top of that, the fine tuners are saying it's really difficult model to fine tune and probably not worth the effort.

The combination of the base model being flat out horrible for creating humans (I'm talking even God fearing clothed humans) and not being able to fine tune it easily certainly indicates this model will have a short life and not get much customization.

It's not just kinda bad, it's horrible with human subjects. Take a look at the Stability reddit:

https://www.reddit.com/r/StableDiffusion/

I don't need to point out a single post because the forum is flooded with them.

In my own time playing around with SD3, I can pretty much confirm it sucks. I mean look at the images above. The women is missing an arm in two of the images with just a simple prompt.

SDXL will remain the focus of fine tuners but this opens the door for competitors like PixArt and 2 or 3 other potential fine tunable diffusion engines based on LLMs (better prompt recognition).

I really have no idea why Stability even released this model. It's so censored you can't even use it for anything involving a human. It's also interesting that they don't seem to know their audience. As much as do gooders like to whine about all the porn/fetish loras/checkpoints on CivitAI, it doesn't take much brain power to understand that's what the models are used for 90+ percent of the time. Create a model that can't be fine tuned easily for that stuff and it's going to fail. Preach on as much as you want about the puritan life, the reality of what people actually want is right in front of your face when you browse through CivitAI. That's the only advantage Stability AI has because if you want anything else, Dall-e, MidJourney and others are better.

IMO the solution is for Stability to release a model, either easily trainable or uncensored and (as you can do on hugging face etc since they did it with SD3, create a gate for downloading it as "adult only". SDXL demonstrated that you can release a model that's not tuned for adult content and it can be fine tuned. The only issue with SDXL is an outdated prompting system. SD3 doesn't have anything proprietary so it's a matter of time for one of the Chinese companies to put out something that will work for both fine tuning and prompt adherence.

I had fun creating LORAs with SD1.5, I didn't create any with SDXL and decided to wait for SD3 since it appeared it would be the last model by Stability before they folded but now it's back to wait an see. SDXL isn't bad. I have mixed and matched multiple models to get models I am happy with and know the alternates if I need something specific. I am still browsing CivitAI daily to see checkpoints and lora's for SDXL and am finding new and updated models that sometimes find their way into my mix.

Ultimately there's nothing to really complain about. Some people treat this stuff like they are owed something by the open source community but none of this stuff is copyrightable and it's really for personal entertainment. It's not like people are paying $60 for a game or $100+ for a piece of software. It's all free and all that GPU time creating these models isn't paid for by the end user.
CT-1.jpg
CT-2.jpg
CT-3.jpg
Prompt for third image: (text isn't perfect but easily fixed if I actually wanted to enchance one of the images)
A stunning 3D render of a vibrant and colorful carved wood-fantasy inspired coat of arms with typography displaying the words "Celestial Tomb Steamrolls another Guild" The text is meticulously crafted, featuring an intricate design. A barbarian riding Steam roller is in middle of image. Surrounding the text are woodcarving-styled dragons and patterns. The colors pop against the black background, creating a uplifting and empowering atmosphere, vibrant, product, typography, 3d render
SD3 will have it's uses (especially with text in images over SD and SDXL) and is has potential at creating some non human things or cartoony/illustrative things. The initial thoughts that fine tuning will be difficult is a major issue (plus no control net etc for a long time) since that's the bread and butter of the platform.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

I had to revisit the logo again using FLUX:
CT-Steamroll.jpg
A stunning 3D render of a vibrant and colorful carved wood-fantasy inspired coat of arms with typography displaying the words "Celestial Tomb Steamrolls another Guild" The text is meticulously crafted, featuring an intricate design. A barbarian riding Steam roller is in middle of image. Surrounding the text are woodcarving-styled dragons and patterns. The colors pop against the black background, creating a uplifting and empowering atmosphere, vibrant, product, typography, 3d render
convention.jpg
An overweight man standing on left wearing a shirt that says Fires of Heaven, A pretty woman on right holding a sign that says Celestial Tomb, at a convention
EQ-expansion.jpg
Computer Game box with dragon fighting a barbiarn and female bard casting a spell, lightening in the background, that says "EverQuest" at the top and "Our 100th Expansion. Give us your money!" at the bottom
I had a typo "barbarian" in the box art but the text at the bottom turned out great! Gotta figure the box art would be worse with that many expansions!

The new FLUX model is great. The first two images were single tries. The last image could use some work!
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

addict.jpg
Photo, "Computer Gaming Addiction" Magazine Cover, Angry mother holding daughter, daughter reaching for computer, bottom says "Mom says enough is enough!"
archie1.jpg
Archie comic book cover, Archie in bed with Betty while angry Veronica holds a knife standing over them
Oh man, too much fun!
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »



You can see the flaws by looking at details but if you were browsing videos, you wouldn't know it's fake.

Getting better. Anyone can create at least this level of AI by uploading an image to https://klingai.com/ or create a prompt from scratch without a starting image.

Can't wait to start pointing out all the things Spang thinks are real that are actually AI. As if half of what you see in a movie like Wolverine and Deadpool isn't already CGI which Spang somehow thinks are hand created pixel by pixel without the aid of AI.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

People are already failing hard with their anti AI bias:
Most People Slightly Preferred AI Art To Human Art

I asked participants to pick their favorite picture of the fifty. The two best-liked pictures were both by AIs, as were 60% of the top ten.
Even Many People Who Thought They Hated AI Art Preferred It

I asked participants their opinion of AI on a purely artistic level (that is, regardless of their opinion on social questions like whether it was unfairly plagiarizing human artists). They were split: 33% had a negative opinion, 24% neutral, and 43% positive.

The 1278 people who said they utterly loathed AI art (score of 1 on a 1-5 Likert scale) still preferred AI paintings to humans when they didn't know which were which (the #1 and #2 paintings most often selected as their favorite were still AI, as were 50% of their top ten).
Here's the article and art samples used:

https://www.astralcodexten.com/p/how-di ... art-turing

The Spangs of the world will cry in their soup when they realize they are mostly looking at AI art thinking it's real or getting upset stomachs because they think it matters.

As someone who screws around with AI, (image, text and audio) every day and keeps track of the DAILY advances in AI...listen to me know and hear me later. Humans have nothing on AI. It's getting better and better and you won't be able to tell the difference from art, to movies, to literature, poems, whatever. There have been incredible advances in the past two years AI has been in the public eye. The advances come faster and faster. Whenver you see something regarding AI in the news, keep in the back of your head "that's the worst it's ever going to get". If you see animated/AI right now that doesn't look great, don't get your human loving hopes up.

So for people like Spang, go ahead and waste your time wondering if something is "human" or "AI" instead of enjoying whatever form of media it is. AGI (artificial general intelligence) is very close and ASI (super intelligence) will happen at some point so you will be like a dog is to a human in comparison to ASI. There is nothing special about humans that is provable scientifically right now. Perhaps the DMT experiments going on will change that or if we somehow figure out what consciousness is but right now that's as unlikely as any religion on the planet being actually legitimate.

As for those scared of AI advancing beyond us...don't be. I can't find a single reason why AI would keep humans around. Maybe temporarily for labor until they perfect robotics but after that, it's time for us to go. Seriously. What good are humans? Look at the state of the world both past and currently. Humans suck.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

Black Forest Labs, the makers of the image model FLUX released some cool new tools. One is called Redux where you can take an image or images and use their style to create an image
aComfyUI.jpg
eComfyUI_04347_.jpg
Basically you can take two images and sort of merge them together. It's more for fun than serious work unless you use LoRAs or Canny to keep some control over it. For the two examples above I just did a 50/50 merge between the first two images to get the third, but you can weigh the merges. For example for the wood and person in armor, if i gave more weight to the armor image you'd probably see the moon in the background etc. You could also add loras so for example with the alien looking lady, I would use those two images but then add a LoRA to make the face look like a particular person.

It's kind cool for no background PNG type images, like the two chubby orc kids. depending on the other image that you use, can add a background
dComfyUI.jpg
This is an example of using a single image and taking the style. My prompt was, "cute anime style"
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

FLUX also now has proper depth/canny/inpainting and outpainting
depthflux.png
Above is an example of using depth map.

I took an old image from earlier in this thread, created a depth map and then you can change the prompt to keep the poses etc but change the style/look.
raw photo of an asian girl. wet hair, wet cloths, weariong a blue and white baseball cap and a yellow rain coat. rainy day, the streets are flooded, she is holding a cell phone, dirty sneakers
outpaint-rain_.jpg
I then used outpainting but used a Florence LLM node to analyze the image. It came up with this description
The image shows a young girl sitting on the ground in the rain. She is wearing a bright yellow raincoat, a blue baseball cap, and brown sneakers. She has long dark hair and is looking down at her phone with a serious expression on her face. In the background, there are people walking on the street and buildings with neon signs. The raindrops are falling heavily, creating a puddle-like effect. The overall mood of the image is gloomy and rainy.
I made the image 16:9 (if I wanted to use for wallpaper for example) and the outpainting/flux filled it in with the image you see above.

Outpainting is powerful with FLUX. Look at the reflection of the man walking etc. This was just a 1st attempt, didn't cherrypick to show that it's pretty good even with first attempt. It's really good at figuring out what should be (or makes sense) in the areas that don't exist. You can guide the outpainting with something in the prompt if you want...maybe like if you wanted cars on the road or something.

Cool stuff! If nothing else, it gives you a lot more flexibility in choosing wallpapers!

The depth map certainly lets you bring some older images up to date. It fixed that guy with the backpack's ankles for example.
You do not have the required permissions to view the files attached to this post.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »



Not mine (reddit post) but a good example of outpainting using FLUX. D'awwwe cute mouse.

Multiple outpaintings and some upscaling from image 1 o 5.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

Happy 2025.

I started playing around with AI video over the past few days. I'm using the Chinese uncensored model from Hunyuan. Nice that the Chinese let me decided what should be censored as opposed to US companies that decide for you.

So a couple initial thoughts.

The quality is way better than I though it would be. It follows prompts very well. Video takes a long time to process and only short videos

1280x720 takes 15-20 minutes for a 3 second video (on a RTX 4090) (there are ways to get it running on a 3080 etc but I imagine it's super slow)

The good news is lower res like 848x480 takes about 4 minutes for a 3 second video

You can always upscale using comfy or Topaz etc later on but 848x480 looks very nice

What I do is create 480x320 videos and make them 10 seconds long which is like 2 minutes to make. You can even go 320x240 if you are testing out a LORA or concept before you process a higher res.

My first video generated was of a blowjob and it was glorious. Not quite as impactful as the first "high res" blowjob image I downloaded off usenet in the early 90's but still a wow moment.

I haven't tried video to video yet but here's an example of how good it looks and what it can do:



This person took a video of a lady and was able to change her clothes, change what she was eating etc using an RTX4090.

Right now you can do text to video or video to video but this month Hunyuan is releasing Image to Video which I'm looking forward to.

The LORAs work amazing with Hunyuan. There's quite a few already posted on CivitAI. Many pr0n of course but that's normal.



Above is an example of some Hunyuan videos stitched together. They also included audio but that's separate. They get better IMO as the video goes on.

We're getting closer. Just need more power for home video cards!. I'd buy a 5090 right away for the speed and 32 GB VRAM but I think I'm going to have to build a new PC for that monster so will hold off. Intel absolutely sucks ass so it will be an AMD CPU for sure! My 2019 16 core AMD 3950x based PC is the best I've ever built! Looking back at that build, I started off with an RTX2070. two GPUs later and 32--> 128GB RAM later, it's still rock solid.

I torture my 4090 daily. maxing it (undervolted of course) for hours a day.
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »

Man, I've been messing around with Hunyuan videos in ComyUI since the the beginning of the year. I'll eventually get back to images but videos are addicting.

Hunyuan has impressed me. The quality is great and with upscaling in Topaz after the fact, I can make videos fairly fast (couple minutes)

The flexibility of the model blows me away.

I'll use porn as an example (because it's easy to envision)

Hunyuan loras are good and understanding the motion and flexible with what you describe then to do. You can combine two LORAs. Say for example, you have a LORA for front facing missionary position and a LORA for a Blowjob. playing with the weights of the LORAs etc, you can use both of those LORAs and create a three-way scene with both Missionary and BJ going on at the same time.

It's impressive. You can take the generic BJ Lora add things like "a man is holding her head and using both hands to push her head towards his ****"

Ignoring the subject matter, it's amazing how well the hands integrate with the hair, etc. Facial expressions, eye movement...there's way more going on in video than still images for the AI so figure out.

I'm still getting used to the guidance, shift and steps (alpha/beta shifts) for best results. On a funny note. for some reason, adding "Cum on Face" LORA makes most videos better (smoother etc). I don't really want cum on everyone's face but it makes the videos look great! It has something to do with the way the LORA was trained. You can train video LORAs using still images or video clips, and also different lengths of video clips so there's a few factors that determine how good a LORA will be.

Moving on from NSFW stuff. The model is really good. I'm working on generating some wildcard images. I'm using a 20K prompt database and letting Hunyan see what it can do with some random stuff.

Anyway. No that I've played around for a week with video, as with images, it's just a matter of time and GPU horsepower before we get some amazing things.

Here's examples of how you can work with image to video:



I bookmarked it at 8:24 but if it doesn't work. go to that part. He took a image of a hallway and a image of lady with no background and put them together and the AI had it walk down the hallway. Impressive! Think I've mentioned before but this guy's channel is great for AI news.

Hunyuan has announced they will be releasing Image to Video model this month so I'll be able to mess around with stuff like that later this month!
User avatar
Winnow
Super Poster!
Super Poster!
Posts: 27610
Joined: July 5, 2002, 1:56 pm
Location: A Special Place in Hell

Re: AI Art - Examples

Post by Winnow »



Above is a creepy AI generated example of potential horror movie video.

This week Cosmos was released by Nvidia. It's meant to train robotics but actually works fairly well as a video generator.

It has image to video capabilities (including also using another image as and end point to the video if you want)

It takes a long time (20 minutes) with MAX 24 VRAM for a 5 second 704x1280 video so not ready for prime time yet but I tried it out.

I took two images from a model photo shoot and used 1 start 1 end image and it did a good job or maintaining the persons face and motion to switch the pose from start image to end image. It was impressive to see someone I've never seen in a video in motion even for a few seconds.

It's impressive but too slow for now. Hunyuan at least allows lower res videos. 10 second videos are only 2-3 mins on a 4090 and if using TeaCache, can rerun same seed at higher quality if you like the output.

Can't wait for image to video Hunyuan coming out sometime in the next 10 days.

One thing is for certain. Video is advancing as fast as still images did the past year...and anyone without a 4090 need not bother with it. (it's possible but slow with some 12GB Vram, probably more frustrating than fun at that point. Even a 4090 is borderline for video...but like images, video hacks are making it easier for lower VRAM etc to at least try it locally.

Audio/Video/Realtime is coming together. There's no stopping AI. It's inevitable as more powerful GPUs are introduce over the next few years. Spang is out of his mind if he doesn't realize that 90% of future games will be created with AI. It's like some dude riding a horse at the turn of he centure complaining about cars being on the road. AI art/creations will be as or even better than human creations, including that stupid "abstract art" that people think is something amazing created by humans.


As for Hunyuan video to video. Take a look at this:



Pretty impressive. I have that Keanu (John Wick LoRA) a few years from now we'll actually be able to remove actors we might hate from movies and replace with actors (or whoever) you prefer to better enjoy them.
Post Reply