One of the largest drawback to image generating AIs is that it is hard to get a consistent character. In this article I’ll explain how I have used a set of tools to overcome that impediment and make an AI character generator.
Getting a set of six views of the same character
Below is a set of images I generated using Stable Diffusion 1.5, Char Turner V2, and Control Net with Open Pose. These images have not been edited, and were generated during a single run. I’m using runpod.io servers, as it is much easier than trying to run Stable Diffusion locally.
For these images I used the model dreamlike-photoreal-2.0. There are still issues, and I tried to push the flow to the limits. The character doesn’t need to be truly human, but it helps. The tattoos also don’t match perfectly, and things tend to swap from left side to right side when the characters turn. There are an impressive number of details that do match, such as the necklace and earrings on the queen character. You can also see that the feet get easily turned around, which is driven by the openpose I’m using not being as well laid out as it could be. Openpose doesn’t give joints angles, so when legs are straight they can easily turn around.
Settings for a Good AI Character Generator
The important settings here are:
- Hires. fix, upscaled by 2
- Width : 1024
- Height : 512
- Prompt includes (charturnerv2) and the description of the character
- CFG Scale : 14
- Enable Control Net with Open Pose
- Grab an image of some characters standing in clear and non-overlapping poses, I used the ones from the charturnerv2 examples
- Tiling checked
I found that using Hires. fix led to lots of the differences between the characters being smoothed out, and the upscaling helped with the consistency of the fine details. Tiling also makes the backgrounds more likely to b