How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

So-calledLive-action comic adaptation, which is to generate a new two-dimensional picture from a real person's picture. In AI painting, this is a very common application scenario. Regarding live-action comic adaptations, several production methods have been shared in the previous advanced series, but the implementation methods are relatively simple. It is difficult to achieve consistency between real people and corresponding two-dimensional effect pictures, whether it is character clothing, background elements, picture colors, etc. Today, I will share a live-action comic adaptation production method that can basically achieve commercial delivery. Let's first look at the renderings of live-action comic adaptations.

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation
How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

1. Production methods of live-action comic adaptations

Next, we will use the real-life picture below as an example to explain the production method of real-life comic adaptations.

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

[Step 1]: Selection of large model

Live-action comic adaptations require the generation of two-dimensional images, so the large model needs to be a two-dimensional large model.

Used here: AWPainting

Model download address(You can also get the network disk address at the end of the article)

https://www.liblib.art/modelinfo/1fd281cf6bcf01b95033c03b471d8fd8

【Step 2】:Writing prompt words

We use the WD1.4 tagger to obtain the prompt word information of the image.

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

The prompt words obtained through the WD1.4 labeler plug-in are:

1girl, jewelry, solo, bracelet, hair ornament, hair flower, braid, flower, earrings, long hair, necklace, navy, looking at viewer, realistic, breasts, strapless, skirt, outdoors, midriff, blurry, bare shoulders, brown hair , single braid, blurry background, tube top, medium breasts, parted lips, crop top, brown eyes, pink skirt, black hair

Because we need to generate corresponding two-dimensional images from real-life comics, we need to check the prompt words. It is recommended to use translation software to translate the prompt words and check them in detail, and remove the prompt words that do not meet the requirements. For example, the prompt word realistic needs to be removed. Some of the prompt words inferred from the pictures include freckles, mole, and other keywords that affect the beauty of the face, which can also be removed.

Positive prompt words

Prompt:1girl, jewelry, solo, bracelet, hair ornament, hair flower, braid, flower, earrings, long hair, necklace, navy, looking at viewer, breasts, strapless, skirt, outdoors, midriff, blurry, bare shoulders, brown hair, single braid, blurry background, tube top, medium breasts, parted lips, crop top, brown eyes, pink skirt, black hair

Prompt word: girl, jewelry, solo, bracelet, hair accessories, hair flowers, braids, flowers, earrings, long hair, necklace, belly button, looking at viewer, breasts, strapless, skirt, outdoors, abdomen, blurred, bare shoulders, brown hair, single braid, blurred background, halter top, medium breasts, parted lips, belly top, brown eyes, pink skirt, black hair

Reverse prompt word

ng_deepnegative_v1_75t,(badhandv4:1.2),(worst quality:2),(low quality:2),(normal quality:2),lowres,bad anatomy,(bad hands),((monochrome)),((grayscale)) watermark,moles,many fingers,(broken hands),nsfw,

Related parameter settings

  • Sampler: DPM++ 2M Karras
  • Sampling iteration number: 30
  • Image width and height: 512*768.
  • Prompt word guidance coefficient (CFG): 7

[Step 3]: ControlNet settings

Here we use 3 ControlNet units.

ControlNet Unit 1: The Depth control model is used to control the distance between screen elements to be consistent

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

The relevant parameter settings are as follows:

  • Control Type: Select "Depth"
  • Preprocessor: depth_zoe
  • Model: control_v11f1p_sd15_depth
  • Control weight: 1
  • Guided intervention time: 0
  • Boot termination time: 1

ControlNet Unit 2: Lineart control model is used to control the consistency of the lines of the picture

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

The relevant parameter settings are as follows:

  • Control Type: Select "Lineart"
  • Preprocessor: lineart_realistic (used to identify lines in real-life pictures)
  • Model: control_v11p_sd15_lineart
  • Control weight: 0.6 (The facial features of real-life photos and 2D photos do not match, so the control weight of ControlNet is appropriately reduced here)
  • Guided intervention time: 0
  • Boot termination time: 1

ControlNet Unit 3: The tile control model is used to control the consistency of the screen color

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

The relevant parameter settings are as follows:

  • Control Type: Select "Tile/Blur"
  • Preprocessor: tile_colorfix (used to fix the color of the image in blocks)
  • Model: control_v11f1e_sd15_tile
  • Control weight: 1
  • Guided intervention time: 0
  • Boot termination time: 1

【Step 4】Image generation

Click the [Generate] button and let’s take a look at the final generated image effect.

Original real person picture

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

Second dimension pictures

How to create a comic effect from a real person's picture? Using Stable Diffusion to achieve real person comic adaptation

2. Related instructions

(1) Regarding the choice of large models, you can try various styles of two-dimensional large models.

(2) Three ControlNet control models are used in this example:

  • The Depth control model controls the distance of the image
  • The Lineart control model controls the overall lines of the image
  • The Tile model controls the color of the image.

Since the Tile model has the effect of changing the details of the image, in this example, the background elements of the two-dimensional image are leaves, which will appropriately enrich the background element information of the image.

(3) The production method still needs to be improved in some details. For example, the style and color of the earrings worn by the beauty are not consistent.

Okay, that’s all for today’s sharing. I hope that what I shared today will be helpful to you.

The model is placed in the network disk, and those who are interested can take it!

https://pan.quark.cn/s/b3df771404e2

 

statement:The content is collected from various media platforms such as public websites. If the included content infringes on your rights, please contact us by email and we will deal with it as soon as possible.
Encyclopedia

A practical guide to improving work efficiency with ChatGPT. I changed these 5 ChatGPT settings

2024-7-2 9:39:57

Encyclopedia

How to write self-media copy with Kimi? Prompt word analysis and optimization ideas to become a copywriting expert in seconds

2024-7-2 12:34:09

Search