ebsynth_utility/README.md

4.5 KiB

ebsynth_utility

Overview

AUTOMATIC1111 UI extension for creating videos using img2img and ebsynth.

This extension allows you to output edited videos using ebsynth.(AE is not required)

Example

  • The following sample is raw output of this extension.

sample 1

sample 2

sample 3 blend background

  • person : masterpiece, best quality, masterpiece, 1girl, masterpiece, best quality,anime screencap, anime style
  • background : cyberpunk, factory, room ,anime screencap, anime style It is also possible to blend with your favorite videos.

Installation

Usage

  • Go to [Ebsynth Utility] tab.
  • Create an empty directory somewhere, and fill in the "Project directory" field.
  • Place the video you want to edit from somewhere, and fill in the "Original Movie Path" field. Use short videos of a few seconds at first.
  • Select stage 1 and Generate.
  • Execute in order from stage 1 to 7. Progress during the process is not reflected in webui, so please check the console screen. If you see "completed." in webui, it is completed.
    (In the current latest webui, it seems to cause an error if you do not drop the image on the main screen of img2img.
    Please drop the image as it does not affect the result.)

Note

For reference, here's what I did when I edited a 1280x720 30fps 15sec video based on

Stage 1

There is nothing to configure.
All frames of the video and mask images for all frames are generated.

Stage 2

In the implementation of this extension, the keyframe interval is chosen to be shorter where there is a lot of motion and longer where there is little motion.
If the animation breaks up, increase the keyframe, if it flickers, decrease the keyframe.
First, generate one time with the default settings and go straight ahead without worrying about the result.

Stage 3

Select one of the keyframes, throw it to img2img, and run [Interrogate DeepBooru].
Delete unwanted words such as blur from the displayed prompt.
Fill in the rest of the settings as you would normally do for image generation.

Here is the settings I used.

  • Sampling method : Euler a
  • Sampling Steps : 50
  • Width : 960
  • Height : 512
  • CFG Scale : 20
  • Denoising strength : 0.2

Here is the settings for extension.

  • Mask Mode(Override img2img Mask mode) : Normal
  • Img2Img Repeat Count : 5
  • Add N to seed when repeating : 1
  • use Face Crop img2img : True
  • Face Detection Method : YuNet
  • Max Crop Size : 1024
  • Face Denoising Strength : 0.25
  • Face Area Magnification : 1.5 (The larger the number, the closer to the model's painting style, but the more likely it is to shift when merged with the body.)
  • Enable Face Prompt : False

Trial and error in this process is the most time-consuming part.
Monitor the destination folder and if you do not like results, interrupt and change the settings.
[Prompt][Denoising strength] and [Face Denoising Strength] settings when using Face Crop img2img will greatly affect the result.
For more information on Face Crop img2img, check here

If you have lots of memory to spare, increasing the width and height values while maintaining the aspect ratio may greatly improve results.

This extension may help with the adjustment.
https://github.com/s9roll7/img2img_for_all_method

Stage 4

Scale it up or down and process it to exactly the same size as the original video.
This process should only need to be done once.

  • Width : 1280
  • Height : 720
  • Upscaler 1 : R-ESRGAN 4x+
  • Upscaler 2 : R-ESRGAN 4x+ Anime6B
  • Upscaler 2 visibility : 0.5
  • GFPGAN visibility : 1
  • CodeFormer visibility : 0
  • CodeFormer weight : 0

Stage 5

There is nothing to configure.
.ebs file will be generated.

Stage 6

Run the .ebs file.
I wouldn't change the settings, but you could adjust the .ebs settings.

Stage 7

Finally, output the video.
In my case, the entire process from 1 to 7 took about 30 minutes.

  • Crossfade blend rate : 1.0
  • Export type : mp4