from modules.shared import cmd_opts from modules.processing import get_fixed_seed import modules.shared as sh import modules.paths as ph import os from .frame_interpolation import set_interp_out_fps, gradio_f_interp_get_fps_and_fcount, process_rife_vid_upload_logic from .video_audio_utilities import find_ffmpeg_binary def Root(): device = sh.device models_path = ph.models_path + '/Deforum' half_precision = not cmd_opts.no_half mask_preset_names = ['everywhere','init_mask','video_mask'] p = None frames_cache = [] initial_seed = None initial_info = None first_frame = None outpath_samples = "" animation_prompts = None color_corrections = None initial_clipskip = None return locals() def DeforumAnimArgs(): #@markdown ####**Animation:** animation_mode = '2D' #@param ['None', '2D', '3D', 'Video Input', 'Interpolation'] {type:'string'} max_frames = 120 #@param {type:"number"} border = 'replicate' #@param ['wrap', 'replicate'] {type:'string'} #@markdown ####**Motion Parameters:** angle = "0:(0)"#@param {type:"string"} zoom = "0:(1.02+0.02*sin(2*3.14*t/20))"#@param {type:"string"} translation_x = "0:(0)"#@param {type:"string"} translation_y = "0:(0)"#@param {type:"string"} translation_z = "0:(10)"#@param {type:"string"} rotation_3d_x = "0:(0)"#@param {type:"string"} rotation_3d_y = "0:(0)"#@param {type:"string"} rotation_3d_z = "0:(0)"#@param {type:"string"} # flip_2d_perspective = False #@param {type:"boolean"} enable_perspective_flip = False #@param {type:"boolean"} perspective_flip_theta = "0:(0)"#@param {type:"string"} perspective_flip_phi = "0:(t%15)"#@param {type:"string"} perspective_flip_gamma = "0:(0)"#@param {type:"string"} perspective_flip_fv = "0:(53)"#@param {type:"string"} noise_schedule = "0: (0.08)"#@param {type:"string"} strength_schedule = "0: (0.6)"#@param {type:"string"} contrast_schedule = "0: (1.0)"#@param {type:"string"} cfg_scale_schedule = "0: (7)" enable_steps_scheduling = False#@param {type:"boolean"} steps_schedule = "0: (25)"#@param {type:"string"} fov_schedule = "0: (70)" near_schedule = "0: (200)" far_schedule = "0: (10000)" seed_schedule = "0:(5), 1:(-1), 219:(-1), 220:(5)" # Sampler Scheduling enable_sampler_scheduling = False #@param {type:"boolean"} sampler_schedule = '0: ("Euler a")' # Composable mask scheduling use_noise_mask = False mask_schedule = '0: ("!({everywhere}^({init_mask}|{video_mask}) ) ")' noise_mask_schedule = '0: ("!({everywhere}^({init_mask}|{video_mask}) ) ")' # Checkpoint Scheduling enable_checkpoint_scheduling = False#@param {type:"boolean"} checkpoint_schedule = '0: ("model1.ckpt"), 100: ("model2.ckpt")' # CLIP skip Scheduling enable_clipskip_scheduling = False #@param {type:"boolean"} clipskip_schedule = '0: (2)' # Anti-blur kernel_schedule = "0: (5)" sigma_schedule = "0: (1.0)" amount_schedule = "0: (0.1)" threshold_schedule = "0: (0.0)" # Hybrid video hybrid_comp_alpha_schedule = "0:(1)" #@param {type:"string"} hybrid_comp_mask_blend_alpha_schedule = "0:(0.5)" #@param {type:"string"} hybrid_comp_mask_contrast_schedule = "0:(1)" #@param {type:"string"} hybrid_comp_mask_auto_contrast_cutoff_high_schedule = "0:(100)" #@param {type:"string"} hybrid_comp_mask_auto_contrast_cutoff_low_schedule = "0:(0)" #@param {type:"string"} #@markdown ####**Coherence:** histogram_matching = False #@param {type:"boolean"} color_coherence = 'Match Frame 0 LAB' #@param ['None', 'Match Frame 0 HSV', 'Match Frame 0 LAB', 'Match Frame 0 RGB', 'Video Input'] {type:'string'} color_coherence_video_every_N_frames = 1 #@param {type:"integer"} color_force_grayscale = False #@param {type:"boolean"} diffusion_cadence = '1' #@param ['1','2','3','4','5','6','7','8'] {type:'string'} #@markdown ####**Noise settings:** noise_type = 'perlin' #@param ['uniform', 'perlin'] {type:'string'} # Perlin params perlin_w = 8 #@param {type:"number"} perlin_h = 8 #@param {type:"number"} perlin_octaves = 4 #@param {type:"number"} perlin_persistence = 0.5 #@param {type:"number"} #@markdown ####**3D Depth Warping:** use_depth_warping = True #@param {type:"boolean"} midas_weight = 0.3 #@param {type:"number"} padding_mode = 'border'#@param ['border', 'reflection', 'zeros'] {type:'string'} sampling_mode = 'bicubic'#@param ['bicubic', 'bilinear', 'nearest'] {type:'string'} save_depth_maps = False #@param {type:"boolean"} #@markdown ####**Video Input:** video_init_path ='/content/video_in.mp4'#@param {type:"string"} extract_nth_frame = 1#@param {type:"number"} extract_from_frame = 0 #@param {type:"number"} extract_to_frame = -1 #@param {type:"number"} minus 1 for unlimited frames overwrite_extracted_frames = True #@param {type:"boolean"} use_mask_video = False #@param {type:"boolean"} video_mask_path ='/content/video_in.mp4'#@param {type:"string"} #@markdown ####**Hybrid Video for 2D/3D Animation Mode:** hybrid_generate_inputframes = False #@param {type:"boolean"} hybrid_generate_human_masks = "None" #@param ['None','PNGs','Video', 'Both'] hybrid_use_first_frame_as_init_image = True #@param {type:"boolean"} hybrid_motion = "None" #@param ['None','Optical Flow','Perspective','Affine'] hybrid_motion_use_prev_img = False #@param {type:"boolean"} hybrid_flow_method = "Farneback" #@param ['DIS Medium','Farneback'] hybrid_composite = False #@param {type:"boolean"} hybrid_comp_mask_type = "None" #@param ['None', 'Depth', 'Video Depth', 'Blend', 'Difference'] hybrid_comp_mask_inverse = False #@param {type:"boolean"} hybrid_comp_mask_equalize = "None" #@param ['None','Before','After','Both'] hybrid_comp_mask_auto_contrast = False #@param {type:"boolean"} hybrid_comp_save_extra_frames = False #@param {type:"boolean"} #@markdown ####**Resume Animation:** resume_from_timestring = False #@param {type:"boolean"} resume_timestring = "20220829210106" #@param {type:"string"} return locals() # def DeforumPrompts(): # return def DeforumAnimPrompts(): return r"""{ "0": "(scenic countryside:1.0), (cherry:`where(cos(6.28*t/10)>0, 1.8*cos(6.28*t/10), 0.001)`), (strawberry:`where(cos(6.28*t/10)<0, -1.8*cos(6.28*t/10), 0.001)`), snow, detailed painting by greg rutkowski --neg (cherry:`where(cos(6.28*t/10)<0, -1.8*cos(6.28*t/10), 0.001)`), (strawberry:`where(cos(6.28*t/10)>0, 1.8*cos(6.28*t/10), 0.001)`)", "60": "a beautiful (((banana))), trending on Artstation", "80": "a beautiful coconut --neg photo, realistic", "100": "a beautiful durian, trending on Artstation" } """ def DeforumArgs(): #@markdown **Image Settings** W = 512 #@param H = 512 #@param W, H = map(lambda x: x - x % 64, (W, H)) # resize to integer multiple of 64 #@markdonw **Webui stuff** tiling = False # firstphase_width = 0 # firstphase_height = 0 seed_enable_extras = False subseed = -1 subseed_strength = 0 seed_resize_from_w = 0 seed_resize_from_h = 0 #@markdown **Sampling Settings** seed = -1 #@param sampler = 'euler_ancestral' #@param ["klms","dpm2","dpm2_ancestral","heun","euler","euler_ancestral","plms", "ddim"] steps = 25 #@param scale = 7 #@param ddim_eta = 0.0 #@param dynamic_threshold = None static_threshold = None #@markdown **Save & Display Settings** save_samples = True #@param {type:"boolean"} save_settings = True #@param {type:"boolean"} display_samples = True #@param {type:"boolean"} save_sample_per_step = False #@param {type:"boolean"} show_sample_per_step = False #@param {type:"boolean"} #@markdown **Prompt Settings** prompt_weighting = False #@param {type:"boolean"} normalize_prompt_weights = True #@param {type:"boolean"} log_weighted_subprompts = False #@param {type:"boolean"} #@markdown **Batch Settings** n_batch = 1 #@param batch_name = "Deforum" #@param {type:"string"} filename_format = "{timestring}_{index}_{prompt}.png" #@param ["{timestring}_{index}_{seed}.png","{timestring}_{index}_{prompt}.png"] seed_behavior = "iter" #@param ["iter","fixed","random","ladder","alternate","schedule"] seed_iter_N = 1 #@param {type:'integer'} # make_grid = False #@param {type:"boolean"} # grid_rows = 2 #@param outdir = ""#get_output_folder(output_path, batch_name) #@markdown **Init Settings** use_init = False #@param {type:"boolean"} strength = 0.0 #@param {type:"number"} strength_0_no_init = True # Set the strength to 0 automatically when no init image is used init_image = "https://github.com/hithereai/d/releases/download/m/kaba.png" #@param {type:"string"} # Whiter areas of the mask are areas that change more use_mask = False #@param {type:"boolean"} use_alpha_as_mask = False # use the alpha channel of the init image as the mask mask_file = "https://github.com/hithereai/d/releases/download/m/mask.jpg" #@param {type:"string"} invert_mask = False #@param {type:"boolean"} # Adjust mask image, 1.0 is no adjustment. Should be positive numbers. mask_contrast_adjust = 1.0 #@param {type:"number"} mask_brightness_adjust = 1.0 #@param {type:"number"} # Overlay the masked image at the end of the generation so it does not get degraded by encoding and decoding overlay_mask = True # {type:"boolean"} # Blur edges of final overlay mask, if used. Minimum = 0 (no blur) mask_overlay_blur = 5 # {type:"number"} fill = 1 #MASKARGSEXPANSION Todo : Rename and convert to same formatting as used in img2img masked content full_res_mask = True full_res_mask_padding = 4 reroll_blank_frames = 'reroll' # reroll, interrupt, or ignore n_samples = 1 # doesnt do anything precision = 'autocast' C = 4 f = 8 prompt = "" timestring = "" init_latent = None init_sample = None init_c = None mask_image = None noise_mask = None seed_internal = 0 return locals() def keyframeExamples(): return '''{ "0": "https://user-images.githubusercontent.com/121192995/215279228-1673df8a-f919-4380-b04c-19379b2041ff.png", "50": "https://user-images.githubusercontent.com/121192995/215279281-7989fd6f-4b9b-4d90-9887-b7960edd59f8.png", "100": "https://user-images.githubusercontent.com/121192995/215279284-afc14543-d220-4142-bbf4-503776ca2b8b.png", "150": "https://user-images.githubusercontent.com/121192995/215279286-23378635-85b3-4457-b248-23e62c048049.jpg", "200": "https://user-images.githubusercontent.com/121192995/215279228-1673df8a-f919-4380-b04c-19379b2041ff.png" }''' def LoopArgs(): use_looper = False init_images = keyframeExamples() image_strength_schedule = "0:(0.6)" blendFactorMax = "0:(0.6)" blendFactorSlope = "0:(0.25)" tweening_frames_schedule = "0:(0.25)" color_correction_factor = "0:(0.075)" return locals() def ParseqArgs(): parseq_manifest = None parseq_use_deltas = True return locals() def DeforumOutputArgs(): skip_video_for_run_all = False #@param {type: 'boolean'} fps = 12 #@param {type:"number"} #@markdown **Manual Settings** use_manual_settings = False #@param {type:"boolean"} image_path = "/content/drive/MyDrive/AI/StableDiffusion/2022-09/20220903000939_%05d.png" #@param {type:"string"} mp4_path = "/content/drive/MyDrive/AI/StableDiffusion/content/drive/MyDrive/AI/StableDiffusion/2022-09/kabachuha/2022-09/20220903000939.mp4" #@param {type:"string"} ffmpeg_location = find_ffmpeg_binary() ffmpeg_crf = '17' ffmpeg_preset = 'veryslow' add_soundtrack = 'None' #@param ["File","Init Video"] soundtrack_path = "https://freetestdata.com/wp-content/uploads/2021/09/Free_Test_Data_1MB_MP3.mp3" render_steps = False #@param {type: 'boolean'} path_name_modifier = "x0_pred" #@param ["x0_pred","x"] max_video_frames = 200 #@param {type:"string"} store_frames_in_ram = False #@param {type: 'boolean'} frame_interpolation_engine = "RIFE v4.6" #@param ["RIFE v4.0","RIFE v4.3","RIFE v4.6"] frame_interpolation_x_amount = "Disabled" #"Disabled" #@param ["Disabled" + all values from x2 to x10] frame_interpolation_slow_mo_amount = "Disabled" #@param ["Disabled","x2","x4","x8"] frame_interpolation_keep_imgs = False #@param {type: 'boolean'} return locals() import gradio as gr import os import time from types import SimpleNamespace i1_store_backup = "

Deforum extension for auto1111 — version 2.0b

" i1_store = i1_store_backup mask_fill_choices=['fill', 'original', 'latent noise', 'latent nothing'] def setup_deforum_setting_dictionary(self, is_img2img, is_extension = True): d = SimpleNamespace(**DeforumArgs()) #default args da = SimpleNamespace(**DeforumAnimArgs()) #default anim args dp = SimpleNamespace(**ParseqArgs()) #default parseq ars dv = SimpleNamespace(**DeforumOutputArgs()) #default video args dloopArgs = SimpleNamespace(**LoopArgs()) if not is_extension: with gr.Row(): btn = gr.Button("Click here after the generation to show the video") with gr.Row(): i1 = gr.HTML(i1_store, elem_id='deforum_header') else: btn = i1 = gr.HTML("") with gr.Accordion("Info, Links and Help", open=False): gr.HTML("""Made by deforum.github.io, port for AUTOMATIC1111's webui maintained by kabachuha""") gr.HTML(""" If you liked this extension, please give it a star on GitHub! 😊""") if not is_extension: def show_vid(): return { i1: gr.update(value=i1_store, visible=True) } btn.click( show_vid, [], [i1] ) with gr.Tab('Run'): # Sampling settings START with gr.Accordion('General Image Sampling Settings', open=True): with gr.Row().style(equal_height=False): # with gr.Column(variant='compact'): with gr.Column(): from modules.sd_samplers import samplers_for_img2img with gr.Row(): sampler = gr.Dropdown(label="sampler", choices=[x.name for x in samplers_for_img2img], value=samplers_for_img2img[0].name, type="value", elem_id="sampler", interactive=True) steps = gr.Slider(label="steps", minimum=0, maximum=200, step=1, value=d.steps, interactive=True) with gr.Row(variant='compat'): with gr.Column(scale=4): # W = gr.Slider(label="Width", minimum=64, maximum=2048, step=64, value=d.W, interactive=True) # H = gr.Slider(label="Height", minimum=64, maximum=2048, step=64, value=d.W, interactive=True) W = gr.Slider(label="Width", minimum=64, maximum=2048, step=64, value=d.W, interactive=True) H = gr.Slider(label="Height", minimum=64, maximum=2048, step=64, value=d.H, interactive=True) with gr.Column(scale=4): seed = gr.Number(label="seed", value=d.seed, interactive=True, precision=0) batch_name = gr.Textbox(label="batch_name", lines=1, interactive=True, value = d.batch_name) with gr.Row(visible=False): # batch_name = gr.Textbox(label="batch_name", lines=1, interactive=True, value = d.batch_name) filename_format = gr.Textbox(label="filename_format", lines=1, interactive=True, value = d.filename_format, visible=False) with gr.Accordion('Subseed controls & More', open=False): with gr.Row(): seed_enable_extras = gr.Checkbox(label="Enable subseed controls", value=False) subseed = gr.Number(label="subseed", value=d.subseed, interactive=True, precision=0) subseed_strength = gr.Slider(label="subseed_strength", minimum=0, maximum=1, step=0.01, value=d.subseed_strength, interactive=True) with gr.Row(): seed_resize_from_w = gr.Slider(minimum=0, maximum=2048, step=64, label="Resize seed from width", value=0) seed_resize_from_h = gr.Slider(minimum=0, maximum=2048, step=64, label="Resize seed from height", value=0) with gr.Row(): ddim_eta = gr.Number(label="ddim_eta", value=d.ddim_eta, interactive=True) tiling = gr.Checkbox(label='Tiling', value=False) n_batch = gr.Number(label="n_batch", value=d.n_batch, interactive=True, precision=0, visible=False) # NOT VISIBLE IN THE UI! with gr.Row(visible=False): save_settings = gr.Checkbox(label="save_settings", value=d.save_settings, interactive=True) # NOT VISIBLE IN THE UI! with gr.Row(visible=False): save_samples = gr.Checkbox(label="save_samples", value=d.save_samples, interactive=True) display_samples = gr.Checkbox(label="display_samples", value=False, interactive=False) with gr.Row(visible=False): save_sample_per_step = gr.Checkbox(label="save_sample_per_step", value=d.save_sample_per_step, interactive=True) show_sample_per_step = gr.Checkbox(label="show_sample_per_step", value=False, interactive=False) with gr.Accordion('Run from Settings file', open=False): with gr.Row(): override_settings_with_file = gr.Checkbox(label="Override settings", value=False, interactive=True) custom_settings_file = gr.Textbox(label="Custom settings file", lines=1, interactive=True) # Animation settings 'Key' tab with gr.Tab('Keyframes'): #TODO make a some sort of the original dictionary parsing # Main top animation settings with gr.Accordion('Animation Mode, Max Frames and Border', open=True): with gr.Row(): animation_mode = gr.Dropdown(label="animation_mode", choices=['2D', '3D', 'Video Input', 'Interpolation'], value=da.animation_mode, type="value", elem_id="animation_mode", interactive=True) max_frames = gr.Number(label="max_frames", value=da.max_frames, interactive=True, precision=0) border = gr.Dropdown(label="border", choices=['replicate', 'wrap'], value=da.border, type="value", elem_id="border", interactive=True) # loopArgs with gr.Accordion('Guided Images', open=False): gr.HTML("""You can use this as a guided image tool or as a looper depending on your settings in the keyframe images field. Set the keyframes and the images that you want to show up. Note: the number of frames between each keyframe should be greater than the tweening frames.""") # In later versions this should be also in the strength schedule, but for now you need to set it. gr.HTML("""Prerequisites: """) gr.HTML("""Looping recommendations: """) with gr.Row(): use_looper = gr.Checkbox(label="Use guided images for the next run", value=False, interactive=True) with gr.Row(): init_images = gr.Textbox(label="Images to use for keyframe guidance", lines=9, value = keyframeExamples(), interactive=True) gr.HTML("""strength schedule might be better if this is higher, around .75 during the keyfames you want to switch on""") with gr.Row(): image_strength_schedule = gr.Textbox(label="Image strength schedule", lines=1, value = "0:(.75)", interactive=True) gr.HTML("""blendFactor = blendFactorMax - blendFactorSlope * cos((frame % tweening_frames_schedule) / (tweening_frames_schedule / 2))""") with gr.Row(): blendFactorMax = gr.Textbox(label="blendFactorMax", lines=1, value = "0:(.35)", interactive=True) with gr.Row(): blendFactorSlope = gr.Textbox(label="blendFactorSlope", lines=1, value = "0:(.25)", interactive=True) with gr.Row(): gr.HTML("""number of frames this will calculated over. After each insersion frame.""") with gr.Row(): tweening_frames_schedule = gr.Textbox(label="tweening frames schedule", lines=1, value = "0:(20)", interactive=True) with gr.Row(): gr.HTML("""the amount each frame during a tweening step to use the new images colors""") with gr.Row(): color_correction_factor = gr.Textbox(label="color correction factor", lines=1, value = "0:(.075)", interactive=True) # Seed Scheduling with gr.Accordion('Seed Scheduling', open=False): with gr.Row(): seed_behavior = gr.Dropdown(label="seed_behavior", choices=['iter', 'fixed', 'random', 'ladder', 'alternate', 'schedule'], value=d.seed_behavior, type="value", elem_id="seed_behavior", interactive=True) seed_iter_N = gr.Number(label="seed_iter_N", value=d.seed_iter_N, interactive=True, precision=0) with gr.Row(): gr.HTML("Please set seed_behavior to 'Schedule' for this option to become active") with gr.Row(): seed_schedule = gr.Textbox(label="seed_schedule", lines=1, value = da.seed_schedule, interactive=True) # 2D + 3D Motion with gr.Accordion('2D + 3D Motion', open=True): with gr.Row(): angle = gr.Textbox(label="angle", lines=1, value = da.angle, interactive=True) with gr.Row(): zoom = gr.Textbox(label="zoom", lines=1, value = da.zoom, interactive=True) with gr.Row(): translation_x = gr.Textbox(label="translation_x", lines=1, value = da.translation_x, interactive=True) with gr.Row(): translation_y = gr.Textbox(label="translation_y", lines=1, value = da.translation_y, interactive=True) # 3D-only Motion with gr.Accordion('3D-Only Motion', open=True): with gr.Row(): translation_z = gr.Textbox(label="translation_z", lines=1, value = da.translation_z, interactive=True) with gr.Row(): rotation_3d_x = gr.Textbox(label="rotation_3d_x", lines=1, value = da.rotation_3d_x, interactive=True) with gr.Row(): rotation_3d_y = gr.Textbox(label="rotation_3d_y", lines=1, value = da.rotation_3d_y, interactive=True) with gr.Row(): rotation_3d_z = gr.Textbox(label="rotation_3d_z", lines=1, value = da.rotation_3d_z, interactive=True) # General Settings with gr.Accordion('General Settings', open=True): with gr.Row(): strength_schedule = gr.Textbox(label="strength_schedule", lines=1, value = da.strength_schedule, interactive=True) with gr.Row(): contrast_schedule = gr.Textbox(label="contrast_schedule", lines=1, value = da.contrast_schedule, interactive=True) with gr.Row(): cfg_scale_schedule = gr.Textbox(label="cfg_scale_schedule", lines=1, value = da.cfg_scale_schedule, interactive=True) # Coherence with gr.Accordion('Coherence', open=True): with gr.Row(): histogram_matching = gr.Checkbox(label="Force all frames to match initial frame's colors. Overrides a1111 settings. NOT RECOMMENDED, enable only for backwards compatibility.", value=da.histogram_matching, interactive=True) with gr.Row(): with gr.Column(variant="compact"): color_coherence = gr.Dropdown(label="color_coherence", choices=['None', 'Match Frame 0 HSV', 'Match Frame 0 LAB', 'Match Frame 0 RGB', 'Video Input'], value=da.color_coherence, type="value", elem_id="color_coherence", interactive=True) with gr.Column(variant="compact"): color_coherence_video_every_N_frames = gr.Number(label="color_coherence_video_every_N_frames", value=1, interactive=True) with gr.Column(variant="compact"): color_force_grayscale = gr.Checkbox(label="color_force_grayscale", value=da.color_force_grayscale, interactive=True) with gr.Column(variant="compact"): diffusion_cadence = gr.Number(label="diffusion_cadence", value=1, interactive=True) with gr.Row(): # what to do with blank frames (they may result from glitches or the NSFW filter being turned on): reroll with +1 seed, interrupt the animation generation, or do nothing reroll_blank_frames = gr.Dropdown(label="reroll_blank_frames", choices=['reroll', 'interrupt', 'ignore'], value=d.reroll_blank_frames, type="value", elem_id="reroll_blank_frames", interactive=True) # Noise with gr.Accordion('Noise', open=True): with gr.Row(): noise_schedule = gr.Textbox(label="noise_schedule", lines=1, value = da.noise_schedule, interactive=True) with gr.Row(): noise_type = gr.Dropdown(label="noise_type", choices=['uniform', 'perlin'], value=da.noise_type, type="value", elem_id="noise_type", interactive=True) with gr.Row(): gr.HTML("

Perlin noise params, if selected:

") perlin_w = gr.Number(label="perlin_w", value=da.perlin_w, interactive=True) perlin_h = gr.Number(label="perlin_h", value=da.perlin_h, interactive=True) with gr.Row(): perlin_octaves = gr.Slider(label="perlin_octaves", minimum=1, maximum=7, value=da.perlin_octaves, step=1, interactive=True) perlin_persistence = gr.Slider(label="perlin_persistence", minimum=0, maximum=1, value=da.perlin_persistence, step=0.02, interactive=True) # Anti-blur with gr.Accordion('Anti Blur', open=True): with gr.Row(): kernel_schedule = gr.Textbox(label="kernel_schedule", lines=1, value = da.kernel_schedule, interactive=True) with gr.Row(): sigma_schedule = gr.Textbox(label="sigma_schedule", lines=1, value = da.sigma_schedule, interactive=True) with gr.Row(): amount_schedule = gr.Textbox(label="amount_schedule", lines=1, value = da.amount_schedule, interactive=True) with gr.Row(): threshold_schedule = gr.Textbox(label="threshold_schedule", lines=1, value = da.threshold_schedule, interactive=True) # 3D Depth Warping with gr.Accordion('3D Depth Warping', open=False): with gr.Row(): use_depth_warping = gr.Checkbox(label="use_depth_warping", value=da.use_depth_warping, interactive=True) with gr.Row(): midas_weight = gr.Number(label="midas_weight", value=da.midas_weight, interactive=True) padding_mode = gr.Dropdown(label="padding_mode", choices=['border', 'reflection', 'zeros'], value=da.padding_mode, type="value", elem_id="padding_mode", interactive=True) sampling_mode = gr.Dropdown(label="sampling_mode", choices=['bicubic', 'bilinear', 'nearest'], value=da.sampling_mode, type="value", elem_id="sampling_mode", interactive=True) save_depth_maps = gr.Checkbox(label="save_depth_maps", value=da.save_depth_maps, interactive=True) # 3D FOV with gr.Accordion('3D Field Of View (FOV)', open=False): with gr.Row(): fov_schedule = gr.Textbox(label="fov_schedule", lines=1, value = da.fov_schedule, interactive=True) with gr.Row(): near_schedule = gr.Textbox(label="near_schedule", lines=1, value = da.near_schedule, interactive=True) with gr.Row(): far_schedule = gr.Textbox(label="far_schedule", lines=1, value = da.far_schedule, interactive=True) # Perspective Flip with gr.Accordion('Perspective Flip', open=False): with gr.Row(): enable_perspective_flip = gr.Checkbox(label="enable_perspective_flip", value=da.enable_perspective_flip, interactive=True) with gr.Row(): perspective_flip_theta = gr.Textbox(label="perspective_flip_theta", lines=1, value = da.perspective_flip_theta, interactive=True) with gr.Row(): perspective_flip_phi = gr.Textbox(label="perspective_flip_phi", lines=1, value = da.perspective_flip_phi, interactive=True) with gr.Row(): perspective_flip_gamma = gr.Textbox(label="perspective_flip_gamma", lines=1, value = da.perspective_flip_gamma, interactive=True) with gr.Row(): perspective_flip_fv = gr.Textbox(label="perspective_flip_fv", lines=1, value = da.perspective_flip_fv, interactive=True) # Steps Scheduling with gr.Accordion('Steps Scheduling', open=False): with gr.Row(): enable_steps_scheduling = gr.Checkbox(label="enable steps scheduling", value=da.enable_steps_scheduling, interactive=True) with gr.Row(): steps_schedule = gr.Textbox(label="steps_schedule", lines=1, value = da.steps_schedule, interactive=True) # Sampler Scheduling with gr.Accordion('Sampler Scheduling', open=False): with gr.Row(): enable_sampler_scheduling = gr.Checkbox(label="enable sampler scheduling.", value=da.enable_sampler_scheduling, interactive=True) with gr.Row(): sampler_schedule = gr.Textbox(label="sampler_schedule", lines=1, value = da.sampler_schedule, interactive=True) # Checkpoint Scheduling with gr.Accordion('Checkpoint Scheduling', open=True): with gr.Row(): enable_checkpoint_scheduling = gr.Checkbox(label="enable_checkpoint_scheduling", value=da.enable_checkpoint_scheduling, interactive=True) with gr.Row(): checkpoint_schedule = gr.Textbox(label="checkpoint_schedule", lines=1, value = da.checkpoint_schedule, interactive=True) with gr.Accordion('CLIP skip Scheduling', open=False): with gr.Row(): enable_clipskip_scheduling = gr.Checkbox(label="enable_clipskip_scheduling", value=da.enable_clipskip_scheduling, interactive=True) with gr.Row(): clipskip_schedule = gr.Textbox(label="clipskip_schedule", lines=1, value = da.clipskip_schedule, interactive=True) # Animation settings END # Prompts tab START with gr.Tab('Prompts'): gr.HTML("""

Important notes and changes from regular vanilla deforum:

""") with gr.Row(): animation_prompts = gr.Textbox(label="animation_prompts", lines=8, interactive=True, value = DeforumAnimPrompts()) gr.HTML("Positive prompt to be appended to all animation prompts") with gr.Row(): animation_prompts_positive = gr.Textbox(label="animation_prompts_positive", lines=1, interactive=True, value = "hq") gr.HTML("Negative prompt to be appended to all animation prompts, dont add --neg here") with gr.Row(): animation_prompts_negative = gr.Textbox(label="animation_prompts_negative", lines=1, interactive=True, value = "blurry") # Composable Mask scheduling with gr.Accordion('Composable Mask scheduling', open=True): gr.HTML("To enable, check use_mask in the Init tab.
Supports boolean operations (! - negation, & - and, | - or, ^ - xor, \ - difference, () - nested operations);
default variables in \{\}, like \{init_mask\}, \{video_mask\}, \{everywhere\};
masks from files in [], like [mask1.png];
description-based word masks in <>, like <apple>, <hair>") with gr.Row(): mask_schedule = gr.Textbox(label="mask_schedule", lines=1, value = da.mask_schedule, interactive=True) with gr.Row(): use_noise_mask = gr.Checkbox(label="use_noise_mask", value=da.use_noise_mask, interactive=True) with gr.Row(): noise_mask_schedule = gr.Textbox(label="noise_mask_schedule", lines=1, value = da.noise_mask_schedule, interactive=True) # Prompts settings END with gr.Tab('Init'): # Image Init with gr.Accordion('Image Init', open=True): with gr.Row(): use_init = gr.Checkbox(label="use_init", value=d.use_init, interactive=True, visible=True) # Need to REMOVE? from_img2img_instead_of_link = gr.Checkbox(label="from_img2img_instead_of_link", value=False, interactive=False, visible=False) with gr.Row(): strength_0_no_init = gr.Checkbox(label="strength_0_no_init", value=True, interactive=True) strength = gr.Slider(label="strength", minimum=0, maximum=1, step=0.02, value=0, interactive=True) with gr.Row(): init_image = gr.Textbox(label="init_image", lines=1, interactive=True, value = d.init_image) with gr.Accordion('Mask Init', open=True): with gr.Row(): use_mask = gr.Checkbox(label="use_mask", value=d.use_mask, interactive=True) use_alpha_as_mask = gr.Checkbox(label="use_alpha_as_mask", value=d.use_alpha_as_mask, interactive=True) invert_mask = gr.Checkbox(label="invert_mask", value=d.invert_mask, interactive=True) overlay_mask = gr.Checkbox(label="overlay_mask", value=d.overlay_mask, interactive=True) with gr.Row(): mask_file = gr.Textbox(label="mask_file", lines=1, interactive=True, value = d.mask_file) with gr.Row(): mask_contrast_adjust = gr.Number(label="mask_contrast_adjust", value=d.mask_contrast_adjust, interactive=True) mask_brightness_adjust = gr.Number(label="mask_brightness_adjust", value=d.mask_brightness_adjust, interactive=True) mask_overlay_blur = gr.Number(label="mask_overlay_blur", value=d.mask_overlay_blur, interactive=True) with gr.Row(): choice = mask_fill_choices[d.fill] fill = gr.Radio(label='mask_fill', choices=mask_fill_choices, value=choice, type="index") with gr.Row(): full_res_mask = gr.Checkbox(label="full_res_mask", value=d.full_res_mask, interactive=True) full_res_mask_padding = gr.Slider(minimum=0, maximum=512, step=1, label="full_res_mask_padding", value=d.full_res_mask_padding, interactive=True) # Video Init with gr.Accordion('Video Init', open=True): with gr.Row(): video_init_path = gr.Textbox(label="video_init_path", lines=1, value = da.video_init_path, interactive=True) with gr.Row(): extract_nth_frame = gr.Number(label="extract_nth_frame", value=da.extract_nth_frame, interactive=True, precision=0) extract_from_frame = gr.Number(label="extract_from_frame", value=da.extract_from_frame, interactive=True, precision=0) extract_to_frame = gr.Number(label="extract_to_frame", value=da.extract_to_frame, interactive=True, precision=0) overwrite_extracted_frames = gr.Checkbox(label="overwrite_extracted_frames", value=False, interactive=True) use_mask_video = gr.Checkbox(label="use_mask_video", value=False, interactive=True) with gr.Row(): video_mask_path = gr.Textbox(label="video_mask_path", lines=1, value = da.video_mask_path, interactive=True) with gr.Accordion('Resume Animation', open=False): with gr.Row(): resume_from_timestring = gr.Checkbox(label="resume_from_timestring", value=da.resume_from_timestring, interactive=True) resume_timestring = gr.Textbox(label="resume_timestring", lines=1, value = da.resume_timestring, interactive=True) # Parseq with gr.Accordion('Parseq', open=False): gr.HTML(""" Use an sd-parseq manifest for your animation (leave blank to ignore).

Note that parseq overrides:

Parseq does not override:

""") with gr.Row(): parseq_manifest = gr.Textbox(label="Parseq Manifest (JSON or URL)", lines=4, value = dp.parseq_manifest, interactive=True) with gr.Row(): parseq_use_deltas = gr.Checkbox(label="Use delta values for movement parameters", value=dp.parseq_use_deltas, interactive=True) # HYBRID VIDEO tab with gr.Tab('Hybrid Video'): with gr.Accordion("Info & Help", open=False): hybrid_html = "

Hybrid Video Compositing in 2D/3D Modeby reallybigname

" hybrid_html += "" hybrid_html += "Hybrid Video Schedules" hybrid_html += "" hybrid_html += "Click Here for more info/ a Guide." gr.HTML(hybrid_html) with gr.Accordion("Hybrid Settings", open=True): with gr.Row(equal_height=True): with gr.Column(): hybrid_generate_inputframes = gr.Checkbox(label="generate_inputframes", value=False, interactive=True) hybrid_composite = gr.Checkbox(label="hybrid_composite", value=False, interactive=True) hybrid_generate_human_masks = gr.Dropdown(label="generate_human_masks", choices=['None', 'PNGs', 'Video', 'Both'], value=da.hybrid_generate_human_masks, type="value", elem_id="hybrid_generate_human_masks", interactive=True) with gr.Row(equal_height=True): with gr.Column(): hybrid_use_first_frame_as_init_image = gr.Checkbox(label="first_frame_as_init_image", value=False, interactive=True) hybrid_motion_use_prev_img = gr.Checkbox(label="motion_use_prev_img", value=False, interactive=True) hybrid_motion = gr.Dropdown(label="hybrid_motion", choices=['None', 'Optical Flow', 'Perspective', 'Affine'], value=da.hybrid_motion, type="value", elem_id="hybrid_motion", interactive=True) hybrid_flow_method = gr.Dropdown(label="flow_method", choices=['DIS Medium', 'Farneback'], value=da.hybrid_flow_method, type="value", elem_id="hybrid_flow_method", interactive=True) with gr.Row(equal_height=True): hybrid_comp_mask_equalize = gr.Dropdown(label="comp_mask_equalize", choices=['None', 'Before', 'After', 'Both'], value=da.hybrid_comp_mask_equalize, type="value", elem_id="hybrid_comp_mask_equalize", interactive=True) hybrid_comp_mask_type = gr.Dropdown(label="comp_mask_type", choices=['None', 'Depth', 'Video Depth', 'Blend', 'Difference'], value=da.hybrid_comp_mask_type, type="value", elem_id="hybrid_comp_mask_type", interactive=True) with gr.Row(equal_height=True): with gr.Column(): hybrid_comp_mask_auto_contrast = gr.Checkbox(label="comp_mask_auto_contrast", value=False, interactive=True) hybrid_comp_mask_inverse = gr.Checkbox(label="comp_mask_inverse", value=False, interactive=True) hybrid_comp_save_extra_frames = gr.Checkbox(label="comp_save_extra_frames", value=False, interactive=True) with gr.Accordion("Hybrid Schedules", open=False): with gr.Row(): hybrid_comp_alpha_schedule = gr.Textbox(label="comp_alpha_schedule", lines=1, value = da.hybrid_comp_alpha_schedule, interactive=True) with gr.Row(): hybrid_comp_mask_blend_alpha_schedule = gr.Textbox(label="comp_mask_blend_alpha_schedule", lines=1, value = da.hybrid_comp_mask_blend_alpha_schedule, interactive=True) with gr.Row(): hybrid_comp_mask_contrast_schedule = gr.Textbox(label="comp_mask_contrast_schedule", lines=1, value = da.hybrid_comp_mask_contrast_schedule, interactive=True) with gr.Row(): hybrid_comp_mask_auto_contrast_cutoff_high_schedule = gr.Textbox(label="comp_mask_auto_contrast_cutoff_high_schedule", lines=1, value = da.hybrid_comp_mask_auto_contrast_cutoff_high_schedule, interactive=True) with gr.Row(): hybrid_comp_mask_auto_contrast_cutoff_low_schedule = gr.Textbox(label="comp_mask_auto_contrast_cutoff_low_schedule", lines=1, value = da.hybrid_comp_mask_auto_contrast_cutoff_low_schedule, interactive=True) # VIDEO OUTPUT TAB with gr.Tab('Video output'): with gr.Accordion('Video Output Settings', open=True): with gr.Row(): fps = gr.Number(label="fps", value=dv.fps, interactive=True) output_format = gr.Dropdown(label="output_format", choices=['PIL gif', 'FFMPEG mp4'], value='FFMPEG mp4', type="value", elem_id="output_format", interactive=True) with gr.Column(): with gr.Row(): ffmpeg_location = gr.Textbox(label="ffmpeg_location", lines=1, interactive=True, value = dv.ffmpeg_location) ffmpeg_crf = gr.Number(label="ffmpeg_crf", interactive=True, value = dv.ffmpeg_crf) ffmpeg_preset = gr.Dropdown(label="ffmpeg_preset", choices=['veryslow', 'slower', 'slow', 'medium', 'fast', 'faster', 'veryfast', 'superfast', 'ultrafast'], interactive=True, value = dv.ffmpeg_preset, type="value") with gr.Column(): with gr.Row(): add_soundtrack = gr.Dropdown(label="add_soundtrack", choices=['None', 'File', 'Init Video'], value=dv.add_soundtrack, interactive=True, type="value") soundtrack_path = gr.Textbox(label="soundtrack_path", lines=1, interactive=True, value = dv.soundtrack_path) with gr.Row(): skip_video_for_run_all = gr.Checkbox(label="skip_video_for_run_all", value=dv.skip_video_for_run_all, interactive=True) store_frames_in_ram = gr.Checkbox(label="store_frames_in_ram", value=dv.store_frames_in_ram, interactive=True) # with gr.Row(): # store_frames_in_ram = gr.Checkbox(label="store_frames_in_ram", value=dv.store_frames_in_ram, interactive=True) with gr.Accordion('Manual Settings', open=False): with gr.Row(): use_manual_settings = gr.Checkbox(label="use_manual_settings", value=dv.use_manual_settings, interactive=True) # rend_step Never worked - set to visible false 28-1-23 render_steps = gr.Checkbox(label="render_steps", value=dv.render_steps, interactive=True, visible=False) with gr.Row(): max_video_frames = gr.Number(label="max_video_frames", value=200, interactive=True) path_name_modifier = gr.Dropdown(label="path_name_modifier", choices=['x0_pred', 'x'], value=dv.path_name_modifier, type="value", elem_id="path_name_modifier", interactive=True) with gr.Row(): image_path = gr.Textbox(label="image_path", lines=1, interactive=True, value = dv.image_path) with gr.Row(): mp4_path = gr.Textbox(label="mp4_path", lines=1, interactive=True, value = dv.mp4_path) with gr.Accordion('Frame Interpolation (RIFE)', open=True): with gr.Accordion('Important notes and Help', open=False): gr.HTML(""" Use RIFE Frame Interpolation to smooth out, slow-mo (or both) any video.

Supported engines:

Important notes:

""") with gr.Column(): with gr.Row(): # Interpolation Engine frame_interpolation_engine = gr.Dropdown(label="Engine", choices=['RIFE v4.0','RIFE v4.3','RIFE v4.6'], value=dv.frame_interpolation_engine, type="value", elem_id="frame_interpolation_engine", interactive=True) # How many times to interpolate (interp x) frame_interpolation_x_amount = gr.Dropdown(label="Interp x", choices=['Disabled','x2','x3','x4','x5','x6','x7','x8','x9','x10'], value=dv.frame_interpolation_x_amount, type="value", elem_id="frae_interpolation_x_amount", interactive=True) # Interp Slow-Mo (setting final output fps, not really doing anything direclty with RIFE) frame_interpolation_slow_mo_amount = gr.Dropdown(label="Slow-Mo x", choices=['Disabled','x2','x4','x8'], value=dv.frame_interpolation_slow_mo_amount, type="value", elem_id="frame_interpolation_slow_mo_amount", interactive=True) # If this is set to True, we keep all of the interpolated frames in a folder. Default is False - means we delete them at the end of the run frame_interpolation_keep_imgs = gr.Checkbox(label="Keep Imgs", elem_id="frame_interpolation_keep_imgs", value=dv.frame_interpolation_keep_imgs, interactive=True) with gr.Row(): # Intrpolate any existing video from the connected PC with gr.Accordion('Interpolate an existing video', open=False): # A drag-n-drop UI box to which the user uploads a *single* (at this stage) video vid_to_rife_chosen_file = gr.File(label="Video to interpolate", interactive=True, file_count="single", file_types=["video"]) with gr.Row(): # Non interactive textbox showing uploaded input vid total Frame Count in_vid_frame_count_window = gr.Textbox(label="In Frame Count", lines=1, interactive=False, value='---') # Non interactive textbox showing uploaded input vid FPS in_vid_fps_ui_window = gr.Textbox(label="In FPS", lines=1, interactive=False, value='---') # Non interactive textbox showing expected output interpolated video FPS out_interp_vid_estimated_fps = gr.Textbox(label="Interpolated Vid FPS", value='---') # This is the actual button that's pressed to initiate the interpolation: rife_btn = gr.Button(value="*Interpolate uploaded video*") # Show a text about CLI outputs: gr.HTML("* check your CLI for outputs") # make the functin call when the RIFE button is clicked rife_btn.click(upload_vid_to_rife,inputs=[vid_to_rife_chosen_file, frame_interpolation_engine, frame_interpolation_x_amount, frame_interpolation_slow_mo_amount, frame_interpolation_keep_imgs, ffmpeg_location, ffmpeg_crf, ffmpeg_preset, in_vid_fps_ui_window]) # update output fps field upon changing of interp_x and/ or slow_mo_x frame_interpolation_x_amount.change(set_interp_out_fps, inputs=[frame_interpolation_x_amount, frame_interpolation_slow_mo_amount, in_vid_fps_ui_window], outputs=out_interp_vid_estimated_fps) frame_interpolation_slow_mo_amount.change(set_interp_out_fps, inputs=[frame_interpolation_x_amount, frame_interpolation_slow_mo_amount, in_vid_fps_ui_window], outputs=out_interp_vid_estimated_fps) # Populate the above FPS and FCount values as soon as a video is uploaded to the FileUploadBox (vid_to_rife_chosen_file) vid_to_rife_chosen_file.change(gradio_f_interp_get_fps_and_fcount,inputs=[vid_to_rife_chosen_file, frame_interpolation_x_amount, frame_interpolation_slow_mo_amount],outputs=[in_vid_fps_ui_window,in_vid_frame_count_window, out_interp_vid_estimated_fps]) # END OF UI TABS return locals() ### SETTINGS STORAGE UPDATE! 2023-01-27 ### To Reduce The Number Of Settings Overrides, ### They Are Being Passed As Dictionaries ### It Would Have Been Also Nice To Retrieve Them ### From Functions Like Deforumoutputargs(), ### But Over Time There Was Some Cross-Polination, ### So They Are Now Hardcoded As 'List'-Strings Below ### If you're adding a new setting, add it to one of the lists ### besides writing it in the setup functions above anim_args_names = str(r'''animation_mode, max_frames, border, angle, zoom, translation_x, translation_y, translation_z, rotation_3d_x, rotation_3d_y, rotation_3d_z, enable_perspective_flip, perspective_flip_theta, perspective_flip_phi, perspective_flip_gamma, perspective_flip_fv, noise_schedule, strength_schedule, contrast_schedule, cfg_scale_schedule, enable_steps_scheduling, steps_schedule, fov_schedule, near_schedule, far_schedule, seed_schedule, enable_sampler_scheduling, sampler_schedule, mask_schedule, use_noise_mask, noise_mask_schedule, enable_checkpoint_scheduling, checkpoint_schedule, enable_clipskip_scheduling, clipskip_schedule, kernel_schedule, sigma_schedule, amount_schedule, threshold_schedule, histogram_matching, color_coherence, color_coherence_video_every_N_frames, color_force_grayscale, diffusion_cadence, noise_type, perlin_w, perlin_h, perlin_octaves, perlin_persistence, use_depth_warping, midas_weight, padding_mode, sampling_mode, save_depth_maps, video_init_path, extract_nth_frame, extract_from_frame, extract_to_frame, overwrite_extracted_frames, use_mask_video, video_mask_path, resume_from_timestring, resume_timestring''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') hybrid_args_names = str(r'''hybrid_generate_inputframes, hybrid_generate_human_masks, hybrid_use_first_frame_as_init_image, hybrid_motion, hybrid_motion_use_prev_img, hybrid_flow_method, hybrid_composite, hybrid_comp_mask_type, hybrid_comp_mask_inverse, hybrid_comp_mask_equalize, hybrid_comp_mask_auto_contrast, hybrid_comp_save_extra_frames, hybrid_comp_alpha_schedule, hybrid_comp_mask_blend_alpha_schedule, hybrid_comp_mask_contrast_schedule, hybrid_comp_mask_auto_contrast_cutoff_high_schedule, hybrid_comp_mask_auto_contrast_cutoff_low_schedule''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') args_names = str(r'''W, H, tiling, seed, sampler, seed_enable_extras, subseed, subseed_strength, seed_resize_from_w, seed_resize_from_h, steps, ddim_eta, n_batch, save_settings, save_samples, display_samples, save_sample_per_step, show_sample_per_step, batch_name, filename_format, seed_behavior, seed_iter_N, use_init, from_img2img_instead_of_link, strength_0_no_init, strength, init_image, use_mask, use_alpha_as_mask, invert_mask, overlay_mask, mask_file, mask_contrast_adjust, mask_brightness_adjust, mask_overlay_blur, fill, full_res_mask, full_res_mask_padding, reroll_blank_frames''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') video_args_names = str(r'''skip_video_for_run_all, fps, output_format, ffmpeg_location, ffmpeg_crf, ffmpeg_preset, add_soundtrack, soundtrack_path, use_manual_settings, render_steps, max_video_frames, path_name_modifier, image_path, mp4_path, store_frames_in_ram, frame_interpolation_engine, frame_interpolation_x_amount, frame_interpolation_slow_mo_amount, frame_interpolation_keep_imgs''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') parseq_args_names = str(r'''parseq_manifest, parseq_use_deltas''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') loop_args_names = str(r'''use_looper, init_images, image_strength_schedule, blendFactorMax, blendFactorSlope, tweening_frames_schedule, color_correction_factor''' ).replace("\n", "").replace("\r", "").replace(" ", "").split(',') component_names = ['override_settings_with_file', 'custom_settings_file'] + anim_args_names +['animation_prompts', 'animation_prompts_positive', 'animation_prompts_negative'] + args_names + video_args_names + parseq_args_names + hybrid_args_names + loop_args_names settings_component_names = [name for name in component_names if name not in video_args_names] def setup_deforum_setting_ui(self, is_img2img, is_extension = True): ds = setup_deforum_setting_dictionary(self, is_img2img, is_extension) return [ds[name] for name in (['btn'] + component_names)] def pack_anim_args(args_dict): return {name: args_dict[name] for name in (anim_args_names + hybrid_args_names)} def pack_args(args_dict): args_dict = {name: args_dict[name] for name in args_names} args_dict['precision'] = 'autocast' args_dict['scale'] = 7 args_dict['C'] = 4 args_dict['f'] = 8 args_dict['timestring'] = "" args_dict['init_latent'] = None args_dict['init_sample'] = None args_dict['init_c'] = None args_dict['noise_mask'] = None args_dict['seed_internal'] = 0 return args_dict def pack_video_args(args_dict): return {name: args_dict[name] for name in video_args_names} def pack_parseq_args(args_dict): return {name: args_dict[name] for name in parseq_args_names} def pack_loop_args(args_dict): return {name: args_dict[name] for name in loop_args_names} def process_args(args_dict_main): override_settings_with_file = args_dict_main['override_settings_with_file'] custom_settings_file = args_dict_main['custom_settings_file'] args_dict = pack_args(args_dict_main) anim_args_dict = pack_anim_args(args_dict_main) video_args_dict = pack_video_args(args_dict_main) parseq_args_dict = pack_parseq_args(args_dict_main) loop_args_dict = pack_loop_args(args_dict_main) import json root = SimpleNamespace(**Root()) root.p = args_dict_main['p'] p = root.p root.animation_prompts = json.loads(args_dict_main['animation_prompts']) positive_prompts = args_dict_main['animation_prompts_positive'] negative_prompts = args_dict_main['animation_prompts_negative'] for key in root.animation_prompts: animationPromptCurr = root.animation_prompts[key] root.animation_prompts[key] = f"{positive_prompts} {animationPromptCurr} {'' if '--neg' in animationPromptCurr else '--neg'} {negative_prompts}" from deforum_helpers.settings import load_args if override_settings_with_file: load_args(args_dict, anim_args_dict, parseq_args_dict, loop_args_dict, custom_settings_file, root) print(f"Additional models path: {root.models_path}") if not os.path.exists(root.models_path): os.mkdir(root.models_path) args = SimpleNamespace(**args_dict) anim_args = SimpleNamespace(**anim_args_dict) video_args = SimpleNamespace(**video_args_dict) parseq_args = SimpleNamespace(**parseq_args_dict) loop_args = SimpleNamespace(**loop_args_dict) p.width, p.height = map(lambda x: x - x % 64, (args.W, args.H)) p.steps = args.steps p.seed = args.seed p.sampler_name = args.sampler p.batch_size = args.n_batch p.tiling = args.tiling p.seed_enable_extras = args.seed_enable_extras p.subseed = args.subseed p.subseed_strength = args.subseed_strength p.seed_resize_from_w = args.seed_resize_from_w p.seed_resize_from_h = args.seed_resize_from_h p.fill = args.fill p.ddim_eta = args.ddim_eta # TODO: Handle batch name dynamically? current_arg_list = [args, anim_args, video_args, parseq_args] args.outdir = os.path.join(p.outpath_samples, args.batch_name) root.outpath_samples = args.outdir args.outdir = os.path.join(os.getcwd(), args.outdir) if not os.path.exists(args.outdir): os.makedirs(args.outdir) args.seed = get_fixed_seed(args.seed) args.timestring = time.strftime('%Y%m%d%H%M%S') args.strength = max(0.0, min(1.0, args.strength)) if not args.use_init: args.init_image = None if anim_args.animation_mode == 'None': anim_args.max_frames = 1 elif anim_args.animation_mode == 'Video Input': args.use_init = True return root, args, anim_args, video_args, parseq_args, loop_args def print_args(args): print("ARGS: /n") for key, value in args.__dict__.items(): print(f"{key}: {value}") # Local gradio-to-rife function. *Needs* to stay here since we do Root() and use gradio elements directly, to be changed in the future def upload_vid_to_rife(file, engine, x_am, sl_am, keep_imgs, f_location, f_crf, f_preset, in_vid_fps): # print msg and do nothing if vid not uploaded or interp_x not provided if not file or x_am == 'Disabled': return print("Please upload a video and set a proper value for 'Interp x'. Can't interpolate x0 times :)") root_params = Root() f_models_path = root_params['models_path'] process_rife_vid_upload_logic(file, engine, x_am, sl_am, keep_imgs, f_location, f_crf, f_preset, in_vid_fps, f_models_path, file.orig_name)