Flat Color - Style

0
0
0
0
Anime StyleGirl
Recently Updated: First Published:
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info
Anime Style,Girl,LoRA,HunyuanDiT v1.2Image info

Flat Color - Style

Trained on images without visible lineart, flat colors, and little to no indication of depth.

This is a small style LoRA I thought would be interesting to try with a v-pred model (noobai v-pred), for the reduced color bleeding and strong blacks in particular.

The effect is quite nice, so I've extended the dataset in following versions, including Hunyuan Video and Wan.

ℹ️ LoRA work best when applied to the base models on which they are trained. Please check the versions for the appropriate base models.

Recommended prompt structure:

Positive prompt:

flat color, no lineart, blending, negative space,
{{tags}}
masterpiece, best quality, very aesthetic, newest

For Noobai V-Pred, a ComfyUI workflow reference with DynamicThresholding, Upscaling, and FaceDetailer can be found here: https://civitai.com/posts/11457095

Trained with https://github.com/tdrussell/diffusion-pipe

Training data consists of:

  • 42 images as a combination of

    • Images used from other versions this model card

    • Images extracted as keyframes from several videos

  • 19 short video clips ~40 frames each

Training configs:

dataset.toml

# Aspect ratio bucketing settings
enable_ar_bucket = true
min_ar = 0.5
max_ar = 2.0
num_ar_buckets = 7

[[directory]] # IMAGES
# Path to the directory containing images and their corresponding caption files.
path = '/mnt/d/huanvideo/training_data/images'
num_repeats = 5
resolutions = [1024]
frame_buckets = [1] # Use 1 frame for images.


[[directory]] # VIDEOS
# Path to the directory containing videos and their corresponding caption files.
path = '/mnt/d/huanvideo/training_data/videos'
num_repeats = 5
resolutions = [256] # Set video resolution to 256 (e.g., 244p).
frame_buckets = [33, 49, 81] # Define frame buckets for videos.

config.toml

# Dataset config file.
output_dir = '/mnt/d/huanvideo/training_output'
dataset = 'dataset.toml'

# Training settings
epochs = 50
micro_batch_size_per_gpu = 1
pipeline_stages = 1
gradient_accumulation_steps = 4
gradient_clipping = 1.0
warmup_steps = 100

# eval settings
eval_every_n_epochs = 5
eval_before_first_step = true
eval_micro_batch_size_per_gpu = 1
eval_gradient_accumulation_steps = 1

# misc settings
save_every_n_epochs = 15
checkpoint_every_n_minutes = 30
activation_checkpointing = true
partition_method = 'parameters'
save_dtype = 'bfloat16'
caching_batch_size = 1
steps_per_print = 1
video_clip_mode = 'single_middle'

[model]
type = 'hunyuan-video'

transformer_path = '/mnt/d/huanvideo/models/diffusion_models/hunyuan_video_720_cfgdistill_fp8_e4m3fn.safetensors'
vae_path = '/mnt/d/huanvideo/models/vae/hunyuan_video_vae_bf16.safetensors'
llm_path = '/mnt/d/huanvideo/models/llm'
clip_path = '/mnt/d/huanvideo/models/clip'

dtype = 'bfloat16'
transformer_dtype = 'float8'
timestep_sample_method = 'logit_normal'

[adapter]
type = 'lora'
rank = 32
dtype = 'bfloat16'

[optimizer]
type = 'adamw_optimi'
lr = 5e-5
betas = [0.9, 0.99]
weight_decay = 0.02
eps = 1e-8

Discussion

Most popular
|
Newest
Send
Coming Soon
Download
(0.00KB)
Version Details
Type
Generation Count
0
Downloads
0
Recommended ParametersAdd

Gallery

Most popular
|
Newest