Creates videos in a modern HD anime style.
The primary trigger is "An1meStyl3," but it is recommended to also include the tag "AnimeStyle" alongside it. When describing subjects, it helps to specify that they are anime characters, so instead of "a man speaking" it is better to use "an anime man speaking."
WAN 2.2 appears to have a strong bias towards realistic generations, and training this LoRA with more than 2000 steps resulted in overfit and yielded a large amount of noise, even with 6 low-noise steps. As a result, more consistent generation can be achieved by bumping the model strength up to 1.2 or adding "((realistic))" to your negative prompt. In the event you are using one or more concept/motion/character LoRAs with a bias towards realism, use a stronger negative like "(((realistic))), ((photograph))"
This is intended to be a general-use anime LoRA, so it was trained from stills on a diverse set of shows that have a common general art style, rather than fixating on a particular show or creator, so that it doesn't closely resemble any particular illustrator's signature style or creative choices.
Version Notes:
Version 1: Creates characters and basic motions. Trained on a set of images
Version 2: Creates characters and supports more dynamic character and camera motions (previous version would inject a high amount of background noise if the camera changed angles). Trained a set of images and a set of video clips
California AB 2013 Training Data Disclosure
This LoRA was fine-tuned using visual data consisting primarily of still images sourced from animated television series, along with a limited amount of publicly available fan-created renderings and AI-generated images. The training data includes copyrighted material owned by third parties, including animation studios, production committees, distributors, and individual artists. No training data was licensed or purchased. This LoRA is provided for non-commercial use only under the terms of its distribution.
The dataset for all versions consists of over 1,500 images collected from publicly accessible sources across more than a dozen animated series between approximately 2000 and 2025. Image data was processed through standard resizing, cropping, normalization, and labeling steps. Synthetic images were included as part of the training dataset.
A second version of this model additionally incorporates approximately 200–250 video clips sourced from more than half a dozen animated series. In that version, clips were used as video sequences to support motion and temporal consistency, and were processed and labeled for video training in addition to the image preprocessing described above.
This model is intended for non-commercial, experimental, and educational use. Generated outputs may reflect copyrighted visual styles or themes associated with the underlying training data. Users are responsible for ensuring compliance with applicable copyright law, other intellectual property laws, and all other applicable laws.
Description
Better compatibility with camera motion and character LoRAs