CivArchive
    Nausicae and the AI of the wind - NauSKaaValOfWi V0.1
    Preview 170463
    Preview 170482
    Preview 170481
    Preview 170480
    Preview 170479
    Preview 170478
    Preview 170477
    Preview 170476
    Preview 170475
    Preview 170474
    Preview 170473
    Preview 170472
    Preview 170471
    Preview 170470
    Preview 170469
    Preview 170468
    Preview 170467
    Preview 170466
    Preview 170465
    Preview 170464

    I can remember that over 30 years ago I did read Nausicaa and enjoyed the unveiling of the characters and world over the story's progression. While the attention stays throughout the works and closely follows characters and their development, it drastically zoomed out and faded away, Leaving the audience wondering what happened to the rest of the story and the world afterwards. This checkpoint in its final version, will have taken in all there is to know about the visuals and style of Nausicaa. It should be, therefore, able to be used by the community to continue telling the story in the actual art style with the ability to recall every character and location and the correct way and extent upon it. All the material for the training data is retrieved from the available public library archives and images that are freely available on the Internet.

    • Update 1: now also LoRa of version 0.1 as safetensors

    The current version is based on the first book's first half as the training data. There's so far no manual tagging to improve the understanding of characters and locations. This version is prerelease for testing; feel free to play around and give feedback. You can also contact me "ClassicRPGArt -at- proton.me" if you want to help with the project.

    Description

    FAQ

    Comments (11)

    sdf34gdssdg4Feb 28, 2023· 2 reactions
    CivitAI

    That's awesome, thanks so much!

    Edit: As stated in the description, this checkpoint is still in its early stage and it shows. Shapes, elements, characters, structures, etc can have a high level of abstraction. I'd love to be able to mix this with LoRA to give a more modern look along with controlnet :)

    ClassicRPGArt
    Author
    Mar 1, 2023

    i did add a LoRa

    Lupin3rdMar 1, 2023· 1 reaction
    CivitAI

    I want to help, tell me what I need to do :)

    ClassicRPGArt
    Author
    Mar 1, 2023

    Send me an email to ClassicRPGArt -At- proton.me

    JustMaierMar 1, 2023
    CivitAI

    I see that this seems to include a config file but it says it’s a 1.5 based model. Is that incorrect? Should this be marked as an SD 2 model?

    BTW I love these sketch style models. Thanks for making this!

    ClassicRPGArt
    Author
    Mar 1, 2023

    it is, for now, 1.5 as a base but once all the training data is in I can also do a 2.X version, 1.5 is better, for now, to work with during development and has a wider range of data in it.

    infeztatorMar 1, 2023· 1 reaction
    CivitAI

    I'm an artist (Marvel back in 1991) and a 3d Guy for the last 20 years but I still draw. I'm interested in training a checkpoint in some of my styles using A1111 dreambooth and am having a hard time finding good information on settings for style, How many images, epochs, etc... Any tips would be very welcome! :) You have done a great job with Nausica.

    ClassicRPGArt
    Author
    Mar 1, 2023

    These are the main points that you have two keep in mind when working with training subjects outside of the photorealistic realm.

    1. make sure that your dataset is consistent in regard to brightness and contrast. If you're working with line art drawings, black should be black, and white should be white; otherwise, you are training the model to generate the noise of your paper and the imperfections of the ink as part of the learning data, which is an artefact that later on will inherently reduce the usefulness of the model.

    2. Clean up all images that you work with. Images shouldn't contain anything that is not part of what you like a model to learn. A typical mistake to avoid is having lines of image borders, speech bubbles, or several images from the comic together on one training data page. Work either with a white or black background and simply photoshop everything away that you don’t want to have as part of the training data. it’s better to have empty space on a training dataset image than any kind of artefact because blank background spaces (hex value FF.FF.FF or 00.00.00 ) will not be learned if your tag description mentions like a white background. While artefacts will compromise your network.

    3. remove noisy and low resolution images from your dataset. Either by not including them or using AI to upress your image (SD batch scaler in A1111 or Topaz Gigapixel AI). I would recommend that the images in your dataset should be two times the size that you are training the model at. This means if your model is working at 512x512 you should have a dataset of 1024x1024 images.

    4. Make sure that every one of your images has a correct set of tag’s. you can automatically generate tags with the smart process add-on for A1111. I would recommend setting the threshold to something like 0.25; the default is 0.85 and by far too high for working with intricate art . don't be afraid to use your text editor or the tools in A1111 Two added to the text files remove tags that you don't like can definitely add tags of the subject domain. Meaning names of places, persons or events that are visible in the images so that you can later on recall those concepts from your promps. as an example, if you have an image of Arnold Schwarzenegger and don't tag him as terminator you won't be able to use the word terminator and the picture of Mr Schwarzenegger will just have been trained as muscular body build up on a movie set. Which is completely useless as training information if you want to have useful prompts.

    5. Always work with classification images in order to not overtrain and break the checkpoint with your training. Have at least 10 classification images for every image in the dataset if your dataset is large and has a great amount of tags make sure that you have between 10 and 50 times the images in your classification directory geenerated for each of your images. Make sure that your classification images have a correct instance prompt ( Instance Prompt ) always use ,[filewords] in order to extract the description and make it part of the training do that for Instance Prompt, Class Prompt and Sample Image Prompt

    6. make sure that your instance prompt is a continuous string that is not part of the actual clip of your checkpoint otherwise, you destroying parts of the checkpoint with your training data, in most cases a combination of the first two letters of a sentence that describes your training data gets a rather random string that is not overwriting anything. This is really important if you make a mistake at this point; your model will only show you back your training data and not allow you to mix it with anything else it knows in the correct ratio.

    7. use negative classification image prompt is two avoid art styles or artists or concepts that are counter-indicated to what you’re trying to train for example, if you are training a western art style in the marvel look make sure that you have, for example, manga as a negative prompt in order to not get the training results confused.

    8. Make sure that you don’t overtrain your checkpoint. What you should do is let’s say for 300 training images set the “Training Steps Per Image (Epochs)” two 500 iterations per image and put the model saving frequency to something like 50 and the preview frequency also. Make sure that you say in the setting that it should savour checkpoint every time while training. And activate the checkbox for safe training data for later use by setting “Save separate diffusers snapshots when training completes.” what will happen now is it will go from 0 to 500 iterations and create 10 checkpoints on the way. You should also set the preview sample number to something like eight images so that you see something. Now you just let it run and you will see in the sample images at what point your network gets overtrain by your results degrading again. Then take the checkpoint that looked good before the one that you saw the degeneration and try to merge both if the resut odd looks better than each one of them then you have the best result that you can possibly get if the result looks worse, just go with the last good looking checkpoint and that is it. You’ve trained your checkpoint on your data in the correct way and have the “Best” result for the time being.

    9. Very important last message info make sure that you have gigabytes of free hard drive space each checkpoint is between 4 and 8 GB and in the current setup of the example this will result in about 40 GB of checkpoints if you run out of disk space while training everything crashes and fails!!!!

    10. don’t be afraid to hit the cancel button if anything looks odd or the model training is not progressing in the sample images. Just cancel everything and start over fresh there is no point letting your GPU waste energy for eight hours on an already failing set up! I mean it if something looks rather ood, it probably is wrong start over. However, don’t do this before you had at least two checkpoints look at but if you don’t see the improvements in the smaple images, just have a look at the setup and do it again with better values.

    if you have any further questions feel free to write me an email “ClassicRPGArt -at- proton . me”

     

    infeztatorMar 1, 2023

    Fantastic Response! Thank you very much. You may indeed get some more question, but to limit them, If you know any good documentation site that is up to date... The youtube examples use a different interface and also have different options.

    I'm going to have to rebuild my training data, but that's fine I have the original scans. I'm interested in having it generate what looks like my pencils, and then train image2image to ink them in different styles. I've evolved my style in the last 24 years of doing 3D work...

    Once I get this working A dream AI project would be to train a 3D model to create 3d human models from portrait photos - using photoreal renders of my models and my standard human topology. It would be fairly simple to create a huge training set by randomizing morph targets and texture layers to generate a large set of virtual people with associated vertex positions to train it against.

    Thank you again.

    snydxyp813Mar 1, 2023
    CivitAI

    看上去像是直接哪漫画训练的?

    ClassicRPGArt
    Author
    Mar 1, 2023· 1 reaction

    風の谷のナウシカ (宮崎 駿)