-
I'm glad the pivotal tuning is ready to test in universal embeddings branch. But there's few clear documentation how to use this feature or the caption. Do I need special condition in the caption?, and is additional embedding are identical to pivotal tuning?. For example, one of my caption is "Ryan Reynolds from Selfless as Damian wearing a black suit with red shoe, Damian standing in front of a bus stop reading a newspaper". My intention is I want to train Ryan Reynolds as himself and Damian as a character from Selfless with his signature appearance. Based on wiki, base embedding is blank in default and I think I should keep them blank because I want to train a new embedding, so it's clear. My confusion start from here, when I put custom placeholder (trigger word), let's say And what "initial embedding text" does?, did it will change the And if I have thousands of concept that I should train, this would be hassle because I should insert the additional embedding one by one and no preset load. My captioning structure is [Character name][Appearance][Activity], [The Background], [additional non descriptive caption]. So if I use Keep tag count (3), the trainer will focus on the first three tags and shuffle the rest, this will avoid the trunctation. If pivotal tuning can be trained based on the Caption emergence (which will focus to learn the character because character name is always in the front of the caption/tags) than it would be good |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
I am not sure why you want to use the additional embedding for this, instead of just a standard finetune, but the initial implementation is in no way going to be useful as you will have thousands of additional text embeddings to manage. Why not start with one concept first and see how it goes? As far as I know, this is uncharted territory, there is not much guidance to give. Nero can give input on how their implementation works, but I do not think anyone has seen how far this can go. I started the additional embedding page on the wiki, check it out (https://github.com/Nerogar/OneTrainer/wiki/Additional-Embeddings). |
Beta Was this translation helpful? Give feedback.
I am not sure why you want to use the additional embedding for this, instead of just a standard finetune, but the initial implementation is in no way going to be useful as you will have thousands of additional text embeddings to manage.
Why not start with one concept first and see how it goes? As far as I know, this is uncharted territory, there is not much guidance to give. Nero can give input on how their implementation works, but I do not think anyone has seen how far this can go.
I started the additional embedding page on the wiki, check it out (https://github.com/Nerogar/OneTrainer/wiki/Additional-Embeddings).