Stable Diffusion training not going great so far. I think a specific, unchanging physical item like a nostr:npub17tyke9lkgxd98ruyeul6wt3pj3s9uxzgp9hxu5tsenjmweue6sqq4y3mgl is harder for it to learn than it is to teach it a new face since it already has a super-deep training set on what a face is.

Could be this LORA approach is more well-suited to learning new variations of a known concept rather than trying to develop a completely new SeedSigner-ness concept.

Still tons of different settings to try.

Reply to this note

Please Login to reply.

Discussion

Progress! Doesn't look like much, but it's preserving a good amount of the @npub17tyke9lkgxd98ruyeul6wt3pj3s9uxzgp9hxu5tsenjmweue6sqq4y3mgl essence AND is able to place it in a totally new context that's unrelated to the training images!

I haven't done much training yet but you might try a textual inversion embedding instead of lora.

I've seen those but haven't read up on them. The Stable Diffusion world is pretty overwhelming!

Good captioning can help with LoRA training too.

Yeah, taking another pass at the captioning. Also resizing everything to 768x768 and not using buckets.

I dunno, I'm seeing some interesting new design directions here... 🤣