Hypernetwork-MonkeyPatch-Extension icon indicating copy to clipboard operation
Hypernetwork-MonkeyPatch-Extension copied to clipboard

Is it possible to make a skip connection hypernetwork?

Open nekoyama32767 opened this issue 3 years ago • 5 comments

It is possible to make a totally manual mode which allows to make connection and activation function every layer? Maybe resblock like architecture or skip connection is useful for hypernetwork

nekoyama32767 avatar Dec 19 '22 02:12 nekoyama32767

Yes, but the problem is UI, I could add some way to load model but creating it is different story.... I'll try making it to save and load any arbitrary model structure, but creating it would be homework for users.

aria1th avatar Dec 21 '22 03:12 aria1th

https://github.com/aria1th/Hypernetwork-MonkeyPatch-Extension/tree/residual-connection

Currently skip-connection, ResBlock architecture for normal tensors are available here.

There are few things you need to know: image

Since it 'interpolates' attention vector, setting value lower than 1 would make loss in vector, because it cannot recover vectors from compressed form.

(It could be solved by having attention-compressing and decompressing model somehow?)

Default standard deviation is too big for attention transformation. I suggest using lower stdev.

Original webui does not support these types of hypernetworks. image image

Training works well anyway, but you might have to struggle with setting.

aria1th avatar Jan 17 '23 06:01 aria1th

Thank you for working on this. I will try it

nekoyama32767 avatar Jan 17 '23 12:01 nekoyama32767

It looks work well Epoch 0 (Cycle 0 after warm up step 298) aokana_side_drop-596 Epoch 39 (Cycle 9 step 11920) aokana_side_drop-11920

nekoyama32767 avatar Jan 19 '23 15:01 nekoyama32767

Skip-connection hypernetworks train on vainilla NovelAI model works excellent with a pre-trained dreambooth model PD5ZSKY
Z690F%4}(VC9ZQ

nekoyama32767 avatar Jan 20 '23 19:01 nekoyama32767