IMDiffusion icon indicating copy to clipboard operation
IMDiffusion copied to clipboard

SWaT processing

Open WSake opened this issue 1 year ago • 3 comments

各位大佬,请问一下有没有SWaT数据集的处理代码~ 用其他论文的代码,最后结果很低!

WSake avatar Dec 06 '24 07:12 WSake

The paper said that SWaT dataset is ran with 51 channels, but in the code, it is ran with 45 channels (refer to line 123 in exe_machine.py). I attached a snippet from their paper as follows.

Image

My team notices this and change the code in line 123 to 51 and ran the code again and attempt to replicate their results. However, we encounter gradient explosion at the first epoch training. The input is normal, but the weights of the models are nan. We tried gradient clipping in the train loop, the convergence become very slow and eventually we are faced with gradient explosion again.

Can the author of this paper enlighten us the differences in code implementation and your paper?

SinChee avatar Mar 26 '25 06:03 SinChee

Hi, the original SWaT dataset contains 51 channels. However, we noticed that 6 of these channels have completely constant values (equal to 1) in both the training and test sets. To optimize computational efficiency, we removed these channels.

2025-03-26 14:25:59 "Sin Chee" @.***> 写道:

The paper said that SWaT dataset is ran with 51 channels, but in the code, it is ran with 45 channels (refer to line 123 in exe_machine.py). I attached a snippet from their paper as follows.

Screenshot.2025-03-26.at.14.19.12.png (view on web)

My team notices this and change the code in line 123 to 51 and ran the code again and attempt to replicate their results. However, we encounter gradient explosion at the first epoch training. The input is normal, but the weights of the models are nan. We tried gradient clipping in the train loop, the convergence become very slow and eventually we are faced with gradient explosion again.

Can the author of this paper enlighten us the differences in code implementation and your paper?

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you are subscribed to this thread.Message ID: @.***>

SinChee left a comment (17000cyh/IMDiffusion#7)

The paper said that SWaT dataset is ran with 51 channels, but in the code, it is ran with 45 channels (refer to line 123 in exe_machine.py). I attached a snippet from their paper as follows.

Screenshot.2025-03-26.at.14.19.12.png (view on web)

My team notices this and change the code in line 123 to 51 and ran the code again and attempt to replicate their results. However, we encounter gradient explosion at the first epoch training. The input is normal, but the weights of the models are nan. We tried gradient clipping in the train loop, the convergence become very slow and eventually we are faced with gradient explosion again.

Can the author of this paper enlighten us the differences in code implementation and your paper?

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you are subscribed to this thread.Message ID: @.***>

17000cyh avatar Mar 27 '25 02:03 17000cyh

Thank you for replying, does this mean that the rest of the comparison in your paper is also processed this way?

SinChee avatar Mar 27 '25 05:03 SinChee