Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some technical questions. #5

Open
lyf1212 opened this issue Jan 8, 2025 · 2 comments
Open

Some technical questions. #5

lyf1212 opened this issue Jan 8, 2025 · 2 comments

Comments

@lyf1212
Copy link

lyf1212 commented Jan 8, 2025

Thank you for your brilliant work! I have some questions after reading:

  1. Why not LayerNorm? As revealed by continuous exposure learning for low-light enhancement using Neural ODE, LayerNorm generalize diverse illumination distribution inherently, probably benefit for your IIM.
  2. Why initial IIM can provide invariant features? As depicted in your code, all kernels are initialized to Gaussian distribution, but not ones. As I thought, those $w_i$ should be 1 in Eq.(7) to ensure illumination-invariant from the beginning.
  3. Why not zero-shot generalization? You retrain IIM on COCO to examine generalizability, but have you trailed on zero-shot manner without training? That is, directly leverage the fused results generated by object detection task, to segmentation task.

Looking forward for your reply!

[1] Continuous exposure learning for low-light enhancement using Neural ODE, submission to ICLR 25.

@MingboHong
Copy link
Owner

3. ou retrain IIM on COCO to exam

Hi,Thank you for your valuable questions and for showing interest in our project :)

Q:Why not LayerNorm?
R: : I align with your point, many researchers would suggest not to use BN on image restoration tasks. It seems that LN would be a better choice for such a task. However, LN typically needs a fixed input shape, which may impede the flexibility of detection tasks (such as Multi-scale training/testing).

Q:Why initial IIM can provide invariant features?
R: : The initialization method probably would not affect the IIM. As you can see in Eq.7:
image

This equation only contains intrinsic property (aka illuminant invariant items), which means you can use various initialization methods for IIM at the beginning. (But zero initialization may not be a good choice)

Q:Why not zero-shot generalization?
R: : Actually, we did some zero-shot experiments, as shown in OpenReview. But, it is not remarkable.

@lyf1212
Copy link
Author

lyf1212 commented Jan 18, 2025

Thank you for such enthusiastic response!

Yes, LayerNorm requires for fixed spatial shape, hampering further adaptability for variable image inputs. Btw, InstanceNorm which only modulate along the channel axis seems to be a better choice? Since it normalize all channels into a similar distribution, which is probably more variant and stable under diverse illumination. Inside a deep CNN, each channel represent basic pixel information of inputs.

As for the initial issue, I made a mistake. Under the settings of Eq.(5) and "zero-mean", it indeed only contains basic surface properties which is illumination-invariant.
However, Eq.(5) is still confusing, the straightforward derivation should be

Image,
although it seems reasonable in Eq.(5) in terms of final result.

As for zero-shot generalization, I really mean "training with only normal-lit images, test on low-light images". Since it is easy to access large-scale high-quality annotated normal-lit images, if it is possible to construct a mapping from "illumination-invariant features" to segmentation/detection results, the robustness should be better intuitively.

In brief, your work has indeed made significant contributions to low-light downstream tasks. Thank you for your insightful thinking and diligent efforts!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants