Hi, I was noticing that when using a normalized float descriptor like Superpoint, there were some issues when training. The weights were different based on how the descriptor was scaled. Upon debugging, I noticed that when we do the tranform when computing the nodeWeights, we copy a floating point value (the distance of the descriptor from the node) into a uint64 variable. This forces the value to be 0 or 1 for a float descriptor and was reducing the resolution. Is there a reason why this was done ? Thanks!