-
Notifications
You must be signed in to change notification settings - Fork 47
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference: own data #42
Comments
Thanks for your interest!
Feel free to let me know if you have further questions. |
Thank you for your quick reply.
Current setting: sparse depth map has 5% density of all pixels, but I use the same for supervising the network. So if I understood it correctly, the supervision of the groundtruth depth maps need to be at least 15% correct? Details of inference: model - Here is an example result --> do you think this is the expected result?
|
Using the same maps for supervision is not sufficient. To alleviate this problem, you need to generate denser GT
|
Okay I understand regarding the sparsity for training. But I am still unsure about what the positional encodings are because I am preparing my own data, so is there a flag for evaluation? I understand --co to std is for training right? I can see in |
You could refer to [An intriguing failing of convolutional neural networks and the CoordConv solution] by Liu for more details about positional encoding. In our default settings, we use the geometric encoding (ie. 3d coordianates) described in our paper. And the evaluation and training process should share consistent settings. |
I get it now, I was able to create the positional encoding (u, v coordinates using the camera intrinsic). I still get a patchy result like this when I evaluate, are you able to analyze what else may possibly go wrong or maybe sensitive? (I am just using the pretrained weights to evaluate on this data) |
Intuitively I guess the GT maps are not dense enough for supervised depth completion. GT maps are reuqired to be much denser than the sparse inputs. |
But we dont need GT maps in this case where we just evaluate. |
I think two points could be taken into consideration:
|
Hi,
I just had a few questions regarding using our own data and running inference using PENet pretrained weights.
How sparse can the depth map be?
Currently, my inference image is from the Kitti360 dataset which is quite similar to the previous kitti that the network was trained on. But there is no GT depth to sample the depth from. So my sparse depth map is quite sparse.
When I run inference on this image, the prediction is also sparse i.e I have prediction only in the regions covered by the sparse depth map. Is this an expected behaviour?
What should my input be for 'positions' (i.e the cropped image), I don't want to crop the images for running inference, so should I just set
input['positions'] = input['rgb']
?It would be great if you can answer these questions when time permits :)
Regards,
Shrisha
The text was updated successfully, but these errors were encountered: